{ "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 11035, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00045310376076121433, "grad_norm": 6.794378137076671, "learning_rate": 7.246376811594204e-08, "loss": 1.3257, "step": 1 }, { "epoch": 0.0009062075215224287, "grad_norm": 3.3173936440747274, "learning_rate": 1.4492753623188408e-07, "loss": 1.3361, "step": 2 }, { "epoch": 0.001359311282283643, "grad_norm": 3.3370587040043995, "learning_rate": 2.173913043478261e-07, "loss": 1.3378, "step": 3 }, { "epoch": 0.0018124150430448573, "grad_norm": 3.322622411169202, "learning_rate": 2.8985507246376816e-07, "loss": 1.327, "step": 4 }, { "epoch": 0.0022655188038060714, "grad_norm": 3.330304660990402, "learning_rate": 3.623188405797102e-07, "loss": 1.3126, "step": 5 }, { "epoch": 0.002718622564567286, "grad_norm": 10.417254964260536, "learning_rate": 4.347826086956522e-07, "loss": 1.3134, "step": 6 }, { "epoch": 0.0031717263253285004, "grad_norm": 3.2713953098833883, "learning_rate": 5.072463768115942e-07, "loss": 1.3072, "step": 7 }, { "epoch": 0.0036248300860897147, "grad_norm": 3.248939492646187, "learning_rate": 5.797101449275363e-07, "loss": 1.3148, "step": 8 }, { "epoch": 0.0040779338468509285, "grad_norm": 3.1715226231430305, "learning_rate": 6.521739130434783e-07, "loss": 1.307, "step": 9 }, { "epoch": 0.004531037607612143, "grad_norm": 7.835435351477772, "learning_rate": 7.246376811594204e-07, "loss": 1.3347, "step": 10 }, { "epoch": 0.004984141368373357, "grad_norm": 6.371494953134731, "learning_rate": 7.971014492753623e-07, "loss": 1.3345, "step": 11 }, { "epoch": 0.005437245129134572, "grad_norm": 3.0854500103445908, "learning_rate": 8.695652173913044e-07, "loss": 1.3212, "step": 12 }, { "epoch": 0.0058903488898957865, "grad_norm": 2.6828762867567777, "learning_rate": 9.420289855072465e-07, "loss": 1.3168, "step": 13 }, { "epoch": 0.006343452650657001, "grad_norm": 2.6536790607059055, "learning_rate": 1.0144927536231885e-06, "loss": 1.3112, "step": 14 }, { "epoch": 0.006796556411418215, "grad_norm": 2.602587565105106, "learning_rate": 1.0869565217391306e-06, "loss": 1.336, "step": 15 }, { "epoch": 0.007249660172179429, "grad_norm": 2.5407809781392197, "learning_rate": 1.1594202898550726e-06, "loss": 1.3261, "step": 16 }, { "epoch": 0.007702763932940644, "grad_norm": 4.081748083745577, "learning_rate": 1.2318840579710147e-06, "loss": 1.2795, "step": 17 }, { "epoch": 0.008155867693701857, "grad_norm": 1.8845768057621888, "learning_rate": 1.3043478260869566e-06, "loss": 1.2772, "step": 18 }, { "epoch": 0.008608971454463071, "grad_norm": 1.8744443024518935, "learning_rate": 1.3768115942028987e-06, "loss": 1.2884, "step": 19 }, { "epoch": 0.009062075215224286, "grad_norm": 2.6695594407691114, "learning_rate": 1.4492753623188408e-06, "loss": 1.273, "step": 20 }, { "epoch": 0.0095151789759855, "grad_norm": 1.6950410365832382, "learning_rate": 1.521739130434783e-06, "loss": 1.2856, "step": 21 }, { "epoch": 0.009968282736746714, "grad_norm": 1.5977078584363975, "learning_rate": 1.5942028985507246e-06, "loss": 1.2553, "step": 22 }, { "epoch": 0.01042138649750793, "grad_norm": 1.6098037970194796, "learning_rate": 1.6666666666666667e-06, "loss": 1.2405, "step": 23 }, { "epoch": 0.010874490258269144, "grad_norm": 1.544671593399167, "learning_rate": 1.7391304347826088e-06, "loss": 1.2457, "step": 24 }, { "epoch": 0.011327594019030359, "grad_norm": 1.683925178933572, "learning_rate": 1.8115942028985508e-06, "loss": 1.2263, "step": 25 }, { "epoch": 0.011780697779791573, "grad_norm": 1.6498974667671953, "learning_rate": 1.884057971014493e-06, "loss": 1.2258, "step": 26 }, { "epoch": 0.012233801540552787, "grad_norm": 1.3765445210507907, "learning_rate": 1.956521739130435e-06, "loss": 1.2266, "step": 27 }, { "epoch": 0.012686905301314002, "grad_norm": 1.0931782183648786, "learning_rate": 2.028985507246377e-06, "loss": 1.202, "step": 28 }, { "epoch": 0.013140009062075216, "grad_norm": 2.407754586243597, "learning_rate": 2.101449275362319e-06, "loss": 1.21, "step": 29 }, { "epoch": 0.01359311282283643, "grad_norm": 1.6096913522939786, "learning_rate": 2.173913043478261e-06, "loss": 1.1822, "step": 30 }, { "epoch": 0.014046216583597644, "grad_norm": 2.126761175730882, "learning_rate": 2.246376811594203e-06, "loss": 1.1912, "step": 31 }, { "epoch": 0.014499320344358859, "grad_norm": 1.884331534269006, "learning_rate": 2.3188405797101453e-06, "loss": 1.1659, "step": 32 }, { "epoch": 0.014952424105120073, "grad_norm": 1.3640637701370182, "learning_rate": 2.391304347826087e-06, "loss": 1.1761, "step": 33 }, { "epoch": 0.015405527865881287, "grad_norm": 1.0501306463986333, "learning_rate": 2.4637681159420295e-06, "loss": 1.1599, "step": 34 }, { "epoch": 0.0158586316266425, "grad_norm": 1.2163823217823881, "learning_rate": 2.5362318840579714e-06, "loss": 1.1732, "step": 35 }, { "epoch": 0.016311735387403714, "grad_norm": 942.0142766230022, "learning_rate": 2.6086956521739132e-06, "loss": 1.1829, "step": 36 }, { "epoch": 0.01676483914816493, "grad_norm": 1.8759289106294637, "learning_rate": 2.6811594202898555e-06, "loss": 1.1384, "step": 37 }, { "epoch": 0.017217942908926143, "grad_norm": 1.5815779423065799, "learning_rate": 2.7536231884057974e-06, "loss": 1.1661, "step": 38 }, { "epoch": 0.017671046669687357, "grad_norm": 4.899452311955387, "learning_rate": 2.8260869565217393e-06, "loss": 1.168, "step": 39 }, { "epoch": 0.01812415043044857, "grad_norm": 1.0384930416013471, "learning_rate": 2.8985507246376816e-06, "loss": 1.1383, "step": 40 }, { "epoch": 0.018577254191209785, "grad_norm": 0.90171374660459, "learning_rate": 2.9710144927536235e-06, "loss": 1.1353, "step": 41 }, { "epoch": 0.019030357951971, "grad_norm": 0.8400130015849284, "learning_rate": 3.043478260869566e-06, "loss": 1.1233, "step": 42 }, { "epoch": 0.019483461712732214, "grad_norm": 0.8143776639684308, "learning_rate": 3.1159420289855073e-06, "loss": 1.1209, "step": 43 }, { "epoch": 0.01993656547349343, "grad_norm": 0.7591274149365826, "learning_rate": 3.188405797101449e-06, "loss": 1.1304, "step": 44 }, { "epoch": 0.020389669234254643, "grad_norm": 0.8652202510877746, "learning_rate": 3.2608695652173914e-06, "loss": 1.1174, "step": 45 }, { "epoch": 0.02084277299501586, "grad_norm": 0.8544279557432947, "learning_rate": 3.3333333333333333e-06, "loss": 1.1143, "step": 46 }, { "epoch": 0.021295876755777075, "grad_norm": 0.6030998639551536, "learning_rate": 3.4057971014492756e-06, "loss": 1.1112, "step": 47 }, { "epoch": 0.02174898051653829, "grad_norm": 0.4843206697772058, "learning_rate": 3.4782608695652175e-06, "loss": 1.1172, "step": 48 }, { "epoch": 0.022202084277299503, "grad_norm": 0.6204615182618454, "learning_rate": 3.55072463768116e-06, "loss": 1.0931, "step": 49 }, { "epoch": 0.022655188038060717, "grad_norm": 0.5944248956836564, "learning_rate": 3.6231884057971017e-06, "loss": 1.098, "step": 50 }, { "epoch": 0.023108291798821932, "grad_norm": 0.5828314714219248, "learning_rate": 3.6956521739130436e-06, "loss": 1.0959, "step": 51 }, { "epoch": 0.023561395559583146, "grad_norm": 0.5999412122770981, "learning_rate": 3.768115942028986e-06, "loss": 1.1089, "step": 52 }, { "epoch": 0.02401449932034436, "grad_norm": 0.46690887958448196, "learning_rate": 3.840579710144928e-06, "loss": 1.1172, "step": 53 }, { "epoch": 0.024467603081105575, "grad_norm": 0.4819721399518448, "learning_rate": 3.91304347826087e-06, "loss": 1.0898, "step": 54 }, { "epoch": 0.02492070684186679, "grad_norm": 0.5430220837479383, "learning_rate": 3.9855072463768115e-06, "loss": 1.0822, "step": 55 }, { "epoch": 0.025373810602628003, "grad_norm": 0.3990739060980434, "learning_rate": 4.057971014492754e-06, "loss": 1.0874, "step": 56 }, { "epoch": 0.025826914363389217, "grad_norm": 0.4818875916232059, "learning_rate": 4.130434782608696e-06, "loss": 1.1004, "step": 57 }, { "epoch": 0.02628001812415043, "grad_norm": 0.49924350521137656, "learning_rate": 4.202898550724638e-06, "loss": 1.0796, "step": 58 }, { "epoch": 0.026733121884911646, "grad_norm": 0.3399810053561973, "learning_rate": 4.27536231884058e-06, "loss": 1.0857, "step": 59 }, { "epoch": 0.02718622564567286, "grad_norm": 0.3693297566671355, "learning_rate": 4.347826086956522e-06, "loss": 1.1021, "step": 60 }, { "epoch": 0.027639329406434075, "grad_norm": 0.37307106876463947, "learning_rate": 4.4202898550724645e-06, "loss": 1.0685, "step": 61 }, { "epoch": 0.02809243316719529, "grad_norm": 0.3556204677881556, "learning_rate": 4.492753623188406e-06, "loss": 1.0811, "step": 62 }, { "epoch": 0.028545536927956503, "grad_norm": 0.34304762475440925, "learning_rate": 4.565217391304348e-06, "loss": 1.1029, "step": 63 }, { "epoch": 0.028998640688717717, "grad_norm": 0.33511613786182287, "learning_rate": 4.637681159420291e-06, "loss": 1.0919, "step": 64 }, { "epoch": 0.02945174444947893, "grad_norm": 0.26547548907066015, "learning_rate": 4.710144927536232e-06, "loss": 1.0857, "step": 65 }, { "epoch": 0.029904848210240146, "grad_norm": 0.30515698877986136, "learning_rate": 4.782608695652174e-06, "loss": 1.1082, "step": 66 }, { "epoch": 0.03035795197100136, "grad_norm": 0.34677642538319214, "learning_rate": 4.855072463768117e-06, "loss": 1.0671, "step": 67 }, { "epoch": 0.030811055731762575, "grad_norm": 0.2793046266213334, "learning_rate": 4.927536231884059e-06, "loss": 1.0525, "step": 68 }, { "epoch": 0.031264159492523785, "grad_norm": 0.3019819183448341, "learning_rate": 5e-06, "loss": 1.0478, "step": 69 }, { "epoch": 0.031717263253285, "grad_norm": 0.2974706242530045, "learning_rate": 5.072463768115943e-06, "loss": 1.0762, "step": 70 }, { "epoch": 0.032170367014046214, "grad_norm": 0.3010779111095866, "learning_rate": 5.144927536231884e-06, "loss": 1.0414, "step": 71 }, { "epoch": 0.03262347077480743, "grad_norm": 0.3110926928939937, "learning_rate": 5.2173913043478265e-06, "loss": 1.0527, "step": 72 }, { "epoch": 0.03307657453556864, "grad_norm": 0.30160481677490203, "learning_rate": 5.289855072463769e-06, "loss": 1.0401, "step": 73 }, { "epoch": 0.03352967829632986, "grad_norm": 0.3504131961910516, "learning_rate": 5.362318840579711e-06, "loss": 1.0475, "step": 74 }, { "epoch": 0.03398278205709107, "grad_norm": 0.29950143101777443, "learning_rate": 5.4347826086956525e-06, "loss": 1.0537, "step": 75 }, { "epoch": 0.034435885817852285, "grad_norm": 0.32391818202298717, "learning_rate": 5.507246376811595e-06, "loss": 1.044, "step": 76 }, { "epoch": 0.0348889895786135, "grad_norm": 0.2693264894257604, "learning_rate": 5.579710144927537e-06, "loss": 1.0594, "step": 77 }, { "epoch": 0.035342093339374714, "grad_norm": 0.2755971528817587, "learning_rate": 5.652173913043479e-06, "loss": 1.056, "step": 78 }, { "epoch": 0.03579519710013593, "grad_norm": 0.2992010008011741, "learning_rate": 5.724637681159421e-06, "loss": 1.0398, "step": 79 }, { "epoch": 0.03624830086089714, "grad_norm": 0.2638836595147477, "learning_rate": 5.797101449275363e-06, "loss": 1.0518, "step": 80 }, { "epoch": 0.03670140462165836, "grad_norm": 0.22338086539083346, "learning_rate": 5.8695652173913055e-06, "loss": 1.0599, "step": 81 }, { "epoch": 0.03715450838241957, "grad_norm": 0.28051273579739283, "learning_rate": 5.942028985507247e-06, "loss": 1.0402, "step": 82 }, { "epoch": 0.037607612143180785, "grad_norm": 0.23789965305865307, "learning_rate": 6.014492753623189e-06, "loss": 1.048, "step": 83 }, { "epoch": 0.038060715903942, "grad_norm": 0.25558298057117784, "learning_rate": 6.086956521739132e-06, "loss": 1.025, "step": 84 }, { "epoch": 0.038513819664703214, "grad_norm": 0.24053853319773327, "learning_rate": 6.159420289855072e-06, "loss": 1.0477, "step": 85 }, { "epoch": 0.03896692342546443, "grad_norm": 0.22008973027493617, "learning_rate": 6.2318840579710145e-06, "loss": 1.0527, "step": 86 }, { "epoch": 0.03942002718622564, "grad_norm": 0.22966024542034746, "learning_rate": 6.304347826086958e-06, "loss": 1.0391, "step": 87 }, { "epoch": 0.03987313094698686, "grad_norm": 0.203355271836529, "learning_rate": 6.376811594202898e-06, "loss": 1.0375, "step": 88 }, { "epoch": 0.04032623470774807, "grad_norm": 0.20016076723631662, "learning_rate": 6.449275362318841e-06, "loss": 1.0367, "step": 89 }, { "epoch": 0.040779338468509285, "grad_norm": 0.24985942663828514, "learning_rate": 6.521739130434783e-06, "loss": 1.0349, "step": 90 }, { "epoch": 0.0412324422292705, "grad_norm": 0.2021671880559238, "learning_rate": 6.594202898550725e-06, "loss": 1.0238, "step": 91 }, { "epoch": 0.04168554599003172, "grad_norm": 0.2619038937214967, "learning_rate": 6.666666666666667e-06, "loss": 1.0435, "step": 92 }, { "epoch": 0.042138649750792935, "grad_norm": 0.2235919262174926, "learning_rate": 6.739130434782609e-06, "loss": 1.0329, "step": 93 }, { "epoch": 0.04259175351155415, "grad_norm": 0.2741558321896316, "learning_rate": 6.811594202898551e-06, "loss": 1.0317, "step": 94 }, { "epoch": 0.043044857272315364, "grad_norm": 0.25770363120478235, "learning_rate": 6.884057971014493e-06, "loss": 1.0417, "step": 95 }, { "epoch": 0.04349796103307658, "grad_norm": 0.2887337912104059, "learning_rate": 6.956521739130435e-06, "loss": 1.0294, "step": 96 }, { "epoch": 0.04395106479383779, "grad_norm": 0.30926199312015634, "learning_rate": 7.028985507246377e-06, "loss": 1.0265, "step": 97 }, { "epoch": 0.044404168554599006, "grad_norm": 0.3643497568697651, "learning_rate": 7.10144927536232e-06, "loss": 1.0166, "step": 98 }, { "epoch": 0.04485727231536022, "grad_norm": 0.35106531109974115, "learning_rate": 7.173913043478261e-06, "loss": 1.0147, "step": 99 }, { "epoch": 0.045310376076121435, "grad_norm": 0.27963009629700647, "learning_rate": 7.246376811594203e-06, "loss": 1.0424, "step": 100 }, { "epoch": 0.04576347983688265, "grad_norm": 0.3514784464007313, "learning_rate": 7.318840579710146e-06, "loss": 1.0248, "step": 101 }, { "epoch": 0.046216583597643864, "grad_norm": 0.36544110442478833, "learning_rate": 7.391304347826087e-06, "loss": 1.0285, "step": 102 }, { "epoch": 0.04666968735840508, "grad_norm": 0.3684310621727419, "learning_rate": 7.4637681159420295e-06, "loss": 1.0318, "step": 103 }, { "epoch": 0.04712279111916629, "grad_norm": 0.2936568839412294, "learning_rate": 7.536231884057972e-06, "loss": 1.0202, "step": 104 }, { "epoch": 0.047575894879927506, "grad_norm": 0.22116713938878638, "learning_rate": 7.608695652173914e-06, "loss": 1.0215, "step": 105 }, { "epoch": 0.04802899864068872, "grad_norm": 0.2933045867404151, "learning_rate": 7.681159420289856e-06, "loss": 1.0231, "step": 106 }, { "epoch": 0.048482102401449935, "grad_norm": 0.26904108901213936, "learning_rate": 7.753623188405797e-06, "loss": 1.032, "step": 107 }, { "epoch": 0.04893520616221115, "grad_norm": 0.25783883415427267, "learning_rate": 7.82608695652174e-06, "loss": 1.0168, "step": 108 }, { "epoch": 0.049388309922972363, "grad_norm": 0.3065447354398125, "learning_rate": 7.898550724637682e-06, "loss": 1.0242, "step": 109 }, { "epoch": 0.04984141368373358, "grad_norm": 0.3564346175478964, "learning_rate": 7.971014492753623e-06, "loss": 1.0273, "step": 110 }, { "epoch": 0.05029451744449479, "grad_norm": 0.37642999939461513, "learning_rate": 8.043478260869566e-06, "loss": 1.0225, "step": 111 }, { "epoch": 0.050747621205256006, "grad_norm": 0.5080645400966854, "learning_rate": 8.115942028985508e-06, "loss": 1.039, "step": 112 }, { "epoch": 0.05120072496601722, "grad_norm": 0.7134811111174192, "learning_rate": 8.188405797101449e-06, "loss": 0.9964, "step": 113 }, { "epoch": 0.051653828726778435, "grad_norm": 0.9520685593574333, "learning_rate": 8.260869565217392e-06, "loss": 1.0179, "step": 114 }, { "epoch": 0.05210693248753965, "grad_norm": 0.9369337832905413, "learning_rate": 8.333333333333334e-06, "loss": 1.0315, "step": 115 }, { "epoch": 0.05256003624830086, "grad_norm": 0.6114229123162317, "learning_rate": 8.405797101449275e-06, "loss": 0.9965, "step": 116 }, { "epoch": 0.05301314000906208, "grad_norm": 0.7650813565070868, "learning_rate": 8.478260869565218e-06, "loss": 1.0334, "step": 117 }, { "epoch": 0.05346624376982329, "grad_norm": 0.7663582280626271, "learning_rate": 8.55072463768116e-06, "loss": 1.009, "step": 118 }, { "epoch": 0.053919347530584506, "grad_norm": 0.6799252291969013, "learning_rate": 8.623188405797103e-06, "loss": 1.0272, "step": 119 }, { "epoch": 0.05437245129134572, "grad_norm": 0.7359182700911933, "learning_rate": 8.695652173913044e-06, "loss": 1.0208, "step": 120 }, { "epoch": 0.054825555052106935, "grad_norm": 0.6805686200184049, "learning_rate": 8.768115942028986e-06, "loss": 1.0122, "step": 121 }, { "epoch": 0.05527865881286815, "grad_norm": 0.7232599231876528, "learning_rate": 8.840579710144929e-06, "loss": 1.0047, "step": 122 }, { "epoch": 0.05573176257362936, "grad_norm": 0.6240422890646705, "learning_rate": 8.91304347826087e-06, "loss": 1.0205, "step": 123 }, { "epoch": 0.05618486633439058, "grad_norm": 0.4675406843020053, "learning_rate": 8.985507246376812e-06, "loss": 0.9973, "step": 124 }, { "epoch": 0.05663797009515179, "grad_norm": 0.7104359828768393, "learning_rate": 9.057971014492755e-06, "loss": 1.0244, "step": 125 }, { "epoch": 0.057091073855913006, "grad_norm": 0.6450997041676377, "learning_rate": 9.130434782608697e-06, "loss": 0.9959, "step": 126 }, { "epoch": 0.05754417761667422, "grad_norm": 0.47612880804918073, "learning_rate": 9.202898550724638e-06, "loss": 1.0134, "step": 127 }, { "epoch": 0.057997281377435435, "grad_norm": 0.607245110915646, "learning_rate": 9.275362318840581e-06, "loss": 1.0102, "step": 128 }, { "epoch": 0.05845038513819665, "grad_norm": 0.6205492389799672, "learning_rate": 9.347826086956523e-06, "loss": 1.0101, "step": 129 }, { "epoch": 0.05890348889895786, "grad_norm": 0.4612121219284199, "learning_rate": 9.420289855072464e-06, "loss": 0.9887, "step": 130 }, { "epoch": 0.05935659265971908, "grad_norm": 0.48698317743207264, "learning_rate": 9.492753623188407e-06, "loss": 1.0005, "step": 131 }, { "epoch": 0.05980969642048029, "grad_norm": 0.4254697937719653, "learning_rate": 9.565217391304349e-06, "loss": 1.0053, "step": 132 }, { "epoch": 0.060262800181241506, "grad_norm": 0.5422186743635874, "learning_rate": 9.63768115942029e-06, "loss": 1.0202, "step": 133 }, { "epoch": 0.06071590394200272, "grad_norm": 0.3678611151696531, "learning_rate": 9.710144927536233e-06, "loss": 1.0307, "step": 134 }, { "epoch": 0.061169007702763935, "grad_norm": 0.509628691476903, "learning_rate": 9.782608695652175e-06, "loss": 1.0185, "step": 135 }, { "epoch": 0.06162211146352515, "grad_norm": 0.49746291019597216, "learning_rate": 9.855072463768118e-06, "loss": 1.0131, "step": 136 }, { "epoch": 0.06207521522428636, "grad_norm": 0.4536071294798016, "learning_rate": 9.927536231884058e-06, "loss": 0.9878, "step": 137 }, { "epoch": 0.06252831898504757, "grad_norm": 0.5178670223965404, "learning_rate": 1e-05, "loss": 1.006, "step": 138 }, { "epoch": 0.06298142274580878, "grad_norm": 0.4722105506655772, "learning_rate": 1.0072463768115944e-05, "loss": 1.015, "step": 139 }, { "epoch": 0.06343452650657, "grad_norm": 0.4945728963984982, "learning_rate": 1.0144927536231885e-05, "loss": 1.0303, "step": 140 }, { "epoch": 0.06388763026733121, "grad_norm": 0.47583102438588937, "learning_rate": 1.0217391304347829e-05, "loss": 1.0109, "step": 141 }, { "epoch": 0.06434073402809243, "grad_norm": 0.6084478422028222, "learning_rate": 1.0289855072463768e-05, "loss": 1.006, "step": 142 }, { "epoch": 0.06479383778885364, "grad_norm": 0.6248888756048953, "learning_rate": 1.036231884057971e-05, "loss": 1.0145, "step": 143 }, { "epoch": 0.06524694154961486, "grad_norm": 0.5077896709362946, "learning_rate": 1.0434782608695653e-05, "loss": 1.0319, "step": 144 }, { "epoch": 0.06570004531037607, "grad_norm": 0.4634842240830667, "learning_rate": 1.0507246376811594e-05, "loss": 0.9833, "step": 145 }, { "epoch": 0.06615314907113728, "grad_norm": 0.5396173233387138, "learning_rate": 1.0579710144927538e-05, "loss": 0.9982, "step": 146 }, { "epoch": 0.0666062528318985, "grad_norm": 0.607614744068741, "learning_rate": 1.0652173913043479e-05, "loss": 1.0166, "step": 147 }, { "epoch": 0.06705935659265971, "grad_norm": 0.71843058554036, "learning_rate": 1.0724637681159422e-05, "loss": 1.0116, "step": 148 }, { "epoch": 0.06751246035342093, "grad_norm": 1.0515056726281744, "learning_rate": 1.0797101449275362e-05, "loss": 1.0084, "step": 149 }, { "epoch": 0.06796556411418214, "grad_norm": 1.2091850176459475, "learning_rate": 1.0869565217391305e-05, "loss": 0.9851, "step": 150 }, { "epoch": 0.06841866787494336, "grad_norm": 0.6630308603410107, "learning_rate": 1.0942028985507247e-05, "loss": 1.009, "step": 151 }, { "epoch": 0.06887177163570457, "grad_norm": 0.628505197676281, "learning_rate": 1.101449275362319e-05, "loss": 1.0043, "step": 152 }, { "epoch": 0.06932487539646578, "grad_norm": 0.9569955044379507, "learning_rate": 1.1086956521739131e-05, "loss": 1.0096, "step": 153 }, { "epoch": 0.069777979157227, "grad_norm": 1.1871578037508035, "learning_rate": 1.1159420289855074e-05, "loss": 1.0113, "step": 154 }, { "epoch": 0.07023108291798821, "grad_norm": 0.7183282275246085, "learning_rate": 1.1231884057971016e-05, "loss": 0.997, "step": 155 }, { "epoch": 0.07068418667874943, "grad_norm": 0.624025993069793, "learning_rate": 1.1304347826086957e-05, "loss": 0.9916, "step": 156 }, { "epoch": 0.07113729043951064, "grad_norm": 0.6947602439538473, "learning_rate": 1.1376811594202899e-05, "loss": 0.9854, "step": 157 }, { "epoch": 0.07159039420027186, "grad_norm": 0.8907145071259218, "learning_rate": 1.1449275362318842e-05, "loss": 1.0124, "step": 158 }, { "epoch": 0.07204349796103307, "grad_norm": 0.9835195584632055, "learning_rate": 1.1521739130434783e-05, "loss": 0.9954, "step": 159 }, { "epoch": 0.07249660172179428, "grad_norm": 1.0116180319086214, "learning_rate": 1.1594202898550726e-05, "loss": 0.9986, "step": 160 }, { "epoch": 0.0729497054825555, "grad_norm": 0.9386907762030736, "learning_rate": 1.1666666666666668e-05, "loss": 1.0021, "step": 161 }, { "epoch": 0.07340280924331671, "grad_norm": 0.6978238233859871, "learning_rate": 1.1739130434782611e-05, "loss": 0.9974, "step": 162 }, { "epoch": 0.07385591300407793, "grad_norm": 0.7174248954005895, "learning_rate": 1.181159420289855e-05, "loss": 1.0042, "step": 163 }, { "epoch": 0.07430901676483914, "grad_norm": 0.938917004370938, "learning_rate": 1.1884057971014494e-05, "loss": 0.9905, "step": 164 }, { "epoch": 0.07476212052560036, "grad_norm": 0.9787066543500267, "learning_rate": 1.1956521739130435e-05, "loss": 0.9909, "step": 165 }, { "epoch": 0.07521522428636157, "grad_norm": 0.7735407647982535, "learning_rate": 1.2028985507246379e-05, "loss": 1.0215, "step": 166 }, { "epoch": 0.07566832804712278, "grad_norm": 0.5049582764996957, "learning_rate": 1.210144927536232e-05, "loss": 0.9995, "step": 167 }, { "epoch": 0.076121431807884, "grad_norm": 0.6107311118374608, "learning_rate": 1.2173913043478263e-05, "loss": 0.9823, "step": 168 }, { "epoch": 0.07657453556864521, "grad_norm": 0.5697421121857996, "learning_rate": 1.2246376811594205e-05, "loss": 0.9859, "step": 169 }, { "epoch": 0.07702763932940643, "grad_norm": 0.5742666171300774, "learning_rate": 1.2318840579710144e-05, "loss": 0.9774, "step": 170 }, { "epoch": 0.07748074309016764, "grad_norm": 0.822177271874778, "learning_rate": 1.2391304347826088e-05, "loss": 0.9766, "step": 171 }, { "epoch": 0.07793384685092886, "grad_norm": 1.0168902969793943, "learning_rate": 1.2463768115942029e-05, "loss": 1.0098, "step": 172 }, { "epoch": 0.07838695061169007, "grad_norm": 0.8453511031688763, "learning_rate": 1.2536231884057972e-05, "loss": 1.0074, "step": 173 }, { "epoch": 0.07884005437245128, "grad_norm": 0.569480171679202, "learning_rate": 1.2608695652173915e-05, "loss": 0.9925, "step": 174 }, { "epoch": 0.0792931581332125, "grad_norm": 0.6255582098895063, "learning_rate": 1.2681159420289857e-05, "loss": 0.9939, "step": 175 }, { "epoch": 0.07974626189397371, "grad_norm": 0.9065721815582469, "learning_rate": 1.2753623188405797e-05, "loss": 0.972, "step": 176 }, { "epoch": 0.08019936565473493, "grad_norm": 1.0870635291585562, "learning_rate": 1.282608695652174e-05, "loss": 0.9918, "step": 177 }, { "epoch": 0.08065246941549614, "grad_norm": 0.7584198993012229, "learning_rate": 1.2898550724637681e-05, "loss": 0.9891, "step": 178 }, { "epoch": 0.08110557317625736, "grad_norm": 0.5016038890610166, "learning_rate": 1.2971014492753624e-05, "loss": 0.988, "step": 179 }, { "epoch": 0.08155867693701857, "grad_norm": 0.6997178039550256, "learning_rate": 1.3043478260869566e-05, "loss": 0.9732, "step": 180 }, { "epoch": 0.08201178069777978, "grad_norm": 0.9575883305806213, "learning_rate": 1.3115942028985509e-05, "loss": 0.9995, "step": 181 }, { "epoch": 0.082464884458541, "grad_norm": 1.098891399276748, "learning_rate": 1.318840579710145e-05, "loss": 0.9854, "step": 182 }, { "epoch": 0.08291798821930221, "grad_norm": 0.7940362257207068, "learning_rate": 1.3260869565217392e-05, "loss": 0.9664, "step": 183 }, { "epoch": 0.08337109198006344, "grad_norm": 0.622575693952322, "learning_rate": 1.3333333333333333e-05, "loss": 1.0086, "step": 184 }, { "epoch": 0.08382419574082466, "grad_norm": 0.596092314218331, "learning_rate": 1.3405797101449276e-05, "loss": 0.9849, "step": 185 }, { "epoch": 0.08427729950158587, "grad_norm": 0.5745761171632152, "learning_rate": 1.3478260869565218e-05, "loss": 0.9977, "step": 186 }, { "epoch": 0.08473040326234708, "grad_norm": 0.7657284641472454, "learning_rate": 1.3550724637681161e-05, "loss": 0.9904, "step": 187 }, { "epoch": 0.0851835070231083, "grad_norm": 0.9494961164827017, "learning_rate": 1.3623188405797103e-05, "loss": 1.0028, "step": 188 }, { "epoch": 0.08563661078386951, "grad_norm": 0.849632487526782, "learning_rate": 1.3695652173913046e-05, "loss": 1.0101, "step": 189 }, { "epoch": 0.08608971454463073, "grad_norm": 0.7661671333173663, "learning_rate": 1.3768115942028985e-05, "loss": 0.995, "step": 190 }, { "epoch": 0.08654281830539194, "grad_norm": 1.0732029280055653, "learning_rate": 1.3840579710144929e-05, "loss": 0.998, "step": 191 }, { "epoch": 0.08699592206615316, "grad_norm": 1.107306494806081, "learning_rate": 1.391304347826087e-05, "loss": 1.0002, "step": 192 }, { "epoch": 0.08744902582691437, "grad_norm": 0.7325083136138134, "learning_rate": 1.3985507246376813e-05, "loss": 0.9714, "step": 193 }, { "epoch": 0.08790212958767558, "grad_norm": 0.5870440462089405, "learning_rate": 1.4057971014492755e-05, "loss": 0.9724, "step": 194 }, { "epoch": 0.0883552333484368, "grad_norm": 0.7616440307262718, "learning_rate": 1.4130434782608698e-05, "loss": 0.9855, "step": 195 }, { "epoch": 0.08880833710919801, "grad_norm": 0.7531165764878514, "learning_rate": 1.420289855072464e-05, "loss": 0.9733, "step": 196 }, { "epoch": 0.08926144086995923, "grad_norm": 0.7603764621047993, "learning_rate": 1.427536231884058e-05, "loss": 0.999, "step": 197 }, { "epoch": 0.08971454463072044, "grad_norm": 1.0363060598424276, "learning_rate": 1.4347826086956522e-05, "loss": 0.9787, "step": 198 }, { "epoch": 0.09016764839148166, "grad_norm": 1.4217107204291115, "learning_rate": 1.4420289855072465e-05, "loss": 0.9808, "step": 199 }, { "epoch": 0.09062075215224287, "grad_norm": 0.4516933511047247, "learning_rate": 1.4492753623188407e-05, "loss": 0.9952, "step": 200 }, { "epoch": 0.09107385591300408, "grad_norm": 0.9051683340068594, "learning_rate": 1.456521739130435e-05, "loss": 0.9919, "step": 201 }, { "epoch": 0.0915269596737653, "grad_norm": 1.691495027231651, "learning_rate": 1.4637681159420291e-05, "loss": 0.9817, "step": 202 }, { "epoch": 0.09198006343452651, "grad_norm": 0.4515045544135162, "learning_rate": 1.4710144927536235e-05, "loss": 0.9888, "step": 203 }, { "epoch": 0.09243316719528773, "grad_norm": 1.7474747357474603, "learning_rate": 1.4782608695652174e-05, "loss": 0.9773, "step": 204 }, { "epoch": 0.09288627095604894, "grad_norm": 0.6055946207648063, "learning_rate": 1.4855072463768116e-05, "loss": 0.961, "step": 205 }, { "epoch": 0.09333937471681016, "grad_norm": 1.3767568207252592, "learning_rate": 1.4927536231884059e-05, "loss": 0.9738, "step": 206 }, { "epoch": 0.09379247847757137, "grad_norm": 1.0369979953955883, "learning_rate": 1.5000000000000002e-05, "loss": 0.9813, "step": 207 }, { "epoch": 0.09424558223833258, "grad_norm": 1.2656072173461081, "learning_rate": 1.5072463768115944e-05, "loss": 0.9981, "step": 208 }, { "epoch": 0.0946986859990938, "grad_norm": 0.641351204135351, "learning_rate": 1.5144927536231887e-05, "loss": 0.9805, "step": 209 }, { "epoch": 0.09515178975985501, "grad_norm": 1.0643631530013913, "learning_rate": 1.5217391304347828e-05, "loss": 0.988, "step": 210 }, { "epoch": 0.09560489352061623, "grad_norm": 0.681878676875636, "learning_rate": 1.5289855072463768e-05, "loss": 0.9535, "step": 211 }, { "epoch": 0.09605799728137744, "grad_norm": 0.9849427591910358, "learning_rate": 1.536231884057971e-05, "loss": 0.9758, "step": 212 }, { "epoch": 0.09651110104213866, "grad_norm": 1.0029830266542648, "learning_rate": 1.5434782608695654e-05, "loss": 0.997, "step": 213 }, { "epoch": 0.09696420480289987, "grad_norm": 0.8670683430614528, "learning_rate": 1.5507246376811594e-05, "loss": 0.977, "step": 214 }, { "epoch": 0.09741730856366108, "grad_norm": 1.0119684347525018, "learning_rate": 1.5579710144927537e-05, "loss": 0.9512, "step": 215 }, { "epoch": 0.0978704123244223, "grad_norm": 1.1410754505531968, "learning_rate": 1.565217391304348e-05, "loss": 0.9851, "step": 216 }, { "epoch": 0.09832351608518351, "grad_norm": 1.2224607074461233, "learning_rate": 1.5724637681159423e-05, "loss": 0.9776, "step": 217 }, { "epoch": 0.09877661984594473, "grad_norm": 0.8402048418960246, "learning_rate": 1.5797101449275363e-05, "loss": 0.9749, "step": 218 }, { "epoch": 0.09922972360670594, "grad_norm": 0.7196372110827022, "learning_rate": 1.5869565217391306e-05, "loss": 0.9801, "step": 219 }, { "epoch": 0.09968282736746716, "grad_norm": 1.1283148052401974, "learning_rate": 1.5942028985507246e-05, "loss": 0.9946, "step": 220 }, { "epoch": 0.10013593112822837, "grad_norm": 1.302482878640589, "learning_rate": 1.601449275362319e-05, "loss": 0.9828, "step": 221 }, { "epoch": 0.10058903488898958, "grad_norm": 0.6691268006772679, "learning_rate": 1.6086956521739132e-05, "loss": 0.9731, "step": 222 }, { "epoch": 0.1010421386497508, "grad_norm": 0.9772256087376912, "learning_rate": 1.6159420289855076e-05, "loss": 0.9744, "step": 223 }, { "epoch": 0.10149524241051201, "grad_norm": 1.312677993484747, "learning_rate": 1.6231884057971015e-05, "loss": 0.9702, "step": 224 }, { "epoch": 0.10194834617127323, "grad_norm": 0.9304852193374167, "learning_rate": 1.630434782608696e-05, "loss": 0.9884, "step": 225 }, { "epoch": 0.10240144993203444, "grad_norm": 1.095577499387557, "learning_rate": 1.6376811594202898e-05, "loss": 0.972, "step": 226 }, { "epoch": 0.10285455369279566, "grad_norm": 0.8362591350579461, "learning_rate": 1.644927536231884e-05, "loss": 0.9641, "step": 227 }, { "epoch": 0.10330765745355687, "grad_norm": 0.8341805082977936, "learning_rate": 1.6521739130434785e-05, "loss": 0.9886, "step": 228 }, { "epoch": 0.10376076121431808, "grad_norm": 0.8924121102882958, "learning_rate": 1.6594202898550728e-05, "loss": 0.981, "step": 229 }, { "epoch": 0.1042138649750793, "grad_norm": 1.233172684190328, "learning_rate": 1.6666666666666667e-05, "loss": 0.9878, "step": 230 }, { "epoch": 0.10466696873584051, "grad_norm": 0.8436416903507178, "learning_rate": 1.673913043478261e-05, "loss": 0.9878, "step": 231 }, { "epoch": 0.10512007249660173, "grad_norm": 0.8859963488452689, "learning_rate": 1.681159420289855e-05, "loss": 0.9844, "step": 232 }, { "epoch": 0.10557317625736294, "grad_norm": 0.7428460730494619, "learning_rate": 1.6884057971014494e-05, "loss": 0.9693, "step": 233 }, { "epoch": 0.10602628001812416, "grad_norm": 0.8243427022321984, "learning_rate": 1.6956521739130437e-05, "loss": 0.9815, "step": 234 }, { "epoch": 0.10647938377888537, "grad_norm": 0.8419257445019829, "learning_rate": 1.702898550724638e-05, "loss": 0.9853, "step": 235 }, { "epoch": 0.10693248753964658, "grad_norm": 1.020466366593606, "learning_rate": 1.710144927536232e-05, "loss": 0.9688, "step": 236 }, { "epoch": 0.1073855913004078, "grad_norm": 1.179081989925465, "learning_rate": 1.7173913043478263e-05, "loss": 0.9813, "step": 237 }, { "epoch": 0.10783869506116901, "grad_norm": 0.720169300869158, "learning_rate": 1.7246376811594206e-05, "loss": 0.9978, "step": 238 }, { "epoch": 0.10829179882193023, "grad_norm": 0.8630607874919379, "learning_rate": 1.7318840579710146e-05, "loss": 0.983, "step": 239 }, { "epoch": 0.10874490258269144, "grad_norm": 1.3286379198678078, "learning_rate": 1.739130434782609e-05, "loss": 0.9749, "step": 240 }, { "epoch": 0.10919800634345266, "grad_norm": 0.6146122873274804, "learning_rate": 1.746376811594203e-05, "loss": 0.978, "step": 241 }, { "epoch": 0.10965111010421387, "grad_norm": 1.0139400716277533, "learning_rate": 1.7536231884057972e-05, "loss": 0.9645, "step": 242 }, { "epoch": 0.11010421386497508, "grad_norm": 1.0811299878760665, "learning_rate": 1.7608695652173915e-05, "loss": 0.9635, "step": 243 }, { "epoch": 0.1105573176257363, "grad_norm": 0.8412697957004113, "learning_rate": 1.7681159420289858e-05, "loss": 0.9672, "step": 244 }, { "epoch": 0.11101042138649751, "grad_norm": 0.774380737179075, "learning_rate": 1.7753623188405798e-05, "loss": 0.9862, "step": 245 }, { "epoch": 0.11146352514725873, "grad_norm": 0.8664210583237614, "learning_rate": 1.782608695652174e-05, "loss": 0.9685, "step": 246 }, { "epoch": 0.11191662890801994, "grad_norm": 1.232233527750389, "learning_rate": 1.789855072463768e-05, "loss": 0.9779, "step": 247 }, { "epoch": 0.11236973266878116, "grad_norm": 0.7761899423285926, "learning_rate": 1.7971014492753624e-05, "loss": 0.9823, "step": 248 }, { "epoch": 0.11282283642954237, "grad_norm": 0.6927820298962599, "learning_rate": 1.8043478260869567e-05, "loss": 0.9871, "step": 249 }, { "epoch": 0.11327594019030358, "grad_norm": 0.8210402367090289, "learning_rate": 1.811594202898551e-05, "loss": 0.9907, "step": 250 }, { "epoch": 0.1137290439510648, "grad_norm": 1.188464021328183, "learning_rate": 1.818840579710145e-05, "loss": 0.9499, "step": 251 }, { "epoch": 0.11418214771182601, "grad_norm": 0.8834977893318479, "learning_rate": 1.8260869565217393e-05, "loss": 0.9792, "step": 252 }, { "epoch": 0.11463525147258723, "grad_norm": 1.0168791049822445, "learning_rate": 1.8333333333333333e-05, "loss": 0.9845, "step": 253 }, { "epoch": 0.11508835523334844, "grad_norm": 1.1252827368963574, "learning_rate": 1.8405797101449276e-05, "loss": 0.9874, "step": 254 }, { "epoch": 0.11554145899410966, "grad_norm": 1.0280993482926937, "learning_rate": 1.847826086956522e-05, "loss": 0.9547, "step": 255 }, { "epoch": 0.11599456275487087, "grad_norm": 1.2857909629973028, "learning_rate": 1.8550724637681162e-05, "loss": 0.971, "step": 256 }, { "epoch": 0.11644766651563208, "grad_norm": 0.7819860420551438, "learning_rate": 1.8623188405797102e-05, "loss": 0.9887, "step": 257 }, { "epoch": 0.1169007702763933, "grad_norm": 0.774104602410304, "learning_rate": 1.8695652173913045e-05, "loss": 0.9609, "step": 258 }, { "epoch": 0.11735387403715451, "grad_norm": 0.7003992307446028, "learning_rate": 1.8768115942028985e-05, "loss": 0.9878, "step": 259 }, { "epoch": 0.11780697779791573, "grad_norm": 0.8021790332755897, "learning_rate": 1.8840579710144928e-05, "loss": 0.9797, "step": 260 }, { "epoch": 0.11826008155867694, "grad_norm": 1.06061282094802, "learning_rate": 1.891304347826087e-05, "loss": 0.9812, "step": 261 }, { "epoch": 0.11871318531943816, "grad_norm": 1.1382847529537776, "learning_rate": 1.8985507246376814e-05, "loss": 0.9873, "step": 262 }, { "epoch": 0.11916628908019937, "grad_norm": 1.0064320853003477, "learning_rate": 1.9057971014492754e-05, "loss": 0.9756, "step": 263 }, { "epoch": 0.11961939284096058, "grad_norm": 1.339616134881972, "learning_rate": 1.9130434782608697e-05, "loss": 0.9928, "step": 264 }, { "epoch": 0.1200724966017218, "grad_norm": 0.8498668377096184, "learning_rate": 1.920289855072464e-05, "loss": 0.9873, "step": 265 }, { "epoch": 0.12052560036248301, "grad_norm": 0.9288838049695854, "learning_rate": 1.927536231884058e-05, "loss": 0.9763, "step": 266 }, { "epoch": 0.12097870412324423, "grad_norm": 1.0841929267934154, "learning_rate": 1.9347826086956523e-05, "loss": 0.9803, "step": 267 }, { "epoch": 0.12143180788400544, "grad_norm": 1.1613621565943448, "learning_rate": 1.9420289855072467e-05, "loss": 0.9598, "step": 268 }, { "epoch": 0.12188491164476666, "grad_norm": 1.2132922639961328, "learning_rate": 1.9492753623188406e-05, "loss": 0.9897, "step": 269 }, { "epoch": 0.12233801540552787, "grad_norm": 1.0907058739551005, "learning_rate": 1.956521739130435e-05, "loss": 0.9702, "step": 270 }, { "epoch": 0.12279111916628908, "grad_norm": 1.0684277854981317, "learning_rate": 1.9637681159420293e-05, "loss": 0.9712, "step": 271 }, { "epoch": 0.1232442229270503, "grad_norm": 0.657754534804156, "learning_rate": 1.9710144927536236e-05, "loss": 0.9853, "step": 272 }, { "epoch": 0.12369732668781151, "grad_norm": 0.5748987766230957, "learning_rate": 1.9782608695652176e-05, "loss": 0.9812, "step": 273 }, { "epoch": 0.12415043044857273, "grad_norm": 0.5849201651490491, "learning_rate": 1.9855072463768115e-05, "loss": 0.9519, "step": 274 }, { "epoch": 0.12460353420933394, "grad_norm": 0.5346946502182941, "learning_rate": 1.992753623188406e-05, "loss": 0.9553, "step": 275 }, { "epoch": 0.12505663797009514, "grad_norm": 0.7001143077253504, "learning_rate": 2e-05, "loss": 0.9927, "step": 276 }, { "epoch": 0.12550974173085636, "grad_norm": 1.0338580720920587, "learning_rate": 2.0072463768115945e-05, "loss": 0.9672, "step": 277 }, { "epoch": 0.12596284549161757, "grad_norm": 1.6898903387723694, "learning_rate": 2.0144927536231888e-05, "loss": 0.979, "step": 278 }, { "epoch": 0.12641594925237878, "grad_norm": 0.6218489326083152, "learning_rate": 2.0217391304347828e-05, "loss": 0.9797, "step": 279 }, { "epoch": 0.12686905301314, "grad_norm": 1.4059674957473993, "learning_rate": 2.028985507246377e-05, "loss": 0.977, "step": 280 }, { "epoch": 0.1273221567739012, "grad_norm": 1.343025111225242, "learning_rate": 2.0362318840579714e-05, "loss": 0.9701, "step": 281 }, { "epoch": 0.12777526053466243, "grad_norm": 1.0263843007692721, "learning_rate": 2.0434782608695657e-05, "loss": 0.972, "step": 282 }, { "epoch": 0.12822836429542364, "grad_norm": 1.167956384471407, "learning_rate": 2.0507246376811594e-05, "loss": 0.9858, "step": 283 }, { "epoch": 0.12868146805618486, "grad_norm": 0.9454861169265084, "learning_rate": 2.0579710144927537e-05, "loss": 0.9717, "step": 284 }, { "epoch": 0.12913457181694607, "grad_norm": 1.2027521738405051, "learning_rate": 2.065217391304348e-05, "loss": 0.9677, "step": 285 }, { "epoch": 0.12958767557770728, "grad_norm": 1.1263993301719233, "learning_rate": 2.072463768115942e-05, "loss": 0.9611, "step": 286 }, { "epoch": 0.1300407793384685, "grad_norm": 1.5800953654486012, "learning_rate": 2.0797101449275363e-05, "loss": 0.9793, "step": 287 }, { "epoch": 0.1304938830992297, "grad_norm": 0.6923334927513635, "learning_rate": 2.0869565217391306e-05, "loss": 0.98, "step": 288 }, { "epoch": 0.13094698685999093, "grad_norm": 1.9143953904761424, "learning_rate": 2.094202898550725e-05, "loss": 0.9627, "step": 289 }, { "epoch": 0.13140009062075214, "grad_norm": 0.8676453921459047, "learning_rate": 2.101449275362319e-05, "loss": 0.9514, "step": 290 }, { "epoch": 0.13185319438151336, "grad_norm": 1.9222131155595794, "learning_rate": 2.1086956521739132e-05, "loss": 0.978, "step": 291 }, { "epoch": 0.13230629814227457, "grad_norm": 1.5722948008500515, "learning_rate": 2.1159420289855075e-05, "loss": 0.9864, "step": 292 }, { "epoch": 0.13275940190303578, "grad_norm": 1.577862236784929, "learning_rate": 2.1231884057971018e-05, "loss": 0.9929, "step": 293 }, { "epoch": 0.133212505663797, "grad_norm": 1.465884112462037, "learning_rate": 2.1304347826086958e-05, "loss": 0.97, "step": 294 }, { "epoch": 0.1336656094245582, "grad_norm": 1.3774579902666506, "learning_rate": 2.13768115942029e-05, "loss": 0.9735, "step": 295 }, { "epoch": 0.13411871318531943, "grad_norm": 1.209680183810372, "learning_rate": 2.1449275362318844e-05, "loss": 0.9906, "step": 296 }, { "epoch": 0.13457181694608064, "grad_norm": 1.31964172741505, "learning_rate": 2.152173913043478e-05, "loss": 0.9852, "step": 297 }, { "epoch": 0.13502492070684186, "grad_norm": 1.2876786287783664, "learning_rate": 2.1594202898550724e-05, "loss": 0.9626, "step": 298 }, { "epoch": 0.13547802446760307, "grad_norm": 1.180357708864339, "learning_rate": 2.1666666666666667e-05, "loss": 0.9729, "step": 299 }, { "epoch": 0.13593112822836428, "grad_norm": 1.1219372475373657, "learning_rate": 2.173913043478261e-05, "loss": 0.9756, "step": 300 }, { "epoch": 0.1363842319891255, "grad_norm": 0.9751022646280695, "learning_rate": 2.1811594202898553e-05, "loss": 0.9756, "step": 301 }, { "epoch": 0.1368373357498867, "grad_norm": 1.174181314240781, "learning_rate": 2.1884057971014493e-05, "loss": 0.9713, "step": 302 }, { "epoch": 0.13729043951064793, "grad_norm": 0.8244316699239077, "learning_rate": 2.1956521739130436e-05, "loss": 0.9773, "step": 303 }, { "epoch": 0.13774354327140914, "grad_norm": 0.9202108539708974, "learning_rate": 2.202898550724638e-05, "loss": 0.9832, "step": 304 }, { "epoch": 0.13819664703217036, "grad_norm": 1.4051606675482864, "learning_rate": 2.2101449275362323e-05, "loss": 0.9564, "step": 305 }, { "epoch": 0.13864975079293157, "grad_norm": 1.055160806013612, "learning_rate": 2.2173913043478262e-05, "loss": 0.9709, "step": 306 }, { "epoch": 0.13910285455369278, "grad_norm": 1.206811230743539, "learning_rate": 2.2246376811594205e-05, "loss": 0.9541, "step": 307 }, { "epoch": 0.139555958314454, "grad_norm": 1.0166442598057905, "learning_rate": 2.231884057971015e-05, "loss": 0.9555, "step": 308 }, { "epoch": 0.1400090620752152, "grad_norm": 1.0885214295942593, "learning_rate": 2.2391304347826092e-05, "loss": 0.9697, "step": 309 }, { "epoch": 0.14046216583597643, "grad_norm": 1.2001753189859246, "learning_rate": 2.246376811594203e-05, "loss": 0.9685, "step": 310 }, { "epoch": 0.14091526959673764, "grad_norm": 1.2979008240589889, "learning_rate": 2.253623188405797e-05, "loss": 0.9685, "step": 311 }, { "epoch": 0.14136837335749886, "grad_norm": 1.1385236652747799, "learning_rate": 2.2608695652173914e-05, "loss": 0.9564, "step": 312 }, { "epoch": 0.14182147711826007, "grad_norm": 0.8150610506201985, "learning_rate": 2.2681159420289854e-05, "loss": 0.9818, "step": 313 }, { "epoch": 0.14227458087902128, "grad_norm": 0.7686360411685415, "learning_rate": 2.2753623188405797e-05, "loss": 0.9743, "step": 314 }, { "epoch": 0.1427276846397825, "grad_norm": 0.9088738532216054, "learning_rate": 2.282608695652174e-05, "loss": 0.9517, "step": 315 }, { "epoch": 0.1431807884005437, "grad_norm": 1.4264107152333791, "learning_rate": 2.2898550724637684e-05, "loss": 0.9829, "step": 316 }, { "epoch": 0.14363389216130493, "grad_norm": 1.057676592665108, "learning_rate": 2.2971014492753623e-05, "loss": 0.9694, "step": 317 }, { "epoch": 0.14408699592206614, "grad_norm": 1.0563324922996575, "learning_rate": 2.3043478260869567e-05, "loss": 0.9836, "step": 318 }, { "epoch": 0.14454009968282736, "grad_norm": 1.4779430218508403, "learning_rate": 2.311594202898551e-05, "loss": 0.9564, "step": 319 }, { "epoch": 0.14499320344358857, "grad_norm": 0.7846545186263838, "learning_rate": 2.3188405797101453e-05, "loss": 0.9205, "step": 320 }, { "epoch": 0.14544630720434978, "grad_norm": 1.1432004335408863, "learning_rate": 2.3260869565217396e-05, "loss": 0.9742, "step": 321 }, { "epoch": 0.145899410965111, "grad_norm": 1.3429198786491294, "learning_rate": 2.3333333333333336e-05, "loss": 0.956, "step": 322 }, { "epoch": 0.1463525147258722, "grad_norm": 1.0613929429689732, "learning_rate": 2.340579710144928e-05, "loss": 0.9874, "step": 323 }, { "epoch": 0.14680561848663343, "grad_norm": 1.6610595697231714, "learning_rate": 2.3478260869565222e-05, "loss": 0.9573, "step": 324 }, { "epoch": 0.14725872224739464, "grad_norm": 0.7648858273938999, "learning_rate": 2.355072463768116e-05, "loss": 0.9731, "step": 325 }, { "epoch": 0.14771182600815586, "grad_norm": 1.4682393768634026, "learning_rate": 2.36231884057971e-05, "loss": 0.976, "step": 326 }, { "epoch": 0.14816492976891707, "grad_norm": 1.232355232647433, "learning_rate": 2.3695652173913045e-05, "loss": 0.9768, "step": 327 }, { "epoch": 0.14861803352967828, "grad_norm": 1.38890885600647, "learning_rate": 2.3768115942028988e-05, "loss": 0.9762, "step": 328 }, { "epoch": 0.1490711372904395, "grad_norm": 0.9135890221209729, "learning_rate": 2.3840579710144928e-05, "loss": 0.9734, "step": 329 }, { "epoch": 0.1495242410512007, "grad_norm": 1.0775952756853593, "learning_rate": 2.391304347826087e-05, "loss": 0.9543, "step": 330 }, { "epoch": 0.14997734481196193, "grad_norm": 0.7254767438980483, "learning_rate": 2.3985507246376814e-05, "loss": 0.9641, "step": 331 }, { "epoch": 0.15043044857272314, "grad_norm": 1.3286290743030535, "learning_rate": 2.4057971014492757e-05, "loss": 0.9567, "step": 332 }, { "epoch": 0.15088355233348436, "grad_norm": 0.9381962761444239, "learning_rate": 2.4130434782608697e-05, "loss": 0.9687, "step": 333 }, { "epoch": 0.15133665609424557, "grad_norm": 1.0610640222022414, "learning_rate": 2.420289855072464e-05, "loss": 0.9448, "step": 334 }, { "epoch": 0.15178975985500678, "grad_norm": 1.0965873832222532, "learning_rate": 2.4275362318840583e-05, "loss": 0.9573, "step": 335 }, { "epoch": 0.152242863615768, "grad_norm": 1.4417129036282583, "learning_rate": 2.4347826086956526e-05, "loss": 0.9718, "step": 336 }, { "epoch": 0.1526959673765292, "grad_norm": 1.1583501078452507, "learning_rate": 2.4420289855072466e-05, "loss": 0.9678, "step": 337 }, { "epoch": 0.15314907113729043, "grad_norm": 0.984548615237672, "learning_rate": 2.449275362318841e-05, "loss": 0.9633, "step": 338 }, { "epoch": 0.15360217489805164, "grad_norm": 1.349027221515409, "learning_rate": 2.456521739130435e-05, "loss": 0.9791, "step": 339 }, { "epoch": 0.15405527865881286, "grad_norm": 1.0161831781350714, "learning_rate": 2.463768115942029e-05, "loss": 0.9705, "step": 340 }, { "epoch": 0.15450838241957407, "grad_norm": 0.9675888591623732, "learning_rate": 2.4710144927536232e-05, "loss": 0.9388, "step": 341 }, { "epoch": 0.15496148618033528, "grad_norm": 1.4222384018304886, "learning_rate": 2.4782608695652175e-05, "loss": 0.9532, "step": 342 }, { "epoch": 0.1554145899410965, "grad_norm": 0.922560013759614, "learning_rate": 2.4855072463768118e-05, "loss": 0.9677, "step": 343 }, { "epoch": 0.1558676937018577, "grad_norm": 1.3089102995406219, "learning_rate": 2.4927536231884058e-05, "loss": 0.9698, "step": 344 }, { "epoch": 0.15632079746261893, "grad_norm": 1.3372715279406444, "learning_rate": 2.5e-05, "loss": 0.9648, "step": 345 }, { "epoch": 0.15677390122338014, "grad_norm": 0.585635929975912, "learning_rate": 2.5072463768115944e-05, "loss": 0.9646, "step": 346 }, { "epoch": 0.15722700498414136, "grad_norm": 0.6543804427377383, "learning_rate": 2.5144927536231888e-05, "loss": 0.9688, "step": 347 }, { "epoch": 0.15768010874490257, "grad_norm": 0.792528978398417, "learning_rate": 2.521739130434783e-05, "loss": 0.9489, "step": 348 }, { "epoch": 0.15813321250566378, "grad_norm": 1.5539031502888876, "learning_rate": 2.528985507246377e-05, "loss": 0.9752, "step": 349 }, { "epoch": 0.158586316266425, "grad_norm": 1.2961060922246186, "learning_rate": 2.5362318840579714e-05, "loss": 0.9669, "step": 350 }, { "epoch": 0.1590394200271862, "grad_norm": 0.9695617124365082, "learning_rate": 2.5434782608695657e-05, "loss": 0.9567, "step": 351 }, { "epoch": 0.15949252378794743, "grad_norm": 1.020161804298758, "learning_rate": 2.5507246376811593e-05, "loss": 0.9481, "step": 352 }, { "epoch": 0.15994562754870864, "grad_norm": 1.2556362982468865, "learning_rate": 2.5579710144927536e-05, "loss": 0.974, "step": 353 }, { "epoch": 0.16039873130946986, "grad_norm": 1.072170478728291, "learning_rate": 2.565217391304348e-05, "loss": 0.9632, "step": 354 }, { "epoch": 0.16085183507023107, "grad_norm": 1.1194030714678112, "learning_rate": 2.5724637681159423e-05, "loss": 0.9621, "step": 355 }, { "epoch": 0.16130493883099228, "grad_norm": 1.545625561941806, "learning_rate": 2.5797101449275362e-05, "loss": 0.9624, "step": 356 }, { "epoch": 0.1617580425917535, "grad_norm": 0.9245832071666894, "learning_rate": 2.5869565217391305e-05, "loss": 0.975, "step": 357 }, { "epoch": 0.1622111463525147, "grad_norm": 1.025883413553205, "learning_rate": 2.594202898550725e-05, "loss": 0.9585, "step": 358 }, { "epoch": 0.16266425011327593, "grad_norm": 1.3599829174443274, "learning_rate": 2.6014492753623192e-05, "loss": 0.9694, "step": 359 }, { "epoch": 0.16311735387403714, "grad_norm": 1.2134922907002421, "learning_rate": 2.608695652173913e-05, "loss": 0.9693, "step": 360 }, { "epoch": 0.16357045763479836, "grad_norm": 1.1318604368194871, "learning_rate": 2.6159420289855075e-05, "loss": 0.9776, "step": 361 }, { "epoch": 0.16402356139555957, "grad_norm": 1.2122007458771364, "learning_rate": 2.6231884057971018e-05, "loss": 0.9505, "step": 362 }, { "epoch": 0.16447666515632078, "grad_norm": 1.3190413994479087, "learning_rate": 2.630434782608696e-05, "loss": 0.9487, "step": 363 }, { "epoch": 0.164929768917082, "grad_norm": 0.9319776267152948, "learning_rate": 2.63768115942029e-05, "loss": 0.9614, "step": 364 }, { "epoch": 0.1653828726778432, "grad_norm": 1.413918141029588, "learning_rate": 2.6449275362318844e-05, "loss": 0.9579, "step": 365 }, { "epoch": 0.16583597643860443, "grad_norm": 0.6442589500128013, "learning_rate": 2.6521739130434784e-05, "loss": 0.9522, "step": 366 }, { "epoch": 0.16628908019936564, "grad_norm": 1.5262668289036763, "learning_rate": 2.6594202898550723e-05, "loss": 0.9674, "step": 367 }, { "epoch": 0.16674218396012688, "grad_norm": 0.8409179205374063, "learning_rate": 2.6666666666666667e-05, "loss": 0.9548, "step": 368 }, { "epoch": 0.1671952877208881, "grad_norm": 1.2623482721892414, "learning_rate": 2.673913043478261e-05, "loss": 0.9658, "step": 369 }, { "epoch": 0.1676483914816493, "grad_norm": 0.8579463832041053, "learning_rate": 2.6811594202898553e-05, "loss": 0.9556, "step": 370 }, { "epoch": 0.16810149524241053, "grad_norm": 1.1806955567077415, "learning_rate": 2.6884057971014496e-05, "loss": 0.975, "step": 371 }, { "epoch": 0.16855459900317174, "grad_norm": 0.9412192040771948, "learning_rate": 2.6956521739130436e-05, "loss": 0.9747, "step": 372 }, { "epoch": 0.16900770276393295, "grad_norm": 1.1296590236889303, "learning_rate": 2.702898550724638e-05, "loss": 0.9566, "step": 373 }, { "epoch": 0.16946080652469417, "grad_norm": 1.0962238532800634, "learning_rate": 2.7101449275362322e-05, "loss": 0.9656, "step": 374 }, { "epoch": 0.16991391028545538, "grad_norm": 1.6044125897212802, "learning_rate": 2.7173913043478265e-05, "loss": 0.9636, "step": 375 }, { "epoch": 0.1703670140462166, "grad_norm": 1.1302713629673262, "learning_rate": 2.7246376811594205e-05, "loss": 0.9564, "step": 376 }, { "epoch": 0.1708201178069778, "grad_norm": 1.2696921488821449, "learning_rate": 2.7318840579710148e-05, "loss": 0.9404, "step": 377 }, { "epoch": 0.17127322156773903, "grad_norm": 0.8494310479869464, "learning_rate": 2.739130434782609e-05, "loss": 0.9554, "step": 378 }, { "epoch": 0.17172632532850024, "grad_norm": 1.2543294377049554, "learning_rate": 2.7463768115942034e-05, "loss": 0.9507, "step": 379 }, { "epoch": 0.17217942908926145, "grad_norm": 0.7866728426059864, "learning_rate": 2.753623188405797e-05, "loss": 0.9542, "step": 380 }, { "epoch": 0.17263253285002267, "grad_norm": 0.8128742484448102, "learning_rate": 2.7608695652173914e-05, "loss": 0.9559, "step": 381 }, { "epoch": 0.17308563661078388, "grad_norm": 1.3576665509442158, "learning_rate": 2.7681159420289857e-05, "loss": 0.9532, "step": 382 }, { "epoch": 0.1735387403715451, "grad_norm": 1.6367877150701304, "learning_rate": 2.7753623188405797e-05, "loss": 0.9617, "step": 383 }, { "epoch": 0.1739918441323063, "grad_norm": 0.6975266504755616, "learning_rate": 2.782608695652174e-05, "loss": 0.9431, "step": 384 }, { "epoch": 0.17444494789306753, "grad_norm": 1.0272403552998404, "learning_rate": 2.7898550724637683e-05, "loss": 0.9364, "step": 385 }, { "epoch": 0.17489805165382874, "grad_norm": 0.9452750903054881, "learning_rate": 2.7971014492753626e-05, "loss": 0.9525, "step": 386 }, { "epoch": 0.17535115541458995, "grad_norm": 1.4278618484113221, "learning_rate": 2.8043478260869566e-05, "loss": 0.9694, "step": 387 }, { "epoch": 0.17580425917535117, "grad_norm": 0.8939194710481505, "learning_rate": 2.811594202898551e-05, "loss": 0.9413, "step": 388 }, { "epoch": 0.17625736293611238, "grad_norm": 1.35350916704218, "learning_rate": 2.8188405797101452e-05, "loss": 0.9761, "step": 389 }, { "epoch": 0.1767104666968736, "grad_norm": 0.9723704692478318, "learning_rate": 2.8260869565217396e-05, "loss": 0.9527, "step": 390 }, { "epoch": 0.1771635704576348, "grad_norm": 1.1244635216536845, "learning_rate": 2.833333333333334e-05, "loss": 0.948, "step": 391 }, { "epoch": 0.17761667421839603, "grad_norm": 1.7792535784103782, "learning_rate": 2.840579710144928e-05, "loss": 0.97, "step": 392 }, { "epoch": 0.17806977797915724, "grad_norm": 1.1134550309048823, "learning_rate": 2.847826086956522e-05, "loss": 0.9389, "step": 393 }, { "epoch": 0.17852288173991845, "grad_norm": 1.581872594497218, "learning_rate": 2.855072463768116e-05, "loss": 0.9635, "step": 394 }, { "epoch": 0.17897598550067967, "grad_norm": 0.8950028773705339, "learning_rate": 2.86231884057971e-05, "loss": 0.9632, "step": 395 }, { "epoch": 0.17942908926144088, "grad_norm": 2.024210879898643, "learning_rate": 2.8695652173913044e-05, "loss": 0.9573, "step": 396 }, { "epoch": 0.1798821930222021, "grad_norm": 1.2181122786321519, "learning_rate": 2.8768115942028988e-05, "loss": 0.9694, "step": 397 }, { "epoch": 0.1803352967829633, "grad_norm": 2.1405563437480697, "learning_rate": 2.884057971014493e-05, "loss": 0.9617, "step": 398 }, { "epoch": 0.18078840054372453, "grad_norm": 1.893059599538506, "learning_rate": 2.891304347826087e-05, "loss": 0.9675, "step": 399 }, { "epoch": 0.18124150430448574, "grad_norm": 1.5136272861821247, "learning_rate": 2.8985507246376814e-05, "loss": 0.9671, "step": 400 }, { "epoch": 0.18169460806524695, "grad_norm": 1.5298150577744702, "learning_rate": 2.9057971014492757e-05, "loss": 0.9525, "step": 401 }, { "epoch": 0.18214771182600817, "grad_norm": 1.3175391047687497, "learning_rate": 2.91304347826087e-05, "loss": 0.9604, "step": 402 }, { "epoch": 0.18260081558676938, "grad_norm": 1.314817188764333, "learning_rate": 2.920289855072464e-05, "loss": 0.9624, "step": 403 }, { "epoch": 0.1830539193475306, "grad_norm": 1.1785461883904476, "learning_rate": 2.9275362318840583e-05, "loss": 0.9552, "step": 404 }, { "epoch": 0.1835070231082918, "grad_norm": 0.9681475332938283, "learning_rate": 2.9347826086956526e-05, "loss": 0.9516, "step": 405 }, { "epoch": 0.18396012686905303, "grad_norm": 1.4170148102372608, "learning_rate": 2.942028985507247e-05, "loss": 0.9644, "step": 406 }, { "epoch": 0.18441323062981424, "grad_norm": 1.4376548274521515, "learning_rate": 2.949275362318841e-05, "loss": 0.9621, "step": 407 }, { "epoch": 0.18486633439057545, "grad_norm": 1.1760845935143684, "learning_rate": 2.956521739130435e-05, "loss": 0.9484, "step": 408 }, { "epoch": 0.18531943815133667, "grad_norm": 1.117407726364747, "learning_rate": 2.9637681159420292e-05, "loss": 0.9252, "step": 409 }, { "epoch": 0.18577254191209788, "grad_norm": 0.7335515327979267, "learning_rate": 2.971014492753623e-05, "loss": 0.9441, "step": 410 }, { "epoch": 0.1862256456728591, "grad_norm": 0.9648502786325712, "learning_rate": 2.9782608695652175e-05, "loss": 0.966, "step": 411 }, { "epoch": 0.1866787494336203, "grad_norm": 1.360788047103844, "learning_rate": 2.9855072463768118e-05, "loss": 0.9543, "step": 412 }, { "epoch": 0.18713185319438153, "grad_norm": 1.042316687251947, "learning_rate": 2.992753623188406e-05, "loss": 0.9679, "step": 413 }, { "epoch": 0.18758495695514274, "grad_norm": 1.1935539892995726, "learning_rate": 3.0000000000000004e-05, "loss": 0.9841, "step": 414 }, { "epoch": 0.18803806071590395, "grad_norm": 1.7672567694821284, "learning_rate": 3.0072463768115944e-05, "loss": 0.9637, "step": 415 }, { "epoch": 0.18849116447666517, "grad_norm": 1.0143115621225256, "learning_rate": 3.0144927536231887e-05, "loss": 0.9596, "step": 416 }, { "epoch": 0.18894426823742638, "grad_norm": 1.3749047623903734, "learning_rate": 3.021739130434783e-05, "loss": 0.9336, "step": 417 }, { "epoch": 0.1893973719981876, "grad_norm": 1.7312426315495744, "learning_rate": 3.0289855072463773e-05, "loss": 0.9712, "step": 418 }, { "epoch": 0.1898504757589488, "grad_norm": 0.883568518236119, "learning_rate": 3.0362318840579713e-05, "loss": 0.9554, "step": 419 }, { "epoch": 0.19030357951971003, "grad_norm": 2.067982908724023, "learning_rate": 3.0434782608695656e-05, "loss": 0.9781, "step": 420 }, { "epoch": 0.19075668328047124, "grad_norm": 1.1789138696401025, "learning_rate": 3.0507246376811596e-05, "loss": 0.9502, "step": 421 }, { "epoch": 0.19120978704123245, "grad_norm": 1.9561559852682535, "learning_rate": 3.0579710144927536e-05, "loss": 0.975, "step": 422 }, { "epoch": 0.19166289080199367, "grad_norm": 1.7368025898777355, "learning_rate": 3.065217391304348e-05, "loss": 0.9861, "step": 423 }, { "epoch": 0.19211599456275488, "grad_norm": 1.6358926482066025, "learning_rate": 3.072463768115942e-05, "loss": 0.9551, "step": 424 }, { "epoch": 0.1925690983235161, "grad_norm": 1.8795776091389105, "learning_rate": 3.079710144927536e-05, "loss": 0.9668, "step": 425 }, { "epoch": 0.1930222020842773, "grad_norm": 1.3584754061911595, "learning_rate": 3.086956521739131e-05, "loss": 0.9594, "step": 426 }, { "epoch": 0.19347530584503853, "grad_norm": 1.6867541123795684, "learning_rate": 3.094202898550725e-05, "loss": 0.9543, "step": 427 }, { "epoch": 0.19392840960579974, "grad_norm": 1.246537127655069, "learning_rate": 3.101449275362319e-05, "loss": 0.9685, "step": 428 }, { "epoch": 0.19438151336656095, "grad_norm": 1.5450835395635403, "learning_rate": 3.1086956521739134e-05, "loss": 0.9444, "step": 429 }, { "epoch": 0.19483461712732217, "grad_norm": 1.0886419714979687, "learning_rate": 3.1159420289855074e-05, "loss": 0.9355, "step": 430 }, { "epoch": 0.19528772088808338, "grad_norm": 1.3203471061713647, "learning_rate": 3.123188405797102e-05, "loss": 0.9417, "step": 431 }, { "epoch": 0.1957408246488446, "grad_norm": 1.0869781370787397, "learning_rate": 3.130434782608696e-05, "loss": 0.9544, "step": 432 }, { "epoch": 0.1961939284096058, "grad_norm": 1.3339962795143159, "learning_rate": 3.13768115942029e-05, "loss": 0.9517, "step": 433 }, { "epoch": 0.19664703217036703, "grad_norm": 1.772308726115863, "learning_rate": 3.144927536231885e-05, "loss": 0.9551, "step": 434 }, { "epoch": 0.19710013593112824, "grad_norm": 0.8165365951240647, "learning_rate": 3.152173913043478e-05, "loss": 0.9559, "step": 435 }, { "epoch": 0.19755323969188945, "grad_norm": 1.1740989519316403, "learning_rate": 3.1594202898550726e-05, "loss": 0.9481, "step": 436 }, { "epoch": 0.19800634345265067, "grad_norm": 1.9510312562297796, "learning_rate": 3.1666666666666666e-05, "loss": 0.9295, "step": 437 }, { "epoch": 0.19845944721341188, "grad_norm": 0.99268823022313, "learning_rate": 3.173913043478261e-05, "loss": 0.9512, "step": 438 }, { "epoch": 0.1989125509741731, "grad_norm": 1.8919594863898648, "learning_rate": 3.181159420289855e-05, "loss": 0.9604, "step": 439 }, { "epoch": 0.1993656547349343, "grad_norm": 1.3806575552851446, "learning_rate": 3.188405797101449e-05, "loss": 0.949, "step": 440 }, { "epoch": 0.19981875849569553, "grad_norm": 1.6374677594172187, "learning_rate": 3.195652173913044e-05, "loss": 0.9545, "step": 441 }, { "epoch": 0.20027186225645674, "grad_norm": 1.3547777793086693, "learning_rate": 3.202898550724638e-05, "loss": 0.9613, "step": 442 }, { "epoch": 0.20072496601721795, "grad_norm": 1.6602133881222298, "learning_rate": 3.2101449275362325e-05, "loss": 0.9866, "step": 443 }, { "epoch": 0.20117806977797917, "grad_norm": 1.4401419960341637, "learning_rate": 3.2173913043478265e-05, "loss": 0.9565, "step": 444 }, { "epoch": 0.20163117353874038, "grad_norm": 1.4653408217927926, "learning_rate": 3.2246376811594205e-05, "loss": 0.9585, "step": 445 }, { "epoch": 0.2020842772995016, "grad_norm": 1.2932444250984774, "learning_rate": 3.231884057971015e-05, "loss": 0.957, "step": 446 }, { "epoch": 0.2025373810602628, "grad_norm": 1.2790822374419828, "learning_rate": 3.239130434782609e-05, "loss": 0.9548, "step": 447 }, { "epoch": 0.20299048482102403, "grad_norm": 1.2924302348622416, "learning_rate": 3.246376811594203e-05, "loss": 0.9574, "step": 448 }, { "epoch": 0.20344358858178524, "grad_norm": 1.490398534407255, "learning_rate": 3.253623188405797e-05, "loss": 0.9612, "step": 449 }, { "epoch": 0.20389669234254645, "grad_norm": 1.047883073551461, "learning_rate": 3.260869565217392e-05, "loss": 0.9354, "step": 450 }, { "epoch": 0.20434979610330767, "grad_norm": 1.4123230374446376, "learning_rate": 3.268115942028986e-05, "loss": 0.9589, "step": 451 }, { "epoch": 0.20480289986406888, "grad_norm": 1.4256280533597505, "learning_rate": 3.2753623188405796e-05, "loss": 0.9475, "step": 452 }, { "epoch": 0.2052560036248301, "grad_norm": 1.3490937385982622, "learning_rate": 3.282608695652174e-05, "loss": 0.9755, "step": 453 }, { "epoch": 0.2057091073855913, "grad_norm": 0.9722480885658352, "learning_rate": 3.289855072463768e-05, "loss": 0.9602, "step": 454 }, { "epoch": 0.20616221114635253, "grad_norm": 1.4156733770584662, "learning_rate": 3.297101449275362e-05, "loss": 0.944, "step": 455 }, { "epoch": 0.20661531490711374, "grad_norm": 1.0000381005515937, "learning_rate": 3.304347826086957e-05, "loss": 0.9371, "step": 456 }, { "epoch": 0.20706841866787495, "grad_norm": 1.5221269682130558, "learning_rate": 3.311594202898551e-05, "loss": 0.9627, "step": 457 }, { "epoch": 0.20752152242863617, "grad_norm": 1.645202085923873, "learning_rate": 3.3188405797101455e-05, "loss": 0.9555, "step": 458 }, { "epoch": 0.20797462618939738, "grad_norm": 0.9522263724732904, "learning_rate": 3.3260869565217395e-05, "loss": 0.9633, "step": 459 }, { "epoch": 0.2084277299501586, "grad_norm": 1.4494663419229146, "learning_rate": 3.3333333333333335e-05, "loss": 0.9539, "step": 460 }, { "epoch": 0.2088808337109198, "grad_norm": 1.4566688455807888, "learning_rate": 3.340579710144928e-05, "loss": 0.9333, "step": 461 }, { "epoch": 0.20933393747168103, "grad_norm": 1.3001925311949074, "learning_rate": 3.347826086956522e-05, "loss": 0.9523, "step": 462 }, { "epoch": 0.20978704123244224, "grad_norm": 1.5296209250792716, "learning_rate": 3.355072463768116e-05, "loss": 0.9479, "step": 463 }, { "epoch": 0.21024014499320345, "grad_norm": 1.1285347522527784, "learning_rate": 3.36231884057971e-05, "loss": 0.9419, "step": 464 }, { "epoch": 0.21069324875396467, "grad_norm": 0.9925610275493972, "learning_rate": 3.369565217391305e-05, "loss": 0.9547, "step": 465 }, { "epoch": 0.21114635251472588, "grad_norm": 1.3782581331573478, "learning_rate": 3.376811594202899e-05, "loss": 0.9609, "step": 466 }, { "epoch": 0.2115994562754871, "grad_norm": 1.4984998189680612, "learning_rate": 3.384057971014493e-05, "loss": 0.9642, "step": 467 }, { "epoch": 0.2120525600362483, "grad_norm": 1.3855599395461182, "learning_rate": 3.391304347826087e-05, "loss": 0.9599, "step": 468 }, { "epoch": 0.21250566379700953, "grad_norm": 0.9458889596662057, "learning_rate": 3.398550724637681e-05, "loss": 0.9554, "step": 469 }, { "epoch": 0.21295876755777074, "grad_norm": 1.2686866315826133, "learning_rate": 3.405797101449276e-05, "loss": 0.9534, "step": 470 }, { "epoch": 0.21341187131853195, "grad_norm": 1.6507900546426455, "learning_rate": 3.41304347826087e-05, "loss": 0.9548, "step": 471 }, { "epoch": 0.21386497507929317, "grad_norm": 1.0972571289211492, "learning_rate": 3.420289855072464e-05, "loss": 0.9464, "step": 472 }, { "epoch": 0.21431807884005438, "grad_norm": 1.598738117029447, "learning_rate": 3.4275362318840586e-05, "loss": 0.9471, "step": 473 }, { "epoch": 0.2147711826008156, "grad_norm": 1.0672173062632015, "learning_rate": 3.4347826086956526e-05, "loss": 0.9454, "step": 474 }, { "epoch": 0.2152242863615768, "grad_norm": 1.5035548622056312, "learning_rate": 3.4420289855072465e-05, "loss": 0.9529, "step": 475 }, { "epoch": 0.21567739012233803, "grad_norm": 1.5359699403160225, "learning_rate": 3.449275362318841e-05, "loss": 0.9427, "step": 476 }, { "epoch": 0.21613049388309924, "grad_norm": 0.8205022195779366, "learning_rate": 3.456521739130435e-05, "loss": 0.9488, "step": 477 }, { "epoch": 0.21658359764386045, "grad_norm": 1.2901507096017604, "learning_rate": 3.463768115942029e-05, "loss": 0.933, "step": 478 }, { "epoch": 0.21703670140462167, "grad_norm": 0.8028987417164734, "learning_rate": 3.471014492753623e-05, "loss": 0.9606, "step": 479 }, { "epoch": 0.21748980516538288, "grad_norm": 1.3173802487538007, "learning_rate": 3.478260869565218e-05, "loss": 0.9584, "step": 480 }, { "epoch": 0.2179429089261441, "grad_norm": 0.7822242388697941, "learning_rate": 3.485507246376812e-05, "loss": 0.9398, "step": 481 }, { "epoch": 0.2183960126869053, "grad_norm": 1.61787930347704, "learning_rate": 3.492753623188406e-05, "loss": 0.9542, "step": 482 }, { "epoch": 0.21884911644766653, "grad_norm": 0.9696286419960797, "learning_rate": 3.5000000000000004e-05, "loss": 0.938, "step": 483 }, { "epoch": 0.21930222020842774, "grad_norm": 1.1593391964000508, "learning_rate": 3.5072463768115943e-05, "loss": 0.9545, "step": 484 }, { "epoch": 0.21975532396918895, "grad_norm": 1.5194742881241283, "learning_rate": 3.514492753623189e-05, "loss": 0.9627, "step": 485 }, { "epoch": 0.22020842772995017, "grad_norm": 1.0424874752841082, "learning_rate": 3.521739130434783e-05, "loss": 0.9595, "step": 486 }, { "epoch": 0.22066153149071138, "grad_norm": 1.4270310327340836, "learning_rate": 3.528985507246377e-05, "loss": 0.9381, "step": 487 }, { "epoch": 0.2211146352514726, "grad_norm": 1.1622764018421534, "learning_rate": 3.5362318840579716e-05, "loss": 0.9528, "step": 488 }, { "epoch": 0.2215677390122338, "grad_norm": 1.7834028496516592, "learning_rate": 3.5434782608695656e-05, "loss": 0.9291, "step": 489 }, { "epoch": 0.22202084277299503, "grad_norm": 1.0508457464823526, "learning_rate": 3.5507246376811596e-05, "loss": 0.9468, "step": 490 }, { "epoch": 0.22247394653375624, "grad_norm": 1.5820730789587656, "learning_rate": 3.5579710144927535e-05, "loss": 0.9625, "step": 491 }, { "epoch": 0.22292705029451745, "grad_norm": 1.0758599373260114, "learning_rate": 3.565217391304348e-05, "loss": 0.9515, "step": 492 }, { "epoch": 0.22338015405527867, "grad_norm": 1.7027194371640815, "learning_rate": 3.572463768115942e-05, "loss": 0.9506, "step": 493 }, { "epoch": 0.22383325781603988, "grad_norm": 1.4074688488413574, "learning_rate": 3.579710144927536e-05, "loss": 0.9525, "step": 494 }, { "epoch": 0.2242863615768011, "grad_norm": 1.4175183976559216, "learning_rate": 3.586956521739131e-05, "loss": 0.9556, "step": 495 }, { "epoch": 0.2247394653375623, "grad_norm": 1.1779294533383857, "learning_rate": 3.594202898550725e-05, "loss": 0.9407, "step": 496 }, { "epoch": 0.22519256909832353, "grad_norm": 1.0969919540458883, "learning_rate": 3.6014492753623194e-05, "loss": 0.9611, "step": 497 }, { "epoch": 0.22564567285908474, "grad_norm": 1.0677520655239021, "learning_rate": 3.6086956521739134e-05, "loss": 0.9533, "step": 498 }, { "epoch": 0.22609877661984595, "grad_norm": 1.2570693037367227, "learning_rate": 3.6159420289855074e-05, "loss": 0.9464, "step": 499 }, { "epoch": 0.22655188038060717, "grad_norm": 1.0239710767983294, "learning_rate": 3.623188405797102e-05, "loss": 0.9571, "step": 500 }, { "epoch": 0.22700498414136838, "grad_norm": 2.206901117067917, "learning_rate": 3.630434782608696e-05, "loss": 0.9715, "step": 501 }, { "epoch": 0.2274580879021296, "grad_norm": 1.2715359828403348, "learning_rate": 3.63768115942029e-05, "loss": 0.9568, "step": 502 }, { "epoch": 0.2279111916628908, "grad_norm": 2.413822143352405, "learning_rate": 3.6449275362318846e-05, "loss": 0.9721, "step": 503 }, { "epoch": 0.22836429542365203, "grad_norm": 1.7993701949325775, "learning_rate": 3.6521739130434786e-05, "loss": 0.9537, "step": 504 }, { "epoch": 0.22881739918441324, "grad_norm": 2.2028319720934553, "learning_rate": 3.6594202898550726e-05, "loss": 0.9622, "step": 505 }, { "epoch": 0.22927050294517445, "grad_norm": 1.8120542431479618, "learning_rate": 3.6666666666666666e-05, "loss": 0.9624, "step": 506 }, { "epoch": 0.22972360670593567, "grad_norm": 2.1567975047758488, "learning_rate": 3.673913043478261e-05, "loss": 0.9559, "step": 507 }, { "epoch": 0.23017671046669688, "grad_norm": 1.751114019227653, "learning_rate": 3.681159420289855e-05, "loss": 0.9739, "step": 508 }, { "epoch": 0.2306298142274581, "grad_norm": 2.0020604946335134, "learning_rate": 3.68840579710145e-05, "loss": 0.9637, "step": 509 }, { "epoch": 0.2310829179882193, "grad_norm": 1.8073128105913583, "learning_rate": 3.695652173913044e-05, "loss": 0.9696, "step": 510 }, { "epoch": 0.23153602174898053, "grad_norm": 1.958782680510924, "learning_rate": 3.702898550724638e-05, "loss": 0.9707, "step": 511 }, { "epoch": 0.23198912550974174, "grad_norm": 1.7341975677533372, "learning_rate": 3.7101449275362325e-05, "loss": 0.954, "step": 512 }, { "epoch": 0.23244222927050295, "grad_norm": 1.632960227588907, "learning_rate": 3.7173913043478264e-05, "loss": 0.9465, "step": 513 }, { "epoch": 0.23289533303126417, "grad_norm": 1.4258253562675178, "learning_rate": 3.7246376811594204e-05, "loss": 0.9537, "step": 514 }, { "epoch": 0.23334843679202538, "grad_norm": 1.5951409419664124, "learning_rate": 3.731884057971015e-05, "loss": 0.9581, "step": 515 }, { "epoch": 0.2338015405527866, "grad_norm": 1.299565735680188, "learning_rate": 3.739130434782609e-05, "loss": 0.939, "step": 516 }, { "epoch": 0.2342546443135478, "grad_norm": 1.2085963813751661, "learning_rate": 3.746376811594204e-05, "loss": 0.9652, "step": 517 }, { "epoch": 0.23470774807430902, "grad_norm": 1.4840807667730531, "learning_rate": 3.753623188405797e-05, "loss": 0.9503, "step": 518 }, { "epoch": 0.23516085183507024, "grad_norm": 1.2555580256704544, "learning_rate": 3.7608695652173917e-05, "loss": 0.9386, "step": 519 }, { "epoch": 0.23561395559583145, "grad_norm": 1.2521897632132748, "learning_rate": 3.7681159420289856e-05, "loss": 0.9643, "step": 520 }, { "epoch": 0.23606705935659267, "grad_norm": 1.5749003286617689, "learning_rate": 3.7753623188405796e-05, "loss": 0.9523, "step": 521 }, { "epoch": 0.23652016311735388, "grad_norm": 1.4998013517178532, "learning_rate": 3.782608695652174e-05, "loss": 0.9437, "step": 522 }, { "epoch": 0.2369732668781151, "grad_norm": 1.5024488424154785, "learning_rate": 3.789855072463768e-05, "loss": 0.9608, "step": 523 }, { "epoch": 0.2374263706388763, "grad_norm": 1.3021509876962238, "learning_rate": 3.797101449275363e-05, "loss": 0.9536, "step": 524 }, { "epoch": 0.23787947439963752, "grad_norm": 0.8994452299792001, "learning_rate": 3.804347826086957e-05, "loss": 0.9586, "step": 525 }, { "epoch": 0.23833257816039874, "grad_norm": 1.6305129821640654, "learning_rate": 3.811594202898551e-05, "loss": 0.9477, "step": 526 }, { "epoch": 0.23878568192115995, "grad_norm": 1.1558566489499902, "learning_rate": 3.8188405797101455e-05, "loss": 0.9665, "step": 527 }, { "epoch": 0.23923878568192117, "grad_norm": 1.8159821557330431, "learning_rate": 3.8260869565217395e-05, "loss": 0.9502, "step": 528 }, { "epoch": 0.23969188944268238, "grad_norm": 1.1692795888927126, "learning_rate": 3.833333333333334e-05, "loss": 0.9304, "step": 529 }, { "epoch": 0.2401449932034436, "grad_norm": 1.4725878766305236, "learning_rate": 3.840579710144928e-05, "loss": 0.9231, "step": 530 }, { "epoch": 0.2405980969642048, "grad_norm": 1.2258875983947652, "learning_rate": 3.847826086956522e-05, "loss": 0.9582, "step": 531 }, { "epoch": 0.24105120072496602, "grad_norm": 1.577801679961895, "learning_rate": 3.855072463768116e-05, "loss": 0.9372, "step": 532 }, { "epoch": 0.24150430448572724, "grad_norm": 1.3844046757318353, "learning_rate": 3.86231884057971e-05, "loss": 0.9285, "step": 533 }, { "epoch": 0.24195740824648845, "grad_norm": 1.6249010279429936, "learning_rate": 3.869565217391305e-05, "loss": 0.9384, "step": 534 }, { "epoch": 0.24241051200724967, "grad_norm": 1.352092809863243, "learning_rate": 3.876811594202899e-05, "loss": 0.9514, "step": 535 }, { "epoch": 0.24286361576801088, "grad_norm": 0.9739221347448674, "learning_rate": 3.884057971014493e-05, "loss": 0.9518, "step": 536 }, { "epoch": 0.2433167195287721, "grad_norm": 1.183887940922217, "learning_rate": 3.891304347826087e-05, "loss": 0.9452, "step": 537 }, { "epoch": 0.2437698232895333, "grad_norm": 1.7247688314790206, "learning_rate": 3.898550724637681e-05, "loss": 0.9331, "step": 538 }, { "epoch": 0.24422292705029452, "grad_norm": 1.5029215251052839, "learning_rate": 3.905797101449276e-05, "loss": 0.9619, "step": 539 }, { "epoch": 0.24467603081105574, "grad_norm": 1.2075533763523865, "learning_rate": 3.91304347826087e-05, "loss": 0.9547, "step": 540 }, { "epoch": 0.24512913457181695, "grad_norm": 1.2574432720764974, "learning_rate": 3.920289855072464e-05, "loss": 0.9543, "step": 541 }, { "epoch": 0.24558223833257817, "grad_norm": 1.04770752009234, "learning_rate": 3.9275362318840585e-05, "loss": 0.9228, "step": 542 }, { "epoch": 0.24603534209333938, "grad_norm": 2.3420558721323563, "learning_rate": 3.9347826086956525e-05, "loss": 0.9472, "step": 543 }, { "epoch": 0.2464884458541006, "grad_norm": 1.0874560193300407, "learning_rate": 3.942028985507247e-05, "loss": 0.9587, "step": 544 }, { "epoch": 0.2469415496148618, "grad_norm": 2.4502861952137454, "learning_rate": 3.949275362318841e-05, "loss": 0.9481, "step": 545 }, { "epoch": 0.24739465337562302, "grad_norm": 1.8676593822635237, "learning_rate": 3.956521739130435e-05, "loss": 0.9652, "step": 546 }, { "epoch": 0.24784775713638424, "grad_norm": 2.212234158633956, "learning_rate": 3.963768115942029e-05, "loss": 0.9417, "step": 547 }, { "epoch": 0.24830086089714545, "grad_norm": 1.8911486866476837, "learning_rate": 3.971014492753623e-05, "loss": 0.965, "step": 548 }, { "epoch": 0.24875396465790667, "grad_norm": 1.7978196087476268, "learning_rate": 3.978260869565218e-05, "loss": 0.939, "step": 549 }, { "epoch": 0.24920706841866788, "grad_norm": 1.4358398374085661, "learning_rate": 3.985507246376812e-05, "loss": 0.9676, "step": 550 }, { "epoch": 0.2496601721794291, "grad_norm": 2.1785104655124443, "learning_rate": 3.9927536231884064e-05, "loss": 0.9375, "step": 551 }, { "epoch": 0.2501132759401903, "grad_norm": 1.4917235160664257, "learning_rate": 4e-05, "loss": 0.9598, "step": 552 }, { "epoch": 0.2505663797009515, "grad_norm": 1.8981577794034559, "learning_rate": 4.007246376811594e-05, "loss": 0.9481, "step": 553 }, { "epoch": 0.2510194834617127, "grad_norm": 1.6039725232941155, "learning_rate": 4.014492753623189e-05, "loss": 0.9166, "step": 554 }, { "epoch": 0.25147258722247395, "grad_norm": 1.8629312825882385, "learning_rate": 4.021739130434782e-05, "loss": 0.9553, "step": 555 }, { "epoch": 0.25192569098323514, "grad_norm": 1.1163302548593497, "learning_rate": 4.0289855072463776e-05, "loss": 0.9429, "step": 556 }, { "epoch": 0.2523787947439964, "grad_norm": 2.232355939449555, "learning_rate": 4.036231884057971e-05, "loss": 0.9426, "step": 557 }, { "epoch": 0.25283189850475757, "grad_norm": 1.707574341755059, "learning_rate": 4.0434782608695655e-05, "loss": 0.9456, "step": 558 }, { "epoch": 0.2532850022655188, "grad_norm": 1.6803108482665834, "learning_rate": 4.0507246376811595e-05, "loss": 0.9426, "step": 559 }, { "epoch": 0.25373810602628, "grad_norm": 1.9699271436505157, "learning_rate": 4.057971014492754e-05, "loss": 0.9166, "step": 560 }, { "epoch": 0.25419120978704124, "grad_norm": 1.1863420589250744, "learning_rate": 4.065217391304348e-05, "loss": 0.9541, "step": 561 }, { "epoch": 0.2546443135478024, "grad_norm": 1.4892809268047607, "learning_rate": 4.072463768115943e-05, "loss": 0.9459, "step": 562 }, { "epoch": 0.25509741730856367, "grad_norm": 1.0036606761462499, "learning_rate": 4.079710144927537e-05, "loss": 0.9299, "step": 563 }, { "epoch": 0.25555052106932485, "grad_norm": 1.90925274594905, "learning_rate": 4.0869565217391314e-05, "loss": 0.9597, "step": 564 }, { "epoch": 0.2560036248300861, "grad_norm": 1.7744534467353157, "learning_rate": 4.094202898550725e-05, "loss": 0.9573, "step": 565 }, { "epoch": 0.2564567285908473, "grad_norm": 1.126787515733341, "learning_rate": 4.101449275362319e-05, "loss": 0.9694, "step": 566 }, { "epoch": 0.2569098323516085, "grad_norm": 1.7339668878001462, "learning_rate": 4.1086956521739134e-05, "loss": 0.9473, "step": 567 }, { "epoch": 0.2573629361123697, "grad_norm": 1.422036607214472, "learning_rate": 4.115942028985507e-05, "loss": 0.9563, "step": 568 }, { "epoch": 0.25781603987313095, "grad_norm": 1.7198368676848514, "learning_rate": 4.123188405797102e-05, "loss": 0.944, "step": 569 }, { "epoch": 0.25826914363389214, "grad_norm": 1.2062756643742152, "learning_rate": 4.130434782608696e-05, "loss": 0.9308, "step": 570 }, { "epoch": 0.2587222473946534, "grad_norm": 1.4584519982809159, "learning_rate": 4.1376811594202906e-05, "loss": 0.9572, "step": 571 }, { "epoch": 0.25917535115541457, "grad_norm": 0.9318366107120218, "learning_rate": 4.144927536231884e-05, "loss": 0.9575, "step": 572 }, { "epoch": 0.2596284549161758, "grad_norm": 1.7048998241368907, "learning_rate": 4.1521739130434786e-05, "loss": 0.9448, "step": 573 }, { "epoch": 0.260081558676937, "grad_norm": 1.3193015980055645, "learning_rate": 4.1594202898550726e-05, "loss": 0.9482, "step": 574 }, { "epoch": 0.26053466243769824, "grad_norm": 1.5181331412848595, "learning_rate": 4.166666666666667e-05, "loss": 0.9639, "step": 575 }, { "epoch": 0.2609877661984594, "grad_norm": 1.4036366657065995, "learning_rate": 4.173913043478261e-05, "loss": 0.9456, "step": 576 }, { "epoch": 0.26144086995922067, "grad_norm": 1.4391292448209332, "learning_rate": 4.181159420289856e-05, "loss": 0.948, "step": 577 }, { "epoch": 0.26189397371998185, "grad_norm": 1.1521683547489925, "learning_rate": 4.18840579710145e-05, "loss": 0.955, "step": 578 }, { "epoch": 0.2623470774807431, "grad_norm": 1.5204461310054402, "learning_rate": 4.1956521739130445e-05, "loss": 0.9334, "step": 579 }, { "epoch": 0.2628001812415043, "grad_norm": 1.5036128999656508, "learning_rate": 4.202898550724638e-05, "loss": 0.9476, "step": 580 }, { "epoch": 0.2632532850022655, "grad_norm": 1.178352602143713, "learning_rate": 4.210144927536232e-05, "loss": 0.9228, "step": 581 }, { "epoch": 0.2637063887630267, "grad_norm": 1.4576554654544704, "learning_rate": 4.2173913043478264e-05, "loss": 0.9554, "step": 582 }, { "epoch": 0.26415949252378795, "grad_norm": 1.7550410149662652, "learning_rate": 4.2246376811594204e-05, "loss": 0.9422, "step": 583 }, { "epoch": 0.26461259628454914, "grad_norm": 0.8898443033549837, "learning_rate": 4.231884057971015e-05, "loss": 0.9355, "step": 584 }, { "epoch": 0.2650657000453104, "grad_norm": 1.1317450628408046, "learning_rate": 4.239130434782609e-05, "loss": 0.9542, "step": 585 }, { "epoch": 0.26551880380607157, "grad_norm": 1.7507789855268754, "learning_rate": 4.2463768115942037e-05, "loss": 0.9406, "step": 586 }, { "epoch": 0.2659719075668328, "grad_norm": 1.562509974795605, "learning_rate": 4.253623188405797e-05, "loss": 0.9493, "step": 587 }, { "epoch": 0.266425011327594, "grad_norm": 1.1499164881087813, "learning_rate": 4.2608695652173916e-05, "loss": 0.9475, "step": 588 }, { "epoch": 0.26687811508835524, "grad_norm": 1.3376692864942785, "learning_rate": 4.2681159420289856e-05, "loss": 0.9479, "step": 589 }, { "epoch": 0.2673312188491164, "grad_norm": 1.729753952608737, "learning_rate": 4.27536231884058e-05, "loss": 0.9451, "step": 590 }, { "epoch": 0.26778432260987767, "grad_norm": 1.121911357215556, "learning_rate": 4.282608695652174e-05, "loss": 0.9678, "step": 591 }, { "epoch": 0.26823742637063885, "grad_norm": 1.7641639723059712, "learning_rate": 4.289855072463769e-05, "loss": 0.9567, "step": 592 }, { "epoch": 0.2686905301314001, "grad_norm": 1.115466226401504, "learning_rate": 4.297101449275363e-05, "loss": 0.8954, "step": 593 }, { "epoch": 0.2691436338921613, "grad_norm": 2.0466767455451147, "learning_rate": 4.304347826086956e-05, "loss": 0.9298, "step": 594 }, { "epoch": 0.2695967376529225, "grad_norm": 1.142198597297308, "learning_rate": 4.311594202898551e-05, "loss": 0.9521, "step": 595 }, { "epoch": 0.2700498414136837, "grad_norm": 1.4548020514910855, "learning_rate": 4.318840579710145e-05, "loss": 0.9608, "step": 596 }, { "epoch": 0.27050294517444495, "grad_norm": 1.021464031808508, "learning_rate": 4.3260869565217394e-05, "loss": 0.9363, "step": 597 }, { "epoch": 0.27095604893520614, "grad_norm": 1.5646926563893484, "learning_rate": 4.3333333333333334e-05, "loss": 0.9573, "step": 598 }, { "epoch": 0.2714091526959674, "grad_norm": 1.2034511361820375, "learning_rate": 4.340579710144928e-05, "loss": 0.94, "step": 599 }, { "epoch": 0.27186225645672857, "grad_norm": 1.993177714777034, "learning_rate": 4.347826086956522e-05, "loss": 0.9259, "step": 600 }, { "epoch": 0.2723153602174898, "grad_norm": 1.2928645678552766, "learning_rate": 4.355072463768117e-05, "loss": 0.9672, "step": 601 }, { "epoch": 0.272768463978251, "grad_norm": 1.3499785494317744, "learning_rate": 4.362318840579711e-05, "loss": 0.942, "step": 602 }, { "epoch": 0.27322156773901224, "grad_norm": 1.9542309912992344, "learning_rate": 4.369565217391305e-05, "loss": 0.9455, "step": 603 }, { "epoch": 0.2736746714997734, "grad_norm": 1.445875192150768, "learning_rate": 4.3768115942028986e-05, "loss": 0.9628, "step": 604 }, { "epoch": 0.27412777526053467, "grad_norm": 1.4401402127788205, "learning_rate": 4.384057971014493e-05, "loss": 0.9514, "step": 605 }, { "epoch": 0.27458087902129585, "grad_norm": 0.87594390755773, "learning_rate": 4.391304347826087e-05, "loss": 0.9451, "step": 606 }, { "epoch": 0.2750339827820571, "grad_norm": 1.6998068344654556, "learning_rate": 4.398550724637682e-05, "loss": 0.9368, "step": 607 }, { "epoch": 0.2754870865428183, "grad_norm": 1.2812240545863185, "learning_rate": 4.405797101449276e-05, "loss": 0.9745, "step": 608 }, { "epoch": 0.2759401903035795, "grad_norm": 1.6160437657699844, "learning_rate": 4.41304347826087e-05, "loss": 0.9609, "step": 609 }, { "epoch": 0.2763932940643407, "grad_norm": 1.3134988649801242, "learning_rate": 4.4202898550724645e-05, "loss": 0.96, "step": 610 }, { "epoch": 0.27684639782510195, "grad_norm": 1.3775715598402054, "learning_rate": 4.427536231884058e-05, "loss": 0.955, "step": 611 }, { "epoch": 0.27729950158586314, "grad_norm": 1.373807896251055, "learning_rate": 4.4347826086956525e-05, "loss": 0.9367, "step": 612 }, { "epoch": 0.2777526053466244, "grad_norm": 1.7976714914276, "learning_rate": 4.4420289855072464e-05, "loss": 0.9453, "step": 613 }, { "epoch": 0.27820570910738557, "grad_norm": 1.4890253742405781, "learning_rate": 4.449275362318841e-05, "loss": 0.9195, "step": 614 }, { "epoch": 0.2786588128681468, "grad_norm": 1.2966519649417882, "learning_rate": 4.456521739130435e-05, "loss": 0.926, "step": 615 }, { "epoch": 0.279111916628908, "grad_norm": 1.0502037938964333, "learning_rate": 4.46376811594203e-05, "loss": 0.9378, "step": 616 }, { "epoch": 0.27956502038966924, "grad_norm": 1.8153095270177775, "learning_rate": 4.471014492753624e-05, "loss": 0.952, "step": 617 }, { "epoch": 0.2800181241504304, "grad_norm": 1.0632425876520273, "learning_rate": 4.4782608695652184e-05, "loss": 0.9318, "step": 618 }, { "epoch": 0.28047122791119167, "grad_norm": 2.0100796008099104, "learning_rate": 4.4855072463768117e-05, "loss": 0.9513, "step": 619 }, { "epoch": 0.28092433167195285, "grad_norm": 1.7621762431297525, "learning_rate": 4.492753623188406e-05, "loss": 0.9731, "step": 620 }, { "epoch": 0.2813774354327141, "grad_norm": 1.31773425886026, "learning_rate": 4.5e-05, "loss": 0.9376, "step": 621 }, { "epoch": 0.2818305391934753, "grad_norm": 2.5668327836029072, "learning_rate": 4.507246376811594e-05, "loss": 0.9503, "step": 622 }, { "epoch": 0.2822836429542365, "grad_norm": 1.6513923800907704, "learning_rate": 4.514492753623189e-05, "loss": 0.9563, "step": 623 }, { "epoch": 0.2827367467149977, "grad_norm": 2.3632162703372934, "learning_rate": 4.521739130434783e-05, "loss": 0.947, "step": 624 }, { "epoch": 0.28318985047575895, "grad_norm": 1.8079460032177557, "learning_rate": 4.5289855072463775e-05, "loss": 0.9515, "step": 625 }, { "epoch": 0.28364295423652014, "grad_norm": 2.2036676732334204, "learning_rate": 4.536231884057971e-05, "loss": 0.964, "step": 626 }, { "epoch": 0.2840960579972814, "grad_norm": 2.0190874508178704, "learning_rate": 4.5434782608695655e-05, "loss": 0.9347, "step": 627 }, { "epoch": 0.28454916175804257, "grad_norm": 1.596862999309559, "learning_rate": 4.5507246376811595e-05, "loss": 0.9658, "step": 628 }, { "epoch": 0.2850022655188038, "grad_norm": 1.6458348309300757, "learning_rate": 4.557971014492754e-05, "loss": 0.9478, "step": 629 }, { "epoch": 0.285455369279565, "grad_norm": 1.5111331869704456, "learning_rate": 4.565217391304348e-05, "loss": 0.961, "step": 630 }, { "epoch": 0.28590847304032624, "grad_norm": 1.832593296305825, "learning_rate": 4.572463768115943e-05, "loss": 0.9322, "step": 631 }, { "epoch": 0.2863615768010874, "grad_norm": 1.3311302398247842, "learning_rate": 4.579710144927537e-05, "loss": 0.9566, "step": 632 }, { "epoch": 0.28681468056184867, "grad_norm": 1.8821403922797415, "learning_rate": 4.5869565217391314e-05, "loss": 0.9438, "step": 633 }, { "epoch": 0.28726778432260985, "grad_norm": 1.2648424443609299, "learning_rate": 4.594202898550725e-05, "loss": 0.938, "step": 634 }, { "epoch": 0.2877208880833711, "grad_norm": 1.9545927394307079, "learning_rate": 4.601449275362319e-05, "loss": 0.9247, "step": 635 }, { "epoch": 0.2881739918441323, "grad_norm": 1.3472313957357194, "learning_rate": 4.608695652173913e-05, "loss": 0.9381, "step": 636 }, { "epoch": 0.2886270956048935, "grad_norm": 1.9975030132454488, "learning_rate": 4.615942028985507e-05, "loss": 0.9594, "step": 637 }, { "epoch": 0.2890801993656547, "grad_norm": 1.6821842200183719, "learning_rate": 4.623188405797102e-05, "loss": 0.9669, "step": 638 }, { "epoch": 0.28953330312641595, "grad_norm": 1.7057105772271868, "learning_rate": 4.630434782608696e-05, "loss": 0.9378, "step": 639 }, { "epoch": 0.28998640688717714, "grad_norm": 1.3189630554033869, "learning_rate": 4.6376811594202906e-05, "loss": 0.947, "step": 640 }, { "epoch": 0.2904395106479384, "grad_norm": 1.6690115589844086, "learning_rate": 4.644927536231884e-05, "loss": 0.947, "step": 641 }, { "epoch": 0.29089261440869957, "grad_norm": 1.4403677073191548, "learning_rate": 4.652173913043479e-05, "loss": 0.9304, "step": 642 }, { "epoch": 0.2913457181694608, "grad_norm": 1.5319229773750616, "learning_rate": 4.6594202898550725e-05, "loss": 0.9454, "step": 643 }, { "epoch": 0.291798821930222, "grad_norm": 1.247457551978175, "learning_rate": 4.666666666666667e-05, "loss": 0.956, "step": 644 }, { "epoch": 0.29225192569098324, "grad_norm": 1.5641043864902142, "learning_rate": 4.673913043478261e-05, "loss": 0.9532, "step": 645 }, { "epoch": 0.2927050294517444, "grad_norm": 1.4491802519962915, "learning_rate": 4.681159420289856e-05, "loss": 0.9333, "step": 646 }, { "epoch": 0.29315813321250567, "grad_norm": 1.417532026297118, "learning_rate": 4.68840579710145e-05, "loss": 0.9271, "step": 647 }, { "epoch": 0.29361123697326685, "grad_norm": 1.2445889339461733, "learning_rate": 4.6956521739130444e-05, "loss": 0.9464, "step": 648 }, { "epoch": 0.2940643407340281, "grad_norm": 1.4916869251739664, "learning_rate": 4.7028985507246384e-05, "loss": 0.9533, "step": 649 }, { "epoch": 0.2945174444947893, "grad_norm": 1.7171607064557992, "learning_rate": 4.710144927536232e-05, "loss": 0.9364, "step": 650 }, { "epoch": 0.2949705482555505, "grad_norm": 1.1020643293920975, "learning_rate": 4.7173913043478264e-05, "loss": 0.9338, "step": 651 }, { "epoch": 0.2954236520163117, "grad_norm": 1.7016773193867598, "learning_rate": 4.72463768115942e-05, "loss": 0.9302, "step": 652 }, { "epoch": 0.29587675577707295, "grad_norm": 2.0866400861429892, "learning_rate": 4.731884057971015e-05, "loss": 0.9574, "step": 653 }, { "epoch": 0.29632985953783414, "grad_norm": 0.8078647475581984, "learning_rate": 4.739130434782609e-05, "loss": 0.9569, "step": 654 }, { "epoch": 0.2967829632985954, "grad_norm": 2.076532471649137, "learning_rate": 4.7463768115942036e-05, "loss": 0.945, "step": 655 }, { "epoch": 0.29723606705935657, "grad_norm": 1.8046489122324216, "learning_rate": 4.7536231884057976e-05, "loss": 0.9367, "step": 656 }, { "epoch": 0.2976891708201178, "grad_norm": 1.3728441311970152, "learning_rate": 4.760869565217392e-05, "loss": 0.9606, "step": 657 }, { "epoch": 0.298142274580879, "grad_norm": 1.4519898892652845, "learning_rate": 4.7681159420289855e-05, "loss": 0.9163, "step": 658 }, { "epoch": 0.29859537834164024, "grad_norm": 1.870168338265197, "learning_rate": 4.77536231884058e-05, "loss": 0.945, "step": 659 }, { "epoch": 0.2990484821024014, "grad_norm": 1.69325825284278, "learning_rate": 4.782608695652174e-05, "loss": 0.9479, "step": 660 }, { "epoch": 0.29950158586316267, "grad_norm": 1.0681277280828099, "learning_rate": 4.789855072463769e-05, "loss": 0.9518, "step": 661 }, { "epoch": 0.29995468962392385, "grad_norm": 1.3144966613812135, "learning_rate": 4.797101449275363e-05, "loss": 0.9341, "step": 662 }, { "epoch": 0.3004077933846851, "grad_norm": 1.1276102952241023, "learning_rate": 4.804347826086957e-05, "loss": 0.9062, "step": 663 }, { "epoch": 0.3008608971454463, "grad_norm": 1.9260965739397962, "learning_rate": 4.8115942028985514e-05, "loss": 0.9391, "step": 664 }, { "epoch": 0.3013140009062075, "grad_norm": 1.5747273001275484, "learning_rate": 4.818840579710145e-05, "loss": 0.932, "step": 665 }, { "epoch": 0.3017671046669687, "grad_norm": 1.0844997064519406, "learning_rate": 4.8260869565217394e-05, "loss": 0.9367, "step": 666 }, { "epoch": 0.30222020842772995, "grad_norm": 1.7503347433470544, "learning_rate": 4.8333333333333334e-05, "loss": 0.9479, "step": 667 }, { "epoch": 0.30267331218849114, "grad_norm": 1.5435924939959635, "learning_rate": 4.840579710144928e-05, "loss": 0.9338, "step": 668 }, { "epoch": 0.3031264159492524, "grad_norm": 1.542468735197468, "learning_rate": 4.847826086956522e-05, "loss": 0.9607, "step": 669 }, { "epoch": 0.30357951971001357, "grad_norm": 1.921855962774172, "learning_rate": 4.8550724637681166e-05, "loss": 0.945, "step": 670 }, { "epoch": 0.3040326234707748, "grad_norm": 0.9107625703575939, "learning_rate": 4.8623188405797106e-05, "loss": 0.9399, "step": 671 }, { "epoch": 0.304485727231536, "grad_norm": 2.5248685805560536, "learning_rate": 4.869565217391305e-05, "loss": 0.9568, "step": 672 }, { "epoch": 0.30493883099229724, "grad_norm": 1.6754821639618416, "learning_rate": 4.8768115942028986e-05, "loss": 0.9467, "step": 673 }, { "epoch": 0.3053919347530584, "grad_norm": 2.2235435699425175, "learning_rate": 4.884057971014493e-05, "loss": 0.9407, "step": 674 }, { "epoch": 0.30584503851381967, "grad_norm": 1.7235784099027713, "learning_rate": 4.891304347826087e-05, "loss": 0.943, "step": 675 }, { "epoch": 0.30629814227458085, "grad_norm": 2.231338479992835, "learning_rate": 4.898550724637682e-05, "loss": 0.9572, "step": 676 }, { "epoch": 0.3067512460353421, "grad_norm": 1.9729263380452853, "learning_rate": 4.905797101449276e-05, "loss": 0.9556, "step": 677 }, { "epoch": 0.3072043497961033, "grad_norm": 1.971191078322159, "learning_rate": 4.91304347826087e-05, "loss": 0.9504, "step": 678 }, { "epoch": 0.3076574535568645, "grad_norm": 1.6319784348530717, "learning_rate": 4.9202898550724645e-05, "loss": 0.9316, "step": 679 }, { "epoch": 0.3081105573176257, "grad_norm": 1.6082458519312084, "learning_rate": 4.927536231884058e-05, "loss": 0.9521, "step": 680 }, { "epoch": 0.30856366107838695, "grad_norm": 1.6735418941548863, "learning_rate": 4.9347826086956524e-05, "loss": 0.9611, "step": 681 }, { "epoch": 0.30901676483914814, "grad_norm": 1.3872955781776408, "learning_rate": 4.9420289855072464e-05, "loss": 0.9387, "step": 682 }, { "epoch": 0.3094698685999094, "grad_norm": 1.7944349632512275, "learning_rate": 4.949275362318841e-05, "loss": 0.9256, "step": 683 }, { "epoch": 0.30992297236067057, "grad_norm": 1.544581557211245, "learning_rate": 4.956521739130435e-05, "loss": 0.9612, "step": 684 }, { "epoch": 0.3103760761214318, "grad_norm": 1.9384830705203744, "learning_rate": 4.96376811594203e-05, "loss": 0.9164, "step": 685 }, { "epoch": 0.310829179882193, "grad_norm": 1.1712264204326148, "learning_rate": 4.9710144927536237e-05, "loss": 0.9487, "step": 686 }, { "epoch": 0.31128228364295424, "grad_norm": 2.2172665908714437, "learning_rate": 4.978260869565218e-05, "loss": 0.9552, "step": 687 }, { "epoch": 0.3117353874037154, "grad_norm": 1.8767272133324437, "learning_rate": 4.9855072463768116e-05, "loss": 0.9253, "step": 688 }, { "epoch": 0.31218849116447667, "grad_norm": 1.5519327167860006, "learning_rate": 4.992753623188407e-05, "loss": 0.947, "step": 689 }, { "epoch": 0.31264159492523785, "grad_norm": 1.4466062860481128, "learning_rate": 5e-05, "loss": 0.9425, "step": 690 }, { "epoch": 0.3130946986859991, "grad_norm": 1.5241802459738112, "learning_rate": 5.007246376811594e-05, "loss": 0.9447, "step": 691 }, { "epoch": 0.3135478024467603, "grad_norm": 1.867050296356958, "learning_rate": 5.014492753623189e-05, "loss": 0.9333, "step": 692 }, { "epoch": 0.3140009062075215, "grad_norm": 1.2233458453023953, "learning_rate": 5.021739130434783e-05, "loss": 0.938, "step": 693 }, { "epoch": 0.3144540099682827, "grad_norm": 1.741794690842049, "learning_rate": 5.0289855072463775e-05, "loss": 0.9538, "step": 694 }, { "epoch": 0.31490711372904395, "grad_norm": 1.6084385543144688, "learning_rate": 5.0362318840579715e-05, "loss": 0.9713, "step": 695 }, { "epoch": 0.31536021748980514, "grad_norm": 1.4939443761298068, "learning_rate": 5.043478260869566e-05, "loss": 0.9351, "step": 696 }, { "epoch": 0.3158133212505664, "grad_norm": 1.144974187898062, "learning_rate": 5.0507246376811594e-05, "loss": 0.9476, "step": 697 }, { "epoch": 0.31626642501132757, "grad_norm": 1.4883472883979079, "learning_rate": 5.057971014492754e-05, "loss": 0.9266, "step": 698 }, { "epoch": 0.3167195287720888, "grad_norm": 2.593828854061514, "learning_rate": 5.065217391304348e-05, "loss": 0.9412, "step": 699 }, { "epoch": 0.31717263253285, "grad_norm": 1.2048464569240143, "learning_rate": 5.072463768115943e-05, "loss": 0.9392, "step": 700 }, { "epoch": 0.31762573629361124, "grad_norm": 3.5658996458156347, "learning_rate": 5.079710144927537e-05, "loss": 0.9317, "step": 701 }, { "epoch": 0.3180788400543724, "grad_norm": 3.119804498295699, "learning_rate": 5.0869565217391313e-05, "loss": 0.9396, "step": 702 }, { "epoch": 0.31853194381513367, "grad_norm": 2.5294289042566045, "learning_rate": 5.094202898550725e-05, "loss": 0.9561, "step": 703 }, { "epoch": 0.31898504757589485, "grad_norm": 2.0042330684008154, "learning_rate": 5.1014492753623186e-05, "loss": 0.9305, "step": 704 }, { "epoch": 0.3194381513366561, "grad_norm": 2.726464170279139, "learning_rate": 5.108695652173913e-05, "loss": 0.9622, "step": 705 }, { "epoch": 0.3198912550974173, "grad_norm": 1.8626123218993627, "learning_rate": 5.115942028985507e-05, "loss": 0.9292, "step": 706 }, { "epoch": 0.3203443588581785, "grad_norm": 2.459739999453121, "learning_rate": 5.123188405797102e-05, "loss": 0.9405, "step": 707 }, { "epoch": 0.3207974626189397, "grad_norm": 1.9090656152659051, "learning_rate": 5.130434782608696e-05, "loss": 0.9437, "step": 708 }, { "epoch": 0.32125056637970095, "grad_norm": 2.1245113183944326, "learning_rate": 5.1376811594202905e-05, "loss": 0.9602, "step": 709 }, { "epoch": 0.32170367014046214, "grad_norm": 1.8053077449137631, "learning_rate": 5.1449275362318845e-05, "loss": 0.9617, "step": 710 }, { "epoch": 0.3221567739012234, "grad_norm": 1.3478067900071589, "learning_rate": 5.152173913043479e-05, "loss": 0.9349, "step": 711 }, { "epoch": 0.32260987766198457, "grad_norm": 0.9095642441198134, "learning_rate": 5.1594202898550725e-05, "loss": 0.9243, "step": 712 }, { "epoch": 0.3230629814227458, "grad_norm": 1.681437358941543, "learning_rate": 5.166666666666667e-05, "loss": 0.9551, "step": 713 }, { "epoch": 0.323516085183507, "grad_norm": 1.1325149737790132, "learning_rate": 5.173913043478261e-05, "loss": 0.94, "step": 714 }, { "epoch": 0.32396918894426824, "grad_norm": 1.9639171218597218, "learning_rate": 5.181159420289856e-05, "loss": 0.9133, "step": 715 }, { "epoch": 0.3244222927050294, "grad_norm": 1.719581631295271, "learning_rate": 5.18840579710145e-05, "loss": 0.9311, "step": 716 }, { "epoch": 0.32487539646579067, "grad_norm": 1.3341757083134236, "learning_rate": 5.1956521739130444e-05, "loss": 0.9334, "step": 717 }, { "epoch": 0.32532850022655185, "grad_norm": 1.673935638997859, "learning_rate": 5.2028985507246384e-05, "loss": 0.9376, "step": 718 }, { "epoch": 0.3257816039873131, "grad_norm": 1.3977914230082147, "learning_rate": 5.2101449275362317e-05, "loss": 0.9319, "step": 719 }, { "epoch": 0.3262347077480743, "grad_norm": 1.7262449701898679, "learning_rate": 5.217391304347826e-05, "loss": 0.9164, "step": 720 }, { "epoch": 0.3266878115088355, "grad_norm": 1.3172136235859022, "learning_rate": 5.22463768115942e-05, "loss": 0.9507, "step": 721 }, { "epoch": 0.3271409152695967, "grad_norm": 1.6452351657408357, "learning_rate": 5.231884057971015e-05, "loss": 0.9368, "step": 722 }, { "epoch": 0.32759401903035795, "grad_norm": 1.2420233586167644, "learning_rate": 5.239130434782609e-05, "loss": 0.9319, "step": 723 }, { "epoch": 0.32804712279111914, "grad_norm": 1.746010941132079, "learning_rate": 5.2463768115942036e-05, "loss": 0.9372, "step": 724 }, { "epoch": 0.3285002265518804, "grad_norm": 1.54211691251862, "learning_rate": 5.2536231884057975e-05, "loss": 0.941, "step": 725 }, { "epoch": 0.32895333031264157, "grad_norm": 1.6812814850196083, "learning_rate": 5.260869565217392e-05, "loss": 0.9177, "step": 726 }, { "epoch": 0.3294064340734028, "grad_norm": 1.1756203303085342, "learning_rate": 5.2681159420289855e-05, "loss": 0.9536, "step": 727 }, { "epoch": 0.329859537834164, "grad_norm": 1.6962447510739282, "learning_rate": 5.27536231884058e-05, "loss": 0.9443, "step": 728 }, { "epoch": 0.33031264159492524, "grad_norm": 1.6666225029635953, "learning_rate": 5.282608695652174e-05, "loss": 0.9239, "step": 729 }, { "epoch": 0.3307657453556864, "grad_norm": 1.1725904538929959, "learning_rate": 5.289855072463769e-05, "loss": 0.9393, "step": 730 }, { "epoch": 0.33121884911644767, "grad_norm": 1.4176259487358853, "learning_rate": 5.297101449275363e-05, "loss": 0.9368, "step": 731 }, { "epoch": 0.33167195287720885, "grad_norm": 1.9390887820436906, "learning_rate": 5.304347826086957e-05, "loss": 0.9548, "step": 732 }, { "epoch": 0.3321250566379701, "grad_norm": 1.1270016640086127, "learning_rate": 5.3115942028985514e-05, "loss": 0.9452, "step": 733 }, { "epoch": 0.3325781603987313, "grad_norm": 1.4192801661202998, "learning_rate": 5.318840579710145e-05, "loss": 0.9461, "step": 734 }, { "epoch": 0.3330312641594925, "grad_norm": 1.2957232929400189, "learning_rate": 5.32608695652174e-05, "loss": 0.933, "step": 735 }, { "epoch": 0.33348436792025377, "grad_norm": 1.786647979654304, "learning_rate": 5.333333333333333e-05, "loss": 0.9395, "step": 736 }, { "epoch": 0.33393747168101495, "grad_norm": 1.6138638290006482, "learning_rate": 5.340579710144928e-05, "loss": 0.9388, "step": 737 }, { "epoch": 0.3343905754417762, "grad_norm": 1.1740256533417397, "learning_rate": 5.347826086956522e-05, "loss": 0.9465, "step": 738 }, { "epoch": 0.3348436792025374, "grad_norm": 1.4330828047663264, "learning_rate": 5.3550724637681166e-05, "loss": 0.9241, "step": 739 }, { "epoch": 0.3352967829632986, "grad_norm": 1.611121462776008, "learning_rate": 5.3623188405797106e-05, "loss": 0.9212, "step": 740 }, { "epoch": 0.3357498867240598, "grad_norm": 1.634068610721377, "learning_rate": 5.369565217391305e-05, "loss": 0.9214, "step": 741 }, { "epoch": 0.33620299048482105, "grad_norm": 1.4524798514423534, "learning_rate": 5.376811594202899e-05, "loss": 0.9505, "step": 742 }, { "epoch": 0.33665609424558224, "grad_norm": 1.1553383561310504, "learning_rate": 5.384057971014494e-05, "loss": 0.9332, "step": 743 }, { "epoch": 0.3371091980063435, "grad_norm": 1.7243136697467263, "learning_rate": 5.391304347826087e-05, "loss": 0.9531, "step": 744 }, { "epoch": 0.33756230176710467, "grad_norm": 1.592707186744106, "learning_rate": 5.398550724637682e-05, "loss": 0.9353, "step": 745 }, { "epoch": 0.3380154055278659, "grad_norm": 1.434010255409312, "learning_rate": 5.405797101449276e-05, "loss": 0.951, "step": 746 }, { "epoch": 0.3384685092886271, "grad_norm": 1.4155651223756864, "learning_rate": 5.41304347826087e-05, "loss": 0.9549, "step": 747 }, { "epoch": 0.33892161304938834, "grad_norm": 2.0050299127644484, "learning_rate": 5.4202898550724644e-05, "loss": 0.9397, "step": 748 }, { "epoch": 0.3393747168101495, "grad_norm": 0.8684399450478206, "learning_rate": 5.4275362318840584e-05, "loss": 0.9406, "step": 749 }, { "epoch": 0.33982782057091077, "grad_norm": 1.1930883401015249, "learning_rate": 5.434782608695653e-05, "loss": 0.915, "step": 750 }, { "epoch": 0.34028092433167195, "grad_norm": 1.8380063045410548, "learning_rate": 5.4420289855072464e-05, "loss": 0.9293, "step": 751 }, { "epoch": 0.3407340280924332, "grad_norm": 1.3859286987932946, "learning_rate": 5.449275362318841e-05, "loss": 0.9595, "step": 752 }, { "epoch": 0.3411871318531944, "grad_norm": 1.876822112818167, "learning_rate": 5.456521739130435e-05, "loss": 0.9489, "step": 753 }, { "epoch": 0.3416402356139556, "grad_norm": 0.9335416609978935, "learning_rate": 5.4637681159420296e-05, "loss": 0.9537, "step": 754 }, { "epoch": 0.3420933393747168, "grad_norm": 2.066370744460121, "learning_rate": 5.4710144927536236e-05, "loss": 0.934, "step": 755 }, { "epoch": 0.34254644313547805, "grad_norm": 1.3228605606799382, "learning_rate": 5.478260869565218e-05, "loss": 0.9387, "step": 756 }, { "epoch": 0.34299954689623924, "grad_norm": 2.3432140148794227, "learning_rate": 5.485507246376812e-05, "loss": 0.9372, "step": 757 }, { "epoch": 0.3434526506570005, "grad_norm": 2.179061961402425, "learning_rate": 5.492753623188407e-05, "loss": 0.9502, "step": 758 }, { "epoch": 0.34390575441776167, "grad_norm": 1.2649011394067426, "learning_rate": 5.5e-05, "loss": 0.9458, "step": 759 }, { "epoch": 0.3443588581785229, "grad_norm": 1.8662264482274864, "learning_rate": 5.507246376811594e-05, "loss": 0.9274, "step": 760 }, { "epoch": 0.3448119619392841, "grad_norm": 1.595595855412435, "learning_rate": 5.514492753623189e-05, "loss": 0.9228, "step": 761 }, { "epoch": 0.34526506570004534, "grad_norm": 1.3803054858900203, "learning_rate": 5.521739130434783e-05, "loss": 0.9384, "step": 762 }, { "epoch": 0.3457181694608065, "grad_norm": 1.7109267613043673, "learning_rate": 5.5289855072463775e-05, "loss": 0.9373, "step": 763 }, { "epoch": 0.34617127322156777, "grad_norm": 1.4526900989136082, "learning_rate": 5.5362318840579714e-05, "loss": 0.9273, "step": 764 }, { "epoch": 0.34662437698232895, "grad_norm": 1.2871192080584035, "learning_rate": 5.543478260869566e-05, "loss": 0.9428, "step": 765 }, { "epoch": 0.3470774807430902, "grad_norm": 1.5790769532128932, "learning_rate": 5.5507246376811594e-05, "loss": 0.9351, "step": 766 }, { "epoch": 0.3475305845038514, "grad_norm": 1.9519991759592048, "learning_rate": 5.557971014492754e-05, "loss": 0.974, "step": 767 }, { "epoch": 0.3479836882646126, "grad_norm": 1.213850862532891, "learning_rate": 5.565217391304348e-05, "loss": 0.9178, "step": 768 }, { "epoch": 0.3484367920253738, "grad_norm": 1.6275875157764281, "learning_rate": 5.572463768115943e-05, "loss": 0.9361, "step": 769 }, { "epoch": 0.34888989578613505, "grad_norm": 1.623468945110209, "learning_rate": 5.5797101449275366e-05, "loss": 0.9289, "step": 770 }, { "epoch": 0.34934299954689624, "grad_norm": 2.19380574282515, "learning_rate": 5.586956521739131e-05, "loss": 0.9403, "step": 771 }, { "epoch": 0.3497961033076575, "grad_norm": 0.9355644425581168, "learning_rate": 5.594202898550725e-05, "loss": 0.9296, "step": 772 }, { "epoch": 0.35024920706841867, "grad_norm": 1.8475810517396765, "learning_rate": 5.6014492753623186e-05, "loss": 0.9732, "step": 773 }, { "epoch": 0.3507023108291799, "grad_norm": 1.3221078631324086, "learning_rate": 5.608695652173913e-05, "loss": 0.9359, "step": 774 }, { "epoch": 0.3511554145899411, "grad_norm": 1.9122188619396188, "learning_rate": 5.615942028985507e-05, "loss": 0.9356, "step": 775 }, { "epoch": 0.35160851835070234, "grad_norm": 1.9409477612356576, "learning_rate": 5.623188405797102e-05, "loss": 0.9414, "step": 776 }, { "epoch": 0.3520616221114635, "grad_norm": 1.070465316010791, "learning_rate": 5.630434782608696e-05, "loss": 0.9234, "step": 777 }, { "epoch": 0.35251472587222477, "grad_norm": 3.1150654634514687, "learning_rate": 5.6376811594202905e-05, "loss": 0.9381, "step": 778 }, { "epoch": 0.35296782963298595, "grad_norm": 2.3376400691882138, "learning_rate": 5.6449275362318845e-05, "loss": 0.9499, "step": 779 }, { "epoch": 0.3534209333937472, "grad_norm": 2.6192309220254093, "learning_rate": 5.652173913043479e-05, "loss": 0.9259, "step": 780 }, { "epoch": 0.3538740371545084, "grad_norm": 2.0086977341030936, "learning_rate": 5.6594202898550724e-05, "loss": 0.9525, "step": 781 }, { "epoch": 0.3543271409152696, "grad_norm": 2.23699884007543, "learning_rate": 5.666666666666668e-05, "loss": 0.9254, "step": 782 }, { "epoch": 0.3547802446760308, "grad_norm": 1.9098965771732224, "learning_rate": 5.673913043478261e-05, "loss": 0.9457, "step": 783 }, { "epoch": 0.35523334843679205, "grad_norm": 1.349472921779123, "learning_rate": 5.681159420289856e-05, "loss": 0.933, "step": 784 }, { "epoch": 0.35568645219755324, "grad_norm": 2.503808185160977, "learning_rate": 5.68840579710145e-05, "loss": 0.9519, "step": 785 }, { "epoch": 0.3561395559583145, "grad_norm": 1.7056763076351638, "learning_rate": 5.695652173913044e-05, "loss": 0.9468, "step": 786 }, { "epoch": 0.35659265971907567, "grad_norm": 2.139487391590003, "learning_rate": 5.702898550724638e-05, "loss": 0.9292, "step": 787 }, { "epoch": 0.3570457634798369, "grad_norm": 2.158596118399532, "learning_rate": 5.710144927536232e-05, "loss": 0.9407, "step": 788 }, { "epoch": 0.3574988672405981, "grad_norm": 1.6837238817642566, "learning_rate": 5.717391304347827e-05, "loss": 0.9486, "step": 789 }, { "epoch": 0.35795197100135934, "grad_norm": 3.1457918895910457, "learning_rate": 5.72463768115942e-05, "loss": 0.9834, "step": 790 }, { "epoch": 0.3584050747621205, "grad_norm": 2.992660565056786, "learning_rate": 5.731884057971015e-05, "loss": 0.9356, "step": 791 }, { "epoch": 0.35885817852288177, "grad_norm": 1.9728568468943022, "learning_rate": 5.739130434782609e-05, "loss": 0.9211, "step": 792 }, { "epoch": 0.35931128228364295, "grad_norm": 1.4247683190286529, "learning_rate": 5.7463768115942035e-05, "loss": 0.9429, "step": 793 }, { "epoch": 0.3597643860444042, "grad_norm": 1.4596166969242923, "learning_rate": 5.7536231884057975e-05, "loss": 0.9439, "step": 794 }, { "epoch": 0.3602174898051654, "grad_norm": 2.410648582953631, "learning_rate": 5.760869565217392e-05, "loss": 0.9613, "step": 795 }, { "epoch": 0.3606705935659266, "grad_norm": 1.2751202503658967, "learning_rate": 5.768115942028986e-05, "loss": 0.9704, "step": 796 }, { "epoch": 0.3611236973266878, "grad_norm": 3.0176611396099142, "learning_rate": 5.775362318840581e-05, "loss": 0.9491, "step": 797 }, { "epoch": 0.36157680108744905, "grad_norm": 1.8589828634164278, "learning_rate": 5.782608695652174e-05, "loss": 0.9358, "step": 798 }, { "epoch": 0.36202990484821024, "grad_norm": 3.6909352873434043, "learning_rate": 5.789855072463769e-05, "loss": 0.9391, "step": 799 }, { "epoch": 0.3624830086089715, "grad_norm": 2.778783341955742, "learning_rate": 5.797101449275363e-05, "loss": 0.9245, "step": 800 }, { "epoch": 0.36293611236973267, "grad_norm": 3.2957760864111276, "learning_rate": 5.804347826086957e-05, "loss": 0.9392, "step": 801 }, { "epoch": 0.3633892161304939, "grad_norm": 3.3511920228348133, "learning_rate": 5.8115942028985513e-05, "loss": 0.9405, "step": 802 }, { "epoch": 0.3638423198912551, "grad_norm": 1.5537392224243713, "learning_rate": 5.818840579710145e-05, "loss": 0.9303, "step": 803 }, { "epoch": 0.36429542365201634, "grad_norm": 2.0095147665251387, "learning_rate": 5.82608695652174e-05, "loss": 0.9406, "step": 804 }, { "epoch": 0.3647485274127775, "grad_norm": 2.0858090439057753, "learning_rate": 5.833333333333333e-05, "loss": 0.9437, "step": 805 }, { "epoch": 0.36520163117353877, "grad_norm": 1.1760110666176509, "learning_rate": 5.840579710144928e-05, "loss": 0.9217, "step": 806 }, { "epoch": 0.36565473493429995, "grad_norm": 2.354004446832654, "learning_rate": 5.847826086956522e-05, "loss": 0.9291, "step": 807 }, { "epoch": 0.3661078386950612, "grad_norm": 1.2324968354604697, "learning_rate": 5.8550724637681166e-05, "loss": 0.9359, "step": 808 }, { "epoch": 0.3665609424558224, "grad_norm": 2.533069438481019, "learning_rate": 5.8623188405797105e-05, "loss": 0.9483, "step": 809 }, { "epoch": 0.3670140462165836, "grad_norm": 2.0561751733801086, "learning_rate": 5.869565217391305e-05, "loss": 0.9508, "step": 810 }, { "epoch": 0.3674671499773448, "grad_norm": 1.7901864284839302, "learning_rate": 5.876811594202899e-05, "loss": 0.9309, "step": 811 }, { "epoch": 0.36792025373810605, "grad_norm": 1.9881564251512267, "learning_rate": 5.884057971014494e-05, "loss": 0.9621, "step": 812 }, { "epoch": 0.36837335749886724, "grad_norm": 1.1399873804688112, "learning_rate": 5.891304347826087e-05, "loss": 0.9544, "step": 813 }, { "epoch": 0.3688264612596285, "grad_norm": 1.8117183544969502, "learning_rate": 5.898550724637682e-05, "loss": 0.9284, "step": 814 }, { "epoch": 0.36927956502038967, "grad_norm": 1.6430076048598186, "learning_rate": 5.905797101449276e-05, "loss": 0.9453, "step": 815 }, { "epoch": 0.3697326687811509, "grad_norm": 1.308569165591473, "learning_rate": 5.91304347826087e-05, "loss": 0.9282, "step": 816 }, { "epoch": 0.3701857725419121, "grad_norm": 1.07173590296738, "learning_rate": 5.9202898550724644e-05, "loss": 0.9346, "step": 817 }, { "epoch": 0.37063887630267334, "grad_norm": 1.384051600086882, "learning_rate": 5.9275362318840584e-05, "loss": 0.9472, "step": 818 }, { "epoch": 0.3710919800634345, "grad_norm": 1.280512319732141, "learning_rate": 5.934782608695653e-05, "loss": 0.9433, "step": 819 }, { "epoch": 0.37154508382419577, "grad_norm": 1.939337565520385, "learning_rate": 5.942028985507246e-05, "loss": 0.9493, "step": 820 }, { "epoch": 0.37199818758495695, "grad_norm": 1.3008016881897546, "learning_rate": 5.949275362318841e-05, "loss": 0.9317, "step": 821 }, { "epoch": 0.3724512913457182, "grad_norm": 1.8835361834111444, "learning_rate": 5.956521739130435e-05, "loss": 0.9129, "step": 822 }, { "epoch": 0.3729043951064794, "grad_norm": 1.6943053969089756, "learning_rate": 5.9637681159420296e-05, "loss": 0.9155, "step": 823 }, { "epoch": 0.3733574988672406, "grad_norm": 1.290000131065437, "learning_rate": 5.9710144927536236e-05, "loss": 0.9359, "step": 824 }, { "epoch": 0.3738106026280018, "grad_norm": 2.058038086850322, "learning_rate": 5.978260869565218e-05, "loss": 0.9413, "step": 825 }, { "epoch": 0.37426370638876305, "grad_norm": 1.4433659391987606, "learning_rate": 5.985507246376812e-05, "loss": 0.929, "step": 826 }, { "epoch": 0.37471681014952424, "grad_norm": 2.0124309249293426, "learning_rate": 5.992753623188407e-05, "loss": 0.954, "step": 827 }, { "epoch": 0.3751699139102855, "grad_norm": 1.6517980531245209, "learning_rate": 6.000000000000001e-05, "loss": 0.9339, "step": 828 }, { "epoch": 0.37562301767104667, "grad_norm": 1.6537831288817268, "learning_rate": 6.007246376811594e-05, "loss": 0.9327, "step": 829 }, { "epoch": 0.3760761214318079, "grad_norm": 1.4995714740142811, "learning_rate": 6.014492753623189e-05, "loss": 0.9487, "step": 830 }, { "epoch": 0.3765292251925691, "grad_norm": 1.480094881904478, "learning_rate": 6.021739130434783e-05, "loss": 0.9289, "step": 831 }, { "epoch": 0.37698232895333034, "grad_norm": 1.3592677016528816, "learning_rate": 6.0289855072463774e-05, "loss": 0.9327, "step": 832 }, { "epoch": 0.3774354327140915, "grad_norm": 1.5407937335819588, "learning_rate": 6.0362318840579714e-05, "loss": 0.93, "step": 833 }, { "epoch": 0.37788853647485277, "grad_norm": 1.60179245271006, "learning_rate": 6.043478260869566e-05, "loss": 0.9618, "step": 834 }, { "epoch": 0.37834164023561395, "grad_norm": 1.249493988907184, "learning_rate": 6.05072463768116e-05, "loss": 0.9361, "step": 835 }, { "epoch": 0.3787947439963752, "grad_norm": 1.94131201164259, "learning_rate": 6.057971014492755e-05, "loss": 0.9547, "step": 836 }, { "epoch": 0.3792478477571364, "grad_norm": 1.2069268894055598, "learning_rate": 6.065217391304348e-05, "loss": 0.9324, "step": 837 }, { "epoch": 0.3797009515178976, "grad_norm": 1.8276888864980547, "learning_rate": 6.0724637681159426e-05, "loss": 0.9517, "step": 838 }, { "epoch": 0.3801540552786588, "grad_norm": 1.2973497778627354, "learning_rate": 6.0797101449275366e-05, "loss": 0.9495, "step": 839 }, { "epoch": 0.38060715903942005, "grad_norm": 2.1452793164710395, "learning_rate": 6.086956521739131e-05, "loss": 0.9461, "step": 840 }, { "epoch": 0.38106026280018124, "grad_norm": 1.9363849687793837, "learning_rate": 6.094202898550725e-05, "loss": 0.9572, "step": 841 }, { "epoch": 0.3815133665609425, "grad_norm": 1.5252357062499307, "learning_rate": 6.101449275362319e-05, "loss": 0.9398, "step": 842 }, { "epoch": 0.38196647032170367, "grad_norm": 1.4130141407329584, "learning_rate": 6.108695652173914e-05, "loss": 0.916, "step": 843 }, { "epoch": 0.3824195740824649, "grad_norm": 1.5015629417926037, "learning_rate": 6.115942028985507e-05, "loss": 0.9229, "step": 844 }, { "epoch": 0.3828726778432261, "grad_norm": 1.1943980908639777, "learning_rate": 6.123188405797102e-05, "loss": 0.9151, "step": 845 }, { "epoch": 0.38332578160398734, "grad_norm": 1.6360467195419974, "learning_rate": 6.130434782608696e-05, "loss": 0.924, "step": 846 }, { "epoch": 0.3837788853647485, "grad_norm": 1.3374459334244837, "learning_rate": 6.137681159420291e-05, "loss": 0.9416, "step": 847 }, { "epoch": 0.38423198912550977, "grad_norm": 1.6292289099449635, "learning_rate": 6.144927536231884e-05, "loss": 0.9473, "step": 848 }, { "epoch": 0.38468509288627095, "grad_norm": 1.3697333645146674, "learning_rate": 6.152173913043479e-05, "loss": 0.9292, "step": 849 }, { "epoch": 0.3851381966470322, "grad_norm": 1.5906956403118555, "learning_rate": 6.159420289855072e-05, "loss": 0.9088, "step": 850 }, { "epoch": 0.3855913004077934, "grad_norm": 1.5635803597423978, "learning_rate": 6.166666666666667e-05, "loss": 0.9375, "step": 851 }, { "epoch": 0.3860444041685546, "grad_norm": 1.2308771115489177, "learning_rate": 6.173913043478262e-05, "loss": 0.9394, "step": 852 }, { "epoch": 0.3864975079293158, "grad_norm": 1.1529774210432973, "learning_rate": 6.181159420289856e-05, "loss": 0.9402, "step": 853 }, { "epoch": 0.38695061169007705, "grad_norm": 2.1113604315431287, "learning_rate": 6.18840579710145e-05, "loss": 0.9543, "step": 854 }, { "epoch": 0.38740371545083824, "grad_norm": 1.5152165835187914, "learning_rate": 6.195652173913044e-05, "loss": 0.9279, "step": 855 }, { "epoch": 0.3878568192115995, "grad_norm": 1.008873490584319, "learning_rate": 6.202898550724638e-05, "loss": 0.9432, "step": 856 }, { "epoch": 0.38830992297236067, "grad_norm": 2.063618151425557, "learning_rate": 6.210144927536232e-05, "loss": 0.935, "step": 857 }, { "epoch": 0.3887630267331219, "grad_norm": 1.0778446588476511, "learning_rate": 6.217391304347827e-05, "loss": 0.935, "step": 858 }, { "epoch": 0.3892161304938831, "grad_norm": 2.4160903987805376, "learning_rate": 6.22463768115942e-05, "loss": 0.9549, "step": 859 }, { "epoch": 0.38966923425464434, "grad_norm": 1.677306195360877, "learning_rate": 6.231884057971015e-05, "loss": 0.9503, "step": 860 }, { "epoch": 0.3901223380154055, "grad_norm": 1.9447602622590532, "learning_rate": 6.23913043478261e-05, "loss": 0.9474, "step": 861 }, { "epoch": 0.39057544177616677, "grad_norm": 1.7807351460608432, "learning_rate": 6.246376811594204e-05, "loss": 0.9469, "step": 862 }, { "epoch": 0.39102854553692795, "grad_norm": 1.6577138738129278, "learning_rate": 6.253623188405797e-05, "loss": 0.9654, "step": 863 }, { "epoch": 0.3914816492976892, "grad_norm": 1.002402848465321, "learning_rate": 6.260869565217392e-05, "loss": 0.9245, "step": 864 }, { "epoch": 0.3919347530584504, "grad_norm": 1.3721167197261497, "learning_rate": 6.268115942028985e-05, "loss": 0.9426, "step": 865 }, { "epoch": 0.3923878568192116, "grad_norm": 1.5915764408415682, "learning_rate": 6.27536231884058e-05, "loss": 0.947, "step": 866 }, { "epoch": 0.3928409605799728, "grad_norm": 1.5392301797038366, "learning_rate": 6.282608695652175e-05, "loss": 0.9265, "step": 867 }, { "epoch": 0.39329406434073405, "grad_norm": 1.8909141291394924, "learning_rate": 6.28985507246377e-05, "loss": 0.941, "step": 868 }, { "epoch": 0.39374716810149524, "grad_norm": 1.0344926440241522, "learning_rate": 6.297101449275363e-05, "loss": 0.9437, "step": 869 }, { "epoch": 0.3942002718622565, "grad_norm": 1.6843483901363272, "learning_rate": 6.304347826086956e-05, "loss": 0.9455, "step": 870 }, { "epoch": 0.39465337562301767, "grad_norm": 1.264500625016654, "learning_rate": 6.31159420289855e-05, "loss": 0.9491, "step": 871 }, { "epoch": 0.3951064793837789, "grad_norm": 2.03649675741356, "learning_rate": 6.318840579710145e-05, "loss": 0.9415, "step": 872 }, { "epoch": 0.3955595831445401, "grad_norm": 1.7348022615249474, "learning_rate": 6.32608695652174e-05, "loss": 0.9341, "step": 873 }, { "epoch": 0.39601268690530134, "grad_norm": 1.417120579435153, "learning_rate": 6.333333333333333e-05, "loss": 0.9356, "step": 874 }, { "epoch": 0.3964657906660625, "grad_norm": 1.3227505528432073, "learning_rate": 6.340579710144928e-05, "loss": 0.9204, "step": 875 }, { "epoch": 0.39691889442682377, "grad_norm": 1.449955277037412, "learning_rate": 6.347826086956523e-05, "loss": 0.9194, "step": 876 }, { "epoch": 0.39737199818758495, "grad_norm": 1.5223901881009263, "learning_rate": 6.355072463768117e-05, "loss": 0.9589, "step": 877 }, { "epoch": 0.3978251019483462, "grad_norm": 2.083168438504065, "learning_rate": 6.36231884057971e-05, "loss": 0.9293, "step": 878 }, { "epoch": 0.3982782057091074, "grad_norm": 1.396526966533763, "learning_rate": 6.369565217391305e-05, "loss": 0.9357, "step": 879 }, { "epoch": 0.3987313094698686, "grad_norm": 1.693071442654607, "learning_rate": 6.376811594202898e-05, "loss": 0.9526, "step": 880 }, { "epoch": 0.3991844132306298, "grad_norm": 1.7314602906145682, "learning_rate": 6.384057971014493e-05, "loss": 0.9123, "step": 881 }, { "epoch": 0.39963751699139105, "grad_norm": 1.2042869737402027, "learning_rate": 6.391304347826088e-05, "loss": 0.9325, "step": 882 }, { "epoch": 0.40009062075215224, "grad_norm": 1.9775396209203333, "learning_rate": 6.398550724637682e-05, "loss": 0.9201, "step": 883 }, { "epoch": 0.4005437245129135, "grad_norm": 1.7772781610458723, "learning_rate": 6.405797101449276e-05, "loss": 0.931, "step": 884 }, { "epoch": 0.40099682827367467, "grad_norm": 0.7820554856389218, "learning_rate": 6.41304347826087e-05, "loss": 0.9401, "step": 885 }, { "epoch": 0.4014499320344359, "grad_norm": 2.613958077467502, "learning_rate": 6.420289855072465e-05, "loss": 0.9461, "step": 886 }, { "epoch": 0.4019030357951971, "grad_norm": 1.4402524532567318, "learning_rate": 6.427536231884058e-05, "loss": 0.9378, "step": 887 }, { "epoch": 0.40235613955595834, "grad_norm": 2.4880106714492154, "learning_rate": 6.434782608695653e-05, "loss": 0.9509, "step": 888 }, { "epoch": 0.4028092433167195, "grad_norm": 1.831688510406157, "learning_rate": 6.442028985507246e-05, "loss": 0.9372, "step": 889 }, { "epoch": 0.40326234707748077, "grad_norm": 2.132277572157772, "learning_rate": 6.449275362318841e-05, "loss": 0.9267, "step": 890 }, { "epoch": 0.40371545083824195, "grad_norm": 1.4306500804882745, "learning_rate": 6.456521739130436e-05, "loss": 0.9328, "step": 891 }, { "epoch": 0.4041685545990032, "grad_norm": 2.3390014821251737, "learning_rate": 6.46376811594203e-05, "loss": 0.9265, "step": 892 }, { "epoch": 0.4046216583597644, "grad_norm": 1.7122887600236614, "learning_rate": 6.471014492753624e-05, "loss": 0.9516, "step": 893 }, { "epoch": 0.4050747621205256, "grad_norm": 2.1717625349058496, "learning_rate": 6.478260869565218e-05, "loss": 0.9315, "step": 894 }, { "epoch": 0.4055278658812868, "grad_norm": 1.4815506828709013, "learning_rate": 6.485507246376811e-05, "loss": 0.9254, "step": 895 }, { "epoch": 0.40598096964204805, "grad_norm": 80.53866418587752, "learning_rate": 6.492753623188406e-05, "loss": 0.9644, "step": 896 }, { "epoch": 0.40643407340280924, "grad_norm": 1365.6454401321132, "learning_rate": 6.500000000000001e-05, "loss": 1.5382, "step": 897 }, { "epoch": 0.4068871771635705, "grad_norm": 8.135967227617344, "learning_rate": 6.507246376811594e-05, "loss": 1.0987, "step": 898 }, { "epoch": 0.40734028092433167, "grad_norm": 117.83109854026416, "learning_rate": 6.514492753623189e-05, "loss": 1.251, "step": 899 }, { "epoch": 0.4077933846850929, "grad_norm": 5.167939163783073, "learning_rate": 6.521739130434783e-05, "loss": 1.0297, "step": 900 }, { "epoch": 0.4082464884458541, "grad_norm": 1.7723319248046467, "learning_rate": 6.528985507246378e-05, "loss": 0.9718, "step": 901 }, { "epoch": 0.40869959220661534, "grad_norm": 6.916449755395945, "learning_rate": 6.536231884057971e-05, "loss": 0.9838, "step": 902 }, { "epoch": 0.4091526959673765, "grad_norm": 5.368501700894766, "learning_rate": 6.543478260869566e-05, "loss": 1.0097, "step": 903 }, { "epoch": 0.40960579972813776, "grad_norm": 3.9328646742285565, "learning_rate": 6.550724637681159e-05, "loss": 1.0174, "step": 904 }, { "epoch": 0.41005890348889895, "grad_norm": 2.2571918163584144, "learning_rate": 6.557971014492754e-05, "loss": 0.9734, "step": 905 }, { "epoch": 0.4105120072496602, "grad_norm": 1.8707279476550231, "learning_rate": 6.565217391304349e-05, "loss": 0.9706, "step": 906 }, { "epoch": 0.4109651110104214, "grad_norm": 1.8475705526382034, "learning_rate": 6.572463768115943e-05, "loss": 0.9754, "step": 907 }, { "epoch": 0.4114182147711826, "grad_norm": 1.3148303310379195, "learning_rate": 6.579710144927537e-05, "loss": 0.9626, "step": 908 }, { "epoch": 0.4118713185319438, "grad_norm": 1.484422786847781, "learning_rate": 6.586956521739131e-05, "loss": 0.9558, "step": 909 }, { "epoch": 0.41232442229270505, "grad_norm": 1.669093474071442, "learning_rate": 6.594202898550725e-05, "loss": 0.9549, "step": 910 }, { "epoch": 0.41277752605346624, "grad_norm": 1.3792683702919502, "learning_rate": 6.601449275362319e-05, "loss": 0.9599, "step": 911 }, { "epoch": 0.4132306298142275, "grad_norm": 2.0854716916112563, "learning_rate": 6.608695652173914e-05, "loss": 0.9535, "step": 912 }, { "epoch": 0.41368373357498867, "grad_norm": 1.4667759321841645, "learning_rate": 6.615942028985507e-05, "loss": 0.9535, "step": 913 }, { "epoch": 0.4141368373357499, "grad_norm": 1.7589368493563393, "learning_rate": 6.623188405797102e-05, "loss": 0.9449, "step": 914 }, { "epoch": 0.4145899410965111, "grad_norm": 1.2569856659313055, "learning_rate": 6.630434782608696e-05, "loss": 0.9453, "step": 915 }, { "epoch": 0.41504304485727234, "grad_norm": 1.8629480684192772, "learning_rate": 6.637681159420291e-05, "loss": 0.9496, "step": 916 }, { "epoch": 0.4154961486180335, "grad_norm": 1.2438452987986774, "learning_rate": 6.644927536231884e-05, "loss": 0.9536, "step": 917 }, { "epoch": 0.41594925237879476, "grad_norm": 2.2563782699074886, "learning_rate": 6.652173913043479e-05, "loss": 0.9541, "step": 918 }, { "epoch": 0.41640235613955595, "grad_norm": 1.5551993045041272, "learning_rate": 6.659420289855072e-05, "loss": 0.9437, "step": 919 }, { "epoch": 0.4168554599003172, "grad_norm": 2.0355052906614786, "learning_rate": 6.666666666666667e-05, "loss": 0.9404, "step": 920 }, { "epoch": 0.4173085636610784, "grad_norm": 1.682511294159359, "learning_rate": 6.673913043478262e-05, "loss": 0.9363, "step": 921 }, { "epoch": 0.4177616674218396, "grad_norm": 1.7206675736587544, "learning_rate": 6.681159420289856e-05, "loss": 0.9173, "step": 922 }, { "epoch": 0.4182147711826008, "grad_norm": 1.3517955985197665, "learning_rate": 6.68840579710145e-05, "loss": 0.9562, "step": 923 }, { "epoch": 0.41866787494336205, "grad_norm": 1.4766389289584916, "learning_rate": 6.695652173913044e-05, "loss": 0.9291, "step": 924 }, { "epoch": 0.41912097870412324, "grad_norm": 1.5191436100763074, "learning_rate": 6.702898550724639e-05, "loss": 0.9478, "step": 925 }, { "epoch": 0.4195740824648845, "grad_norm": 1.1456819282324526, "learning_rate": 6.710144927536232e-05, "loss": 0.9478, "step": 926 }, { "epoch": 0.42002718622564567, "grad_norm": 1.788871418539645, "learning_rate": 6.717391304347827e-05, "loss": 0.934, "step": 927 }, { "epoch": 0.4204802899864069, "grad_norm": 1.3715433148641785, "learning_rate": 6.72463768115942e-05, "loss": 0.9592, "step": 928 }, { "epoch": 0.4209333937471681, "grad_norm": 2.2559204630295455, "learning_rate": 6.731884057971015e-05, "loss": 0.9381, "step": 929 }, { "epoch": 0.42138649750792934, "grad_norm": 1.8192111092638497, "learning_rate": 6.73913043478261e-05, "loss": 0.9396, "step": 930 }, { "epoch": 0.4218396012686905, "grad_norm": 1.9105490667473406, "learning_rate": 6.746376811594204e-05, "loss": 0.9335, "step": 931 }, { "epoch": 0.42229270502945176, "grad_norm": 1.6917980285927041, "learning_rate": 6.753623188405797e-05, "loss": 0.9542, "step": 932 }, { "epoch": 0.42274580879021295, "grad_norm": 1.6868996598956045, "learning_rate": 6.760869565217392e-05, "loss": 0.9382, "step": 933 }, { "epoch": 0.4231989125509742, "grad_norm": 1.3477973877511678, "learning_rate": 6.768115942028985e-05, "loss": 0.943, "step": 934 }, { "epoch": 0.4236520163117354, "grad_norm": 1.8308489548304965, "learning_rate": 6.77536231884058e-05, "loss": 0.9431, "step": 935 }, { "epoch": 0.4241051200724966, "grad_norm": 1.6778856883377098, "learning_rate": 6.782608695652175e-05, "loss": 0.9253, "step": 936 }, { "epoch": 0.4245582238332578, "grad_norm": 1.4415304847400303, "learning_rate": 6.78985507246377e-05, "loss": 0.937, "step": 937 }, { "epoch": 0.42501132759401905, "grad_norm": 1.3529932375463645, "learning_rate": 6.797101449275363e-05, "loss": 0.9207, "step": 938 }, { "epoch": 0.42546443135478024, "grad_norm": 1.6826393619607434, "learning_rate": 6.804347826086957e-05, "loss": 0.9149, "step": 939 }, { "epoch": 0.4259175351155415, "grad_norm": 1.3707549228307465, "learning_rate": 6.811594202898552e-05, "loss": 0.9106, "step": 940 }, { "epoch": 0.42637063887630267, "grad_norm": 1.83045020426448, "learning_rate": 6.818840579710145e-05, "loss": 0.9229, "step": 941 }, { "epoch": 0.4268237426370639, "grad_norm": 1.1932482021974216, "learning_rate": 6.82608695652174e-05, "loss": 0.925, "step": 942 }, { "epoch": 0.4272768463978251, "grad_norm": 1.799197321722904, "learning_rate": 6.833333333333333e-05, "loss": 0.9417, "step": 943 }, { "epoch": 0.42772995015858634, "grad_norm": 1.4602475384455642, "learning_rate": 6.840579710144928e-05, "loss": 0.9207, "step": 944 }, { "epoch": 0.4281830539193475, "grad_norm": 1.6199149741205103, "learning_rate": 6.847826086956522e-05, "loss": 0.9325, "step": 945 }, { "epoch": 0.42863615768010876, "grad_norm": 1.2913968310575628, "learning_rate": 6.855072463768117e-05, "loss": 0.9365, "step": 946 }, { "epoch": 0.42908926144086995, "grad_norm": 1.4908197014320408, "learning_rate": 6.86231884057971e-05, "loss": 0.9301, "step": 947 }, { "epoch": 0.4295423652016312, "grad_norm": 1.253268413902169, "learning_rate": 6.869565217391305e-05, "loss": 0.939, "step": 948 }, { "epoch": 0.4299954689623924, "grad_norm": 1.5448221408319804, "learning_rate": 6.876811594202898e-05, "loss": 0.9425, "step": 949 }, { "epoch": 0.4304485727231536, "grad_norm": 1.282706601879041, "learning_rate": 6.884057971014493e-05, "loss": 0.9226, "step": 950 }, { "epoch": 0.4309016764839148, "grad_norm": 1.3099769328907773, "learning_rate": 6.891304347826088e-05, "loss": 0.9308, "step": 951 }, { "epoch": 0.43135478024467605, "grad_norm": 1.9469614365456036, "learning_rate": 6.898550724637682e-05, "loss": 0.9166, "step": 952 }, { "epoch": 0.43180788400543724, "grad_norm": 1.630544983261556, "learning_rate": 6.905797101449276e-05, "loss": 0.9369, "step": 953 }, { "epoch": 0.4322609877661985, "grad_norm": 1.1823524423512757, "learning_rate": 6.91304347826087e-05, "loss": 0.9127, "step": 954 }, { "epoch": 0.43271409152695967, "grad_norm": 1.4610030018326292, "learning_rate": 6.920289855072465e-05, "loss": 0.9214, "step": 955 }, { "epoch": 0.4331671952877209, "grad_norm": 2.167615917113872, "learning_rate": 6.927536231884058e-05, "loss": 0.9523, "step": 956 }, { "epoch": 0.4336202990484821, "grad_norm": 1.1332962465228662, "learning_rate": 6.934782608695653e-05, "loss": 0.9417, "step": 957 }, { "epoch": 0.43407340280924334, "grad_norm": 3.7402003587546084, "learning_rate": 6.942028985507246e-05, "loss": 0.9777, "step": 958 }, { "epoch": 0.4345265065700045, "grad_norm": 1.6868743828850994, "learning_rate": 6.949275362318841e-05, "loss": 0.9516, "step": 959 }, { "epoch": 0.43497961033076576, "grad_norm": 1.5771248855235687, "learning_rate": 6.956521739130436e-05, "loss": 0.9298, "step": 960 }, { "epoch": 0.43543271409152695, "grad_norm": 1.879543482787959, "learning_rate": 6.96376811594203e-05, "loss": 0.928, "step": 961 }, { "epoch": 0.4358858178522882, "grad_norm": 1.803033232549323, "learning_rate": 6.971014492753623e-05, "loss": 0.9554, "step": 962 }, { "epoch": 0.4363389216130494, "grad_norm": 1.226566994598829, "learning_rate": 6.978260869565218e-05, "loss": 0.934, "step": 963 }, { "epoch": 0.4367920253738106, "grad_norm": 1.7873077754931423, "learning_rate": 6.985507246376811e-05, "loss": 0.9304, "step": 964 }, { "epoch": 0.4372451291345718, "grad_norm": 1.1541084262691084, "learning_rate": 6.992753623188407e-05, "loss": 0.9286, "step": 965 }, { "epoch": 0.43769823289533305, "grad_norm": 2.3947133097913222, "learning_rate": 7.000000000000001e-05, "loss": 0.9212, "step": 966 }, { "epoch": 0.43815133665609424, "grad_norm": 1.7021442849413693, "learning_rate": 7.007246376811594e-05, "loss": 0.9365, "step": 967 }, { "epoch": 0.4386044404168555, "grad_norm": 1.789442893156277, "learning_rate": 7.014492753623189e-05, "loss": 0.9262, "step": 968 }, { "epoch": 0.43905754417761667, "grad_norm": 1.7858594296186456, "learning_rate": 7.021739130434783e-05, "loss": 0.9272, "step": 969 }, { "epoch": 0.4395106479383779, "grad_norm": 1.586856065859088, "learning_rate": 7.028985507246378e-05, "loss": 0.9214, "step": 970 }, { "epoch": 0.4399637516991391, "grad_norm": 1.6060148389908928, "learning_rate": 7.036231884057971e-05, "loss": 0.9168, "step": 971 }, { "epoch": 0.44041685545990034, "grad_norm": 1.6693423087000894, "learning_rate": 7.043478260869566e-05, "loss": 0.9344, "step": 972 }, { "epoch": 0.4408699592206615, "grad_norm": 1.1868072270550278, "learning_rate": 7.050724637681159e-05, "loss": 0.9407, "step": 973 }, { "epoch": 0.44132306298142276, "grad_norm": 1.534502914434514, "learning_rate": 7.057971014492754e-05, "loss": 0.9293, "step": 974 }, { "epoch": 0.44177616674218395, "grad_norm": 2.057607098734167, "learning_rate": 7.065217391304349e-05, "loss": 0.9438, "step": 975 }, { "epoch": 0.4422292705029452, "grad_norm": 1.1085356343411727, "learning_rate": 7.072463768115943e-05, "loss": 0.9488, "step": 976 }, { "epoch": 0.4426823742637064, "grad_norm": 2.0793829306622036, "learning_rate": 7.079710144927537e-05, "loss": 0.952, "step": 977 }, { "epoch": 0.4431354780244676, "grad_norm": 1.254086293088225, "learning_rate": 7.086956521739131e-05, "loss": 0.9429, "step": 978 }, { "epoch": 0.4435885817852288, "grad_norm": 2.3730249542358317, "learning_rate": 7.094202898550726e-05, "loss": 0.9337, "step": 979 }, { "epoch": 0.44404168554599005, "grad_norm": 1.9252641334163636, "learning_rate": 7.101449275362319e-05, "loss": 0.931, "step": 980 }, { "epoch": 0.44449478930675124, "grad_norm": 2.1316639658151395, "learning_rate": 7.108695652173914e-05, "loss": 0.9231, "step": 981 }, { "epoch": 0.4449478930675125, "grad_norm": 1.750147951610209, "learning_rate": 7.115942028985507e-05, "loss": 0.9258, "step": 982 }, { "epoch": 0.44540099682827367, "grad_norm": 1.7905150244551946, "learning_rate": 7.123188405797102e-05, "loss": 0.9405, "step": 983 }, { "epoch": 0.4458541005890349, "grad_norm": 1.3251709560885185, "learning_rate": 7.130434782608696e-05, "loss": 0.9291, "step": 984 }, { "epoch": 0.4463072043497961, "grad_norm": 1.7613157967690332, "learning_rate": 7.137681159420291e-05, "loss": 0.9381, "step": 985 }, { "epoch": 0.44676030811055734, "grad_norm": 1.1139186498772446, "learning_rate": 7.144927536231884e-05, "loss": 0.9447, "step": 986 }, { "epoch": 0.4472134118713185, "grad_norm": 1.8498069078750057, "learning_rate": 7.152173913043479e-05, "loss": 0.9339, "step": 987 }, { "epoch": 0.44766651563207976, "grad_norm": 1.4165466026927083, "learning_rate": 7.159420289855072e-05, "loss": 0.9351, "step": 988 }, { "epoch": 0.44811961939284095, "grad_norm": 1.7480034586871962, "learning_rate": 7.166666666666667e-05, "loss": 0.9296, "step": 989 }, { "epoch": 0.4485727231536022, "grad_norm": 1.42196785351496, "learning_rate": 7.173913043478262e-05, "loss": 0.9287, "step": 990 }, { "epoch": 0.4490258269143634, "grad_norm": 1.9369647592301418, "learning_rate": 7.181159420289856e-05, "loss": 0.9204, "step": 991 }, { "epoch": 0.4494789306751246, "grad_norm": 1.5116818700444048, "learning_rate": 7.18840579710145e-05, "loss": 0.9505, "step": 992 }, { "epoch": 0.4499320344358858, "grad_norm": 1.2156969095439616, "learning_rate": 7.195652173913044e-05, "loss": 0.9135, "step": 993 }, { "epoch": 0.45038513819664705, "grad_norm": 1.4505394594595604, "learning_rate": 7.202898550724639e-05, "loss": 0.9441, "step": 994 }, { "epoch": 0.45083824195740824, "grad_norm": 1.2724279367706872, "learning_rate": 7.210144927536232e-05, "loss": 0.9283, "step": 995 }, { "epoch": 0.4512913457181695, "grad_norm": 1.602944162915794, "learning_rate": 7.217391304347827e-05, "loss": 0.9262, "step": 996 }, { "epoch": 0.45174444947893067, "grad_norm": 1.8152393327162428, "learning_rate": 7.22463768115942e-05, "loss": 0.9386, "step": 997 }, { "epoch": 0.4521975532396919, "grad_norm": 1.4385163495940938, "learning_rate": 7.231884057971015e-05, "loss": 0.9557, "step": 998 }, { "epoch": 0.4526506570004531, "grad_norm": 1.5725979959006076, "learning_rate": 7.23913043478261e-05, "loss": 0.9547, "step": 999 }, { "epoch": 0.45310376076121434, "grad_norm": 1.0169784986409975, "learning_rate": 7.246376811594204e-05, "loss": 0.9098, "step": 1000 }, { "epoch": 0.4535568645219755, "grad_norm": 1.5213680224346657, "learning_rate": 7.253623188405797e-05, "loss": 0.9248, "step": 1001 }, { "epoch": 0.45400996828273676, "grad_norm": 1.474537429633555, "learning_rate": 7.260869565217392e-05, "loss": 0.9207, "step": 1002 }, { "epoch": 0.45446307204349795, "grad_norm": 1.6981160475715924, "learning_rate": 7.268115942028985e-05, "loss": 0.9635, "step": 1003 }, { "epoch": 0.4549161758042592, "grad_norm": 1.2649593761671898, "learning_rate": 7.27536231884058e-05, "loss": 0.924, "step": 1004 }, { "epoch": 0.4553692795650204, "grad_norm": 2.293011028419649, "learning_rate": 7.282608695652175e-05, "loss": 0.9206, "step": 1005 }, { "epoch": 0.4558223833257816, "grad_norm": 1.4153113266402753, "learning_rate": 7.289855072463769e-05, "loss": 0.9057, "step": 1006 }, { "epoch": 0.4562754870865428, "grad_norm": 2.124227829194283, "learning_rate": 7.297101449275363e-05, "loss": 0.9276, "step": 1007 }, { "epoch": 0.45672859084730405, "grad_norm": 1.985938178114726, "learning_rate": 7.304347826086957e-05, "loss": 0.9396, "step": 1008 }, { "epoch": 0.45718169460806524, "grad_norm": 1.8856200136211894, "learning_rate": 7.311594202898552e-05, "loss": 0.9325, "step": 1009 }, { "epoch": 0.4576347983688265, "grad_norm": 1.6059225747785522, "learning_rate": 7.318840579710145e-05, "loss": 0.9573, "step": 1010 }, { "epoch": 0.45808790212958767, "grad_norm": 1.2847981756384494, "learning_rate": 7.32608695652174e-05, "loss": 0.9219, "step": 1011 }, { "epoch": 0.4585410058903489, "grad_norm": 2.3434063694975977, "learning_rate": 7.333333333333333e-05, "loss": 0.9391, "step": 1012 }, { "epoch": 0.4589941096511101, "grad_norm": 1.2243500928178161, "learning_rate": 7.340579710144928e-05, "loss": 0.9242, "step": 1013 }, { "epoch": 0.45944721341187134, "grad_norm": 1.418666371165993, "learning_rate": 7.347826086956522e-05, "loss": 0.9426, "step": 1014 }, { "epoch": 0.4599003171726325, "grad_norm": 1.9571814777298087, "learning_rate": 7.355072463768117e-05, "loss": 0.9443, "step": 1015 }, { "epoch": 0.46035342093339376, "grad_norm": 1.0033310590135422, "learning_rate": 7.36231884057971e-05, "loss": 0.9289, "step": 1016 }, { "epoch": 0.46080652469415495, "grad_norm": 2.7444112215763066, "learning_rate": 7.369565217391305e-05, "loss": 0.9599, "step": 1017 }, { "epoch": 0.4612596284549162, "grad_norm": 2.3628667242457047, "learning_rate": 7.3768115942029e-05, "loss": 0.9752, "step": 1018 }, { "epoch": 0.4617127322156774, "grad_norm": 1.8427495984741853, "learning_rate": 7.384057971014494e-05, "loss": 0.9492, "step": 1019 }, { "epoch": 0.4621658359764386, "grad_norm": 2.1592181974062408, "learning_rate": 7.391304347826088e-05, "loss": 0.9126, "step": 1020 }, { "epoch": 0.4626189397371998, "grad_norm": 1.5982907089082694, "learning_rate": 7.398550724637682e-05, "loss": 0.9393, "step": 1021 }, { "epoch": 0.46307204349796105, "grad_norm": 1.6315664581185108, "learning_rate": 7.405797101449276e-05, "loss": 0.9449, "step": 1022 }, { "epoch": 0.46352514725872224, "grad_norm": 1.749472257236361, "learning_rate": 7.41304347826087e-05, "loss": 0.9358, "step": 1023 }, { "epoch": 0.4639782510194835, "grad_norm": 1.7801092810746144, "learning_rate": 7.420289855072465e-05, "loss": 0.9304, "step": 1024 }, { "epoch": 0.46443135478024467, "grad_norm": 0.963200763349096, "learning_rate": 7.427536231884058e-05, "loss": 0.9256, "step": 1025 }, { "epoch": 0.4648844585410059, "grad_norm": 1.4237263367805253, "learning_rate": 7.434782608695653e-05, "loss": 0.9487, "step": 1026 }, { "epoch": 0.4653375623017671, "grad_norm": 1.1760028195585521, "learning_rate": 7.442028985507246e-05, "loss": 0.9282, "step": 1027 }, { "epoch": 0.46579066606252834, "grad_norm": 1.5667699359776726, "learning_rate": 7.449275362318841e-05, "loss": 0.951, "step": 1028 }, { "epoch": 0.4662437698232895, "grad_norm": 1.6861621320648823, "learning_rate": 7.456521739130435e-05, "loss": 0.964, "step": 1029 }, { "epoch": 0.46669687358405076, "grad_norm": 1.78185718067046, "learning_rate": 7.46376811594203e-05, "loss": 0.9364, "step": 1030 }, { "epoch": 0.46714997734481195, "grad_norm": 1.3835472413049803, "learning_rate": 7.471014492753623e-05, "loss": 0.9197, "step": 1031 }, { "epoch": 0.4676030811055732, "grad_norm": 0.8893185026305293, "learning_rate": 7.478260869565218e-05, "loss": 0.9363, "step": 1032 }, { "epoch": 0.4680561848663344, "grad_norm": 1.1085109846892507, "learning_rate": 7.485507246376813e-05, "loss": 0.9148, "step": 1033 }, { "epoch": 0.4685092886270956, "grad_norm": 1.8444156293689467, "learning_rate": 7.492753623188407e-05, "loss": 0.9105, "step": 1034 }, { "epoch": 0.4689623923878568, "grad_norm": 1.5976567054684865, "learning_rate": 7.500000000000001e-05, "loss": 0.9296, "step": 1035 }, { "epoch": 0.46941549614861805, "grad_norm": 1.6813885351151359, "learning_rate": 7.507246376811594e-05, "loss": 0.9175, "step": 1036 }, { "epoch": 0.46986859990937924, "grad_norm": 0.9988847775174384, "learning_rate": 7.514492753623189e-05, "loss": 0.9324, "step": 1037 }, { "epoch": 0.4703217036701405, "grad_norm": 1.9367815499301186, "learning_rate": 7.521739130434783e-05, "loss": 0.9709, "step": 1038 }, { "epoch": 0.47077480743090167, "grad_norm": 1.636965311194167, "learning_rate": 7.528985507246378e-05, "loss": 0.9331, "step": 1039 }, { "epoch": 0.4712279111916629, "grad_norm": 1.6707917494114104, "learning_rate": 7.536231884057971e-05, "loss": 0.952, "step": 1040 }, { "epoch": 0.4716810149524241, "grad_norm": 1.192744751263504, "learning_rate": 7.543478260869566e-05, "loss": 0.9545, "step": 1041 }, { "epoch": 0.47213411871318534, "grad_norm": 1.375918123266685, "learning_rate": 7.550724637681159e-05, "loss": 0.938, "step": 1042 }, { "epoch": 0.4725872224739465, "grad_norm": 1.429530509308327, "learning_rate": 7.557971014492754e-05, "loss": 0.9356, "step": 1043 }, { "epoch": 0.47304032623470776, "grad_norm": 2.0603762643779797, "learning_rate": 7.565217391304349e-05, "loss": 0.9214, "step": 1044 }, { "epoch": 0.47349342999546895, "grad_norm": 1.3461622494527974, "learning_rate": 7.572463768115943e-05, "loss": 0.9504, "step": 1045 }, { "epoch": 0.4739465337562302, "grad_norm": 1.2943611730700801, "learning_rate": 7.579710144927536e-05, "loss": 0.9213, "step": 1046 }, { "epoch": 0.4743996375169914, "grad_norm": 1.0480117033781313, "learning_rate": 7.586956521739131e-05, "loss": 0.9103, "step": 1047 }, { "epoch": 0.4748527412777526, "grad_norm": 1.4849058534530761, "learning_rate": 7.594202898550726e-05, "loss": 0.9342, "step": 1048 }, { "epoch": 0.4753058450385138, "grad_norm": 1.8561096411646143, "learning_rate": 7.601449275362319e-05, "loss": 0.914, "step": 1049 }, { "epoch": 0.47575894879927505, "grad_norm": 1.356435385915571, "learning_rate": 7.608695652173914e-05, "loss": 0.9142, "step": 1050 }, { "epoch": 0.47621205256003624, "grad_norm": 1.7642362441329629, "learning_rate": 7.615942028985507e-05, "loss": 0.9289, "step": 1051 }, { "epoch": 0.4766651563207975, "grad_norm": 1.5416131574271534, "learning_rate": 7.623188405797102e-05, "loss": 0.938, "step": 1052 }, { "epoch": 0.47711826008155866, "grad_norm": 1.494379645607806, "learning_rate": 7.630434782608696e-05, "loss": 0.9174, "step": 1053 }, { "epoch": 0.4775713638423199, "grad_norm": 1.5546777138922556, "learning_rate": 7.637681159420291e-05, "loss": 0.9421, "step": 1054 }, { "epoch": 0.4780244676030811, "grad_norm": 1.2789758002539167, "learning_rate": 7.644927536231884e-05, "loss": 0.9181, "step": 1055 }, { "epoch": 0.47847757136384234, "grad_norm": 2.4330708217230606, "learning_rate": 7.652173913043479e-05, "loss": 0.9299, "step": 1056 }, { "epoch": 0.4789306751246035, "grad_norm": 0.9268739723244087, "learning_rate": 7.659420289855072e-05, "loss": 0.9233, "step": 1057 }, { "epoch": 0.47938377888536476, "grad_norm": 1.5927440712927512, "learning_rate": 7.666666666666668e-05, "loss": 0.9355, "step": 1058 }, { "epoch": 0.47983688264612595, "grad_norm": 2.425798741440126, "learning_rate": 7.673913043478262e-05, "loss": 0.9493, "step": 1059 }, { "epoch": 0.4802899864068872, "grad_norm": 1.212169301103661, "learning_rate": 7.681159420289856e-05, "loss": 0.9257, "step": 1060 }, { "epoch": 0.4807430901676484, "grad_norm": 1.6416176571261896, "learning_rate": 7.68840579710145e-05, "loss": 0.9268, "step": 1061 }, { "epoch": 0.4811961939284096, "grad_norm": 2.571932473396058, "learning_rate": 7.695652173913044e-05, "loss": 0.9354, "step": 1062 }, { "epoch": 0.4816492976891708, "grad_norm": 1.3276555256653833, "learning_rate": 7.702898550724639e-05, "loss": 0.9154, "step": 1063 }, { "epoch": 0.48210240144993205, "grad_norm": 3.002247782160794, "learning_rate": 7.710144927536232e-05, "loss": 0.9138, "step": 1064 }, { "epoch": 0.48255550521069324, "grad_norm": 1.8763538871052796, "learning_rate": 7.717391304347827e-05, "loss": 0.946, "step": 1065 }, { "epoch": 0.4830086089714545, "grad_norm": 3.153846793480665, "learning_rate": 7.72463768115942e-05, "loss": 0.9244, "step": 1066 }, { "epoch": 0.48346171273221566, "grad_norm": 2.6892147239576887, "learning_rate": 7.731884057971015e-05, "loss": 0.9426, "step": 1067 }, { "epoch": 0.4839148164929769, "grad_norm": 2.2559612286450927, "learning_rate": 7.73913043478261e-05, "loss": 0.9461, "step": 1068 }, { "epoch": 0.4843679202537381, "grad_norm": 2.549673395687669, "learning_rate": 7.746376811594204e-05, "loss": 0.933, "step": 1069 }, { "epoch": 0.48482102401449934, "grad_norm": 24.400755262407923, "learning_rate": 7.753623188405797e-05, "loss": 1.0372, "step": 1070 }, { "epoch": 0.4852741277752605, "grad_norm": 2.8248824548589573, "learning_rate": 7.760869565217392e-05, "loss": 0.9542, "step": 1071 }, { "epoch": 0.48572723153602176, "grad_norm": 331.2183447510598, "learning_rate": 7.768115942028987e-05, "loss": 1.7466, "step": 1072 }, { "epoch": 0.48618033529678295, "grad_norm": 1221.7402893115147, "learning_rate": 7.775362318840581e-05, "loss": 7.9063, "step": 1073 }, { "epoch": 0.4866334390575442, "grad_norm": 8.962030622500977, "learning_rate": 7.782608695652175e-05, "loss": 1.0674, "step": 1074 }, { "epoch": 0.4870865428183054, "grad_norm": 5.28239069544447, "learning_rate": 7.789855072463769e-05, "loss": 1.0103, "step": 1075 }, { "epoch": 0.4875396465790666, "grad_norm": 2.0678569642726887, "learning_rate": 7.797101449275363e-05, "loss": 0.9585, "step": 1076 }, { "epoch": 0.4879927503398278, "grad_norm": 2.8880079824056324, "learning_rate": 7.804347826086957e-05, "loss": 0.956, "step": 1077 }, { "epoch": 0.48844585410058905, "grad_norm": 40.073581435970766, "learning_rate": 7.811594202898552e-05, "loss": 1.0302, "step": 1078 }, { "epoch": 0.48889895786135024, "grad_norm": 152.66031417859557, "learning_rate": 7.818840579710145e-05, "loss": 6.3812, "step": 1079 }, { "epoch": 0.4893520616221115, "grad_norm": 163.20024059230335, "learning_rate": 7.82608695652174e-05, "loss": 3.2298, "step": 1080 }, { "epoch": 0.48980516538287266, "grad_norm": 16.14171823229821, "learning_rate": 7.833333333333333e-05, "loss": 1.7019, "step": 1081 }, { "epoch": 0.4902582691436339, "grad_norm": 5.335075881951055, "learning_rate": 7.840579710144928e-05, "loss": 1.1421, "step": 1082 }, { "epoch": 0.4907113729043951, "grad_norm": 2.686172570946294, "learning_rate": 7.847826086956522e-05, "loss": 1.0227, "step": 1083 }, { "epoch": 0.49116447666515634, "grad_norm": 9.72821778210814, "learning_rate": 7.855072463768117e-05, "loss": 1.1333, "step": 1084 }, { "epoch": 0.4916175804259175, "grad_norm": 6.109038435990479, "learning_rate": 7.86231884057971e-05, "loss": 1.2044, "step": 1085 }, { "epoch": 0.49207068418667876, "grad_norm": 5.421673480056457, "learning_rate": 7.869565217391305e-05, "loss": 1.1152, "step": 1086 }, { "epoch": 0.49252378794743995, "grad_norm": 3.75168809810852, "learning_rate": 7.8768115942029e-05, "loss": 1.0697, "step": 1087 }, { "epoch": 0.4929768917082012, "grad_norm": 6.32954462544734, "learning_rate": 7.884057971014494e-05, "loss": 1.1505, "step": 1088 }, { "epoch": 0.4934299954689624, "grad_norm": 4.977516171630413, "learning_rate": 7.891304347826088e-05, "loss": 1.0932, "step": 1089 }, { "epoch": 0.4938830992297236, "grad_norm": 3.345370965456693, "learning_rate": 7.898550724637682e-05, "loss": 1.0271, "step": 1090 }, { "epoch": 0.4943362029904848, "grad_norm": 1.661715974134025, "learning_rate": 7.905797101449276e-05, "loss": 1.0259, "step": 1091 }, { "epoch": 0.49478930675124605, "grad_norm": 2.823510666187744, "learning_rate": 7.91304347826087e-05, "loss": 1.013, "step": 1092 }, { "epoch": 0.49524241051200724, "grad_norm": 1.6334725387871063, "learning_rate": 7.920289855072465e-05, "loss": 1.0153, "step": 1093 }, { "epoch": 0.4956955142727685, "grad_norm": 2.544979152482241, "learning_rate": 7.927536231884058e-05, "loss": 1.0065, "step": 1094 }, { "epoch": 0.49614861803352966, "grad_norm": 1.7363966521000254, "learning_rate": 7.934782608695653e-05, "loss": 0.9827, "step": 1095 }, { "epoch": 0.4966017217942909, "grad_norm": 2.109061917365716, "learning_rate": 7.942028985507246e-05, "loss": 0.987, "step": 1096 }, { "epoch": 0.4970548255550521, "grad_norm": 1.4457865426696972, "learning_rate": 7.949275362318841e-05, "loss": 0.9684, "step": 1097 }, { "epoch": 0.49750792931581334, "grad_norm": 1.8409870453764372, "learning_rate": 7.956521739130435e-05, "loss": 0.9761, "step": 1098 }, { "epoch": 0.4979610330765745, "grad_norm": 1.7059110347673407, "learning_rate": 7.96376811594203e-05, "loss": 0.981, "step": 1099 }, { "epoch": 0.49841413683733576, "grad_norm": 1.632266311150356, "learning_rate": 7.971014492753623e-05, "loss": 0.9503, "step": 1100 }, { "epoch": 0.49886724059809695, "grad_norm": 2.216070232320456, "learning_rate": 7.978260869565218e-05, "loss": 0.9668, "step": 1101 }, { "epoch": 0.4993203443588582, "grad_norm": 1.5735766644639817, "learning_rate": 7.985507246376813e-05, "loss": 0.9666, "step": 1102 }, { "epoch": 0.4997734481196194, "grad_norm": 1.5855848858273036, "learning_rate": 7.992753623188407e-05, "loss": 0.9861, "step": 1103 }, { "epoch": 0.5002265518803806, "grad_norm": 1.4028015300438965, "learning_rate": 8e-05, "loss": 0.9636, "step": 1104 }, { "epoch": 0.5006796556411418, "grad_norm": 1.544871880065963, "learning_rate": 7.999999799855449e-05, "loss": 0.9452, "step": 1105 }, { "epoch": 0.501132759401903, "grad_norm": 9.645614955507725, "learning_rate": 7.999999199421812e-05, "loss": 0.9638, "step": 1106 }, { "epoch": 0.5015858631626643, "grad_norm": 55.19455176979595, "learning_rate": 7.99999819869915e-05, "loss": 2.0149, "step": 1107 }, { "epoch": 0.5020389669234254, "grad_norm": 11.72864820668351, "learning_rate": 7.999996797687565e-05, "loss": 1.3031, "step": 1108 }, { "epoch": 0.5024920706841867, "grad_norm": 2.773896254259438, "learning_rate": 7.999994996387195e-05, "loss": 0.9997, "step": 1109 }, { "epoch": 0.5029451744449479, "grad_norm": 2.333640021151178, "learning_rate": 7.999992794798221e-05, "loss": 0.989, "step": 1110 }, { "epoch": 0.5033982782057091, "grad_norm": 2.210330504627968, "learning_rate": 7.999990192920862e-05, "loss": 1.0049, "step": 1111 }, { "epoch": 0.5038513819664703, "grad_norm": 1.9222613151456123, "learning_rate": 7.999987190755382e-05, "loss": 0.9611, "step": 1112 }, { "epoch": 0.5043044857272315, "grad_norm": 2.098102953345335, "learning_rate": 7.999983788302078e-05, "loss": 0.9773, "step": 1113 }, { "epoch": 0.5047575894879928, "grad_norm": 1.5689691692096535, "learning_rate": 7.999979985561294e-05, "loss": 0.9645, "step": 1114 }, { "epoch": 0.505210693248754, "grad_norm": 2.255967336411544, "learning_rate": 7.999975782533405e-05, "loss": 0.9772, "step": 1115 }, { "epoch": 0.5056637970095151, "grad_norm": 1.6327346105958573, "learning_rate": 7.999971179218836e-05, "loss": 0.961, "step": 1116 }, { "epoch": 0.5061169007702764, "grad_norm": 1.812495002015341, "learning_rate": 7.999966175618047e-05, "loss": 0.9696, "step": 1117 }, { "epoch": 0.5065700045310376, "grad_norm": 1.475119470264895, "learning_rate": 7.99996077173154e-05, "loss": 0.9577, "step": 1118 }, { "epoch": 0.5070231082917989, "grad_norm": 1.8337946019851679, "learning_rate": 7.99995496755985e-05, "loss": 0.9509, "step": 1119 }, { "epoch": 0.50747621205256, "grad_norm": 1.2772351206036352, "learning_rate": 7.999948763103565e-05, "loss": 0.9561, "step": 1120 }, { "epoch": 0.5079293158133212, "grad_norm": 1.8126589897518792, "learning_rate": 7.999942158363303e-05, "loss": 0.9438, "step": 1121 }, { "epoch": 0.5083824195740825, "grad_norm": 1.5086615439419273, "learning_rate": 7.999935153339723e-05, "loss": 0.9539, "step": 1122 }, { "epoch": 0.5088355233348437, "grad_norm": 1.999896943440219, "learning_rate": 7.999927748033528e-05, "loss": 0.9437, "step": 1123 }, { "epoch": 0.5092886270956049, "grad_norm": 1.687902468527497, "learning_rate": 7.999919942445459e-05, "loss": 0.9468, "step": 1124 }, { "epoch": 0.5097417308563661, "grad_norm": 1.7050865669636233, "learning_rate": 7.999911736576296e-05, "loss": 0.9627, "step": 1125 }, { "epoch": 0.5101948346171273, "grad_norm": 2.1597332860478384, "learning_rate": 7.999903130426863e-05, "loss": 0.962, "step": 1126 }, { "epoch": 0.5106479383778886, "grad_norm": 1.6357686593078984, "learning_rate": 7.999894123998018e-05, "loss": 0.9577, "step": 1127 }, { "epoch": 0.5111010421386497, "grad_norm": 5.12754332921398, "learning_rate": 7.999884717290664e-05, "loss": 1.0594, "step": 1128 }, { "epoch": 0.511554145899411, "grad_norm": 2.277943928422361, "learning_rate": 7.999874910305743e-05, "loss": 0.9729, "step": 1129 }, { "epoch": 0.5120072496601722, "grad_norm": 1.089430763691333, "learning_rate": 7.999864703044234e-05, "loss": 0.9491, "step": 1130 }, { "epoch": 0.5124603534209334, "grad_norm": 1.5868621647781929, "learning_rate": 7.999854095507159e-05, "loss": 0.9831, "step": 1131 }, { "epoch": 0.5129134571816946, "grad_norm": 1.658325298749708, "learning_rate": 7.999843087695582e-05, "loss": 0.9939, "step": 1132 }, { "epoch": 0.5133665609424558, "grad_norm": 1.2973229582850836, "learning_rate": 7.999831679610602e-05, "loss": 0.9422, "step": 1133 }, { "epoch": 0.513819664703217, "grad_norm": 1.7659255104555642, "learning_rate": 7.999819871253363e-05, "loss": 0.9355, "step": 1134 }, { "epoch": 0.5142727684639783, "grad_norm": 1.3346182082900901, "learning_rate": 7.999807662625045e-05, "loss": 0.9485, "step": 1135 }, { "epoch": 0.5147258722247394, "grad_norm": 1.742225890389151, "learning_rate": 7.999795053726867e-05, "loss": 0.9614, "step": 1136 }, { "epoch": 0.5151789759855007, "grad_norm": 2.1936280142121154, "learning_rate": 7.999782044560097e-05, "loss": 0.9342, "step": 1137 }, { "epoch": 0.5156320797462619, "grad_norm": 1.0116957419829085, "learning_rate": 7.999768635126033e-05, "loss": 0.9526, "step": 1138 }, { "epoch": 0.5160851835070231, "grad_norm": 1.3095098377188663, "learning_rate": 7.999754825426017e-05, "loss": 0.9622, "step": 1139 }, { "epoch": 0.5165382872677843, "grad_norm": 1.238881808407424, "learning_rate": 7.999740615461432e-05, "loss": 0.9556, "step": 1140 }, { "epoch": 0.5169913910285455, "grad_norm": 1.9701323725169801, "learning_rate": 7.9997260052337e-05, "loss": 0.958, "step": 1141 }, { "epoch": 0.5174444947893068, "grad_norm": 1.440340914626929, "learning_rate": 7.999710994744282e-05, "loss": 0.9538, "step": 1142 }, { "epoch": 0.517897598550068, "grad_norm": 1.2673592422322524, "learning_rate": 7.999695583994682e-05, "loss": 0.9414, "step": 1143 }, { "epoch": 0.5183507023108291, "grad_norm": 1.6635783296030224, "learning_rate": 7.999679772986438e-05, "loss": 0.9453, "step": 1144 }, { "epoch": 0.5188038060715904, "grad_norm": 2.0991660053808476, "learning_rate": 7.999663561721138e-05, "loss": 0.9452, "step": 1145 }, { "epoch": 0.5192569098323516, "grad_norm": 1.4285544498211549, "learning_rate": 7.9996469502004e-05, "loss": 0.9377, "step": 1146 }, { "epoch": 0.5197100135931129, "grad_norm": 1.139176178863125, "learning_rate": 7.999629938425889e-05, "loss": 0.9489, "step": 1147 }, { "epoch": 0.520163117353874, "grad_norm": 2.7148193673535936, "learning_rate": 7.999612526399306e-05, "loss": 0.9406, "step": 1148 }, { "epoch": 0.5206162211146352, "grad_norm": 1.2593827921465417, "learning_rate": 7.999594714122394e-05, "loss": 0.9629, "step": 1149 }, { "epoch": 0.5210693248753965, "grad_norm": 3.399293587459989, "learning_rate": 7.999576501596935e-05, "loss": 0.9625, "step": 1150 }, { "epoch": 0.5215224286361577, "grad_norm": 2.3956491313965915, "learning_rate": 7.99955788882475e-05, "loss": 0.9557, "step": 1151 }, { "epoch": 0.5219755323969189, "grad_norm": 2.9977822852233853, "learning_rate": 7.999538875807705e-05, "loss": 0.9456, "step": 1152 }, { "epoch": 0.5224286361576801, "grad_norm": 2.188589237139018, "learning_rate": 7.999519462547701e-05, "loss": 0.9599, "step": 1153 }, { "epoch": 0.5228817399184413, "grad_norm": 2.5464408279616815, "learning_rate": 7.999499649046682e-05, "loss": 0.9449, "step": 1154 }, { "epoch": 0.5233348436792026, "grad_norm": 1.9592804752432134, "learning_rate": 7.999479435306628e-05, "loss": 0.9346, "step": 1155 }, { "epoch": 0.5237879474399637, "grad_norm": 2.318101129767458, "learning_rate": 7.999458821329565e-05, "loss": 0.9372, "step": 1156 }, { "epoch": 0.524241051200725, "grad_norm": 1.6230807702612973, "learning_rate": 7.999437807117553e-05, "loss": 0.9394, "step": 1157 }, { "epoch": 0.5246941549614862, "grad_norm": 1.6827100405782822, "learning_rate": 7.999416392672696e-05, "loss": 0.9337, "step": 1158 }, { "epoch": 0.5251472587222474, "grad_norm": 2.0949340586672713, "learning_rate": 7.99939457799714e-05, "loss": 0.9302, "step": 1159 }, { "epoch": 0.5256003624830086, "grad_norm": 1.0884198312917868, "learning_rate": 7.999372363093063e-05, "loss": 0.9399, "step": 1160 }, { "epoch": 0.5260534662437698, "grad_norm": 1.7817229098088354, "learning_rate": 7.99934974796269e-05, "loss": 0.9284, "step": 1161 }, { "epoch": 0.526506570004531, "grad_norm": 1.8695934679562092, "learning_rate": 7.999326732608285e-05, "loss": 0.9356, "step": 1162 }, { "epoch": 0.5269596737652923, "grad_norm": 1.1303299887920708, "learning_rate": 7.999303317032152e-05, "loss": 0.9468, "step": 1163 }, { "epoch": 0.5274127775260534, "grad_norm": 2.03877483850827, "learning_rate": 7.999279501236631e-05, "loss": 0.9372, "step": 1164 }, { "epoch": 0.5278658812868147, "grad_norm": 1.4126247627831283, "learning_rate": 7.999255285224108e-05, "loss": 0.9443, "step": 1165 }, { "epoch": 0.5283189850475759, "grad_norm": 2.633963605561864, "learning_rate": 7.999230668997007e-05, "loss": 0.9309, "step": 1166 }, { "epoch": 0.5287720888083371, "grad_norm": 1.882892877230831, "learning_rate": 7.99920565255779e-05, "loss": 0.9541, "step": 1167 }, { "epoch": 0.5292251925690983, "grad_norm": 2.3315606125128383, "learning_rate": 7.999180235908958e-05, "loss": 0.9481, "step": 1168 }, { "epoch": 0.5296782963298595, "grad_norm": 2.106094357515029, "learning_rate": 7.99915441905306e-05, "loss": 0.9338, "step": 1169 }, { "epoch": 0.5301314000906208, "grad_norm": 1.6890238524824466, "learning_rate": 7.999128201992674e-05, "loss": 0.9423, "step": 1170 }, { "epoch": 0.530584503851382, "grad_norm": 1.6927882708728543, "learning_rate": 7.999101584730428e-05, "loss": 0.9083, "step": 1171 }, { "epoch": 0.5310376076121431, "grad_norm": 1.597449817019524, "learning_rate": 7.999074567268984e-05, "loss": 0.9354, "step": 1172 }, { "epoch": 0.5314907113729044, "grad_norm": 1.104980142767925, "learning_rate": 7.999047149611044e-05, "loss": 0.9587, "step": 1173 }, { "epoch": 0.5319438151336656, "grad_norm": 1.7463900581552372, "learning_rate": 7.999019331759354e-05, "loss": 0.9255, "step": 1174 }, { "epoch": 0.5323969188944269, "grad_norm": 1.4317550865785966, "learning_rate": 7.998991113716697e-05, "loss": 0.9216, "step": 1175 }, { "epoch": 0.532850022655188, "grad_norm": 1.849483037767801, "learning_rate": 7.998962495485898e-05, "loss": 0.9284, "step": 1176 }, { "epoch": 0.5333031264159492, "grad_norm": 1.3262930351388855, "learning_rate": 7.998933477069818e-05, "loss": 0.9421, "step": 1177 }, { "epoch": 0.5337562301767105, "grad_norm": 1.7136215257547907, "learning_rate": 7.998904058471365e-05, "loss": 0.9402, "step": 1178 }, { "epoch": 0.5342093339374717, "grad_norm": 1.1975624752478857, "learning_rate": 7.998874239693479e-05, "loss": 0.9146, "step": 1179 }, { "epoch": 0.5346624376982329, "grad_norm": 2.267079367707463, "learning_rate": 7.998844020739146e-05, "loss": 0.9554, "step": 1180 }, { "epoch": 0.5351155414589941, "grad_norm": 2.0796176937908126, "learning_rate": 7.99881340161139e-05, "loss": 0.9216, "step": 1181 }, { "epoch": 0.5355686452197553, "grad_norm": 1.3136409111808756, "learning_rate": 7.998782382313275e-05, "loss": 0.9493, "step": 1182 }, { "epoch": 0.5360217489805166, "grad_norm": 1.4457004753791844, "learning_rate": 7.998750962847904e-05, "loss": 0.9362, "step": 1183 }, { "epoch": 0.5364748527412777, "grad_norm": 1.389672720020194, "learning_rate": 7.998719143218424e-05, "loss": 0.9147, "step": 1184 }, { "epoch": 0.536927956502039, "grad_norm": 1.3049131000140515, "learning_rate": 7.998686923428018e-05, "loss": 0.9427, "step": 1185 }, { "epoch": 0.5373810602628002, "grad_norm": 1.6283351477838164, "learning_rate": 7.998654303479908e-05, "loss": 0.941, "step": 1186 }, { "epoch": 0.5378341640235614, "grad_norm": 1.442536515764588, "learning_rate": 7.998621283377361e-05, "loss": 0.9265, "step": 1187 }, { "epoch": 0.5382872677843226, "grad_norm": 1.6421016274423463, "learning_rate": 7.998587863123682e-05, "loss": 0.9092, "step": 1188 }, { "epoch": 0.5387403715450838, "grad_norm": 1.5273524847868858, "learning_rate": 7.998554042722213e-05, "loss": 0.9405, "step": 1189 }, { "epoch": 0.539193475305845, "grad_norm": 1.2095533728735877, "learning_rate": 7.99851982217634e-05, "loss": 0.907, "step": 1190 }, { "epoch": 0.5396465790666063, "grad_norm": 1.336992849797945, "learning_rate": 7.998485201489487e-05, "loss": 0.941, "step": 1191 }, { "epoch": 0.5400996828273674, "grad_norm": 2.4018103427337647, "learning_rate": 7.998450180665118e-05, "loss": 0.9265, "step": 1192 }, { "epoch": 0.5405527865881287, "grad_norm": 0.9491781378530535, "learning_rate": 7.998414759706739e-05, "loss": 0.9105, "step": 1193 }, { "epoch": 0.5410058903488899, "grad_norm": 3.302643714841159, "learning_rate": 7.998378938617894e-05, "loss": 0.932, "step": 1194 }, { "epoch": 0.5414589941096511, "grad_norm": 2.2841241326180595, "learning_rate": 7.998342717402168e-05, "loss": 0.9468, "step": 1195 }, { "epoch": 0.5419120978704123, "grad_norm": 3.0477894791441065, "learning_rate": 7.998306096063184e-05, "loss": 0.9288, "step": 1196 }, { "epoch": 0.5423652016311735, "grad_norm": 2.256868600411717, "learning_rate": 7.99826907460461e-05, "loss": 0.9365, "step": 1197 }, { "epoch": 0.5428183053919348, "grad_norm": 3.137483980294451, "learning_rate": 7.998231653030148e-05, "loss": 0.9479, "step": 1198 }, { "epoch": 0.543271409152696, "grad_norm": 2.7203301899106065, "learning_rate": 7.998193831343543e-05, "loss": 0.9308, "step": 1199 }, { "epoch": 0.5437245129134571, "grad_norm": 2.6579021809661523, "learning_rate": 7.998155609548581e-05, "loss": 0.9375, "step": 1200 }, { "epoch": 0.5441776166742184, "grad_norm": 2.107077635072986, "learning_rate": 7.998116987649089e-05, "loss": 0.9383, "step": 1201 }, { "epoch": 0.5446307204349796, "grad_norm": 2.6015138571340035, "learning_rate": 7.998077965648927e-05, "loss": 0.9356, "step": 1202 }, { "epoch": 0.5450838241957409, "grad_norm": 1.7763809980418237, "learning_rate": 7.998038543552005e-05, "loss": 0.9369, "step": 1203 }, { "epoch": 0.545536927956502, "grad_norm": 3.472523986743842, "learning_rate": 7.997998721362265e-05, "loss": 0.9463, "step": 1204 }, { "epoch": 0.5459900317172632, "grad_norm": 2.4320890226048726, "learning_rate": 7.997958499083691e-05, "loss": 0.9511, "step": 1205 }, { "epoch": 0.5464431354780245, "grad_norm": 2.84910387539841, "learning_rate": 7.997917876720312e-05, "loss": 0.9446, "step": 1206 }, { "epoch": 0.5468962392387857, "grad_norm": 2.848339526911281, "learning_rate": 7.997876854276191e-05, "loss": 0.9211, "step": 1207 }, { "epoch": 0.5473493429995469, "grad_norm": 1.8351175964132722, "learning_rate": 7.997835431755434e-05, "loss": 0.924, "step": 1208 }, { "epoch": 0.5478024467603081, "grad_norm": 1.4534328243604175, "learning_rate": 7.997793609162184e-05, "loss": 0.9325, "step": 1209 }, { "epoch": 0.5482555505210693, "grad_norm": 2.8979315772079546, "learning_rate": 7.997751386500629e-05, "loss": 0.9167, "step": 1210 }, { "epoch": 0.5487086542818306, "grad_norm": 2.313514995767736, "learning_rate": 7.997708763774993e-05, "loss": 0.9364, "step": 1211 }, { "epoch": 0.5491617580425917, "grad_norm": 2.5708158666504235, "learning_rate": 7.997665740989542e-05, "loss": 0.9166, "step": 1212 }, { "epoch": 0.549614861803353, "grad_norm": 2.306898339042898, "learning_rate": 7.99762231814858e-05, "loss": 0.932, "step": 1213 }, { "epoch": 0.5500679655641142, "grad_norm": 2.027032205940994, "learning_rate": 7.997578495256454e-05, "loss": 0.947, "step": 1214 }, { "epoch": 0.5505210693248754, "grad_norm": 1.670801524025198, "learning_rate": 7.997534272317549e-05, "loss": 0.9506, "step": 1215 }, { "epoch": 0.5509741730856366, "grad_norm": 2.8475631358801325, "learning_rate": 7.99748964933629e-05, "loss": 0.9536, "step": 1216 }, { "epoch": 0.5514272768463978, "grad_norm": 2.08758823584554, "learning_rate": 7.997444626317144e-05, "loss": 0.9177, "step": 1217 }, { "epoch": 0.551880380607159, "grad_norm": 2.4191725752730266, "learning_rate": 7.997399203264615e-05, "loss": 0.9303, "step": 1218 }, { "epoch": 0.5523334843679203, "grad_norm": 2.2295880678552424, "learning_rate": 7.997353380183248e-05, "loss": 0.9172, "step": 1219 }, { "epoch": 0.5527865881286814, "grad_norm": 2.330625515620385, "learning_rate": 7.997307157077628e-05, "loss": 0.9501, "step": 1220 }, { "epoch": 0.5532396918894427, "grad_norm": 1.8923314958286732, "learning_rate": 7.997260533952385e-05, "loss": 0.9104, "step": 1221 }, { "epoch": 0.5536927956502039, "grad_norm": 2.349241271490613, "learning_rate": 7.997213510812182e-05, "loss": 0.9215, "step": 1222 }, { "epoch": 0.5541458994109651, "grad_norm": 1.9426312570396405, "learning_rate": 7.997166087661724e-05, "loss": 0.9252, "step": 1223 }, { "epoch": 0.5545990031717263, "grad_norm": 2.5771926526098645, "learning_rate": 7.997118264505758e-05, "loss": 0.9392, "step": 1224 }, { "epoch": 0.5550521069324875, "grad_norm": 2.2719915390891985, "learning_rate": 7.997070041349068e-05, "loss": 0.9512, "step": 1225 }, { "epoch": 0.5555052106932488, "grad_norm": 2.033374509585558, "learning_rate": 7.997021418196483e-05, "loss": 0.9128, "step": 1226 }, { "epoch": 0.55595831445401, "grad_norm": 1.6414235768549768, "learning_rate": 7.996972395052865e-05, "loss": 0.9312, "step": 1227 }, { "epoch": 0.5564114182147711, "grad_norm": 2.611074391871324, "learning_rate": 7.996922971923122e-05, "loss": 0.9353, "step": 1228 }, { "epoch": 0.5568645219755324, "grad_norm": 2.225319567315068, "learning_rate": 7.996873148812199e-05, "loss": 0.9268, "step": 1229 }, { "epoch": 0.5573176257362936, "grad_norm": 2.205289320928654, "learning_rate": 7.996822925725084e-05, "loss": 0.9363, "step": 1230 }, { "epoch": 0.5577707294970549, "grad_norm": 2.0786458022207452, "learning_rate": 7.996772302666801e-05, "loss": 0.9138, "step": 1231 }, { "epoch": 0.558223833257816, "grad_norm": 1.9627933258326953, "learning_rate": 7.996721279642416e-05, "loss": 0.9379, "step": 1232 }, { "epoch": 0.5586769370185772, "grad_norm": 1.5834081599653576, "learning_rate": 7.996669856657035e-05, "loss": 0.9334, "step": 1233 }, { "epoch": 0.5591300407793385, "grad_norm": 2.5435770868626117, "learning_rate": 7.996618033715805e-05, "loss": 0.8936, "step": 1234 }, { "epoch": 0.5595831445400997, "grad_norm": 2.1796536740298307, "learning_rate": 7.996565810823911e-05, "loss": 0.9425, "step": 1235 }, { "epoch": 0.5600362483008609, "grad_norm": 2.203611012272621, "learning_rate": 7.99651318798658e-05, "loss": 0.9484, "step": 1236 }, { "epoch": 0.5604893520616221, "grad_norm": 2.2038884274168438, "learning_rate": 7.996460165209079e-05, "loss": 0.928, "step": 1237 }, { "epoch": 0.5609424558223833, "grad_norm": 1.6865297450963548, "learning_rate": 7.99640674249671e-05, "loss": 0.928, "step": 1238 }, { "epoch": 0.5613955595831446, "grad_norm": 1.3216872947559455, "learning_rate": 7.996352919854822e-05, "loss": 0.9351, "step": 1239 }, { "epoch": 0.5618486633439057, "grad_norm": 2.3951733317428827, "learning_rate": 7.996298697288802e-05, "loss": 0.9415, "step": 1240 }, { "epoch": 0.562301767104667, "grad_norm": 1.890225027228579, "learning_rate": 7.996244074804076e-05, "loss": 0.9235, "step": 1241 }, { "epoch": 0.5627548708654282, "grad_norm": 2.2593963490891453, "learning_rate": 7.996189052406108e-05, "loss": 0.9384, "step": 1242 }, { "epoch": 0.5632079746261894, "grad_norm": 2.043232128658984, "learning_rate": 7.996133630100406e-05, "loss": 0.9158, "step": 1243 }, { "epoch": 0.5636610783869506, "grad_norm": 1.8411567450961557, "learning_rate": 7.996077807892516e-05, "loss": 0.9084, "step": 1244 }, { "epoch": 0.5641141821477118, "grad_norm": 1.5194575276511555, "learning_rate": 7.996021585788024e-05, "loss": 0.9238, "step": 1245 }, { "epoch": 0.564567285908473, "grad_norm": 2.1628814833013923, "learning_rate": 7.995964963792555e-05, "loss": 0.9285, "step": 1246 }, { "epoch": 0.5650203896692343, "grad_norm": 1.6160496233811508, "learning_rate": 7.995907941911778e-05, "loss": 0.926, "step": 1247 }, { "epoch": 0.5654734934299954, "grad_norm": 2.2017590597264327, "learning_rate": 7.995850520151397e-05, "loss": 0.9301, "step": 1248 }, { "epoch": 0.5659265971907567, "grad_norm": 2.0491233728104334, "learning_rate": 7.995792698517161e-05, "loss": 0.9256, "step": 1249 }, { "epoch": 0.5663797009515179, "grad_norm": 1.5777785639631934, "learning_rate": 7.995734477014854e-05, "loss": 0.9214, "step": 1250 }, { "epoch": 0.5668328047122791, "grad_norm": 1.4946792874290147, "learning_rate": 7.995675855650302e-05, "loss": 0.9492, "step": 1251 }, { "epoch": 0.5672859084730403, "grad_norm": 1.5837562256507123, "learning_rate": 7.995616834429372e-05, "loss": 0.9269, "step": 1252 }, { "epoch": 0.5677390122338015, "grad_norm": 1.1509730640174354, "learning_rate": 7.995557413357972e-05, "loss": 0.9217, "step": 1253 }, { "epoch": 0.5681921159945628, "grad_norm": 2.1281265470174406, "learning_rate": 7.995497592442046e-05, "loss": 0.9379, "step": 1254 }, { "epoch": 0.568645219755324, "grad_norm": 1.8597518710337935, "learning_rate": 7.995437371687583e-05, "loss": 0.9478, "step": 1255 }, { "epoch": 0.5690983235160851, "grad_norm": 1.5978827411432854, "learning_rate": 7.995376751100607e-05, "loss": 0.9144, "step": 1256 }, { "epoch": 0.5695514272768464, "grad_norm": 1.517674038561444, "learning_rate": 7.995315730687186e-05, "loss": 0.9235, "step": 1257 }, { "epoch": 0.5700045310376076, "grad_norm": 1.5336827852396768, "learning_rate": 7.995254310453427e-05, "loss": 0.94, "step": 1258 }, { "epoch": 0.5704576347983689, "grad_norm": 1.2025093035049714, "learning_rate": 7.995192490405473e-05, "loss": 0.9304, "step": 1259 }, { "epoch": 0.57091073855913, "grad_norm": 2.093685708457868, "learning_rate": 7.995130270549513e-05, "loss": 0.9313, "step": 1260 }, { "epoch": 0.5713638423198912, "grad_norm": 1.5212493434893517, "learning_rate": 7.995067650891774e-05, "loss": 0.9466, "step": 1261 }, { "epoch": 0.5718169460806525, "grad_norm": 2.126705261288572, "learning_rate": 7.995004631438522e-05, "loss": 0.9341, "step": 1262 }, { "epoch": 0.5722700498414137, "grad_norm": 1.8129581559258414, "learning_rate": 7.994941212196063e-05, "loss": 0.9428, "step": 1263 }, { "epoch": 0.5727231536021749, "grad_norm": 1.4598829151962498, "learning_rate": 7.994877393170743e-05, "loss": 0.9129, "step": 1264 }, { "epoch": 0.5731762573629361, "grad_norm": 1.6978238701417454, "learning_rate": 7.994813174368951e-05, "loss": 0.9251, "step": 1265 }, { "epoch": 0.5736293611236973, "grad_norm": 1.1136607926069388, "learning_rate": 7.99474855579711e-05, "loss": 0.9452, "step": 1266 }, { "epoch": 0.5740824648844586, "grad_norm": 1.3113956094804229, "learning_rate": 7.994683537461689e-05, "loss": 0.9366, "step": 1267 }, { "epoch": 0.5745355686452197, "grad_norm": 1.1945356521824186, "learning_rate": 7.994618119369195e-05, "loss": 0.9197, "step": 1268 }, { "epoch": 0.574988672405981, "grad_norm": 1.7031268206562058, "learning_rate": 7.994552301526172e-05, "loss": 0.9205, "step": 1269 }, { "epoch": 0.5754417761667422, "grad_norm": 1.0508545688430941, "learning_rate": 7.994486083939208e-05, "loss": 0.9544, "step": 1270 }, { "epoch": 0.5758948799275034, "grad_norm": 1.2676863216311176, "learning_rate": 7.994419466614931e-05, "loss": 0.9158, "step": 1271 }, { "epoch": 0.5763479836882646, "grad_norm": 0.7884228663991826, "learning_rate": 7.994352449560004e-05, "loss": 0.9373, "step": 1272 }, { "epoch": 0.5768010874490258, "grad_norm": 1.1653242806846293, "learning_rate": 7.994285032781137e-05, "loss": 0.924, "step": 1273 }, { "epoch": 0.577254191209787, "grad_norm": 1.3651715435634186, "learning_rate": 7.994217216285075e-05, "loss": 0.946, "step": 1274 }, { "epoch": 0.5777072949705483, "grad_norm": 1.7007773718665096, "learning_rate": 7.994149000078605e-05, "loss": 0.9164, "step": 1275 }, { "epoch": 0.5781603987313094, "grad_norm": 1.298765977987965, "learning_rate": 7.994080384168553e-05, "loss": 0.9281, "step": 1276 }, { "epoch": 0.5786135024920707, "grad_norm": 1.409083600305634, "learning_rate": 7.994011368561784e-05, "loss": 0.8958, "step": 1277 }, { "epoch": 0.5790666062528319, "grad_norm": 0.8293788926624698, "learning_rate": 7.993941953265208e-05, "loss": 0.9268, "step": 1278 }, { "epoch": 0.5795197100135931, "grad_norm": 1.4329888402102562, "learning_rate": 7.99387213828577e-05, "loss": 0.9153, "step": 1279 }, { "epoch": 0.5799728137743543, "grad_norm": 1.5407043191066696, "learning_rate": 7.993801923630456e-05, "loss": 0.9285, "step": 1280 }, { "epoch": 0.5804259175351155, "grad_norm": 1.221677489187981, "learning_rate": 7.993731309306294e-05, "loss": 0.9283, "step": 1281 }, { "epoch": 0.5808790212958768, "grad_norm": 1.564806363341513, "learning_rate": 7.993660295320348e-05, "loss": 0.9335, "step": 1282 }, { "epoch": 0.581332125056638, "grad_norm": 1.338648718317404, "learning_rate": 7.993588881679726e-05, "loss": 0.93, "step": 1283 }, { "epoch": 0.5817852288173991, "grad_norm": 1.1427918091924782, "learning_rate": 7.993517068391576e-05, "loss": 0.9313, "step": 1284 }, { "epoch": 0.5822383325781604, "grad_norm": 1.550888469116869, "learning_rate": 7.993444855463083e-05, "loss": 0.9357, "step": 1285 }, { "epoch": 0.5826914363389216, "grad_norm": 0.8711308677935365, "learning_rate": 7.993372242901472e-05, "loss": 0.9182, "step": 1286 }, { "epoch": 0.5831445400996829, "grad_norm": 1.4408763122153836, "learning_rate": 7.993299230714012e-05, "loss": 0.9153, "step": 1287 }, { "epoch": 0.583597643860444, "grad_norm": 1.6960578122236931, "learning_rate": 7.993225818908008e-05, "loss": 0.9373, "step": 1288 }, { "epoch": 0.5840507476212052, "grad_norm": 0.7972590198246517, "learning_rate": 7.993152007490808e-05, "loss": 0.9166, "step": 1289 }, { "epoch": 0.5845038513819665, "grad_norm": 1.345973189711013, "learning_rate": 7.993077796469798e-05, "loss": 0.9445, "step": 1290 }, { "epoch": 0.5849569551427277, "grad_norm": 1.707629212338845, "learning_rate": 7.993003185852403e-05, "loss": 0.9099, "step": 1291 }, { "epoch": 0.5854100589034889, "grad_norm": 1.2564710922017346, "learning_rate": 7.992928175646091e-05, "loss": 0.9484, "step": 1292 }, { "epoch": 0.5858631626642501, "grad_norm": 1.2423310451903373, "learning_rate": 7.992852765858368e-05, "loss": 0.9185, "step": 1293 }, { "epoch": 0.5863162664250113, "grad_norm": 0.8633569001151772, "learning_rate": 7.99277695649678e-05, "loss": 0.9334, "step": 1294 }, { "epoch": 0.5867693701857726, "grad_norm": 1.0658872680222726, "learning_rate": 7.992700747568913e-05, "loss": 0.9287, "step": 1295 }, { "epoch": 0.5872224739465337, "grad_norm": 1.5220127057788178, "learning_rate": 7.992624139082395e-05, "loss": 0.9347, "step": 1296 }, { "epoch": 0.587675577707295, "grad_norm": 1.6437556549177448, "learning_rate": 7.992547131044893e-05, "loss": 0.9244, "step": 1297 }, { "epoch": 0.5881286814680562, "grad_norm": 1.0978371515470295, "learning_rate": 7.99246972346411e-05, "loss": 0.9267, "step": 1298 }, { "epoch": 0.5885817852288174, "grad_norm": 1.5110572445079047, "learning_rate": 7.992391916347794e-05, "loss": 0.924, "step": 1299 }, { "epoch": 0.5890348889895786, "grad_norm": 0.8010243966788073, "learning_rate": 7.992313709703734e-05, "loss": 0.9198, "step": 1300 }, { "epoch": 0.5894879927503398, "grad_norm": 1.9545931803062733, "learning_rate": 7.992235103539751e-05, "loss": 0.9277, "step": 1301 }, { "epoch": 0.589941096511101, "grad_norm": 1.1192046112743463, "learning_rate": 7.992156097863717e-05, "loss": 0.9394, "step": 1302 }, { "epoch": 0.5903942002718623, "grad_norm": 1.4585466406251317, "learning_rate": 7.992076692683533e-05, "loss": 0.9303, "step": 1303 }, { "epoch": 0.5908473040326234, "grad_norm": 1.1003115270176498, "learning_rate": 7.99199688800715e-05, "loss": 0.9311, "step": 1304 }, { "epoch": 0.5913004077933847, "grad_norm": 1.6733644115115598, "learning_rate": 7.99191668384255e-05, "loss": 0.924, "step": 1305 }, { "epoch": 0.5917535115541459, "grad_norm": 1.2706909280939118, "learning_rate": 7.991836080197763e-05, "loss": 0.9264, "step": 1306 }, { "epoch": 0.5922066153149071, "grad_norm": 1.3055109101324893, "learning_rate": 7.991755077080852e-05, "loss": 0.9202, "step": 1307 }, { "epoch": 0.5926597190756683, "grad_norm": 1.212292224186536, "learning_rate": 7.991673674499925e-05, "loss": 0.9336, "step": 1308 }, { "epoch": 0.5931128228364295, "grad_norm": 1.2708812702146615, "learning_rate": 7.991591872463128e-05, "loss": 0.9359, "step": 1309 }, { "epoch": 0.5935659265971908, "grad_norm": 1.1453810025565954, "learning_rate": 7.991509670978647e-05, "loss": 0.9071, "step": 1310 }, { "epoch": 0.594019030357952, "grad_norm": 1.4374134822867322, "learning_rate": 7.991427070054707e-05, "loss": 0.9407, "step": 1311 }, { "epoch": 0.5944721341187131, "grad_norm": 1.4700604728487296, "learning_rate": 7.991344069699576e-05, "loss": 0.969, "step": 1312 }, { "epoch": 0.5949252378794744, "grad_norm": 1.3264908167075662, "learning_rate": 7.991260669921558e-05, "loss": 0.9165, "step": 1313 }, { "epoch": 0.5953783416402356, "grad_norm": 1.0088560843707903, "learning_rate": 7.991176870729002e-05, "loss": 0.9373, "step": 1314 }, { "epoch": 0.5958314454009969, "grad_norm": 1.7974506618434678, "learning_rate": 7.99109267213029e-05, "loss": 0.9224, "step": 1315 }, { "epoch": 0.596284549161758, "grad_norm": 1.1239031923553082, "learning_rate": 7.99100807413385e-05, "loss": 0.9405, "step": 1316 }, { "epoch": 0.5967376529225192, "grad_norm": 1.0001579458639986, "learning_rate": 7.990923076748148e-05, "loss": 0.9416, "step": 1317 }, { "epoch": 0.5971907566832805, "grad_norm": 1.285269382787375, "learning_rate": 7.99083767998169e-05, "loss": 0.9306, "step": 1318 }, { "epoch": 0.5976438604440417, "grad_norm": 1.4318127665756735, "learning_rate": 7.990751883843022e-05, "loss": 0.9322, "step": 1319 }, { "epoch": 0.5980969642048029, "grad_norm": 1.4896096032078137, "learning_rate": 7.990665688340729e-05, "loss": 0.9259, "step": 1320 }, { "epoch": 0.5985500679655641, "grad_norm": 1.295072008399731, "learning_rate": 7.990579093483437e-05, "loss": 0.9416, "step": 1321 }, { "epoch": 0.5990031717263253, "grad_norm": 1.1560246310875923, "learning_rate": 7.990492099279814e-05, "loss": 0.9203, "step": 1322 }, { "epoch": 0.5994562754870866, "grad_norm": 1.5613927007821882, "learning_rate": 7.99040470573856e-05, "loss": 0.9345, "step": 1323 }, { "epoch": 0.5999093792478477, "grad_norm": 1.3883712943075082, "learning_rate": 7.990316912868427e-05, "loss": 0.959, "step": 1324 }, { "epoch": 0.600362483008609, "grad_norm": 1.3605406096096384, "learning_rate": 7.990228720678197e-05, "loss": 0.9338, "step": 1325 }, { "epoch": 0.6008155867693702, "grad_norm": 1.1230665841686456, "learning_rate": 7.990140129176697e-05, "loss": 0.9057, "step": 1326 }, { "epoch": 0.6012686905301314, "grad_norm": 1.1594029385723215, "learning_rate": 7.99005113837279e-05, "loss": 0.9196, "step": 1327 }, { "epoch": 0.6017217942908926, "grad_norm": 1.4934212935695377, "learning_rate": 7.989961748275387e-05, "loss": 0.9142, "step": 1328 }, { "epoch": 0.6021748980516538, "grad_norm": 1.2836471171554853, "learning_rate": 7.989871958893428e-05, "loss": 0.9388, "step": 1329 }, { "epoch": 0.602628001812415, "grad_norm": 1.4742258288089973, "learning_rate": 7.989781770235901e-05, "loss": 0.9245, "step": 1330 }, { "epoch": 0.6030811055731763, "grad_norm": 1.6579601456726913, "learning_rate": 7.989691182311831e-05, "loss": 0.9218, "step": 1331 }, { "epoch": 0.6035342093339374, "grad_norm": 0.9133883417319762, "learning_rate": 7.989600195130283e-05, "loss": 0.9115, "step": 1332 }, { "epoch": 0.6039873130946987, "grad_norm": 1.2600643442211712, "learning_rate": 7.989508808700361e-05, "loss": 0.9019, "step": 1333 }, { "epoch": 0.6044404168554599, "grad_norm": 1.2603961402618566, "learning_rate": 7.989417023031215e-05, "loss": 0.9166, "step": 1334 }, { "epoch": 0.6048935206162211, "grad_norm": 1.6074839064685331, "learning_rate": 7.989324838132025e-05, "loss": 0.9309, "step": 1335 }, { "epoch": 0.6053466243769823, "grad_norm": 1.5041978784274765, "learning_rate": 7.989232254012018e-05, "loss": 0.9366, "step": 1336 }, { "epoch": 0.6057997281377435, "grad_norm": 1.232015301044206, "learning_rate": 7.98913927068046e-05, "loss": 0.9179, "step": 1337 }, { "epoch": 0.6062528318985048, "grad_norm": 1.0769622286283502, "learning_rate": 7.989045888146654e-05, "loss": 0.922, "step": 1338 }, { "epoch": 0.606705935659266, "grad_norm": 1.884356942381466, "learning_rate": 7.988952106419947e-05, "loss": 0.9177, "step": 1339 }, { "epoch": 0.6071590394200271, "grad_norm": 0.9196025506158864, "learning_rate": 7.988857925509724e-05, "loss": 0.9131, "step": 1340 }, { "epoch": 0.6076121431807884, "grad_norm": 1.576139839142406, "learning_rate": 7.988763345425409e-05, "loss": 0.9302, "step": 1341 }, { "epoch": 0.6080652469415496, "grad_norm": 1.4873465422516248, "learning_rate": 7.988668366176465e-05, "loss": 0.9229, "step": 1342 }, { "epoch": 0.6085183507023109, "grad_norm": 1.152099554944683, "learning_rate": 7.988572987772401e-05, "loss": 0.9214, "step": 1343 }, { "epoch": 0.608971454463072, "grad_norm": 1.3089196108753989, "learning_rate": 7.988477210222758e-05, "loss": 0.9142, "step": 1344 }, { "epoch": 0.6094245582238332, "grad_norm": 1.331846871096959, "learning_rate": 7.988381033537124e-05, "loss": 0.9338, "step": 1345 }, { "epoch": 0.6098776619845945, "grad_norm": 1.5141229633995286, "learning_rate": 7.98828445772512e-05, "loss": 0.9159, "step": 1346 }, { "epoch": 0.6103307657453557, "grad_norm": 1.3443868239381245, "learning_rate": 7.988187482796413e-05, "loss": 0.9091, "step": 1347 }, { "epoch": 0.6107838695061168, "grad_norm": 1.5419302444942233, "learning_rate": 7.988090108760706e-05, "loss": 0.9242, "step": 1348 }, { "epoch": 0.6112369732668781, "grad_norm": 0.6498388411901829, "learning_rate": 7.987992335627746e-05, "loss": 0.9059, "step": 1349 }, { "epoch": 0.6116900770276393, "grad_norm": 1.355351081944455, "learning_rate": 7.987894163407315e-05, "loss": 0.9445, "step": 1350 }, { "epoch": 0.6121431807884006, "grad_norm": 1.3591355622811323, "learning_rate": 7.987795592109237e-05, "loss": 0.9134, "step": 1351 }, { "epoch": 0.6125962845491617, "grad_norm": 1.6571671907565941, "learning_rate": 7.987696621743379e-05, "loss": 0.913, "step": 1352 }, { "epoch": 0.613049388309923, "grad_norm": 1.207413739521993, "learning_rate": 7.987597252319642e-05, "loss": 0.9048, "step": 1353 }, { "epoch": 0.6135024920706842, "grad_norm": 1.3719543504911185, "learning_rate": 7.987497483847972e-05, "loss": 0.9321, "step": 1354 }, { "epoch": 0.6139555958314454, "grad_norm": 0.7774018124954647, "learning_rate": 7.987397316338354e-05, "loss": 0.9107, "step": 1355 }, { "epoch": 0.6144086995922066, "grad_norm": 1.0951529382001948, "learning_rate": 7.987296749800807e-05, "loss": 0.9166, "step": 1356 }, { "epoch": 0.6148618033529678, "grad_norm": 2.0224573070974587, "learning_rate": 7.987195784245402e-05, "loss": 0.9228, "step": 1357 }, { "epoch": 0.615314907113729, "grad_norm": 1.0694789849072488, "learning_rate": 7.987094419682238e-05, "loss": 0.9234, "step": 1358 }, { "epoch": 0.6157680108744903, "grad_norm": 1.4053166613709502, "learning_rate": 7.986992656121461e-05, "loss": 0.9288, "step": 1359 }, { "epoch": 0.6162211146352514, "grad_norm": 1.7415418900514923, "learning_rate": 7.986890493573252e-05, "loss": 0.9409, "step": 1360 }, { "epoch": 0.6166742183960127, "grad_norm": 1.3295704461679008, "learning_rate": 7.986787932047838e-05, "loss": 0.916, "step": 1361 }, { "epoch": 0.6171273221567739, "grad_norm": 1.2257230846696736, "learning_rate": 7.98668497155548e-05, "loss": 0.9277, "step": 1362 }, { "epoch": 0.6175804259175351, "grad_norm": 1.2526671565552752, "learning_rate": 7.986581612106485e-05, "loss": 0.9306, "step": 1363 }, { "epoch": 0.6180335296782963, "grad_norm": 1.7947841033178737, "learning_rate": 7.986477853711191e-05, "loss": 0.9154, "step": 1364 }, { "epoch": 0.6184866334390575, "grad_norm": 0.9925888975754186, "learning_rate": 7.986373696379987e-05, "loss": 0.9111, "step": 1365 }, { "epoch": 0.6189397371998188, "grad_norm": 1.7431540736003857, "learning_rate": 7.986269140123291e-05, "loss": 0.9259, "step": 1366 }, { "epoch": 0.61939284096058, "grad_norm": 1.0105187099658328, "learning_rate": 7.98616418495157e-05, "loss": 0.9376, "step": 1367 }, { "epoch": 0.6198459447213411, "grad_norm": 2.2370522509618787, "learning_rate": 7.986058830875326e-05, "loss": 0.9184, "step": 1368 }, { "epoch": 0.6202990484821024, "grad_norm": 1.8247465012942694, "learning_rate": 7.985953077905102e-05, "loss": 0.9336, "step": 1369 }, { "epoch": 0.6207521522428636, "grad_norm": 1.7005968854507443, "learning_rate": 7.985846926051483e-05, "loss": 0.9321, "step": 1370 }, { "epoch": 0.6212052560036249, "grad_norm": 1.4577476096179078, "learning_rate": 7.985740375325085e-05, "loss": 0.9403, "step": 1371 }, { "epoch": 0.621658359764386, "grad_norm": 1.7633646850406153, "learning_rate": 7.98563342573658e-05, "loss": 0.9405, "step": 1372 }, { "epoch": 0.6221114635251472, "grad_norm": 1.2674467000102905, "learning_rate": 7.985526077296663e-05, "loss": 0.9196, "step": 1373 }, { "epoch": 0.6225645672859085, "grad_norm": 1.7324790382207829, "learning_rate": 7.98541833001608e-05, "loss": 0.9217, "step": 1374 }, { "epoch": 0.6230176710466697, "grad_norm": 1.3973247816029641, "learning_rate": 7.985310183905615e-05, "loss": 0.9072, "step": 1375 }, { "epoch": 0.6234707748074308, "grad_norm": 1.4712679322949538, "learning_rate": 7.985201638976087e-05, "loss": 0.9207, "step": 1376 }, { "epoch": 0.6239238785681921, "grad_norm": 1.2937163860879475, "learning_rate": 7.98509269523836e-05, "loss": 0.9278, "step": 1377 }, { "epoch": 0.6243769823289533, "grad_norm": 1.4126644179190748, "learning_rate": 7.984983352703338e-05, "loss": 0.9192, "step": 1378 }, { "epoch": 0.6248300860897146, "grad_norm": 1.1852761313320577, "learning_rate": 7.98487361138196e-05, "loss": 0.9235, "step": 1379 }, { "epoch": 0.6252831898504757, "grad_norm": 1.7332140942694272, "learning_rate": 7.98476347128521e-05, "loss": 0.9327, "step": 1380 }, { "epoch": 0.625736293611237, "grad_norm": 1.2183268110366403, "learning_rate": 7.98465293242411e-05, "loss": 0.9221, "step": 1381 }, { "epoch": 0.6261893973719982, "grad_norm": 1.9667710657860331, "learning_rate": 7.98454199480972e-05, "loss": 0.9323, "step": 1382 }, { "epoch": 0.6266425011327594, "grad_norm": 1.7347605780541002, "learning_rate": 7.984430658453145e-05, "loss": 0.927, "step": 1383 }, { "epoch": 0.6270956048935206, "grad_norm": 1.261415048621085, "learning_rate": 7.984318923365524e-05, "loss": 0.914, "step": 1384 }, { "epoch": 0.6275487086542818, "grad_norm": 1.453848108702844, "learning_rate": 7.984206789558039e-05, "loss": 0.9434, "step": 1385 }, { "epoch": 0.628001812415043, "grad_norm": 1.0866395486563336, "learning_rate": 7.984094257041912e-05, "loss": 0.9325, "step": 1386 }, { "epoch": 0.6284549161758043, "grad_norm": 1.7516145683722466, "learning_rate": 7.983981325828405e-05, "loss": 0.911, "step": 1387 }, { "epoch": 0.6289080199365654, "grad_norm": 1.2857813318331914, "learning_rate": 7.983867995928818e-05, "loss": 0.9089, "step": 1388 }, { "epoch": 0.6293611236973267, "grad_norm": 1.1301062915550246, "learning_rate": 7.983754267354491e-05, "loss": 0.9507, "step": 1389 }, { "epoch": 0.6298142274580879, "grad_norm": 1.397710217421275, "learning_rate": 7.98364014011681e-05, "loss": 0.9205, "step": 1390 }, { "epoch": 0.6302673312188491, "grad_norm": 1.5235471556584077, "learning_rate": 7.983525614227191e-05, "loss": 0.9156, "step": 1391 }, { "epoch": 0.6307204349796103, "grad_norm": 1.3089488838524146, "learning_rate": 7.983410689697095e-05, "loss": 0.9209, "step": 1392 }, { "epoch": 0.6311735387403715, "grad_norm": 1.0233246148659256, "learning_rate": 7.983295366538027e-05, "loss": 0.9281, "step": 1393 }, { "epoch": 0.6316266425011328, "grad_norm": 1.903559883560404, "learning_rate": 7.983179644761524e-05, "loss": 0.9164, "step": 1394 }, { "epoch": 0.632079746261894, "grad_norm": 0.8554077601500665, "learning_rate": 7.983063524379169e-05, "loss": 0.9489, "step": 1395 }, { "epoch": 0.6325328500226551, "grad_norm": 1.661378487553581, "learning_rate": 7.98294700540258e-05, "loss": 0.9216, "step": 1396 }, { "epoch": 0.6329859537834164, "grad_norm": 0.9642751256827673, "learning_rate": 7.982830087843418e-05, "loss": 0.9181, "step": 1397 }, { "epoch": 0.6334390575441776, "grad_norm": 1.4784097954799926, "learning_rate": 7.982712771713385e-05, "loss": 0.9008, "step": 1398 }, { "epoch": 0.6338921613049389, "grad_norm": 1.6679730368140295, "learning_rate": 7.982595057024216e-05, "loss": 0.9316, "step": 1399 }, { "epoch": 0.6343452650657, "grad_norm": 1.5011609813086446, "learning_rate": 7.982476943787699e-05, "loss": 0.9301, "step": 1400 }, { "epoch": 0.6347983688264612, "grad_norm": 1.0665419367557876, "learning_rate": 7.982358432015648e-05, "loss": 0.9109, "step": 1401 }, { "epoch": 0.6352514725872225, "grad_norm": 1.54973453443042, "learning_rate": 7.982239521719923e-05, "loss": 0.9335, "step": 1402 }, { "epoch": 0.6357045763479837, "grad_norm": 1.2369014917510495, "learning_rate": 7.982120212912426e-05, "loss": 0.9543, "step": 1403 }, { "epoch": 0.6361576801087448, "grad_norm": 1.7519820363561864, "learning_rate": 7.982000505605096e-05, "loss": 0.8995, "step": 1404 }, { "epoch": 0.6366107838695061, "grad_norm": 1.155697174632588, "learning_rate": 7.981880399809911e-05, "loss": 0.9052, "step": 1405 }, { "epoch": 0.6370638876302673, "grad_norm": 1.2828250394294503, "learning_rate": 7.98175989553889e-05, "loss": 0.9321, "step": 1406 }, { "epoch": 0.6375169913910286, "grad_norm": 0.9967781324268324, "learning_rate": 7.981638992804094e-05, "loss": 0.9356, "step": 1407 }, { "epoch": 0.6379700951517897, "grad_norm": 1.2750656786253551, "learning_rate": 7.981517691617622e-05, "loss": 0.9291, "step": 1408 }, { "epoch": 0.638423198912551, "grad_norm": 1.0986744164088091, "learning_rate": 7.98139599199161e-05, "loss": 0.9511, "step": 1409 }, { "epoch": 0.6388763026733122, "grad_norm": 2.0521475721259246, "learning_rate": 7.981273893938241e-05, "loss": 0.9324, "step": 1410 }, { "epoch": 0.6393294064340734, "grad_norm": 1.2918750699852461, "learning_rate": 7.98115139746973e-05, "loss": 0.921, "step": 1411 }, { "epoch": 0.6397825101948346, "grad_norm": 1.3557729421915055, "learning_rate": 7.981028502598338e-05, "loss": 0.9013, "step": 1412 }, { "epoch": 0.6402356139555958, "grad_norm": 2.716897387794949, "learning_rate": 7.980905209336361e-05, "loss": 0.9072, "step": 1413 }, { "epoch": 0.640688717716357, "grad_norm": 1.6137976487823147, "learning_rate": 7.980781517696139e-05, "loss": 0.9409, "step": 1414 }, { "epoch": 0.6411418214771183, "grad_norm": 3.4510532724428713, "learning_rate": 7.98065742769005e-05, "loss": 0.9119, "step": 1415 }, { "epoch": 0.6415949252378794, "grad_norm": 3.3003085134140737, "learning_rate": 7.980532939330511e-05, "loss": 0.9273, "step": 1416 }, { "epoch": 0.6420480289986407, "grad_norm": 2.0280199364898586, "learning_rate": 7.980408052629982e-05, "loss": 0.9401, "step": 1417 }, { "epoch": 0.6425011327594019, "grad_norm": 1.9522724395301807, "learning_rate": 7.980282767600956e-05, "loss": 0.9209, "step": 1418 }, { "epoch": 0.6429542365201631, "grad_norm": 2.1686387466231105, "learning_rate": 7.980157084255978e-05, "loss": 0.9458, "step": 1419 }, { "epoch": 0.6434073402809243, "grad_norm": 1.5144416436673427, "learning_rate": 7.980031002607618e-05, "loss": 0.9304, "step": 1420 }, { "epoch": 0.6438604440416855, "grad_norm": 2.9842934289362315, "learning_rate": 7.979904522668497e-05, "loss": 0.9324, "step": 1421 }, { "epoch": 0.6443135478024468, "grad_norm": 2.519439415866549, "learning_rate": 7.979777644451272e-05, "loss": 0.9191, "step": 1422 }, { "epoch": 0.644766651563208, "grad_norm": 2.237308529037639, "learning_rate": 7.979650367968638e-05, "loss": 0.9107, "step": 1423 }, { "epoch": 0.6452197553239691, "grad_norm": 2.1669052105898636, "learning_rate": 7.979522693233336e-05, "loss": 0.9115, "step": 1424 }, { "epoch": 0.6456728590847304, "grad_norm": 1.9192508540681263, "learning_rate": 7.979394620258137e-05, "loss": 0.9359, "step": 1425 }, { "epoch": 0.6461259628454916, "grad_norm": 1.674930688517555, "learning_rate": 7.979266149055863e-05, "loss": 0.8941, "step": 1426 }, { "epoch": 0.6465790666062529, "grad_norm": 2.183176388281653, "learning_rate": 7.979137279639369e-05, "loss": 0.9229, "step": 1427 }, { "epoch": 0.647032170367014, "grad_norm": 1.5577666140382562, "learning_rate": 7.979008012021547e-05, "loss": 0.9312, "step": 1428 }, { "epoch": 0.6474852741277752, "grad_norm": 2.4961246103967536, "learning_rate": 7.978878346215338e-05, "loss": 0.8966, "step": 1429 }, { "epoch": 0.6479383778885365, "grad_norm": 2.1772720466326585, "learning_rate": 7.978748282233717e-05, "loss": 0.9405, "step": 1430 }, { "epoch": 0.6483914816492977, "grad_norm": 2.054307249824813, "learning_rate": 7.978617820089698e-05, "loss": 0.94, "step": 1431 }, { "epoch": 0.6488445854100588, "grad_norm": 1.6854261877982202, "learning_rate": 7.978486959796339e-05, "loss": 0.9175, "step": 1432 }, { "epoch": 0.6492976891708201, "grad_norm": 2.2065796813389613, "learning_rate": 7.978355701366733e-05, "loss": 0.9176, "step": 1433 }, { "epoch": 0.6497507929315813, "grad_norm": 1.901761070849703, "learning_rate": 7.978224044814017e-05, "loss": 0.9394, "step": 1434 }, { "epoch": 0.6502038966923426, "grad_norm": 2.4293938441867673, "learning_rate": 7.978091990151366e-05, "loss": 0.9161, "step": 1435 }, { "epoch": 0.6506570004531037, "grad_norm": 2.3362134809377464, "learning_rate": 7.977959537391995e-05, "loss": 0.9403, "step": 1436 }, { "epoch": 0.651110104213865, "grad_norm": 1.7917847153659117, "learning_rate": 7.977826686549159e-05, "loss": 0.9279, "step": 1437 }, { "epoch": 0.6515632079746262, "grad_norm": 1.6366159384506602, "learning_rate": 7.977693437636151e-05, "loss": 0.9057, "step": 1438 }, { "epoch": 0.6520163117353874, "grad_norm": 2.1054944492585808, "learning_rate": 7.977559790666308e-05, "loss": 0.9272, "step": 1439 }, { "epoch": 0.6524694154961486, "grad_norm": 1.6992254180748607, "learning_rate": 7.977425745653001e-05, "loss": 0.9022, "step": 1440 }, { "epoch": 0.6529225192569098, "grad_norm": 2.0492319241459196, "learning_rate": 7.97729130260965e-05, "loss": 0.9314, "step": 1441 }, { "epoch": 0.653375623017671, "grad_norm": 1.5972369628576721, "learning_rate": 7.977156461549702e-05, "loss": 0.9318, "step": 1442 }, { "epoch": 0.6538287267784323, "grad_norm": 2.4203832606045963, "learning_rate": 7.977021222486656e-05, "loss": 0.929, "step": 1443 }, { "epoch": 0.6542818305391934, "grad_norm": 2.015909104569495, "learning_rate": 7.976885585434044e-05, "loss": 0.923, "step": 1444 }, { "epoch": 0.6547349342999547, "grad_norm": 2.2146894961719368, "learning_rate": 7.97674955040544e-05, "loss": 0.9414, "step": 1445 }, { "epoch": 0.6551880380607159, "grad_norm": 1.952829022433904, "learning_rate": 7.976613117414454e-05, "loss": 0.94, "step": 1446 }, { "epoch": 0.6556411418214771, "grad_norm": 2.0445640530529703, "learning_rate": 7.976476286474745e-05, "loss": 0.921, "step": 1447 }, { "epoch": 0.6560942455822383, "grad_norm": 1.7498040522438674, "learning_rate": 7.976339057600001e-05, "loss": 0.9366, "step": 1448 }, { "epoch": 0.6565473493429995, "grad_norm": 2.172685606977037, "learning_rate": 7.976201430803958e-05, "loss": 0.9391, "step": 1449 }, { "epoch": 0.6570004531037608, "grad_norm": 1.8005029694581391, "learning_rate": 7.976063406100386e-05, "loss": 0.9313, "step": 1450 }, { "epoch": 0.657453556864522, "grad_norm": 2.11670706853161, "learning_rate": 7.9759249835031e-05, "loss": 0.9095, "step": 1451 }, { "epoch": 0.6579066606252831, "grad_norm": 1.6416300714177234, "learning_rate": 7.975786163025951e-05, "loss": 0.9072, "step": 1452 }, { "epoch": 0.6583597643860444, "grad_norm": 2.226561049095517, "learning_rate": 7.975646944682831e-05, "loss": 0.9121, "step": 1453 }, { "epoch": 0.6588128681468056, "grad_norm": 1.780193967042574, "learning_rate": 7.975507328487671e-05, "loss": 0.9366, "step": 1454 }, { "epoch": 0.6592659719075669, "grad_norm": 1.9979563331826375, "learning_rate": 7.975367314454445e-05, "loss": 0.907, "step": 1455 }, { "epoch": 0.659719075668328, "grad_norm": 1.517168896644689, "learning_rate": 7.975226902597163e-05, "loss": 0.9053, "step": 1456 }, { "epoch": 0.6601721794290892, "grad_norm": 2.2433884026149267, "learning_rate": 7.975086092929877e-05, "loss": 0.9004, "step": 1457 }, { "epoch": 0.6606252831898505, "grad_norm": 1.7578429888547933, "learning_rate": 7.974944885466677e-05, "loss": 0.9136, "step": 1458 }, { "epoch": 0.6610783869506117, "grad_norm": 2.1283515477661155, "learning_rate": 7.974803280221695e-05, "loss": 0.9054, "step": 1459 }, { "epoch": 0.6615314907113728, "grad_norm": 1.9118559608039538, "learning_rate": 7.974661277209102e-05, "loss": 0.9336, "step": 1460 }, { "epoch": 0.6619845944721341, "grad_norm": 1.7815296904690974, "learning_rate": 7.974518876443106e-05, "loss": 0.8885, "step": 1461 }, { "epoch": 0.6624376982328953, "grad_norm": 1.681526846964537, "learning_rate": 7.974376077937962e-05, "loss": 0.9138, "step": 1462 }, { "epoch": 0.6628908019936566, "grad_norm": 1.41643755593007, "learning_rate": 7.974232881707956e-05, "loss": 0.9207, "step": 1463 }, { "epoch": 0.6633439057544177, "grad_norm": 1.2131463483910536, "learning_rate": 7.97408928776742e-05, "loss": 0.9313, "step": 1464 }, { "epoch": 0.663797009515179, "grad_norm": 1.9359957541322972, "learning_rate": 7.973945296130723e-05, "loss": 0.9373, "step": 1465 }, { "epoch": 0.6642501132759402, "grad_norm": 0.9714081867384213, "learning_rate": 7.973800906812276e-05, "loss": 0.9444, "step": 1466 }, { "epoch": 0.6647032170367014, "grad_norm": 2.724035054214807, "learning_rate": 7.973656119826526e-05, "loss": 0.9437, "step": 1467 }, { "epoch": 0.6651563207974626, "grad_norm": 2.19841149029362, "learning_rate": 7.973510935187964e-05, "loss": 0.9342, "step": 1468 }, { "epoch": 0.6656094245582238, "grad_norm": 1.9391369798658704, "learning_rate": 7.973365352911119e-05, "loss": 0.916, "step": 1469 }, { "epoch": 0.666062528318985, "grad_norm": 2.0302660223118676, "learning_rate": 7.973219373010558e-05, "loss": 0.974, "step": 1470 }, { "epoch": 0.6665156320797463, "grad_norm": 2.088291312081727, "learning_rate": 7.973072995500891e-05, "loss": 0.9636, "step": 1471 }, { "epoch": 0.6669687358405075, "grad_norm": 1.491271528619439, "learning_rate": 7.972926220396765e-05, "loss": 0.9405, "step": 1472 }, { "epoch": 0.6674218396012687, "grad_norm": 1.2024416505311402, "learning_rate": 7.972779047712872e-05, "loss": 0.9288, "step": 1473 }, { "epoch": 0.6678749433620299, "grad_norm": 1.7711044581128248, "learning_rate": 7.972631477463935e-05, "loss": 0.9361, "step": 1474 }, { "epoch": 0.6683280471227911, "grad_norm": 1.1190554089603597, "learning_rate": 7.972483509664724e-05, "loss": 0.9341, "step": 1475 }, { "epoch": 0.6687811508835524, "grad_norm": 1.7406668655457365, "learning_rate": 7.972335144330046e-05, "loss": 0.9425, "step": 1476 }, { "epoch": 0.6692342546443135, "grad_norm": 1.2847870931980732, "learning_rate": 7.972186381474748e-05, "loss": 0.939, "step": 1477 }, { "epoch": 0.6696873584050748, "grad_norm": 1.4535016329224644, "learning_rate": 7.972037221113719e-05, "loss": 0.9301, "step": 1478 }, { "epoch": 0.670140462165836, "grad_norm": 1.2547209363295808, "learning_rate": 7.971887663261882e-05, "loss": 0.9323, "step": 1479 }, { "epoch": 0.6705935659265972, "grad_norm": 1.2209176277429672, "learning_rate": 7.971737707934208e-05, "loss": 0.9102, "step": 1480 }, { "epoch": 0.6710466696873584, "grad_norm": 1.2664867719254287, "learning_rate": 7.9715873551457e-05, "loss": 0.9236, "step": 1481 }, { "epoch": 0.6714997734481196, "grad_norm": 1.0876041566991859, "learning_rate": 7.971436604911406e-05, "loss": 0.9422, "step": 1482 }, { "epoch": 0.6719528772088809, "grad_norm": 1.3364177425032955, "learning_rate": 7.971285457246412e-05, "loss": 0.9265, "step": 1483 }, { "epoch": 0.6724059809696421, "grad_norm": 1.395404981684534, "learning_rate": 7.971133912165842e-05, "loss": 0.92, "step": 1484 }, { "epoch": 0.6728590847304032, "grad_norm": 1.23667726011024, "learning_rate": 7.970981969684862e-05, "loss": 0.9173, "step": 1485 }, { "epoch": 0.6733121884911645, "grad_norm": 1.086474922186941, "learning_rate": 7.97082962981868e-05, "loss": 0.9265, "step": 1486 }, { "epoch": 0.6737652922519257, "grad_norm": 1.0575883123908978, "learning_rate": 7.970676892582535e-05, "loss": 0.9261, "step": 1487 }, { "epoch": 0.674218396012687, "grad_norm": 1.2860222337357887, "learning_rate": 7.970523757991719e-05, "loss": 0.9301, "step": 1488 }, { "epoch": 0.6746714997734481, "grad_norm": 1.0395672278780517, "learning_rate": 7.970370226061551e-05, "loss": 0.918, "step": 1489 }, { "epoch": 0.6751246035342093, "grad_norm": 1.6655724954195152, "learning_rate": 7.970216296807399e-05, "loss": 0.9203, "step": 1490 }, { "epoch": 0.6755777072949706, "grad_norm": 1.024480495146598, "learning_rate": 7.970061970244664e-05, "loss": 0.9216, "step": 1491 }, { "epoch": 0.6760308110557318, "grad_norm": 1.3150453793273844, "learning_rate": 7.969907246388791e-05, "loss": 0.9389, "step": 1492 }, { "epoch": 0.676483914816493, "grad_norm": 1.134606530820478, "learning_rate": 7.969752125255264e-05, "loss": 0.934, "step": 1493 }, { "epoch": 0.6769370185772542, "grad_norm": 1.3392071606314806, "learning_rate": 7.969596606859607e-05, "loss": 0.9133, "step": 1494 }, { "epoch": 0.6773901223380154, "grad_norm": 1.0195830971379938, "learning_rate": 7.969440691217381e-05, "loss": 0.9306, "step": 1495 }, { "epoch": 0.6778432260987767, "grad_norm": 1.1462883338833707, "learning_rate": 7.96928437834419e-05, "loss": 0.9272, "step": 1496 }, { "epoch": 0.6782963298595378, "grad_norm": 1.9772083118636852, "learning_rate": 7.969127668255677e-05, "loss": 0.9643, "step": 1497 }, { "epoch": 0.678749433620299, "grad_norm": 0.7953918414697047, "learning_rate": 7.968970560967524e-05, "loss": 0.914, "step": 1498 }, { "epoch": 0.6792025373810603, "grad_norm": 1.4143736049752327, "learning_rate": 7.968813056495452e-05, "loss": 0.9428, "step": 1499 }, { "epoch": 0.6796556411418215, "grad_norm": 1.993840935480901, "learning_rate": 7.968655154855224e-05, "loss": 0.9247, "step": 1500 }, { "epoch": 0.6801087449025827, "grad_norm": 0.8679644127090915, "learning_rate": 7.968496856062643e-05, "loss": 0.9294, "step": 1501 }, { "epoch": 0.6805618486633439, "grad_norm": 2.5648891125788147, "learning_rate": 7.968338160133547e-05, "loss": 0.9142, "step": 1502 }, { "epoch": 0.6810149524241051, "grad_norm": 1.6781899789359822, "learning_rate": 7.96817906708382e-05, "loss": 0.9464, "step": 1503 }, { "epoch": 0.6814680561848664, "grad_norm": 2.61349264134114, "learning_rate": 7.96801957692938e-05, "loss": 0.9322, "step": 1504 }, { "epoch": 0.6819211599456275, "grad_norm": 2.022494549772285, "learning_rate": 7.96785968968619e-05, "loss": 0.9611, "step": 1505 }, { "epoch": 0.6823742637063888, "grad_norm": 2.2090833311635834, "learning_rate": 7.96769940537025e-05, "loss": 0.9205, "step": 1506 }, { "epoch": 0.68282736746715, "grad_norm": 1.4978341834333424, "learning_rate": 7.967538723997598e-05, "loss": 0.9377, "step": 1507 }, { "epoch": 0.6832804712279112, "grad_norm": 1.876089835221747, "learning_rate": 7.967377645584316e-05, "loss": 0.8906, "step": 1508 }, { "epoch": 0.6837335749886724, "grad_norm": 1.5189238828769376, "learning_rate": 7.967216170146523e-05, "loss": 0.9284, "step": 1509 }, { "epoch": 0.6841866787494336, "grad_norm": 1.4286563752597221, "learning_rate": 7.967054297700377e-05, "loss": 0.9285, "step": 1510 }, { "epoch": 0.6846397825101949, "grad_norm": 1.7932988316092071, "learning_rate": 7.966892028262079e-05, "loss": 0.9594, "step": 1511 }, { "epoch": 0.6850928862709561, "grad_norm": 0.9437300810409268, "learning_rate": 7.966729361847865e-05, "loss": 0.9009, "step": 1512 }, { "epoch": 0.6855459900317172, "grad_norm": 2.5359760157199087, "learning_rate": 7.966566298474015e-05, "loss": 0.9279, "step": 1513 }, { "epoch": 0.6859990937924785, "grad_norm": 1.8664761369028635, "learning_rate": 7.966402838156849e-05, "loss": 0.9324, "step": 1514 }, { "epoch": 0.6864521975532397, "grad_norm": 2.1224670855862384, "learning_rate": 7.96623898091272e-05, "loss": 0.9393, "step": 1515 }, { "epoch": 0.686905301314001, "grad_norm": 1.576692864355473, "learning_rate": 7.96607472675803e-05, "loss": 0.9146, "step": 1516 }, { "epoch": 0.6873584050747621, "grad_norm": 1.9782383964564167, "learning_rate": 7.965910075709214e-05, "loss": 0.9058, "step": 1517 }, { "epoch": 0.6878115088355233, "grad_norm": 1.2447179819064564, "learning_rate": 7.965745027782749e-05, "loss": 0.935, "step": 1518 }, { "epoch": 0.6882646125962846, "grad_norm": 2.013677721456696, "learning_rate": 7.965579582995154e-05, "loss": 0.9113, "step": 1519 }, { "epoch": 0.6887177163570458, "grad_norm": 1.1988293094977762, "learning_rate": 7.965413741362984e-05, "loss": 0.9232, "step": 1520 }, { "epoch": 0.689170820117807, "grad_norm": 1.5969577034256823, "learning_rate": 7.965247502902834e-05, "loss": 0.9109, "step": 1521 }, { "epoch": 0.6896239238785682, "grad_norm": 1.3388117183647692, "learning_rate": 7.96508086763134e-05, "loss": 0.9373, "step": 1522 }, { "epoch": 0.6900770276393294, "grad_norm": 1.225432152479181, "learning_rate": 7.96491383556518e-05, "loss": 0.9419, "step": 1523 }, { "epoch": 0.6905301314000907, "grad_norm": 1.6563534047638473, "learning_rate": 7.964746406721068e-05, "loss": 0.9319, "step": 1524 }, { "epoch": 0.6909832351608518, "grad_norm": 1.2803712281575839, "learning_rate": 7.964578581115757e-05, "loss": 0.9338, "step": 1525 }, { "epoch": 0.691436338921613, "grad_norm": 1.4594905120791821, "learning_rate": 7.964410358766045e-05, "loss": 0.936, "step": 1526 }, { "epoch": 0.6918894426823743, "grad_norm": 1.279021521033072, "learning_rate": 7.964241739688764e-05, "loss": 0.8871, "step": 1527 }, { "epoch": 0.6923425464431355, "grad_norm": 1.3570640848991316, "learning_rate": 7.964072723900788e-05, "loss": 0.9345, "step": 1528 }, { "epoch": 0.6927956502038967, "grad_norm": 1.1460493246613885, "learning_rate": 7.963903311419033e-05, "loss": 0.9254, "step": 1529 }, { "epoch": 0.6932487539646579, "grad_norm": 1.3879611090122517, "learning_rate": 7.963733502260452e-05, "loss": 0.9096, "step": 1530 }, { "epoch": 0.6937018577254191, "grad_norm": 0.9914277055362373, "learning_rate": 7.963563296442034e-05, "loss": 0.9257, "step": 1531 }, { "epoch": 0.6941549614861804, "grad_norm": 1.4097621874794533, "learning_rate": 7.96339269398082e-05, "loss": 0.9231, "step": 1532 }, { "epoch": 0.6946080652469415, "grad_norm": 1.1237395728274806, "learning_rate": 7.963221694893874e-05, "loss": 0.9318, "step": 1533 }, { "epoch": 0.6950611690077028, "grad_norm": 1.5206947323463877, "learning_rate": 7.963050299198313e-05, "loss": 0.9051, "step": 1534 }, { "epoch": 0.695514272768464, "grad_norm": 1.242904816877862, "learning_rate": 7.962878506911289e-05, "loss": 0.9352, "step": 1535 }, { "epoch": 0.6959673765292252, "grad_norm": 1.3856134031688778, "learning_rate": 7.962706318049991e-05, "loss": 0.9206, "step": 1536 }, { "epoch": 0.6964204802899864, "grad_norm": 1.1325944989210173, "learning_rate": 7.962533732631652e-05, "loss": 0.9137, "step": 1537 }, { "epoch": 0.6968735840507476, "grad_norm": 1.3887637616172588, "learning_rate": 7.962360750673545e-05, "loss": 0.9311, "step": 1538 }, { "epoch": 0.6973266878115089, "grad_norm": 1.148921200851106, "learning_rate": 7.962187372192978e-05, "loss": 0.921, "step": 1539 }, { "epoch": 0.6977797915722701, "grad_norm": 1.19069769158511, "learning_rate": 7.962013597207302e-05, "loss": 0.9306, "step": 1540 }, { "epoch": 0.6982328953330312, "grad_norm": 1.1221342894034125, "learning_rate": 7.961839425733907e-05, "loss": 0.9272, "step": 1541 }, { "epoch": 0.6986859990937925, "grad_norm": 1.3103085857879477, "learning_rate": 7.961664857790222e-05, "loss": 0.9251, "step": 1542 }, { "epoch": 0.6991391028545537, "grad_norm": 1.2444885761652855, "learning_rate": 7.961489893393718e-05, "loss": 0.9298, "step": 1543 }, { "epoch": 0.699592206615315, "grad_norm": 1.0511379911730347, "learning_rate": 7.961314532561905e-05, "loss": 0.9203, "step": 1544 }, { "epoch": 0.7000453103760761, "grad_norm": 1.2913930538342206, "learning_rate": 7.961138775312326e-05, "loss": 0.916, "step": 1545 }, { "epoch": 0.7004984141368373, "grad_norm": 1.504966232062948, "learning_rate": 7.960962621662578e-05, "loss": 0.8982, "step": 1546 }, { "epoch": 0.7009515178975986, "grad_norm": 1.11498615609885, "learning_rate": 7.960786071630282e-05, "loss": 0.9329, "step": 1547 }, { "epoch": 0.7014046216583598, "grad_norm": 1.0153193250145243, "learning_rate": 7.96060912523311e-05, "loss": 0.9229, "step": 1548 }, { "epoch": 0.701857725419121, "grad_norm": 1.1273656272147592, "learning_rate": 7.960431782488767e-05, "loss": 0.9179, "step": 1549 }, { "epoch": 0.7023108291798822, "grad_norm": 1.7375581608509032, "learning_rate": 7.960254043415002e-05, "loss": 0.9097, "step": 1550 }, { "epoch": 0.7027639329406434, "grad_norm": 0.8931552046853515, "learning_rate": 7.9600759080296e-05, "loss": 0.9363, "step": 1551 }, { "epoch": 0.7032170367014047, "grad_norm": 1.3505523557345995, "learning_rate": 7.959897376350388e-05, "loss": 0.9218, "step": 1552 }, { "epoch": 0.7036701404621658, "grad_norm": 1.6862759031600096, "learning_rate": 7.959718448395232e-05, "loss": 0.9452, "step": 1553 }, { "epoch": 0.704123244222927, "grad_norm": 1.2373592010275136, "learning_rate": 7.959539124182038e-05, "loss": 0.9203, "step": 1554 }, { "epoch": 0.7045763479836883, "grad_norm": 1.7381710367269645, "learning_rate": 7.959359403728752e-05, "loss": 0.9162, "step": 1555 }, { "epoch": 0.7050294517444495, "grad_norm": 0.9003610303961207, "learning_rate": 7.959179287053357e-05, "loss": 0.9202, "step": 1556 }, { "epoch": 0.7054825555052107, "grad_norm": 2.340549152266063, "learning_rate": 7.958998774173881e-05, "loss": 0.9498, "step": 1557 }, { "epoch": 0.7059356592659719, "grad_norm": 1.656677274905496, "learning_rate": 7.958817865108383e-05, "loss": 0.9223, "step": 1558 }, { "epoch": 0.7063887630267331, "grad_norm": 2.371032361197372, "learning_rate": 7.958636559874974e-05, "loss": 0.943, "step": 1559 }, { "epoch": 0.7068418667874944, "grad_norm": 29.736243340415434, "learning_rate": 7.958454858491792e-05, "loss": 1.0324, "step": 1560 }, { "epoch": 0.7072949705482555, "grad_norm": 14.135932267458571, "learning_rate": 7.958272760977023e-05, "loss": 1.096, "step": 1561 }, { "epoch": 0.7077480743090168, "grad_norm": 3.3080254889852494, "learning_rate": 7.958090267348889e-05, "loss": 0.9999, "step": 1562 }, { "epoch": 0.708201178069778, "grad_norm": 1.2746771970471613, "learning_rate": 7.957907377625653e-05, "loss": 0.9532, "step": 1563 }, { "epoch": 0.7086542818305392, "grad_norm": 1.789338175653644, "learning_rate": 7.957724091825616e-05, "loss": 0.9401, "step": 1564 }, { "epoch": 0.7091073855913004, "grad_norm": 1.6062907655331868, "learning_rate": 7.957540409967122e-05, "loss": 0.9361, "step": 1565 }, { "epoch": 0.7095604893520616, "grad_norm": 1.2994203341692163, "learning_rate": 7.95735633206855e-05, "loss": 0.9344, "step": 1566 }, { "epoch": 0.7100135931128229, "grad_norm": 1.533326411466143, "learning_rate": 7.957171858148323e-05, "loss": 0.93, "step": 1567 }, { "epoch": 0.7104666968735841, "grad_norm": 2.0179562191045632, "learning_rate": 7.9569869882249e-05, "loss": 0.9304, "step": 1568 }, { "epoch": 0.7109198006343452, "grad_norm": 1.0570148444145693, "learning_rate": 7.956801722316783e-05, "loss": 0.9568, "step": 1569 }, { "epoch": 0.7113729043951065, "grad_norm": 2.0945111615212464, "learning_rate": 7.956616060442512e-05, "loss": 0.9476, "step": 1570 }, { "epoch": 0.7118260081558677, "grad_norm": 1.5494536125084135, "learning_rate": 7.956430002620665e-05, "loss": 0.9213, "step": 1571 }, { "epoch": 0.712279111916629, "grad_norm": 1.4274792046927332, "learning_rate": 7.956243548869863e-05, "loss": 0.9426, "step": 1572 }, { "epoch": 0.7127322156773901, "grad_norm": 2.062410024037755, "learning_rate": 7.956056699208763e-05, "loss": 0.9274, "step": 1573 }, { "epoch": 0.7131853194381513, "grad_norm": 1.7059219722690793, "learning_rate": 7.955869453656066e-05, "loss": 0.9612, "step": 1574 }, { "epoch": 0.7136384231989126, "grad_norm": 1.904724378999752, "learning_rate": 7.955681812230507e-05, "loss": 0.9192, "step": 1575 }, { "epoch": 0.7140915269596738, "grad_norm": 0.9705482086616483, "learning_rate": 7.955493774950865e-05, "loss": 0.9299, "step": 1576 }, { "epoch": 0.714544630720435, "grad_norm": 1.8079412485935107, "learning_rate": 7.955305341835961e-05, "loss": 0.9519, "step": 1577 }, { "epoch": 0.7149977344811962, "grad_norm": 1.2362064161412092, "learning_rate": 7.955116512904645e-05, "loss": 0.9496, "step": 1578 }, { "epoch": 0.7154508382419574, "grad_norm": 1.6274653157550751, "learning_rate": 7.954927288175817e-05, "loss": 0.9539, "step": 1579 }, { "epoch": 0.7159039420027187, "grad_norm": 1.184374578612064, "learning_rate": 7.954737667668415e-05, "loss": 0.9161, "step": 1580 }, { "epoch": 0.7163570457634798, "grad_norm": 1.7197192042492877, "learning_rate": 7.954547651401413e-05, "loss": 0.9304, "step": 1581 }, { "epoch": 0.716810149524241, "grad_norm": 1.4434041699166074, "learning_rate": 7.954357239393826e-05, "loss": 0.9341, "step": 1582 }, { "epoch": 0.7172632532850023, "grad_norm": 1.3040217223299566, "learning_rate": 7.954166431664709e-05, "loss": 0.9173, "step": 1583 }, { "epoch": 0.7177163570457635, "grad_norm": 1.2813157963550774, "learning_rate": 7.953975228233157e-05, "loss": 0.9201, "step": 1584 }, { "epoch": 0.7181694608065247, "grad_norm": 1.4860973766812338, "learning_rate": 7.953783629118304e-05, "loss": 0.9045, "step": 1585 }, { "epoch": 0.7186225645672859, "grad_norm": 0.8570683470197646, "learning_rate": 7.953591634339324e-05, "loss": 0.9338, "step": 1586 }, { "epoch": 0.7190756683280471, "grad_norm": 0.9925369142125093, "learning_rate": 7.95339924391543e-05, "loss": 0.9363, "step": 1587 }, { "epoch": 0.7195287720888084, "grad_norm": 1.0723405228836504, "learning_rate": 7.953206457865876e-05, "loss": 0.9005, "step": 1588 }, { "epoch": 0.7199818758495695, "grad_norm": 1.1581986073579267, "learning_rate": 7.953013276209952e-05, "loss": 0.9255, "step": 1589 }, { "epoch": 0.7204349796103308, "grad_norm": 1.1985421306794328, "learning_rate": 7.952819698966991e-05, "loss": 0.9255, "step": 1590 }, { "epoch": 0.720888083371092, "grad_norm": 0.9661081332298359, "learning_rate": 7.952625726156368e-05, "loss": 0.9088, "step": 1591 }, { "epoch": 0.7213411871318532, "grad_norm": 1.0888773109894145, "learning_rate": 7.952431357797491e-05, "loss": 0.9286, "step": 1592 }, { "epoch": 0.7217942908926144, "grad_norm": 1.1094147866765955, "learning_rate": 7.952236593909812e-05, "loss": 0.9249, "step": 1593 }, { "epoch": 0.7222473946533756, "grad_norm": 1.954847086367977, "learning_rate": 7.952041434512819e-05, "loss": 0.9178, "step": 1594 }, { "epoch": 0.7227004984141369, "grad_norm": 0.7357547759744629, "learning_rate": 7.951845879626046e-05, "loss": 0.9317, "step": 1595 }, { "epoch": 0.7231536021748981, "grad_norm": 2.0423314578823484, "learning_rate": 7.951649929269061e-05, "loss": 0.9294, "step": 1596 }, { "epoch": 0.7236067059356592, "grad_norm": 1.3145439687053246, "learning_rate": 7.951453583461473e-05, "loss": 0.9413, "step": 1597 }, { "epoch": 0.7240598096964205, "grad_norm": 1.4471747211419124, "learning_rate": 7.95125684222293e-05, "loss": 0.9364, "step": 1598 }, { "epoch": 0.7245129134571817, "grad_norm": 1.5018542941437723, "learning_rate": 7.951059705573122e-05, "loss": 0.9089, "step": 1599 }, { "epoch": 0.724966017217943, "grad_norm": 1.421217907298113, "learning_rate": 7.950862173531776e-05, "loss": 0.9314, "step": 1600 }, { "epoch": 0.7254191209787041, "grad_norm": 1.3878240115788554, "learning_rate": 7.950664246118658e-05, "loss": 0.9313, "step": 1601 }, { "epoch": 0.7258722247394653, "grad_norm": 1.0456820697202744, "learning_rate": 7.950465923353576e-05, "loss": 0.9215, "step": 1602 }, { "epoch": 0.7263253285002266, "grad_norm": 1.2024737696248686, "learning_rate": 7.95026720525638e-05, "loss": 0.9237, "step": 1603 }, { "epoch": 0.7267784322609878, "grad_norm": 1.7780585267416678, "learning_rate": 7.950068091846953e-05, "loss": 0.9155, "step": 1604 }, { "epoch": 0.727231536021749, "grad_norm": 0.6856866558889602, "learning_rate": 7.949868583145218e-05, "loss": 0.9338, "step": 1605 }, { "epoch": 0.7276846397825102, "grad_norm": 1.087678943333294, "learning_rate": 7.949668679171146e-05, "loss": 0.9, "step": 1606 }, { "epoch": 0.7281377435432714, "grad_norm": 1.8294095989597894, "learning_rate": 7.949468379944738e-05, "loss": 0.916, "step": 1607 }, { "epoch": 0.7285908473040327, "grad_norm": 1.0507236685230854, "learning_rate": 7.949267685486041e-05, "loss": 0.93, "step": 1608 }, { "epoch": 0.7290439510647938, "grad_norm": 1.495711624776704, "learning_rate": 7.949066595815136e-05, "loss": 0.9057, "step": 1609 }, { "epoch": 0.729497054825555, "grad_norm": 0.8173453270375447, "learning_rate": 7.94886511095215e-05, "loss": 0.9253, "step": 1610 }, { "epoch": 0.7299501585863163, "grad_norm": 1.6910866189152207, "learning_rate": 7.948663230917243e-05, "loss": 0.9378, "step": 1611 }, { "epoch": 0.7304032623470775, "grad_norm": 1.079455379233352, "learning_rate": 7.948460955730619e-05, "loss": 0.9067, "step": 1612 }, { "epoch": 0.7308563661078387, "grad_norm": 1.8658315189875714, "learning_rate": 7.948258285412519e-05, "loss": 0.9246, "step": 1613 }, { "epoch": 0.7313094698685999, "grad_norm": 1.4072995422784378, "learning_rate": 7.948055219983226e-05, "loss": 0.9154, "step": 1614 }, { "epoch": 0.7317625736293611, "grad_norm": 1.777046661806918, "learning_rate": 7.947851759463061e-05, "loss": 0.9141, "step": 1615 }, { "epoch": 0.7322156773901224, "grad_norm": 1.5365732762021493, "learning_rate": 7.947647903872385e-05, "loss": 0.9345, "step": 1616 }, { "epoch": 0.7326687811508835, "grad_norm": 1.6002063628025618, "learning_rate": 7.947443653231597e-05, "loss": 0.9339, "step": 1617 }, { "epoch": 0.7331218849116448, "grad_norm": 1.3651531907316057, "learning_rate": 7.947239007561138e-05, "loss": 0.9234, "step": 1618 }, { "epoch": 0.733574988672406, "grad_norm": 1.3253452303484963, "learning_rate": 7.947033966881485e-05, "loss": 0.9289, "step": 1619 }, { "epoch": 0.7340280924331672, "grad_norm": 1.0019820566715991, "learning_rate": 7.94682853121316e-05, "loss": 0.9017, "step": 1620 }, { "epoch": 0.7344811961939284, "grad_norm": 1.3400682495615617, "learning_rate": 7.946622700576723e-05, "loss": 0.903, "step": 1621 }, { "epoch": 0.7349342999546896, "grad_norm": 0.9980843187910761, "learning_rate": 7.946416474992767e-05, "loss": 0.9217, "step": 1622 }, { "epoch": 0.7353874037154509, "grad_norm": 1.5733842878718594, "learning_rate": 7.946209854481931e-05, "loss": 0.9295, "step": 1623 }, { "epoch": 0.7358405074762121, "grad_norm": 1.2566615003586763, "learning_rate": 7.946002839064892e-05, "loss": 0.9303, "step": 1624 }, { "epoch": 0.7362936112369732, "grad_norm": 1.331161672804207, "learning_rate": 7.94579542876237e-05, "loss": 0.8905, "step": 1625 }, { "epoch": 0.7367467149977345, "grad_norm": 1.1216185379997974, "learning_rate": 7.945587623595118e-05, "loss": 0.8963, "step": 1626 }, { "epoch": 0.7371998187584957, "grad_norm": 1.2285445865778672, "learning_rate": 7.945379423583929e-05, "loss": 0.9239, "step": 1627 }, { "epoch": 0.737652922519257, "grad_norm": 1.4307358379865243, "learning_rate": 7.945170828749643e-05, "loss": 0.9303, "step": 1628 }, { "epoch": 0.7381060262800181, "grad_norm": 1.2493910294637256, "learning_rate": 7.944961839113132e-05, "loss": 0.9338, "step": 1629 }, { "epoch": 0.7385591300407793, "grad_norm": 0.94078867424509, "learning_rate": 7.944752454695309e-05, "loss": 0.9128, "step": 1630 }, { "epoch": 0.7390122338015406, "grad_norm": 1.5080266620952327, "learning_rate": 7.944542675517129e-05, "loss": 0.9296, "step": 1631 }, { "epoch": 0.7394653375623018, "grad_norm": 1.4934132173531194, "learning_rate": 7.944332501599587e-05, "loss": 0.9296, "step": 1632 }, { "epoch": 0.739918441323063, "grad_norm": 0.5906929139271772, "learning_rate": 7.944121932963714e-05, "loss": 0.9078, "step": 1633 }, { "epoch": 0.7403715450838242, "grad_norm": 1.4960187522237354, "learning_rate": 7.943910969630579e-05, "loss": 0.9258, "step": 1634 }, { "epoch": 0.7408246488445854, "grad_norm": 1.3347162929184622, "learning_rate": 7.943699611621297e-05, "loss": 0.9252, "step": 1635 }, { "epoch": 0.7412777526053467, "grad_norm": 1.0138392282738025, "learning_rate": 7.94348785895702e-05, "loss": 0.9062, "step": 1636 }, { "epoch": 0.7417308563661078, "grad_norm": 1.4901175654433148, "learning_rate": 7.943275711658936e-05, "loss": 0.9363, "step": 1637 }, { "epoch": 0.742183960126869, "grad_norm": 1.073585486629276, "learning_rate": 7.943063169748275e-05, "loss": 0.9027, "step": 1638 }, { "epoch": 0.7426370638876303, "grad_norm": 1.1086876018692526, "learning_rate": 7.94285023324631e-05, "loss": 0.9343, "step": 1639 }, { "epoch": 0.7430901676483915, "grad_norm": 1.2714064553254494, "learning_rate": 7.942636902174346e-05, "loss": 0.9435, "step": 1640 }, { "epoch": 0.7435432714091527, "grad_norm": 1.26254764488911, "learning_rate": 7.942423176553733e-05, "loss": 0.9231, "step": 1641 }, { "epoch": 0.7439963751699139, "grad_norm": 1.3610013218470323, "learning_rate": 7.94220905640586e-05, "loss": 0.9108, "step": 1642 }, { "epoch": 0.7444494789306751, "grad_norm": 1.0673991033113976, "learning_rate": 7.941994541752152e-05, "loss": 0.9076, "step": 1643 }, { "epoch": 0.7449025826914364, "grad_norm": 1.4554531558408619, "learning_rate": 7.94177963261408e-05, "loss": 0.9135, "step": 1644 }, { "epoch": 0.7453556864521975, "grad_norm": 1.0545420162005537, "learning_rate": 7.941564329013146e-05, "loss": 0.9036, "step": 1645 }, { "epoch": 0.7458087902129588, "grad_norm": 1.229022157294899, "learning_rate": 7.9413486309709e-05, "loss": 0.9093, "step": 1646 }, { "epoch": 0.74626189397372, "grad_norm": 1.2563250069409335, "learning_rate": 7.941132538508924e-05, "loss": 0.9161, "step": 1647 }, { "epoch": 0.7467149977344812, "grad_norm": 1.3021197509365945, "learning_rate": 7.940916051648843e-05, "loss": 0.9138, "step": 1648 }, { "epoch": 0.7471681014952424, "grad_norm": 1.108303573107042, "learning_rate": 7.940699170412325e-05, "loss": 0.9026, "step": 1649 }, { "epoch": 0.7476212052560036, "grad_norm": 1.3747608657267085, "learning_rate": 7.940481894821069e-05, "loss": 0.9407, "step": 1650 }, { "epoch": 0.7480743090167649, "grad_norm": 1.1819809304757796, "learning_rate": 7.940264224896822e-05, "loss": 0.9056, "step": 1651 }, { "epoch": 0.7485274127775261, "grad_norm": 1.3469516806140651, "learning_rate": 7.940046160661367e-05, "loss": 0.9232, "step": 1652 }, { "epoch": 0.7489805165382872, "grad_norm": 1.1839427167788623, "learning_rate": 7.939827702136521e-05, "loss": 0.9243, "step": 1653 }, { "epoch": 0.7494336202990485, "grad_norm": 1.3086145831816578, "learning_rate": 7.939608849344151e-05, "loss": 0.9098, "step": 1654 }, { "epoch": 0.7498867240598097, "grad_norm": 1.162430032201807, "learning_rate": 7.939389602306156e-05, "loss": 0.9039, "step": 1655 }, { "epoch": 0.750339827820571, "grad_norm": 1.3350031391257355, "learning_rate": 7.939169961044478e-05, "loss": 0.9159, "step": 1656 }, { "epoch": 0.7507929315813321, "grad_norm": 1.1371748396506922, "learning_rate": 7.938949925581095e-05, "loss": 0.9268, "step": 1657 }, { "epoch": 0.7512460353420933, "grad_norm": 1.4419155381400732, "learning_rate": 7.938729495938028e-05, "loss": 0.9361, "step": 1658 }, { "epoch": 0.7516991391028546, "grad_norm": 1.1460881940291805, "learning_rate": 7.938508672137334e-05, "loss": 0.9012, "step": 1659 }, { "epoch": 0.7521522428636158, "grad_norm": 1.014080426251179, "learning_rate": 7.938287454201112e-05, "loss": 0.9235, "step": 1660 }, { "epoch": 0.752605346624377, "grad_norm": 1.4962755521453661, "learning_rate": 7.938065842151501e-05, "loss": 0.9215, "step": 1661 }, { "epoch": 0.7530584503851382, "grad_norm": 1.1441254497139168, "learning_rate": 7.937843836010678e-05, "loss": 0.9052, "step": 1662 }, { "epoch": 0.7535115541458994, "grad_norm": 1.304028563680805, "learning_rate": 7.937621435800859e-05, "loss": 0.9047, "step": 1663 }, { "epoch": 0.7539646579066607, "grad_norm": 1.099104424912616, "learning_rate": 7.937398641544301e-05, "loss": 0.9066, "step": 1664 }, { "epoch": 0.7544177616674218, "grad_norm": 1.3996260717129383, "learning_rate": 7.937175453263297e-05, "loss": 0.9207, "step": 1665 }, { "epoch": 0.754870865428183, "grad_norm": 1.0226309259430966, "learning_rate": 7.936951870980186e-05, "loss": 0.9036, "step": 1666 }, { "epoch": 0.7553239691889443, "grad_norm": 1.240574175725854, "learning_rate": 7.936727894717337e-05, "loss": 0.9174, "step": 1667 }, { "epoch": 0.7557770729497055, "grad_norm": 1.440185288958416, "learning_rate": 7.93650352449717e-05, "loss": 0.9193, "step": 1668 }, { "epoch": 0.7562301767104667, "grad_norm": 0.9061026061469134, "learning_rate": 7.936278760342134e-05, "loss": 0.9227, "step": 1669 }, { "epoch": 0.7566832804712279, "grad_norm": 1.4540979716334974, "learning_rate": 7.936053602274722e-05, "loss": 0.9312, "step": 1670 }, { "epoch": 0.7571363842319891, "grad_norm": 6.120086089658885, "learning_rate": 7.935828050317468e-05, "loss": 0.9795, "step": 1671 }, { "epoch": 0.7575894879927504, "grad_norm": 1.9792800012198768, "learning_rate": 7.935602104492942e-05, "loss": 0.9561, "step": 1672 }, { "epoch": 0.7580425917535115, "grad_norm": 1.5623981049145281, "learning_rate": 7.935375764823756e-05, "loss": 0.934, "step": 1673 }, { "epoch": 0.7584956955142728, "grad_norm": 0.7851760745623121, "learning_rate": 7.935149031332559e-05, "loss": 0.9275, "step": 1674 }, { "epoch": 0.758948799275034, "grad_norm": 1.3550946572660878, "learning_rate": 7.934921904042042e-05, "loss": 0.9307, "step": 1675 }, { "epoch": 0.7594019030357952, "grad_norm": 1.382164540609309, "learning_rate": 7.934694382974932e-05, "loss": 0.9344, "step": 1676 }, { "epoch": 0.7598550067965564, "grad_norm": 1.254329351919454, "learning_rate": 7.934466468154002e-05, "loss": 0.9429, "step": 1677 }, { "epoch": 0.7603081105573176, "grad_norm": 1.3073378303658396, "learning_rate": 7.934238159602056e-05, "loss": 0.9217, "step": 1678 }, { "epoch": 0.7607612143180789, "grad_norm": 1.0912911816238922, "learning_rate": 7.93400945734194e-05, "loss": 0.9348, "step": 1679 }, { "epoch": 0.7612143180788401, "grad_norm": 1.6154144616613706, "learning_rate": 7.933780361396545e-05, "loss": 0.9332, "step": 1680 }, { "epoch": 0.7616674218396012, "grad_norm": 1.0330386384255639, "learning_rate": 7.933550871788796e-05, "loss": 0.9415, "step": 1681 }, { "epoch": 0.7621205256003625, "grad_norm": 1.3299706746205888, "learning_rate": 7.933320988541658e-05, "loss": 0.9212, "step": 1682 }, { "epoch": 0.7625736293611237, "grad_norm": 0.7347299950380057, "learning_rate": 7.933090711678134e-05, "loss": 0.9201, "step": 1683 }, { "epoch": 0.763026733121885, "grad_norm": 0.9255185200803742, "learning_rate": 7.932860041221272e-05, "loss": 0.9352, "step": 1684 }, { "epoch": 0.7634798368826461, "grad_norm": 1.0938612661789087, "learning_rate": 7.932628977194153e-05, "loss": 0.9359, "step": 1685 }, { "epoch": 0.7639329406434073, "grad_norm": 1.2503601841727772, "learning_rate": 7.932397519619902e-05, "loss": 0.9427, "step": 1686 }, { "epoch": 0.7643860444041686, "grad_norm": 1.5705619620563847, "learning_rate": 7.932165668521679e-05, "loss": 0.9218, "step": 1687 }, { "epoch": 0.7648391481649298, "grad_norm": 1.039130732357991, "learning_rate": 7.931933423922686e-05, "loss": 0.9448, "step": 1688 }, { "epoch": 0.765292251925691, "grad_norm": 1.6697804040166242, "learning_rate": 7.931700785846168e-05, "loss": 0.9346, "step": 1689 }, { "epoch": 0.7657453556864522, "grad_norm": 0.8288620816138257, "learning_rate": 7.931467754315403e-05, "loss": 0.9162, "step": 1690 }, { "epoch": 0.7661984594472134, "grad_norm": 1.2575581948028394, "learning_rate": 7.931234329353711e-05, "loss": 0.9353, "step": 1691 }, { "epoch": 0.7666515632079747, "grad_norm": 1.2443526179860087, "learning_rate": 7.931000510984451e-05, "loss": 0.9463, "step": 1692 }, { "epoch": 0.7671046669687358, "grad_norm": 1.5724807978626516, "learning_rate": 7.930766299231023e-05, "loss": 0.9355, "step": 1693 }, { "epoch": 0.767557770729497, "grad_norm": 1.2529092744228383, "learning_rate": 7.930531694116864e-05, "loss": 0.9066, "step": 1694 }, { "epoch": 0.7680108744902583, "grad_norm": 1.0491112889918064, "learning_rate": 7.930296695665451e-05, "loss": 0.9019, "step": 1695 }, { "epoch": 0.7684639782510195, "grad_norm": 0.9359208089813401, "learning_rate": 7.930061303900303e-05, "loss": 0.9086, "step": 1696 }, { "epoch": 0.7689170820117807, "grad_norm": 0.979189420944679, "learning_rate": 7.929825518844974e-05, "loss": 0.9038, "step": 1697 }, { "epoch": 0.7693701857725419, "grad_norm": 0.9955773901269453, "learning_rate": 7.92958934052306e-05, "loss": 0.9148, "step": 1698 }, { "epoch": 0.7698232895333031, "grad_norm": 1.8284588736908713, "learning_rate": 7.929352768958199e-05, "loss": 0.9189, "step": 1699 }, { "epoch": 0.7702763932940644, "grad_norm": 1.0313678811751394, "learning_rate": 7.929115804174061e-05, "loss": 0.9419, "step": 1700 }, { "epoch": 0.7707294970548255, "grad_norm": 1.653009820291984, "learning_rate": 7.928878446194361e-05, "loss": 0.9405, "step": 1701 }, { "epoch": 0.7711826008155868, "grad_norm": 0.942871680922851, "learning_rate": 7.928640695042851e-05, "loss": 0.9381, "step": 1702 }, { "epoch": 0.771635704576348, "grad_norm": 1.6833771988564978, "learning_rate": 7.928402550743325e-05, "loss": 0.9221, "step": 1703 }, { "epoch": 0.7720888083371092, "grad_norm": 1.2307313629448724, "learning_rate": 7.928164013319615e-05, "loss": 0.9415, "step": 1704 }, { "epoch": 0.7725419120978704, "grad_norm": 1.2740755857591537, "learning_rate": 7.927925082795591e-05, "loss": 0.9261, "step": 1705 }, { "epoch": 0.7729950158586316, "grad_norm": 1.0699556611764622, "learning_rate": 7.927685759195163e-05, "loss": 0.9337, "step": 1706 }, { "epoch": 0.7734481196193929, "grad_norm": 1.3575340292445424, "learning_rate": 7.927446042542282e-05, "loss": 0.9228, "step": 1707 }, { "epoch": 0.7739012233801541, "grad_norm": 1.0509967392383608, "learning_rate": 7.927205932860935e-05, "loss": 0.9187, "step": 1708 }, { "epoch": 0.7743543271409152, "grad_norm": 1.458709962883859, "learning_rate": 7.926965430175151e-05, "loss": 0.9102, "step": 1709 }, { "epoch": 0.7748074309016765, "grad_norm": 1.3888654831275487, "learning_rate": 7.926724534508997e-05, "loss": 0.9158, "step": 1710 }, { "epoch": 0.7752605346624377, "grad_norm": 1.0836487978426184, "learning_rate": 7.926483245886584e-05, "loss": 0.9162, "step": 1711 }, { "epoch": 0.775713638423199, "grad_norm": 1.2375716945154323, "learning_rate": 7.926241564332054e-05, "loss": 0.9042, "step": 1712 }, { "epoch": 0.7761667421839601, "grad_norm": 1.5783053796358861, "learning_rate": 7.925999489869593e-05, "loss": 0.9207, "step": 1713 }, { "epoch": 0.7766198459447213, "grad_norm": 0.923765423848734, "learning_rate": 7.925757022523428e-05, "loss": 0.913, "step": 1714 }, { "epoch": 0.7770729497054826, "grad_norm": 1.5441853852457839, "learning_rate": 7.925514162317822e-05, "loss": 0.915, "step": 1715 }, { "epoch": 0.7775260534662438, "grad_norm": 1.1174367147688762, "learning_rate": 7.925270909277078e-05, "loss": 0.9136, "step": 1716 }, { "epoch": 0.777979157227005, "grad_norm": 1.0195595240821824, "learning_rate": 7.92502726342554e-05, "loss": 0.9141, "step": 1717 }, { "epoch": 0.7784322609877662, "grad_norm": 1.5949181513741704, "learning_rate": 7.92478322478759e-05, "loss": 0.9449, "step": 1718 }, { "epoch": 0.7788853647485274, "grad_norm": 0.9794667950449146, "learning_rate": 7.924538793387649e-05, "loss": 0.9175, "step": 1719 }, { "epoch": 0.7793384685092887, "grad_norm": 1.6877237353867491, "learning_rate": 7.924293969250179e-05, "loss": 0.905, "step": 1720 }, { "epoch": 0.7797915722700498, "grad_norm": 1.1587116730574252, "learning_rate": 7.924048752399677e-05, "loss": 0.926, "step": 1721 }, { "epoch": 0.780244676030811, "grad_norm": 1.41702888343175, "learning_rate": 7.923803142860686e-05, "loss": 0.919, "step": 1722 }, { "epoch": 0.7806977797915723, "grad_norm": 1.1442022984867486, "learning_rate": 7.923557140657783e-05, "loss": 0.9218, "step": 1723 }, { "epoch": 0.7811508835523335, "grad_norm": 1.5917673618474943, "learning_rate": 7.923310745815586e-05, "loss": 0.9275, "step": 1724 }, { "epoch": 0.7816039873130947, "grad_norm": 1.073181974916838, "learning_rate": 7.923063958358753e-05, "loss": 0.9395, "step": 1725 }, { "epoch": 0.7820570910738559, "grad_norm": 1.6349856509446412, "learning_rate": 7.92281677831198e-05, "loss": 0.9181, "step": 1726 }, { "epoch": 0.7825101948346171, "grad_norm": 1.1042404093844298, "learning_rate": 7.922569205700003e-05, "loss": 0.923, "step": 1727 }, { "epoch": 0.7829632985953784, "grad_norm": 1.7005454281409376, "learning_rate": 7.922321240547599e-05, "loss": 0.9305, "step": 1728 }, { "epoch": 0.7834164023561395, "grad_norm": 1.3897781586286229, "learning_rate": 7.922072882879578e-05, "loss": 0.9431, "step": 1729 }, { "epoch": 0.7838695061169008, "grad_norm": 1.2766132158692967, "learning_rate": 7.921824132720797e-05, "loss": 0.9098, "step": 1730 }, { "epoch": 0.784322609877662, "grad_norm": 1.4662086930559242, "learning_rate": 7.92157499009615e-05, "loss": 0.8952, "step": 1731 }, { "epoch": 0.7847757136384232, "grad_norm": 1.264223014018469, "learning_rate": 7.921325455030567e-05, "loss": 0.9353, "step": 1732 }, { "epoch": 0.7852288173991844, "grad_norm": 1.5232519057301992, "learning_rate": 7.921075527549019e-05, "loss": 0.9141, "step": 1733 }, { "epoch": 0.7856819211599456, "grad_norm": 1.001500734430694, "learning_rate": 7.920825207676517e-05, "loss": 0.9181, "step": 1734 }, { "epoch": 0.7861350249207069, "grad_norm": 1.3453701656817847, "learning_rate": 7.920574495438114e-05, "loss": 0.9306, "step": 1735 }, { "epoch": 0.7865881286814681, "grad_norm": 1.5230421602532815, "learning_rate": 7.920323390858897e-05, "loss": 0.9219, "step": 1736 }, { "epoch": 0.7870412324422292, "grad_norm": 1.1152846125477798, "learning_rate": 7.920071893963995e-05, "loss": 0.8955, "step": 1737 }, { "epoch": 0.7874943362029905, "grad_norm": 1.2907554072488472, "learning_rate": 7.919820004778576e-05, "loss": 0.9008, "step": 1738 }, { "epoch": 0.7879474399637517, "grad_norm": 0.9647253457899815, "learning_rate": 7.919567723327847e-05, "loss": 0.923, "step": 1739 }, { "epoch": 0.788400543724513, "grad_norm": 1.4619004895165553, "learning_rate": 7.919315049637055e-05, "loss": 0.9323, "step": 1740 }, { "epoch": 0.7888536474852741, "grad_norm": 1.1981486948756248, "learning_rate": 7.919061983731484e-05, "loss": 0.9231, "step": 1741 }, { "epoch": 0.7893067512460353, "grad_norm": 0.9054438994783046, "learning_rate": 7.918808525636459e-05, "loss": 0.9303, "step": 1742 }, { "epoch": 0.7897598550067966, "grad_norm": 1.1916653293426982, "learning_rate": 7.918554675377346e-05, "loss": 0.9266, "step": 1743 }, { "epoch": 0.7902129587675578, "grad_norm": 1.6460295306856652, "learning_rate": 7.918300432979548e-05, "loss": 0.9188, "step": 1744 }, { "epoch": 0.790666062528319, "grad_norm": 0.9051561603575855, "learning_rate": 7.918045798468506e-05, "loss": 0.9223, "step": 1745 }, { "epoch": 0.7911191662890802, "grad_norm": 0.8899491516140119, "learning_rate": 7.917790771869705e-05, "loss": 0.9186, "step": 1746 }, { "epoch": 0.7915722700498414, "grad_norm": 0.960922544208992, "learning_rate": 7.917535353208661e-05, "loss": 0.8942, "step": 1747 }, { "epoch": 0.7920253738106027, "grad_norm": 1.278316304781657, "learning_rate": 7.917279542510939e-05, "loss": 0.9301, "step": 1748 }, { "epoch": 0.7924784775713638, "grad_norm": 1.354751041528369, "learning_rate": 7.917023339802136e-05, "loss": 0.9295, "step": 1749 }, { "epoch": 0.792931581332125, "grad_norm": 1.6103934870273433, "learning_rate": 7.916766745107891e-05, "loss": 0.8935, "step": 1750 }, { "epoch": 0.7933846850928863, "grad_norm": 0.8764750492023812, "learning_rate": 7.916509758453884e-05, "loss": 0.923, "step": 1751 }, { "epoch": 0.7938377888536475, "grad_norm": 1.175892913866176, "learning_rate": 7.916252379865832e-05, "loss": 0.9378, "step": 1752 }, { "epoch": 0.7942908926144087, "grad_norm": 1.7005509480780905, "learning_rate": 7.915994609369489e-05, "loss": 0.9261, "step": 1753 }, { "epoch": 0.7947439963751699, "grad_norm": 1.0749156154720225, "learning_rate": 7.915736446990651e-05, "loss": 0.9057, "step": 1754 }, { "epoch": 0.7951971001359311, "grad_norm": 1.5935857600526084, "learning_rate": 7.915477892755155e-05, "loss": 0.9403, "step": 1755 }, { "epoch": 0.7956502038966924, "grad_norm": 0.7406222562979111, "learning_rate": 7.915218946688874e-05, "loss": 0.9161, "step": 1756 }, { "epoch": 0.7961033076574535, "grad_norm": 1.3042310053294368, "learning_rate": 7.914959608817722e-05, "loss": 0.9239, "step": 1757 }, { "epoch": 0.7965564114182148, "grad_norm": 1.5252936664586199, "learning_rate": 7.91469987916765e-05, "loss": 0.9499, "step": 1758 }, { "epoch": 0.797009515178976, "grad_norm": 1.129625367208865, "learning_rate": 7.914439757764649e-05, "loss": 0.9449, "step": 1759 }, { "epoch": 0.7974626189397372, "grad_norm": 1.254839760408182, "learning_rate": 7.914179244634755e-05, "loss": 0.9127, "step": 1760 }, { "epoch": 0.7979157227004984, "grad_norm": 0.8874892378057997, "learning_rate": 7.913918339804031e-05, "loss": 0.9183, "step": 1761 }, { "epoch": 0.7983688264612596, "grad_norm": 1.382397662912848, "learning_rate": 7.913657043298593e-05, "loss": 0.9329, "step": 1762 }, { "epoch": 0.7988219302220209, "grad_norm": 1.1148464013020083, "learning_rate": 7.913395355144585e-05, "loss": 0.9125, "step": 1763 }, { "epoch": 0.7992750339827821, "grad_norm": 1.4995513322093859, "learning_rate": 7.913133275368197e-05, "loss": 0.9195, "step": 1764 }, { "epoch": 0.7997281377435432, "grad_norm": 0.9249296760124036, "learning_rate": 7.912870803995656e-05, "loss": 0.9314, "step": 1765 }, { "epoch": 0.8001812415043045, "grad_norm": 1.1873987946784397, "learning_rate": 7.912607941053227e-05, "loss": 0.9287, "step": 1766 }, { "epoch": 0.8006343452650657, "grad_norm": 0.9282947742622331, "learning_rate": 7.912344686567213e-05, "loss": 0.9107, "step": 1767 }, { "epoch": 0.801087449025827, "grad_norm": 1.03525700391265, "learning_rate": 7.912081040563964e-05, "loss": 0.9204, "step": 1768 }, { "epoch": 0.8015405527865881, "grad_norm": 0.8191221731205075, "learning_rate": 7.911817003069861e-05, "loss": 0.9259, "step": 1769 }, { "epoch": 0.8019936565473493, "grad_norm": 0.9070945621437749, "learning_rate": 7.911552574111325e-05, "loss": 0.9005, "step": 1770 }, { "epoch": 0.8024467603081106, "grad_norm": 0.9219164527558567, "learning_rate": 7.91128775371482e-05, "loss": 0.9196, "step": 1771 }, { "epoch": 0.8028998640688718, "grad_norm": 1.0841929386729627, "learning_rate": 7.911022541906847e-05, "loss": 0.919, "step": 1772 }, { "epoch": 0.803352967829633, "grad_norm": 1.7796842376675206, "learning_rate": 7.910756938713946e-05, "loss": 0.9311, "step": 1773 }, { "epoch": 0.8038060715903942, "grad_norm": 0.959117220878085, "learning_rate": 7.910490944162697e-05, "loss": 0.9141, "step": 1774 }, { "epoch": 0.8042591753511554, "grad_norm": 0.9005365856777418, "learning_rate": 7.910224558279718e-05, "loss": 0.9033, "step": 1775 }, { "epoch": 0.8047122791119167, "grad_norm": 1.1023823970126387, "learning_rate": 7.909957781091668e-05, "loss": 0.9261, "step": 1776 }, { "epoch": 0.8051653828726778, "grad_norm": 1.0930597015385468, "learning_rate": 7.909690612625242e-05, "loss": 0.9018, "step": 1777 }, { "epoch": 0.805618486633439, "grad_norm": 1.5875113859827468, "learning_rate": 7.909423052907177e-05, "loss": 0.9332, "step": 1778 }, { "epoch": 0.8060715903942003, "grad_norm": 1.1743952902416748, "learning_rate": 7.909155101964249e-05, "loss": 0.9205, "step": 1779 }, { "epoch": 0.8065246941549615, "grad_norm": 1.0115588584030366, "learning_rate": 7.908886759823272e-05, "loss": 0.9152, "step": 1780 }, { "epoch": 0.8069777979157227, "grad_norm": 1.4082683173986517, "learning_rate": 7.9086180265111e-05, "loss": 0.9064, "step": 1781 }, { "epoch": 0.8074309016764839, "grad_norm": 1.0221168098692175, "learning_rate": 7.908348902054625e-05, "loss": 0.8999, "step": 1782 }, { "epoch": 0.8078840054372451, "grad_norm": 1.8885563552094968, "learning_rate": 7.90807938648078e-05, "loss": 0.9068, "step": 1783 }, { "epoch": 0.8083371091980064, "grad_norm": 0.8727483652947631, "learning_rate": 7.907809479816536e-05, "loss": 0.9115, "step": 1784 }, { "epoch": 0.8087902129587675, "grad_norm": 1.610807190301351, "learning_rate": 7.9075391820889e-05, "loss": 0.935, "step": 1785 }, { "epoch": 0.8092433167195288, "grad_norm": 1.7584570030103646, "learning_rate": 7.907268493324927e-05, "loss": 0.9326, "step": 1786 }, { "epoch": 0.80969642048029, "grad_norm": 1.0631456428882067, "learning_rate": 7.9069974135517e-05, "loss": 0.9268, "step": 1787 }, { "epoch": 0.8101495242410512, "grad_norm": 1.5535098158976692, "learning_rate": 7.906725942796352e-05, "loss": 0.9027, "step": 1788 }, { "epoch": 0.8106026280018124, "grad_norm": 1.4954477564921043, "learning_rate": 7.906454081086044e-05, "loss": 0.9307, "step": 1789 }, { "epoch": 0.8110557317625736, "grad_norm": 1.0671450785149008, "learning_rate": 7.906181828447984e-05, "loss": 0.8985, "step": 1790 }, { "epoch": 0.8115088355233349, "grad_norm": 1.0173138941450286, "learning_rate": 7.905909184909418e-05, "loss": 0.8925, "step": 1791 }, { "epoch": 0.8119619392840961, "grad_norm": 1.1723871322465886, "learning_rate": 7.90563615049763e-05, "loss": 0.9157, "step": 1792 }, { "epoch": 0.8124150430448572, "grad_norm": 1.5962686329726647, "learning_rate": 7.905362725239943e-05, "loss": 0.8993, "step": 1793 }, { "epoch": 0.8128681468056185, "grad_norm": 0.9426538481064299, "learning_rate": 7.905088909163717e-05, "loss": 0.9198, "step": 1794 }, { "epoch": 0.8133212505663797, "grad_norm": 1.4242401220356657, "learning_rate": 7.904814702296357e-05, "loss": 0.9254, "step": 1795 }, { "epoch": 0.813774354327141, "grad_norm": 1.1681353488744177, "learning_rate": 7.904540104665301e-05, "loss": 0.9249, "step": 1796 }, { "epoch": 0.8142274580879021, "grad_norm": 0.9951719379411494, "learning_rate": 7.904265116298029e-05, "loss": 0.8906, "step": 1797 }, { "epoch": 0.8146805618486633, "grad_norm": 1.427400214946877, "learning_rate": 7.903989737222061e-05, "loss": 0.9176, "step": 1798 }, { "epoch": 0.8151336656094246, "grad_norm": 1.2714178292549803, "learning_rate": 7.903713967464953e-05, "loss": 0.9058, "step": 1799 }, { "epoch": 0.8155867693701858, "grad_norm": 1.5241887282748636, "learning_rate": 7.903437807054303e-05, "loss": 0.9353, "step": 1800 }, { "epoch": 0.816039873130947, "grad_norm": 0.9974886109831907, "learning_rate": 7.903161256017746e-05, "loss": 0.9046, "step": 1801 }, { "epoch": 0.8164929768917082, "grad_norm": 1.3020628231145757, "learning_rate": 7.90288431438296e-05, "loss": 0.9178, "step": 1802 }, { "epoch": 0.8169460806524694, "grad_norm": 0.8880093667874915, "learning_rate": 7.902606982177656e-05, "loss": 0.9345, "step": 1803 }, { "epoch": 0.8173991844132307, "grad_norm": 1.4290378935121837, "learning_rate": 7.90232925942959e-05, "loss": 0.9238, "step": 1804 }, { "epoch": 0.8178522881739918, "grad_norm": 0.8990067038359386, "learning_rate": 7.90205114616655e-05, "loss": 0.9099, "step": 1805 }, { "epoch": 0.818305391934753, "grad_norm": 1.7720506449214675, "learning_rate": 7.901772642416371e-05, "loss": 0.9171, "step": 1806 }, { "epoch": 0.8187584956955143, "grad_norm": 1.2415812915582125, "learning_rate": 7.901493748206923e-05, "loss": 0.9224, "step": 1807 }, { "epoch": 0.8192115994562755, "grad_norm": 1.5268201636572651, "learning_rate": 7.901214463566115e-05, "loss": 0.9276, "step": 1808 }, { "epoch": 0.8196647032170367, "grad_norm": 1.352640771447672, "learning_rate": 7.900934788521895e-05, "loss": 0.918, "step": 1809 }, { "epoch": 0.8201178069777979, "grad_norm": 1.3079713842053466, "learning_rate": 7.900654723102254e-05, "loss": 0.9221, "step": 1810 }, { "epoch": 0.8205709107385591, "grad_norm": 1.1359639592845872, "learning_rate": 7.900374267335214e-05, "loss": 0.9242, "step": 1811 }, { "epoch": 0.8210240144993204, "grad_norm": 1.084407925964234, "learning_rate": 7.900093421248845e-05, "loss": 0.9031, "step": 1812 }, { "epoch": 0.8214771182600815, "grad_norm": 1.339313227952597, "learning_rate": 7.899812184871248e-05, "loss": 0.9291, "step": 1813 }, { "epoch": 0.8219302220208428, "grad_norm": 1.3333596002603372, "learning_rate": 7.89953055823057e-05, "loss": 0.9406, "step": 1814 }, { "epoch": 0.822383325781604, "grad_norm": 1.028800039139251, "learning_rate": 7.899248541354994e-05, "loss": 0.9271, "step": 1815 }, { "epoch": 0.8228364295423652, "grad_norm": 1.108158632863253, "learning_rate": 7.898966134272739e-05, "loss": 0.9264, "step": 1816 }, { "epoch": 0.8232895333031264, "grad_norm": 1.2190671068564365, "learning_rate": 7.898683337012071e-05, "loss": 0.9139, "step": 1817 }, { "epoch": 0.8237426370638876, "grad_norm": 1.696896200499147, "learning_rate": 7.898400149601286e-05, "loss": 0.912, "step": 1818 }, { "epoch": 0.8241957408246489, "grad_norm": 0.9337590201769559, "learning_rate": 7.898116572068725e-05, "loss": 0.9039, "step": 1819 }, { "epoch": 0.8246488445854101, "grad_norm": 1.5774053485466515, "learning_rate": 7.897832604442766e-05, "loss": 0.9309, "step": 1820 }, { "epoch": 0.8251019483461712, "grad_norm": 1.815754519995014, "learning_rate": 7.897548246751826e-05, "loss": 0.921, "step": 1821 }, { "epoch": 0.8255550521069325, "grad_norm": 1.0042828793086418, "learning_rate": 7.89726349902436e-05, "loss": 0.9264, "step": 1822 }, { "epoch": 0.8260081558676937, "grad_norm": 2.297883113182183, "learning_rate": 7.896978361288867e-05, "loss": 0.9346, "step": 1823 }, { "epoch": 0.826461259628455, "grad_norm": 1.5592501170675122, "learning_rate": 7.896692833573878e-05, "loss": 0.921, "step": 1824 }, { "epoch": 0.8269143633892161, "grad_norm": 2.1366744771696475, "learning_rate": 7.896406915907968e-05, "loss": 0.9065, "step": 1825 }, { "epoch": 0.8273674671499773, "grad_norm": 1.8416468397931072, "learning_rate": 7.896120608319751e-05, "loss": 0.9231, "step": 1826 }, { "epoch": 0.8278205709107386, "grad_norm": 2.1361626465464396, "learning_rate": 7.895833910837873e-05, "loss": 0.9207, "step": 1827 }, { "epoch": 0.8282736746714998, "grad_norm": 1.758392289474006, "learning_rate": 7.89554682349103e-05, "loss": 0.9083, "step": 1828 }, { "epoch": 0.828726778432261, "grad_norm": 1.9336967419470057, "learning_rate": 7.89525934630795e-05, "loss": 0.9092, "step": 1829 }, { "epoch": 0.8291798821930222, "grad_norm": 1.86673592883596, "learning_rate": 7.894971479317401e-05, "loss": 0.93, "step": 1830 }, { "epoch": 0.8296329859537834, "grad_norm": 1.66302678773131, "learning_rate": 7.894683222548189e-05, "loss": 0.925, "step": 1831 }, { "epoch": 0.8300860897145447, "grad_norm": 1.3498926598896202, "learning_rate": 7.894394576029163e-05, "loss": 0.9058, "step": 1832 }, { "epoch": 0.8305391934753058, "grad_norm": 2.091067162219251, "learning_rate": 7.894105539789207e-05, "loss": 0.923, "step": 1833 }, { "epoch": 0.830992297236067, "grad_norm": 1.5627006283036935, "learning_rate": 7.893816113857246e-05, "loss": 0.9224, "step": 1834 }, { "epoch": 0.8314454009968283, "grad_norm": 2.17328640632478, "learning_rate": 7.893526298262244e-05, "loss": 0.9368, "step": 1835 }, { "epoch": 0.8318985047575895, "grad_norm": 1.887647663686509, "learning_rate": 7.893236093033203e-05, "loss": 0.9078, "step": 1836 }, { "epoch": 0.8323516085183507, "grad_norm": 1.6435633969682024, "learning_rate": 7.892945498199166e-05, "loss": 0.8906, "step": 1837 }, { "epoch": 0.8328047122791119, "grad_norm": 1.6766581925068147, "learning_rate": 7.89265451378921e-05, "loss": 0.8999, "step": 1838 }, { "epoch": 0.8332578160398731, "grad_norm": 1.332370649320404, "learning_rate": 7.892363139832456e-05, "loss": 0.9174, "step": 1839 }, { "epoch": 0.8337109198006344, "grad_norm": 1.394317996871938, "learning_rate": 7.892071376358065e-05, "loss": 0.9053, "step": 1840 }, { "epoch": 0.8341640235613955, "grad_norm": 1.5467783625118023, "learning_rate": 7.891779223395231e-05, "loss": 0.931, "step": 1841 }, { "epoch": 0.8346171273221568, "grad_norm": 1.0550849543705043, "learning_rate": 7.891486680973191e-05, "loss": 0.9324, "step": 1842 }, { "epoch": 0.835070231082918, "grad_norm": 1.4590136439838826, "learning_rate": 7.891193749121224e-05, "loss": 0.9079, "step": 1843 }, { "epoch": 0.8355233348436792, "grad_norm": 0.6781886360178417, "learning_rate": 7.890900427868639e-05, "loss": 0.9135, "step": 1844 }, { "epoch": 0.8359764386044404, "grad_norm": 1.5370161732882832, "learning_rate": 7.890606717244794e-05, "loss": 0.9219, "step": 1845 }, { "epoch": 0.8364295423652016, "grad_norm": 0.8091502787817443, "learning_rate": 7.890312617279078e-05, "loss": 0.9177, "step": 1846 }, { "epoch": 0.8368826461259629, "grad_norm": 2.0886442028446752, "learning_rate": 7.890018128000924e-05, "loss": 0.9024, "step": 1847 }, { "epoch": 0.8373357498867241, "grad_norm": 1.5758914921049614, "learning_rate": 7.8897232494398e-05, "loss": 0.905, "step": 1848 }, { "epoch": 0.8377888536474852, "grad_norm": 1.66391013344371, "learning_rate": 7.889427981625218e-05, "loss": 0.9172, "step": 1849 }, { "epoch": 0.8382419574082465, "grad_norm": 1.6059617728985744, "learning_rate": 7.889132324586725e-05, "loss": 0.9145, "step": 1850 }, { "epoch": 0.8386950611690077, "grad_norm": 1.3700682872073866, "learning_rate": 7.888836278353908e-05, "loss": 0.9211, "step": 1851 }, { "epoch": 0.839148164929769, "grad_norm": 1.232491280033139, "learning_rate": 7.888539842956393e-05, "loss": 0.9143, "step": 1852 }, { "epoch": 0.8396012686905301, "grad_norm": 1.482307711709211, "learning_rate": 7.888243018423845e-05, "loss": 0.912, "step": 1853 }, { "epoch": 0.8400543724512913, "grad_norm": 1.0553983088544558, "learning_rate": 7.887945804785967e-05, "loss": 0.9416, "step": 1854 }, { "epoch": 0.8405074762120526, "grad_norm": 2.117543628676045, "learning_rate": 7.887648202072503e-05, "loss": 0.9226, "step": 1855 }, { "epoch": 0.8409605799728138, "grad_norm": 1.9693671002332453, "learning_rate": 7.887350210313236e-05, "loss": 0.9286, "step": 1856 }, { "epoch": 0.841413683733575, "grad_norm": 1.2356287886592714, "learning_rate": 7.887051829537984e-05, "loss": 0.921, "step": 1857 }, { "epoch": 0.8418667874943362, "grad_norm": 1.2844894345714517, "learning_rate": 7.886753059776608e-05, "loss": 0.8919, "step": 1858 }, { "epoch": 0.8423198912550974, "grad_norm": 1.5558507458760145, "learning_rate": 7.886453901059008e-05, "loss": 0.8956, "step": 1859 }, { "epoch": 0.8427729950158587, "grad_norm": 1.2409478771349276, "learning_rate": 7.886154353415118e-05, "loss": 0.922, "step": 1860 }, { "epoch": 0.8432260987766198, "grad_norm": 1.225624258873219, "learning_rate": 7.885854416874916e-05, "loss": 0.8987, "step": 1861 }, { "epoch": 0.843679202537381, "grad_norm": 0.9008930392279068, "learning_rate": 7.885554091468418e-05, "loss": 0.9044, "step": 1862 }, { "epoch": 0.8441323062981423, "grad_norm": 1.6709341038944834, "learning_rate": 7.885253377225679e-05, "loss": 0.8928, "step": 1863 }, { "epoch": 0.8445854100589035, "grad_norm": 1.5482110053019846, "learning_rate": 7.884952274176791e-05, "loss": 0.9406, "step": 1864 }, { "epoch": 0.8450385138196647, "grad_norm": 1.1364566224269699, "learning_rate": 7.884650782351886e-05, "loss": 0.9098, "step": 1865 }, { "epoch": 0.8454916175804259, "grad_norm": 0.7460064230387393, "learning_rate": 7.884348901781134e-05, "loss": 0.9112, "step": 1866 }, { "epoch": 0.8459447213411871, "grad_norm": 1.1773933527254643, "learning_rate": 7.884046632494747e-05, "loss": 0.8928, "step": 1867 }, { "epoch": 0.8463978251019484, "grad_norm": 1.3508700918185792, "learning_rate": 7.883743974522972e-05, "loss": 0.9101, "step": 1868 }, { "epoch": 0.8468509288627095, "grad_norm": 1.6766193871920598, "learning_rate": 7.883440927896098e-05, "loss": 0.9037, "step": 1869 }, { "epoch": 0.8473040326234708, "grad_norm": 0.8073158674103862, "learning_rate": 7.883137492644452e-05, "loss": 0.9182, "step": 1870 }, { "epoch": 0.847757136384232, "grad_norm": 0.9380441464747004, "learning_rate": 7.882833668798397e-05, "loss": 0.9245, "step": 1871 }, { "epoch": 0.8482102401449932, "grad_norm": 1.1650124416681347, "learning_rate": 7.88252945638834e-05, "loss": 0.9316, "step": 1872 }, { "epoch": 0.8486633439057544, "grad_norm": 1.3062091499877526, "learning_rate": 7.882224855444723e-05, "loss": 0.8949, "step": 1873 }, { "epoch": 0.8491164476665156, "grad_norm": 1.7472503057468867, "learning_rate": 7.881919865998026e-05, "loss": 0.9438, "step": 1874 }, { "epoch": 0.8495695514272769, "grad_norm": 0.9811090772255636, "learning_rate": 7.881614488078775e-05, "loss": 0.9357, "step": 1875 }, { "epoch": 0.8500226551880381, "grad_norm": 1.4083389322153606, "learning_rate": 7.881308721717525e-05, "loss": 0.9314, "step": 1876 }, { "epoch": 0.8504757589487992, "grad_norm": 1.1571287358701074, "learning_rate": 7.881002566944878e-05, "loss": 0.9055, "step": 1877 }, { "epoch": 0.8509288627095605, "grad_norm": 2.0357990755006417, "learning_rate": 7.88069602379147e-05, "loss": 0.9151, "step": 1878 }, { "epoch": 0.8513819664703217, "grad_norm": 1.138707199614341, "learning_rate": 7.880389092287977e-05, "loss": 0.9032, "step": 1879 }, { "epoch": 0.851835070231083, "grad_norm": 2.049219925663314, "learning_rate": 7.880081772465117e-05, "loss": 0.9055, "step": 1880 }, { "epoch": 0.8522881739918441, "grad_norm": 1.7750639078552357, "learning_rate": 7.879774064353641e-05, "loss": 0.9197, "step": 1881 }, { "epoch": 0.8527412777526053, "grad_norm": 1.2171729741495967, "learning_rate": 7.879465967984343e-05, "loss": 0.9147, "step": 1882 }, { "epoch": 0.8531943815133666, "grad_norm": 2.71224242443362, "learning_rate": 7.879157483388055e-05, "loss": 0.9125, "step": 1883 }, { "epoch": 0.8536474852741278, "grad_norm": 2.1672726802969664, "learning_rate": 7.87884861059565e-05, "loss": 0.9154, "step": 1884 }, { "epoch": 0.854100589034889, "grad_norm": 2.321301348860796, "learning_rate": 7.878539349638035e-05, "loss": 0.9157, "step": 1885 }, { "epoch": 0.8545536927956502, "grad_norm": 1.822387501757745, "learning_rate": 7.878229700546159e-05, "loss": 0.9327, "step": 1886 }, { "epoch": 0.8550067965564114, "grad_norm": 2.548952129644567, "learning_rate": 7.877919663351009e-05, "loss": 0.9332, "step": 1887 }, { "epoch": 0.8554599003171727, "grad_norm": 1.697873783079806, "learning_rate": 7.877609238083611e-05, "loss": 0.9282, "step": 1888 }, { "epoch": 0.8559130040779338, "grad_norm": 3.346476646146668, "learning_rate": 7.877298424775032e-05, "loss": 0.9407, "step": 1889 }, { "epoch": 0.856366107838695, "grad_norm": 3.0645815174309985, "learning_rate": 7.876987223456371e-05, "loss": 0.9407, "step": 1890 }, { "epoch": 0.8568192115994563, "grad_norm": 1.4686426921330804, "learning_rate": 7.876675634158779e-05, "loss": 0.9331, "step": 1891 }, { "epoch": 0.8572723153602175, "grad_norm": 1.840169079128824, "learning_rate": 7.876363656913428e-05, "loss": 0.9122, "step": 1892 }, { "epoch": 0.8577254191209787, "grad_norm": 1.683913717382428, "learning_rate": 7.876051291751545e-05, "loss": 0.9362, "step": 1893 }, { "epoch": 0.8581785228817399, "grad_norm": 1.2075115646111176, "learning_rate": 7.875738538704387e-05, "loss": 0.9026, "step": 1894 }, { "epoch": 0.8586316266425011, "grad_norm": 2.02920787921519, "learning_rate": 7.87542539780325e-05, "loss": 0.9332, "step": 1895 }, { "epoch": 0.8590847304032624, "grad_norm": 1.3814895971336176, "learning_rate": 7.875111869079472e-05, "loss": 0.9261, "step": 1896 }, { "epoch": 0.8595378341640235, "grad_norm": 2.479498400631507, "learning_rate": 7.874797952564431e-05, "loss": 0.9131, "step": 1897 }, { "epoch": 0.8599909379247848, "grad_norm": 1.8378122255017977, "learning_rate": 7.874483648289538e-05, "loss": 0.9069, "step": 1898 }, { "epoch": 0.860444041685546, "grad_norm": 2.368316042687256, "learning_rate": 7.874168956286248e-05, "loss": 0.9197, "step": 1899 }, { "epoch": 0.8608971454463072, "grad_norm": 2.0918572901121943, "learning_rate": 7.873853876586052e-05, "loss": 0.9361, "step": 1900 }, { "epoch": 0.8613502492070684, "grad_norm": 1.8117636321920358, "learning_rate": 7.873538409220482e-05, "loss": 0.9143, "step": 1901 }, { "epoch": 0.8618033529678296, "grad_norm": 1.9110051477691556, "learning_rate": 7.873222554221105e-05, "loss": 0.9077, "step": 1902 }, { "epoch": 0.8622564567285909, "grad_norm": 1.8283869445521934, "learning_rate": 7.87290631161953e-05, "loss": 0.8992, "step": 1903 }, { "epoch": 0.8627095604893521, "grad_norm": 1.6373278530384128, "learning_rate": 7.872589681447407e-05, "loss": 0.9308, "step": 1904 }, { "epoch": 0.8631626642501132, "grad_norm": 1.6166948455348582, "learning_rate": 7.87227266373642e-05, "loss": 0.9048, "step": 1905 }, { "epoch": 0.8636157680108745, "grad_norm": 1.6691755356273732, "learning_rate": 7.871955258518294e-05, "loss": 0.9156, "step": 1906 }, { "epoch": 0.8640688717716357, "grad_norm": 1.0186856904671497, "learning_rate": 7.87163746582479e-05, "loss": 0.9095, "step": 1907 }, { "epoch": 0.864521975532397, "grad_norm": 2.510046075628132, "learning_rate": 7.871319285687715e-05, "loss": 0.9122, "step": 1908 }, { "epoch": 0.8649750792931581, "grad_norm": 1.900639693908416, "learning_rate": 7.871000718138905e-05, "loss": 0.9281, "step": 1909 }, { "epoch": 0.8654281830539193, "grad_norm": 2.378864706493721, "learning_rate": 7.870681763210243e-05, "loss": 0.9297, "step": 1910 }, { "epoch": 0.8658812868146806, "grad_norm": 2.2141634557391856, "learning_rate": 7.870362420933648e-05, "loss": 0.9461, "step": 1911 }, { "epoch": 0.8663343905754418, "grad_norm": 1.8774647080227045, "learning_rate": 7.870042691341074e-05, "loss": 0.9123, "step": 1912 }, { "epoch": 0.866787494336203, "grad_norm": 1.572706745729863, "learning_rate": 7.869722574464521e-05, "loss": 0.9072, "step": 1913 }, { "epoch": 0.8672405980969642, "grad_norm": 1.9102170233451343, "learning_rate": 7.86940207033602e-05, "loss": 0.9072, "step": 1914 }, { "epoch": 0.8676937018577254, "grad_norm": 1.2830599637388849, "learning_rate": 7.869081178987647e-05, "loss": 0.8988, "step": 1915 }, { "epoch": 0.8681468056184867, "grad_norm": 2.0268630373347025, "learning_rate": 7.868759900451514e-05, "loss": 0.9286, "step": 1916 }, { "epoch": 0.8685999093792478, "grad_norm": 1.455781393733942, "learning_rate": 7.868438234759772e-05, "loss": 0.9078, "step": 1917 }, { "epoch": 0.869053013140009, "grad_norm": 1.9236712783285896, "learning_rate": 7.86811618194461e-05, "loss": 0.9199, "step": 1918 }, { "epoch": 0.8695061169007703, "grad_norm": 1.5261250219712525, "learning_rate": 7.867793742038258e-05, "loss": 0.9143, "step": 1919 }, { "epoch": 0.8699592206615315, "grad_norm": 1.6031636647135177, "learning_rate": 7.867470915072982e-05, "loss": 0.9077, "step": 1920 }, { "epoch": 0.8704123244222927, "grad_norm": 1.9493407223545964, "learning_rate": 7.867147701081088e-05, "loss": 0.9149, "step": 1921 }, { "epoch": 0.8708654281830539, "grad_norm": 1.1370644621116706, "learning_rate": 7.866824100094923e-05, "loss": 0.9164, "step": 1922 }, { "epoch": 0.8713185319438151, "grad_norm": 2.9262548115770746, "learning_rate": 7.866500112146867e-05, "loss": 0.9273, "step": 1923 }, { "epoch": 0.8717716357045764, "grad_norm": 2.6838484060430607, "learning_rate": 7.866175737269345e-05, "loss": 0.9438, "step": 1924 }, { "epoch": 0.8722247394653375, "grad_norm": 1.4739422749993587, "learning_rate": 7.865850975494818e-05, "loss": 0.9066, "step": 1925 }, { "epoch": 0.8726778432260988, "grad_norm": 1.5152745576097777, "learning_rate": 7.865525826855782e-05, "loss": 0.9156, "step": 1926 }, { "epoch": 0.87313094698686, "grad_norm": 2.0099144437993663, "learning_rate": 7.86520029138478e-05, "loss": 0.9122, "step": 1927 }, { "epoch": 0.8735840507476212, "grad_norm": 1.1861970000626596, "learning_rate": 7.864874369114386e-05, "loss": 0.9227, "step": 1928 }, { "epoch": 0.8740371545083824, "grad_norm": 2.5749945831825363, "learning_rate": 7.864548060077218e-05, "loss": 0.9076, "step": 1929 }, { "epoch": 0.8744902582691436, "grad_norm": 2.117519101807773, "learning_rate": 7.86422136430593e-05, "loss": 0.9216, "step": 1930 }, { "epoch": 0.8749433620299049, "grad_norm": 2.0296451707241356, "learning_rate": 7.863894281833212e-05, "loss": 0.9054, "step": 1931 }, { "epoch": 0.8753964657906661, "grad_norm": 1.9117190351345312, "learning_rate": 7.863566812691799e-05, "loss": 0.9177, "step": 1932 }, { "epoch": 0.8758495695514272, "grad_norm": 1.9807455222011623, "learning_rate": 7.863238956914462e-05, "loss": 0.9114, "step": 1933 }, { "epoch": 0.8763026733121885, "grad_norm": 1.4803785060464387, "learning_rate": 7.862910714534008e-05, "loss": 0.9148, "step": 1934 }, { "epoch": 0.8767557770729497, "grad_norm": 2.062688028728617, "learning_rate": 7.862582085583286e-05, "loss": 0.9231, "step": 1935 }, { "epoch": 0.877208880833711, "grad_norm": 1.6516560288755513, "learning_rate": 7.862253070095185e-05, "loss": 0.9278, "step": 1936 }, { "epoch": 0.8776619845944721, "grad_norm": 2.1607913870994735, "learning_rate": 7.861923668102626e-05, "loss": 0.9072, "step": 1937 }, { "epoch": 0.8781150883552333, "grad_norm": 1.837376492050144, "learning_rate": 7.861593879638577e-05, "loss": 0.923, "step": 1938 }, { "epoch": 0.8785681921159946, "grad_norm": 1.9503151332234496, "learning_rate": 7.861263704736038e-05, "loss": 0.9128, "step": 1939 }, { "epoch": 0.8790212958767558, "grad_norm": 1.7404196182234672, "learning_rate": 7.860933143428051e-05, "loss": 0.9201, "step": 1940 }, { "epoch": 0.879474399637517, "grad_norm": 1.9533370707622764, "learning_rate": 7.860602195747696e-05, "loss": 0.8991, "step": 1941 }, { "epoch": 0.8799275033982782, "grad_norm": 1.7064531946341213, "learning_rate": 7.860270861728092e-05, "loss": 0.9034, "step": 1942 }, { "epoch": 0.8803806071590394, "grad_norm": 1.8665701219208526, "learning_rate": 7.859939141402397e-05, "loss": 0.92, "step": 1943 }, { "epoch": 0.8808337109198007, "grad_norm": 1.5780474053643918, "learning_rate": 7.859607034803805e-05, "loss": 0.9138, "step": 1944 }, { "epoch": 0.8812868146805618, "grad_norm": 2.0023886140499654, "learning_rate": 7.859274541965554e-05, "loss": 0.9021, "step": 1945 }, { "epoch": 0.881739918441323, "grad_norm": 1.6555413239127916, "learning_rate": 7.858941662920914e-05, "loss": 0.9052, "step": 1946 }, { "epoch": 0.8821930222020843, "grad_norm": 2.0666292489075166, "learning_rate": 7.858608397703198e-05, "loss": 0.9171, "step": 1947 }, { "epoch": 0.8826461259628455, "grad_norm": 1.6931110304896562, "learning_rate": 7.858274746345757e-05, "loss": 0.9079, "step": 1948 }, { "epoch": 0.8830992297236067, "grad_norm": 2.015689249735896, "learning_rate": 7.85794070888198e-05, "loss": 0.9171, "step": 1949 }, { "epoch": 0.8835523334843679, "grad_norm": 1.6944561056808516, "learning_rate": 7.857606285345295e-05, "loss": 0.9004, "step": 1950 }, { "epoch": 0.8840054372451291, "grad_norm": 1.9769855452632445, "learning_rate": 7.857271475769169e-05, "loss": 0.9083, "step": 1951 }, { "epoch": 0.8844585410058904, "grad_norm": 1.6185902783382586, "learning_rate": 7.856936280187108e-05, "loss": 0.9188, "step": 1952 }, { "epoch": 0.8849116447666515, "grad_norm": 2.0290515001437406, "learning_rate": 7.856600698632651e-05, "loss": 0.9223, "step": 1953 }, { "epoch": 0.8853647485274128, "grad_norm": 1.8268387777175341, "learning_rate": 7.856264731139387e-05, "loss": 0.8938, "step": 1954 }, { "epoch": 0.885817852288174, "grad_norm": 1.812460776504522, "learning_rate": 7.855928377740932e-05, "loss": 0.9152, "step": 1955 }, { "epoch": 0.8862709560489352, "grad_norm": 1.5031272446114559, "learning_rate": 7.855591638470948e-05, "loss": 0.9176, "step": 1956 }, { "epoch": 0.8867240598096964, "grad_norm": 2.0321678559790244, "learning_rate": 7.855254513363134e-05, "loss": 0.9043, "step": 1957 }, { "epoch": 0.8871771635704576, "grad_norm": 1.6887121879117963, "learning_rate": 7.854917002451225e-05, "loss": 0.9231, "step": 1958 }, { "epoch": 0.8876302673312189, "grad_norm": 1.913087216798141, "learning_rate": 7.854579105768995e-05, "loss": 0.8942, "step": 1959 }, { "epoch": 0.8880833710919801, "grad_norm": 1.6113699545447673, "learning_rate": 7.854240823350262e-05, "loss": 0.9138, "step": 1960 }, { "epoch": 0.8885364748527412, "grad_norm": 1.8659649253812722, "learning_rate": 7.853902155228878e-05, "loss": 0.8839, "step": 1961 }, { "epoch": 0.8889895786135025, "grad_norm": 1.558349577179699, "learning_rate": 7.853563101438733e-05, "loss": 0.9113, "step": 1962 }, { "epoch": 0.8894426823742637, "grad_norm": 1.9880707473298633, "learning_rate": 7.853223662013756e-05, "loss": 0.9313, "step": 1963 }, { "epoch": 0.889895786135025, "grad_norm": 1.5923121953750954, "learning_rate": 7.852883836987917e-05, "loss": 0.8997, "step": 1964 }, { "epoch": 0.8903488898957861, "grad_norm": 2.056303176849062, "learning_rate": 7.852543626395222e-05, "loss": 0.9412, "step": 1965 }, { "epoch": 0.8908019936565473, "grad_norm": 1.8626575356798187, "learning_rate": 7.852203030269718e-05, "loss": 0.9059, "step": 1966 }, { "epoch": 0.8912550974173086, "grad_norm": 1.687534302160946, "learning_rate": 7.851862048645488e-05, "loss": 0.9011, "step": 1967 }, { "epoch": 0.8917082011780698, "grad_norm": 1.533739974511095, "learning_rate": 7.851520681556656e-05, "loss": 0.9176, "step": 1968 }, { "epoch": 0.892161304938831, "grad_norm": 1.8494639923824667, "learning_rate": 7.851178929037383e-05, "loss": 0.9304, "step": 1969 }, { "epoch": 0.8926144086995922, "grad_norm": 1.5518060726468352, "learning_rate": 7.850836791121869e-05, "loss": 0.9038, "step": 1970 }, { "epoch": 0.8930675124603534, "grad_norm": 1.8677307154293223, "learning_rate": 7.85049426784435e-05, "loss": 0.9077, "step": 1971 }, { "epoch": 0.8935206162211147, "grad_norm": 1.5440106881457496, "learning_rate": 7.850151359239107e-05, "loss": 0.9172, "step": 1972 }, { "epoch": 0.8939737199818758, "grad_norm": 1.96482047522077, "learning_rate": 7.849808065340453e-05, "loss": 0.8936, "step": 1973 }, { "epoch": 0.894426823742637, "grad_norm": 1.8008482998299291, "learning_rate": 7.849464386182744e-05, "loss": 0.9163, "step": 1974 }, { "epoch": 0.8948799275033983, "grad_norm": 1.7269162289316213, "learning_rate": 7.849120321800372e-05, "loss": 0.8996, "step": 1975 }, { "epoch": 0.8953330312641595, "grad_norm": 1.5026987729369818, "learning_rate": 7.848775872227768e-05, "loss": 0.9128, "step": 1976 }, { "epoch": 0.8957861350249207, "grad_norm": 1.8159364664763775, "learning_rate": 7.848431037499402e-05, "loss": 0.9028, "step": 1977 }, { "epoch": 0.8962392387856819, "grad_norm": 1.5674380857881254, "learning_rate": 7.848085817649782e-05, "loss": 0.9424, "step": 1978 }, { "epoch": 0.8966923425464431, "grad_norm": 1.8816391559481571, "learning_rate": 7.847740212713456e-05, "loss": 0.9199, "step": 1979 }, { "epoch": 0.8971454463072044, "grad_norm": 1.5596539836381347, "learning_rate": 7.847394222725009e-05, "loss": 0.9068, "step": 1980 }, { "epoch": 0.8975985500679655, "grad_norm": 1.6910809144978936, "learning_rate": 7.847047847719063e-05, "loss": 0.9016, "step": 1981 }, { "epoch": 0.8980516538287268, "grad_norm": 1.5447586260685477, "learning_rate": 7.846701087730285e-05, "loss": 0.9187, "step": 1982 }, { "epoch": 0.898504757589488, "grad_norm": 1.7528021493307042, "learning_rate": 7.846353942793372e-05, "loss": 0.9233, "step": 1983 }, { "epoch": 0.8989578613502492, "grad_norm": 1.416304766602211, "learning_rate": 7.846006412943065e-05, "loss": 0.9318, "step": 1984 }, { "epoch": 0.8994109651110104, "grad_norm": 1.7194909184491016, "learning_rate": 7.845658498214143e-05, "loss": 0.8979, "step": 1985 }, { "epoch": 0.8998640688717716, "grad_norm": 1.5194546592784004, "learning_rate": 7.84531019864142e-05, "loss": 0.9217, "step": 1986 }, { "epoch": 0.9003171726325329, "grad_norm": 1.854586831251275, "learning_rate": 7.844961514259755e-05, "loss": 0.9029, "step": 1987 }, { "epoch": 0.9007702763932941, "grad_norm": 1.5689314097214724, "learning_rate": 7.844612445104039e-05, "loss": 0.8747, "step": 1988 }, { "epoch": 0.9012233801540552, "grad_norm": 1.6106767803013247, "learning_rate": 7.844262991209204e-05, "loss": 0.8887, "step": 1989 }, { "epoch": 0.9016764839148165, "grad_norm": 1.4219272452221878, "learning_rate": 7.843913152610222e-05, "loss": 0.9098, "step": 1990 }, { "epoch": 0.9021295876755777, "grad_norm": 1.7134936395254854, "learning_rate": 7.8435629293421e-05, "loss": 0.8919, "step": 1991 }, { "epoch": 0.902582691436339, "grad_norm": 1.4450262390297879, "learning_rate": 7.843212321439889e-05, "loss": 0.894, "step": 1992 }, { "epoch": 0.9030357951971001, "grad_norm": 1.9136827084838675, "learning_rate": 7.842861328938671e-05, "loss": 0.9118, "step": 1993 }, { "epoch": 0.9034888989578613, "grad_norm": 1.6517659217674314, "learning_rate": 7.842509951873576e-05, "loss": 0.9138, "step": 1994 }, { "epoch": 0.9039420027186226, "grad_norm": 1.4600776327060137, "learning_rate": 7.842158190279762e-05, "loss": 0.9274, "step": 1995 }, { "epoch": 0.9043951064793838, "grad_norm": 1.2082486177730156, "learning_rate": 7.841806044192433e-05, "loss": 0.921, "step": 1996 }, { "epoch": 0.904848210240145, "grad_norm": 1.8715690486677112, "learning_rate": 7.841453513646829e-05, "loss": 0.8973, "step": 1997 }, { "epoch": 0.9053013140009062, "grad_norm": 1.5984572483215604, "learning_rate": 7.841100598678227e-05, "loss": 0.8981, "step": 1998 }, { "epoch": 0.9057544177616674, "grad_norm": 1.5081336000460495, "learning_rate": 7.840747299321946e-05, "loss": 0.8995, "step": 1999 }, { "epoch": 0.9062075215224287, "grad_norm": 1.331031192301732, "learning_rate": 7.84039361561334e-05, "loss": 0.9114, "step": 2000 }, { "epoch": 0.9066606252831898, "grad_norm": 1.6475845429046898, "learning_rate": 7.840039547587803e-05, "loss": 0.8951, "step": 2001 }, { "epoch": 0.907113729043951, "grad_norm": 1.3372693642985132, "learning_rate": 7.839685095280769e-05, "loss": 0.9133, "step": 2002 }, { "epoch": 0.9075668328047123, "grad_norm": 1.634708955971805, "learning_rate": 7.839330258727707e-05, "loss": 0.9114, "step": 2003 }, { "epoch": 0.9080199365654735, "grad_norm": 1.3431482247918902, "learning_rate": 7.838975037964127e-05, "loss": 0.9026, "step": 2004 }, { "epoch": 0.9084730403262347, "grad_norm": 1.6736403485634392, "learning_rate": 7.838619433025578e-05, "loss": 0.8898, "step": 2005 }, { "epoch": 0.9089261440869959, "grad_norm": 1.4392352641486075, "learning_rate": 7.838263443947646e-05, "loss": 0.9195, "step": 2006 }, { "epoch": 0.9093792478477571, "grad_norm": 1.6212695766966356, "learning_rate": 7.837907070765951e-05, "loss": 0.9338, "step": 2007 }, { "epoch": 0.9098323516085184, "grad_norm": 1.4447343059307034, "learning_rate": 7.837550313516163e-05, "loss": 0.9269, "step": 2008 }, { "epoch": 0.9102854553692795, "grad_norm": 1.2239025005971687, "learning_rate": 7.83719317223398e-05, "loss": 0.9091, "step": 2009 }, { "epoch": 0.9107385591300408, "grad_norm": 1.0376136096633277, "learning_rate": 7.83683564695514e-05, "loss": 0.93, "step": 2010 }, { "epoch": 0.911191662890802, "grad_norm": 1.3168860293505966, "learning_rate": 7.836477737715426e-05, "loss": 0.8988, "step": 2011 }, { "epoch": 0.9116447666515632, "grad_norm": 0.7782370068003444, "learning_rate": 7.836119444550651e-05, "loss": 0.9089, "step": 2012 }, { "epoch": 0.9120978704123244, "grad_norm": 1.4901653662607166, "learning_rate": 7.835760767496672e-05, "loss": 0.9064, "step": 2013 }, { "epoch": 0.9125509741730856, "grad_norm": 0.8914793614965695, "learning_rate": 7.835401706589381e-05, "loss": 0.8907, "step": 2014 }, { "epoch": 0.9130040779338469, "grad_norm": 1.6233110654057212, "learning_rate": 7.835042261864713e-05, "loss": 0.9229, "step": 2015 }, { "epoch": 0.9134571816946081, "grad_norm": 1.3415816990081664, "learning_rate": 7.834682433358636e-05, "loss": 0.9059, "step": 2016 }, { "epoch": 0.9139102854553692, "grad_norm": 1.2134782579112127, "learning_rate": 7.83432222110716e-05, "loss": 0.9075, "step": 2017 }, { "epoch": 0.9143633892161305, "grad_norm": 1.3341562054277525, "learning_rate": 7.833961625146331e-05, "loss": 0.9165, "step": 2018 }, { "epoch": 0.9148164929768917, "grad_norm": 1.1186105204708068, "learning_rate": 7.833600645512236e-05, "loss": 0.9213, "step": 2019 }, { "epoch": 0.915269596737653, "grad_norm": 0.8809031660744174, "learning_rate": 7.833239282240998e-05, "loss": 0.8932, "step": 2020 }, { "epoch": 0.9157227004984141, "grad_norm": 0.9769853742557405, "learning_rate": 7.832877535368782e-05, "loss": 0.9223, "step": 2021 }, { "epoch": 0.9161758042591753, "grad_norm": 0.7790840497656042, "learning_rate": 7.832515404931784e-05, "loss": 0.9027, "step": 2022 }, { "epoch": 0.9166289080199366, "grad_norm": 0.8364199090114728, "learning_rate": 7.832152890966246e-05, "loss": 0.8894, "step": 2023 }, { "epoch": 0.9170820117806978, "grad_norm": 1.0515855518227046, "learning_rate": 7.831789993508448e-05, "loss": 0.9293, "step": 2024 }, { "epoch": 0.917535115541459, "grad_norm": 1.1408552929634543, "learning_rate": 7.831426712594702e-05, "loss": 0.9152, "step": 2025 }, { "epoch": 0.9179882193022202, "grad_norm": 0.9490436230277801, "learning_rate": 7.831063048261364e-05, "loss": 0.8941, "step": 2026 }, { "epoch": 0.9184413230629814, "grad_norm": 0.8117909270418809, "learning_rate": 7.830699000544826e-05, "loss": 0.924, "step": 2027 }, { "epoch": 0.9188944268237427, "grad_norm": 0.7383286741455938, "learning_rate": 7.830334569481521e-05, "loss": 0.8831, "step": 2028 }, { "epoch": 0.9193475305845038, "grad_norm": 1.2653992434175025, "learning_rate": 7.829969755107917e-05, "loss": 0.9166, "step": 2029 }, { "epoch": 0.919800634345265, "grad_norm": 1.545758487981436, "learning_rate": 7.829604557460523e-05, "loss": 0.901, "step": 2030 }, { "epoch": 0.9202537381060263, "grad_norm": 0.7980389657497203, "learning_rate": 7.829238976575883e-05, "loss": 0.9203, "step": 2031 }, { "epoch": 0.9207068418667875, "grad_norm": 1.3030182215589208, "learning_rate": 7.828873012490581e-05, "loss": 0.9171, "step": 2032 }, { "epoch": 0.9211599456275487, "grad_norm": 0.920758989409301, "learning_rate": 7.828506665241244e-05, "loss": 0.9201, "step": 2033 }, { "epoch": 0.9216130493883099, "grad_norm": 1.3888920989141054, "learning_rate": 7.82813993486453e-05, "loss": 0.9429, "step": 2034 }, { "epoch": 0.9220661531490711, "grad_norm": 1.2195441520385994, "learning_rate": 7.827772821397139e-05, "loss": 0.9299, "step": 2035 }, { "epoch": 0.9225192569098324, "grad_norm": 1.0800710530579416, "learning_rate": 7.82740532487581e-05, "loss": 0.9237, "step": 2036 }, { "epoch": 0.9229723606705935, "grad_norm": 1.2958317285447924, "learning_rate": 7.827037445337317e-05, "loss": 0.9207, "step": 2037 }, { "epoch": 0.9234254644313548, "grad_norm": 1.0678368573273014, "learning_rate": 7.826669182818476e-05, "loss": 0.9106, "step": 2038 }, { "epoch": 0.923878568192116, "grad_norm": 1.2408977893665956, "learning_rate": 7.826300537356141e-05, "loss": 0.9032, "step": 2039 }, { "epoch": 0.9243316719528772, "grad_norm": 0.8396046040753828, "learning_rate": 7.825931508987202e-05, "loss": 0.9127, "step": 2040 }, { "epoch": 0.9247847757136384, "grad_norm": 1.0531029236077365, "learning_rate": 7.825562097748588e-05, "loss": 0.9018, "step": 2041 }, { "epoch": 0.9252378794743996, "grad_norm": 1.6107049028295495, "learning_rate": 7.825192303677266e-05, "loss": 0.9017, "step": 2042 }, { "epoch": 0.9256909832351609, "grad_norm": 0.7401840100828901, "learning_rate": 7.824822126810245e-05, "loss": 0.893, "step": 2043 }, { "epoch": 0.9261440869959221, "grad_norm": 0.8032920631631217, "learning_rate": 7.824451567184567e-05, "loss": 0.9049, "step": 2044 }, { "epoch": 0.9265971907566832, "grad_norm": 1.248651635379151, "learning_rate": 7.824080624837317e-05, "loss": 0.9094, "step": 2045 }, { "epoch": 0.9270502945174445, "grad_norm": 1.5883490139219565, "learning_rate": 7.823709299805613e-05, "loss": 0.8895, "step": 2046 }, { "epoch": 0.9275033982782057, "grad_norm": 0.8143250550637959, "learning_rate": 7.823337592126617e-05, "loss": 0.9117, "step": 2047 }, { "epoch": 0.927956502038967, "grad_norm": 1.1751882248095062, "learning_rate": 7.822965501837525e-05, "loss": 0.8877, "step": 2048 }, { "epoch": 0.9284096057997281, "grad_norm": 1.6052256827259623, "learning_rate": 7.822593028975574e-05, "loss": 0.8964, "step": 2049 }, { "epoch": 0.9288627095604893, "grad_norm": 1.0541581355821277, "learning_rate": 7.822220173578038e-05, "loss": 0.898, "step": 2050 }, { "epoch": 0.9293158133212506, "grad_norm": 1.3143472155597933, "learning_rate": 7.821846935682228e-05, "loss": 0.9224, "step": 2051 }, { "epoch": 0.9297689170820118, "grad_norm": 0.9318583036659827, "learning_rate": 7.821473315325498e-05, "loss": 0.9119, "step": 2052 }, { "epoch": 0.930222020842773, "grad_norm": 1.3490530116563177, "learning_rate": 7.821099312545233e-05, "loss": 0.9104, "step": 2053 }, { "epoch": 0.9306751246035342, "grad_norm": 0.7831261195459, "learning_rate": 7.820724927378864e-05, "loss": 0.9207, "step": 2054 }, { "epoch": 0.9311282283642954, "grad_norm": 1.2827543493748232, "learning_rate": 7.820350159863854e-05, "loss": 0.9117, "step": 2055 }, { "epoch": 0.9315813321250567, "grad_norm": 0.8996015612187417, "learning_rate": 7.819975010037708e-05, "loss": 0.91, "step": 2056 }, { "epoch": 0.9320344358858178, "grad_norm": 1.3341555366304967, "learning_rate": 7.819599477937967e-05, "loss": 0.9313, "step": 2057 }, { "epoch": 0.932487539646579, "grad_norm": 0.821916064429505, "learning_rate": 7.819223563602214e-05, "loss": 0.9068, "step": 2058 }, { "epoch": 0.9329406434073403, "grad_norm": 1.0924401288724457, "learning_rate": 7.818847267068065e-05, "loss": 0.9186, "step": 2059 }, { "epoch": 0.9333937471681015, "grad_norm": 1.2429335754667665, "learning_rate": 7.818470588373178e-05, "loss": 0.8775, "step": 2060 }, { "epoch": 0.9338468509288627, "grad_norm": 1.429252354363751, "learning_rate": 7.818093527555248e-05, "loss": 0.9365, "step": 2061 }, { "epoch": 0.9342999546896239, "grad_norm": 0.7762927197246124, "learning_rate": 7.817716084652007e-05, "loss": 0.907, "step": 2062 }, { "epoch": 0.9347530584503851, "grad_norm": 1.0921040798279096, "learning_rate": 7.817338259701227e-05, "loss": 0.913, "step": 2063 }, { "epoch": 0.9352061622111464, "grad_norm": 1.1817275349986858, "learning_rate": 7.816960052740721e-05, "loss": 0.9303, "step": 2064 }, { "epoch": 0.9356592659719075, "grad_norm": 1.7204940386583663, "learning_rate": 7.816581463808334e-05, "loss": 0.9258, "step": 2065 }, { "epoch": 0.9361123697326688, "grad_norm": 0.624264577831123, "learning_rate": 7.816202492941952e-05, "loss": 0.8985, "step": 2066 }, { "epoch": 0.93656547349343, "grad_norm": 1.5832754770758268, "learning_rate": 7.8158231401795e-05, "loss": 0.8879, "step": 2067 }, { "epoch": 0.9370185772541912, "grad_norm": 1.2912434617380348, "learning_rate": 7.815443405558942e-05, "loss": 0.9128, "step": 2068 }, { "epoch": 0.9374716810149524, "grad_norm": 0.9481750521231108, "learning_rate": 7.815063289118276e-05, "loss": 0.9184, "step": 2069 }, { "epoch": 0.9379247847757136, "grad_norm": 1.551122111169552, "learning_rate": 7.814682790895544e-05, "loss": 0.918, "step": 2070 }, { "epoch": 0.9383778885364749, "grad_norm": 0.7821948627700802, "learning_rate": 7.814301910928824e-05, "loss": 0.9315, "step": 2071 }, { "epoch": 0.9388309922972361, "grad_norm": 1.9057495067168437, "learning_rate": 7.813920649256227e-05, "loss": 0.913, "step": 2072 }, { "epoch": 0.9392840960579972, "grad_norm": 1.14254590627036, "learning_rate": 7.813539005915911e-05, "loss": 0.9127, "step": 2073 }, { "epoch": 0.9397371998187585, "grad_norm": 2.2506503571541234, "learning_rate": 7.813156980946066e-05, "loss": 0.907, "step": 2074 }, { "epoch": 0.9401903035795197, "grad_norm": 1.976454524526806, "learning_rate": 7.812774574384922e-05, "loss": 0.9231, "step": 2075 }, { "epoch": 0.940643407340281, "grad_norm": 1.6003364108123825, "learning_rate": 7.812391786270748e-05, "loss": 0.8898, "step": 2076 }, { "epoch": 0.9410965111010421, "grad_norm": 1.7249786467003538, "learning_rate": 7.812008616641851e-05, "loss": 0.9018, "step": 2077 }, { "epoch": 0.9415496148618033, "grad_norm": 1.2027422761017446, "learning_rate": 7.811625065536576e-05, "loss": 0.9022, "step": 2078 }, { "epoch": 0.9420027186225646, "grad_norm": 1.750195294486351, "learning_rate": 7.811241132993305e-05, "loss": 0.9033, "step": 2079 }, { "epoch": 0.9424558223833258, "grad_norm": 1.124832860327793, "learning_rate": 7.810856819050457e-05, "loss": 0.9006, "step": 2080 }, { "epoch": 0.942908926144087, "grad_norm": 1.9062540803946175, "learning_rate": 7.810472123746493e-05, "loss": 0.9093, "step": 2081 }, { "epoch": 0.9433620299048482, "grad_norm": 1.479187371297973, "learning_rate": 7.810087047119911e-05, "loss": 0.9312, "step": 2082 }, { "epoch": 0.9438151336656094, "grad_norm": 1.875384803735614, "learning_rate": 7.809701589209246e-05, "loss": 0.9199, "step": 2083 }, { "epoch": 0.9442682374263707, "grad_norm": 1.5973390969933101, "learning_rate": 7.809315750053071e-05, "loss": 0.9216, "step": 2084 }, { "epoch": 0.9447213411871318, "grad_norm": 1.6626672306559171, "learning_rate": 7.808929529689999e-05, "loss": 0.8848, "step": 2085 }, { "epoch": 0.945174444947893, "grad_norm": 1.4070267343046658, "learning_rate": 7.808542928158678e-05, "loss": 0.9116, "step": 2086 }, { "epoch": 0.9456275487086543, "grad_norm": 1.7188928856245425, "learning_rate": 7.808155945497799e-05, "loss": 0.8837, "step": 2087 }, { "epoch": 0.9460806524694155, "grad_norm": 1.3165575574219068, "learning_rate": 7.807768581746086e-05, "loss": 0.9023, "step": 2088 }, { "epoch": 0.9465337562301767, "grad_norm": 1.757567036345046, "learning_rate": 7.807380836942303e-05, "loss": 0.9073, "step": 2089 }, { "epoch": 0.9469868599909379, "grad_norm": 1.47981443737845, "learning_rate": 7.806992711125254e-05, "loss": 0.9034, "step": 2090 }, { "epoch": 0.9474399637516991, "grad_norm": 1.7351307750934721, "learning_rate": 7.80660420433378e-05, "loss": 0.9115, "step": 2091 }, { "epoch": 0.9478930675124604, "grad_norm": 1.2963528020386734, "learning_rate": 7.806215316606758e-05, "loss": 0.9051, "step": 2092 }, { "epoch": 0.9483461712732215, "grad_norm": 1.7438900944818503, "learning_rate": 7.805826047983105e-05, "loss": 0.9025, "step": 2093 }, { "epoch": 0.9487992750339828, "grad_norm": 1.39840618877722, "learning_rate": 7.805436398501778e-05, "loss": 0.9181, "step": 2094 }, { "epoch": 0.949252378794744, "grad_norm": 1.896548002281233, "learning_rate": 7.80504636820177e-05, "loss": 0.9167, "step": 2095 }, { "epoch": 0.9497054825555052, "grad_norm": 1.6096309804969873, "learning_rate": 7.804655957122108e-05, "loss": 0.9245, "step": 2096 }, { "epoch": 0.9501585863162664, "grad_norm": 1.55563453824307, "learning_rate": 7.804265165301866e-05, "loss": 0.9028, "step": 2097 }, { "epoch": 0.9506116900770276, "grad_norm": 1.379700620739658, "learning_rate": 7.80387399278015e-05, "loss": 0.9038, "step": 2098 }, { "epoch": 0.9510647938377889, "grad_norm": 1.727417728369417, "learning_rate": 7.803482439596105e-05, "loss": 0.9042, "step": 2099 }, { "epoch": 0.9515178975985501, "grad_norm": 1.4062723273400082, "learning_rate": 7.803090505788914e-05, "loss": 0.9159, "step": 2100 }, { "epoch": 0.9519710013593112, "grad_norm": 1.756993647129131, "learning_rate": 7.802698191397801e-05, "loss": 0.9138, "step": 2101 }, { "epoch": 0.9524241051200725, "grad_norm": 1.6150219172611289, "learning_rate": 7.802305496462024e-05, "loss": 0.9065, "step": 2102 }, { "epoch": 0.9528772088808337, "grad_norm": 1.4905864417980996, "learning_rate": 7.801912421020881e-05, "loss": 0.9131, "step": 2103 }, { "epoch": 0.953330312641595, "grad_norm": 1.3589677825026223, "learning_rate": 7.801518965113708e-05, "loss": 0.9007, "step": 2104 }, { "epoch": 0.9537834164023561, "grad_norm": 1.4226126449331518, "learning_rate": 7.80112512877988e-05, "loss": 0.8908, "step": 2105 }, { "epoch": 0.9542365201631173, "grad_norm": 1.0783860895661364, "learning_rate": 7.800730912058808e-05, "loss": 0.9144, "step": 2106 }, { "epoch": 0.9546896239238786, "grad_norm": 1.7287671283514656, "learning_rate": 7.800336314989944e-05, "loss": 0.9064, "step": 2107 }, { "epoch": 0.9551427276846398, "grad_norm": 1.2313831189584676, "learning_rate": 7.799941337612773e-05, "loss": 0.8905, "step": 2108 }, { "epoch": 0.955595831445401, "grad_norm": 1.8106857191646692, "learning_rate": 7.799545979966823e-05, "loss": 0.8884, "step": 2109 }, { "epoch": 0.9560489352061622, "grad_norm": 1.608561771572011, "learning_rate": 7.799150242091659e-05, "loss": 0.9022, "step": 2110 }, { "epoch": 0.9565020389669234, "grad_norm": 1.3186811833687933, "learning_rate": 7.798754124026883e-05, "loss": 0.9047, "step": 2111 }, { "epoch": 0.9569551427276847, "grad_norm": 1.1375349599767368, "learning_rate": 7.798357625812136e-05, "loss": 0.9166, "step": 2112 }, { "epoch": 0.9574082464884458, "grad_norm": 1.4767829732369921, "learning_rate": 7.797960747487095e-05, "loss": 0.9054, "step": 2113 }, { "epoch": 0.957861350249207, "grad_norm": 1.1372256776343592, "learning_rate": 7.797563489091477e-05, "loss": 0.9231, "step": 2114 }, { "epoch": 0.9583144540099683, "grad_norm": 1.6655386890253963, "learning_rate": 7.797165850665036e-05, "loss": 0.8982, "step": 2115 }, { "epoch": 0.9587675577707295, "grad_norm": 1.3726952216473824, "learning_rate": 7.796767832247567e-05, "loss": 0.9119, "step": 2116 }, { "epoch": 0.9592206615314907, "grad_norm": 1.4127261661343704, "learning_rate": 7.796369433878898e-05, "loss": 0.9232, "step": 2117 }, { "epoch": 0.9596737652922519, "grad_norm": 1.2314415002827828, "learning_rate": 7.795970655598899e-05, "loss": 0.8864, "step": 2118 }, { "epoch": 0.9601268690530131, "grad_norm": 1.4231658654714534, "learning_rate": 7.795571497447476e-05, "loss": 0.8969, "step": 2119 }, { "epoch": 0.9605799728137744, "grad_norm": 1.1447623905521205, "learning_rate": 7.795171959464574e-05, "loss": 0.9105, "step": 2120 }, { "epoch": 0.9610330765745355, "grad_norm": 1.3866281526669988, "learning_rate": 7.794772041690176e-05, "loss": 0.9049, "step": 2121 }, { "epoch": 0.9614861803352968, "grad_norm": 1.1995746937421004, "learning_rate": 7.794371744164302e-05, "loss": 0.9197, "step": 2122 }, { "epoch": 0.961939284096058, "grad_norm": 1.344798001481972, "learning_rate": 7.793971066927011e-05, "loss": 0.9107, "step": 2123 }, { "epoch": 0.9623923878568192, "grad_norm": 1.0546662906432525, "learning_rate": 7.7935700100184e-05, "loss": 0.9098, "step": 2124 }, { "epoch": 0.9628454916175804, "grad_norm": 1.271621196882647, "learning_rate": 7.793168573478602e-05, "loss": 0.9072, "step": 2125 }, { "epoch": 0.9632985953783416, "grad_norm": 1.0333685539680229, "learning_rate": 7.792766757347793e-05, "loss": 0.9006, "step": 2126 }, { "epoch": 0.9637516991391029, "grad_norm": 1.2852224478015, "learning_rate": 7.79236456166618e-05, "loss": 0.918, "step": 2127 }, { "epoch": 0.9642048028998641, "grad_norm": 1.116415522963132, "learning_rate": 7.791961986474014e-05, "loss": 0.9374, "step": 2128 }, { "epoch": 0.9646579066606252, "grad_norm": 1.1307185023204702, "learning_rate": 7.791559031811582e-05, "loss": 0.8999, "step": 2129 }, { "epoch": 0.9651110104213865, "grad_norm": 1.0493765793855512, "learning_rate": 7.791155697719205e-05, "loss": 0.9064, "step": 2130 }, { "epoch": 0.9655641141821477, "grad_norm": 1.1828479473080167, "learning_rate": 7.790751984237249e-05, "loss": 0.8828, "step": 2131 }, { "epoch": 0.966017217942909, "grad_norm": 1.2018953915550086, "learning_rate": 7.790347891406115e-05, "loss": 0.9162, "step": 2132 }, { "epoch": 0.9664703217036701, "grad_norm": 1.0117686734449252, "learning_rate": 7.789943419266238e-05, "loss": 0.891, "step": 2133 }, { "epoch": 0.9669234254644313, "grad_norm": 1.0678447431419096, "learning_rate": 7.789538567858098e-05, "loss": 0.8949, "step": 2134 }, { "epoch": 0.9673765292251926, "grad_norm": 1.1893589964996882, "learning_rate": 7.789133337222208e-05, "loss": 0.8961, "step": 2135 }, { "epoch": 0.9678296329859538, "grad_norm": 1.3562759812906493, "learning_rate": 7.78872772739912e-05, "loss": 0.9119, "step": 2136 }, { "epoch": 0.968282736746715, "grad_norm": 0.9683286120577437, "learning_rate": 7.788321738429423e-05, "loss": 0.8922, "step": 2137 }, { "epoch": 0.9687358405074762, "grad_norm": 1.22953105146788, "learning_rate": 7.787915370353748e-05, "loss": 0.8868, "step": 2138 }, { "epoch": 0.9691889442682374, "grad_norm": 0.7767889726326562, "learning_rate": 7.787508623212761e-05, "loss": 0.9262, "step": 2139 }, { "epoch": 0.9696420480289987, "grad_norm": 0.9710827929757868, "learning_rate": 7.787101497047164e-05, "loss": 0.9125, "step": 2140 }, { "epoch": 0.9700951517897598, "grad_norm": 1.108029041668736, "learning_rate": 7.7866939918977e-05, "loss": 0.9136, "step": 2141 }, { "epoch": 0.970548255550521, "grad_norm": 1.1790446849083798, "learning_rate": 7.78628610780515e-05, "loss": 0.9083, "step": 2142 }, { "epoch": 0.9710013593112823, "grad_norm": 1.2172463403455263, "learning_rate": 7.78587784481033e-05, "loss": 0.9046, "step": 2143 }, { "epoch": 0.9714544630720435, "grad_norm": 1.611507798851159, "learning_rate": 7.785469202954098e-05, "loss": 0.926, "step": 2144 }, { "epoch": 0.9719075668328047, "grad_norm": 0.8573106783011641, "learning_rate": 7.785060182277346e-05, "loss": 0.9157, "step": 2145 }, { "epoch": 0.9723606705935659, "grad_norm": 0.8227107595058596, "learning_rate": 7.784650782821007e-05, "loss": 0.9161, "step": 2146 }, { "epoch": 0.9728137743543271, "grad_norm": 1.3723913574399584, "learning_rate": 7.784241004626047e-05, "loss": 0.9095, "step": 2147 }, { "epoch": 0.9732668781150884, "grad_norm": 1.3611281140849727, "learning_rate": 7.783830847733479e-05, "loss": 0.8789, "step": 2148 }, { "epoch": 0.9737199818758495, "grad_norm": 1.1536507052298528, "learning_rate": 7.783420312184343e-05, "loss": 0.9325, "step": 2149 }, { "epoch": 0.9741730856366108, "grad_norm": 0.9566807504097435, "learning_rate": 7.783009398019726e-05, "loss": 0.9154, "step": 2150 }, { "epoch": 0.974626189397372, "grad_norm": 1.180126196844664, "learning_rate": 7.782598105280748e-05, "loss": 0.8983, "step": 2151 }, { "epoch": 0.9750792931581332, "grad_norm": 1.5452937521633734, "learning_rate": 7.782186434008568e-05, "loss": 0.8761, "step": 2152 }, { "epoch": 0.9755323969188944, "grad_norm": 0.8496700973208414, "learning_rate": 7.781774384244382e-05, "loss": 0.9032, "step": 2153 }, { "epoch": 0.9759855006796556, "grad_norm": 0.9839392813008203, "learning_rate": 7.781361956029426e-05, "loss": 0.9093, "step": 2154 }, { "epoch": 0.9764386044404169, "grad_norm": 1.3595163082182207, "learning_rate": 7.780949149404972e-05, "loss": 0.9253, "step": 2155 }, { "epoch": 0.9768917082011781, "grad_norm": 1.0909368795351218, "learning_rate": 7.780535964412331e-05, "loss": 0.8994, "step": 2156 }, { "epoch": 0.9773448119619392, "grad_norm": 1.7165384614917305, "learning_rate": 7.780122401092851e-05, "loss": 0.9149, "step": 2157 }, { "epoch": 0.9777979157227005, "grad_norm": 0.6927055096794386, "learning_rate": 7.779708459487916e-05, "loss": 0.9098, "step": 2158 }, { "epoch": 0.9782510194834617, "grad_norm": 1.3883422432751094, "learning_rate": 7.779294139638954e-05, "loss": 0.9031, "step": 2159 }, { "epoch": 0.978704123244223, "grad_norm": 1.3062664592626767, "learning_rate": 7.778879441587426e-05, "loss": 0.9217, "step": 2160 }, { "epoch": 0.9791572270049841, "grad_norm": 1.330169739934037, "learning_rate": 7.778464365374829e-05, "loss": 0.9096, "step": 2161 }, { "epoch": 0.9796103307657453, "grad_norm": 0.774445157958269, "learning_rate": 7.778048911042704e-05, "loss": 0.901, "step": 2162 }, { "epoch": 0.9800634345265066, "grad_norm": 0.8609039984715626, "learning_rate": 7.777633078632625e-05, "loss": 0.9245, "step": 2163 }, { "epoch": 0.9805165382872678, "grad_norm": 1.0619565524513406, "learning_rate": 7.777216868186204e-05, "loss": 0.9147, "step": 2164 }, { "epoch": 0.980969642048029, "grad_norm": 0.8434421046247234, "learning_rate": 7.776800279745096e-05, "loss": 0.9012, "step": 2165 }, { "epoch": 0.9814227458087902, "grad_norm": 0.850943414728808, "learning_rate": 7.776383313350985e-05, "loss": 0.9033, "step": 2166 }, { "epoch": 0.9818758495695514, "grad_norm": 0.7351380801690852, "learning_rate": 7.775965969045601e-05, "loss": 0.8895, "step": 2167 }, { "epoch": 0.9823289533303127, "grad_norm": 0.7163991806558831, "learning_rate": 7.775548246870708e-05, "loss": 0.8889, "step": 2168 }, { "epoch": 0.9827820570910738, "grad_norm": 0.8072357102823753, "learning_rate": 7.775130146868108e-05, "loss": 0.9168, "step": 2169 }, { "epoch": 0.983235160851835, "grad_norm": 1.1539287426122264, "learning_rate": 7.774711669079642e-05, "loss": 0.9072, "step": 2170 }, { "epoch": 0.9836882646125963, "grad_norm": 1.7443900136319843, "learning_rate": 7.774292813547185e-05, "loss": 0.9128, "step": 2171 }, { "epoch": 0.9841413683733575, "grad_norm": 0.5390270216546278, "learning_rate": 7.773873580312658e-05, "loss": 0.9034, "step": 2172 }, { "epoch": 0.9845944721341187, "grad_norm": 1.26136286421233, "learning_rate": 7.773453969418012e-05, "loss": 0.9197, "step": 2173 }, { "epoch": 0.9850475758948799, "grad_norm": 2.0221081546995463, "learning_rate": 7.773033980905237e-05, "loss": 0.8978, "step": 2174 }, { "epoch": 0.9855006796556411, "grad_norm": 0.7842222139796802, "learning_rate": 7.772613614816363e-05, "loss": 0.8957, "step": 2175 }, { "epoch": 0.9859537834164024, "grad_norm": 2.86589192753888, "learning_rate": 7.772192871193458e-05, "loss": 0.9251, "step": 2176 }, { "epoch": 0.9864068871771635, "grad_norm": 2.0679158829216404, "learning_rate": 7.771771750078627e-05, "loss": 0.9176, "step": 2177 }, { "epoch": 0.9868599909379248, "grad_norm": 2.563513913580875, "learning_rate": 7.771350251514011e-05, "loss": 0.9254, "step": 2178 }, { "epoch": 0.987313094698686, "grad_norm": 2.2917997071964615, "learning_rate": 7.770928375541792e-05, "loss": 0.9068, "step": 2179 }, { "epoch": 0.9877661984594472, "grad_norm": 1.7227670639897898, "learning_rate": 7.770506122204186e-05, "loss": 0.9115, "step": 2180 }, { "epoch": 0.9882193022202084, "grad_norm": 1.7596544111562975, "learning_rate": 7.770083491543451e-05, "loss": 0.9149, "step": 2181 }, { "epoch": 0.9886724059809696, "grad_norm": 1.881308883154012, "learning_rate": 7.76966048360188e-05, "loss": 0.8971, "step": 2182 }, { "epoch": 0.9891255097417309, "grad_norm": 1.317867426415812, "learning_rate": 7.769237098421804e-05, "loss": 0.8997, "step": 2183 }, { "epoch": 0.9895786135024921, "grad_norm": 1.5973222196325025, "learning_rate": 7.768813336045592e-05, "loss": 0.9173, "step": 2184 }, { "epoch": 0.9900317172632532, "grad_norm": 1.3599437196607445, "learning_rate": 7.76838919651565e-05, "loss": 0.9167, "step": 2185 }, { "epoch": 0.9904848210240145, "grad_norm": 1.3569612808622193, "learning_rate": 7.767964679874426e-05, "loss": 0.9173, "step": 2186 }, { "epoch": 0.9909379247847757, "grad_norm": 1.4342416046828004, "learning_rate": 7.767539786164397e-05, "loss": 0.8965, "step": 2187 }, { "epoch": 0.991391028545537, "grad_norm": 1.4283144562414665, "learning_rate": 7.767114515428088e-05, "loss": 0.903, "step": 2188 }, { "epoch": 0.9918441323062981, "grad_norm": 1.0709628014535824, "learning_rate": 7.766688867708054e-05, "loss": 0.9158, "step": 2189 }, { "epoch": 0.9922972360670593, "grad_norm": 1.7808012705600376, "learning_rate": 7.766262843046892e-05, "loss": 0.9222, "step": 2190 }, { "epoch": 0.9927503398278206, "grad_norm": 1.3009966565113877, "learning_rate": 7.765836441487235e-05, "loss": 0.9203, "step": 2191 }, { "epoch": 0.9932034435885818, "grad_norm": 1.626153327669958, "learning_rate": 7.765409663071753e-05, "loss": 0.9055, "step": 2192 }, { "epoch": 0.993656547349343, "grad_norm": 1.529658192778311, "learning_rate": 7.764982507843155e-05, "loss": 0.9324, "step": 2193 }, { "epoch": 0.9941096511101042, "grad_norm": 1.5724766662215568, "learning_rate": 7.764554975844189e-05, "loss": 0.9008, "step": 2194 }, { "epoch": 0.9945627548708654, "grad_norm": 1.294210634498605, "learning_rate": 7.764127067117637e-05, "loss": 0.9272, "step": 2195 }, { "epoch": 0.9950158586316267, "grad_norm": 1.807422844848207, "learning_rate": 7.763698781706321e-05, "loss": 0.9116, "step": 2196 }, { "epoch": 0.9954689623923878, "grad_norm": 1.480900035627717, "learning_rate": 7.763270119653102e-05, "loss": 0.8954, "step": 2197 }, { "epoch": 0.995922066153149, "grad_norm": 1.577168515951401, "learning_rate": 7.762841081000878e-05, "loss": 0.8998, "step": 2198 }, { "epoch": 0.9963751699139103, "grad_norm": 1.4812364045447108, "learning_rate": 7.762411665792579e-05, "loss": 0.9095, "step": 2199 }, { "epoch": 0.9968282736746715, "grad_norm": 1.3108340740467772, "learning_rate": 7.761981874071183e-05, "loss": 0.9172, "step": 2200 }, { "epoch": 0.9972813774354327, "grad_norm": 0.9121391660091636, "learning_rate": 7.761551705879698e-05, "loss": 0.9165, "step": 2201 }, { "epoch": 0.9977344811961939, "grad_norm": 1.6745401285041308, "learning_rate": 7.76112116126117e-05, "loss": 0.9245, "step": 2202 }, { "epoch": 0.9981875849569551, "grad_norm": 1.096826032463692, "learning_rate": 7.760690240258689e-05, "loss": 0.9174, "step": 2203 }, { "epoch": 0.9986406887177164, "grad_norm": 1.9268489557118276, "learning_rate": 7.760258942915374e-05, "loss": 0.9096, "step": 2204 }, { "epoch": 0.9990937924784775, "grad_norm": 1.8061211961416608, "learning_rate": 7.759827269274389e-05, "loss": 0.9153, "step": 2205 }, { "epoch": 0.9995468962392388, "grad_norm": 1.3332294443149921, "learning_rate": 7.759395219378928e-05, "loss": 0.9235, "step": 2206 }, { "epoch": 1.0, "grad_norm": 1.2827473227865909, "learning_rate": 7.758962793272233e-05, "loss": 0.9125, "step": 2207 }, { "epoch": 1.0004531037607611, "grad_norm": 1.247611051060619, "learning_rate": 7.758529990997576e-05, "loss": 0.905, "step": 2208 }, { "epoch": 1.0009062075215225, "grad_norm": 0.9439152877865336, "learning_rate": 7.758096812598266e-05, "loss": 0.8938, "step": 2209 }, { "epoch": 1.0013593112822836, "grad_norm": 1.553387192619467, "learning_rate": 7.757663258117655e-05, "loss": 0.8912, "step": 2210 }, { "epoch": 1.001812415043045, "grad_norm": 1.2183620865121854, "learning_rate": 7.757229327599128e-05, "loss": 0.8925, "step": 2211 }, { "epoch": 1.002265518803806, "grad_norm": 1.306623172866436, "learning_rate": 7.756795021086111e-05, "loss": 0.8719, "step": 2212 }, { "epoch": 1.0027186225645672, "grad_norm": 1.004759157297963, "learning_rate": 7.756360338622064e-05, "loss": 0.8747, "step": 2213 }, { "epoch": 1.0031717263253286, "grad_norm": 1.4294965259687145, "learning_rate": 7.75592528025049e-05, "loss": 0.8865, "step": 2214 }, { "epoch": 1.0036248300860897, "grad_norm": 1.0873418514188065, "learning_rate": 7.755489846014921e-05, "loss": 0.8839, "step": 2215 }, { "epoch": 1.0040779338468508, "grad_norm": 1.4808827376606324, "learning_rate": 7.755054035958938e-05, "loss": 0.9095, "step": 2216 }, { "epoch": 1.0045310376076122, "grad_norm": 1.2754917535709127, "learning_rate": 7.75461785012615e-05, "loss": 0.8917, "step": 2217 }, { "epoch": 1.0049841413683733, "grad_norm": 1.3424726470637844, "learning_rate": 7.754181288560206e-05, "loss": 0.8887, "step": 2218 }, { "epoch": 1.0054372451291347, "grad_norm": 1.2794920699366457, "learning_rate": 7.753744351304796e-05, "loss": 0.8983, "step": 2219 }, { "epoch": 1.0058903488898958, "grad_norm": 1.1630247101519136, "learning_rate": 7.753307038403645e-05, "loss": 0.8922, "step": 2220 }, { "epoch": 1.006343452650657, "grad_norm": 0.9970294221519728, "learning_rate": 7.752869349900515e-05, "loss": 0.9163, "step": 2221 }, { "epoch": 1.0067965564114183, "grad_norm": 1.126929899981477, "learning_rate": 7.752431285839208e-05, "loss": 0.8934, "step": 2222 }, { "epoch": 1.0072496601721794, "grad_norm": 1.0683378411752826, "learning_rate": 7.75199284626356e-05, "loss": 0.8934, "step": 2223 }, { "epoch": 1.0077027639329406, "grad_norm": 0.8478168943942406, "learning_rate": 7.751554031217448e-05, "loss": 0.8892, "step": 2224 }, { "epoch": 1.008155867693702, "grad_norm": 0.8966996699746376, "learning_rate": 7.751114840744785e-05, "loss": 0.8848, "step": 2225 }, { "epoch": 1.008608971454463, "grad_norm": 0.9427107446883306, "learning_rate": 7.750675274889523e-05, "loss": 0.8848, "step": 2226 }, { "epoch": 1.0090620752152244, "grad_norm": 1.1722952713151453, "learning_rate": 7.750235333695647e-05, "loss": 0.8942, "step": 2227 }, { "epoch": 1.0095151789759855, "grad_norm": 1.1870223472671966, "learning_rate": 7.749795017207188e-05, "loss": 0.904, "step": 2228 }, { "epoch": 1.0099682827367467, "grad_norm": 1.3262603561926156, "learning_rate": 7.749354325468204e-05, "loss": 0.9011, "step": 2229 }, { "epoch": 1.010421386497508, "grad_norm": 0.8309851108014287, "learning_rate": 7.748913258522798e-05, "loss": 0.8867, "step": 2230 }, { "epoch": 1.0108744902582691, "grad_norm": 1.0854813042664904, "learning_rate": 7.74847181641511e-05, "loss": 0.9108, "step": 2231 }, { "epoch": 1.0113275940190303, "grad_norm": 1.0636890077979129, "learning_rate": 7.748029999189317e-05, "loss": 0.8998, "step": 2232 }, { "epoch": 1.0117806977797916, "grad_norm": 1.390803647051958, "learning_rate": 7.747587806889629e-05, "loss": 0.8919, "step": 2233 }, { "epoch": 1.0122338015405528, "grad_norm": 0.9613026712875695, "learning_rate": 7.747145239560298e-05, "loss": 0.8998, "step": 2234 }, { "epoch": 1.012686905301314, "grad_norm": 1.1577574490947198, "learning_rate": 7.746702297245617e-05, "loss": 0.9195, "step": 2235 }, { "epoch": 1.0131400090620752, "grad_norm": 1.4542049333594165, "learning_rate": 7.746258979989908e-05, "loss": 0.9147, "step": 2236 }, { "epoch": 1.0135931128228364, "grad_norm": 0.8619776322097045, "learning_rate": 7.745815287837534e-05, "loss": 0.8914, "step": 2237 }, { "epoch": 1.0140462165835977, "grad_norm": 1.3559517738700848, "learning_rate": 7.745371220832899e-05, "loss": 0.9065, "step": 2238 }, { "epoch": 1.0144993203443589, "grad_norm": 0.7616310400044419, "learning_rate": 7.744926779020441e-05, "loss": 0.8951, "step": 2239 }, { "epoch": 1.01495242410512, "grad_norm": 0.6698153505964627, "learning_rate": 7.744481962444637e-05, "loss": 0.9029, "step": 2240 }, { "epoch": 1.0154055278658813, "grad_norm": 0.9948818955298652, "learning_rate": 7.744036771149999e-05, "loss": 0.9106, "step": 2241 }, { "epoch": 1.0158586316266425, "grad_norm": 1.147371803157324, "learning_rate": 7.743591205181079e-05, "loss": 0.892, "step": 2242 }, { "epoch": 1.0163117353874038, "grad_norm": 1.0277341747376996, "learning_rate": 7.743145264582467e-05, "loss": 0.9008, "step": 2243 }, { "epoch": 1.016764839148165, "grad_norm": 1.6536236802452249, "learning_rate": 7.742698949398789e-05, "loss": 0.8861, "step": 2244 }, { "epoch": 1.017217942908926, "grad_norm": 0.8618361205495769, "learning_rate": 7.742252259674706e-05, "loss": 0.9045, "step": 2245 }, { "epoch": 1.0176710466696874, "grad_norm": 1.4166136609746616, "learning_rate": 7.741805195454922e-05, "loss": 0.924, "step": 2246 }, { "epoch": 1.0181241504304486, "grad_norm": 0.8212685203699162, "learning_rate": 7.741357756784176e-05, "loss": 0.903, "step": 2247 }, { "epoch": 1.0185772541912097, "grad_norm": 1.7010456282894801, "learning_rate": 7.740909943707243e-05, "loss": 0.895, "step": 2248 }, { "epoch": 1.019030357951971, "grad_norm": 0.9459942758022661, "learning_rate": 7.740461756268937e-05, "loss": 0.8952, "step": 2249 }, { "epoch": 1.0194834617127322, "grad_norm": 1.4030688137354075, "learning_rate": 7.74001319451411e-05, "loss": 0.9066, "step": 2250 }, { "epoch": 1.0199365654734933, "grad_norm": 0.874182213498135, "learning_rate": 7.739564258487649e-05, "loss": 0.8963, "step": 2251 }, { "epoch": 1.0203896692342547, "grad_norm": 1.2947064419338827, "learning_rate": 7.73911494823448e-05, "loss": 0.8946, "step": 2252 }, { "epoch": 1.0208427729950158, "grad_norm": 1.4603959976883971, "learning_rate": 7.738665263799568e-05, "loss": 0.9029, "step": 2253 }, { "epoch": 1.0212958767557772, "grad_norm": 0.99833323733246, "learning_rate": 7.738215205227913e-05, "loss": 0.9022, "step": 2254 }, { "epoch": 1.0217489805165383, "grad_norm": 1.2139809899452523, "learning_rate": 7.737764772564555e-05, "loss": 0.8691, "step": 2255 }, { "epoch": 1.0222020842772994, "grad_norm": 1.0991884423890468, "learning_rate": 7.737313965854568e-05, "loss": 0.9082, "step": 2256 }, { "epoch": 1.0226551880380608, "grad_norm": 1.719304565285944, "learning_rate": 7.736862785143064e-05, "loss": 0.892, "step": 2257 }, { "epoch": 1.023108291798822, "grad_norm": 1.1096231543733035, "learning_rate": 7.736411230475198e-05, "loss": 0.9082, "step": 2258 }, { "epoch": 1.0235613955595833, "grad_norm": 1.4755549501985106, "learning_rate": 7.735959301896155e-05, "loss": 0.9035, "step": 2259 }, { "epoch": 1.0240144993203444, "grad_norm": 1.5313167134958618, "learning_rate": 7.735506999451159e-05, "loss": 0.9232, "step": 2260 }, { "epoch": 1.0244676030811055, "grad_norm": 1.0589459297828445, "learning_rate": 7.735054323185476e-05, "loss": 0.8992, "step": 2261 }, { "epoch": 1.0249207068418669, "grad_norm": 0.9322744055454588, "learning_rate": 7.734601273144405e-05, "loss": 0.8948, "step": 2262 }, { "epoch": 1.025373810602628, "grad_norm": 1.802436087579866, "learning_rate": 7.734147849373285e-05, "loss": 0.9003, "step": 2263 }, { "epoch": 1.0258269143633891, "grad_norm": 0.995581471164582, "learning_rate": 7.733694051917489e-05, "loss": 0.8828, "step": 2264 }, { "epoch": 1.0262800181241505, "grad_norm": 1.4729257431897633, "learning_rate": 7.733239880822432e-05, "loss": 0.8837, "step": 2265 }, { "epoch": 1.0267331218849116, "grad_norm": 1.5648872346803746, "learning_rate": 7.732785336133563e-05, "loss": 0.9206, "step": 2266 }, { "epoch": 1.027186225645673, "grad_norm": 0.8817886212121067, "learning_rate": 7.732330417896368e-05, "loss": 0.9111, "step": 2267 }, { "epoch": 1.027639329406434, "grad_norm": 1.4306412181915384, "learning_rate": 7.731875126156372e-05, "loss": 0.8954, "step": 2268 }, { "epoch": 1.0280924331671952, "grad_norm": 1.2661449537900025, "learning_rate": 7.73141946095914e-05, "loss": 0.8921, "step": 2269 }, { "epoch": 1.0285455369279566, "grad_norm": 1.2399196039285274, "learning_rate": 7.730963422350267e-05, "loss": 0.8872, "step": 2270 }, { "epoch": 1.0289986406887177, "grad_norm": 0.8151737831577236, "learning_rate": 7.730507010375393e-05, "loss": 0.9092, "step": 2271 }, { "epoch": 1.0294517444494788, "grad_norm": 0.7789149720800147, "learning_rate": 7.730050225080191e-05, "loss": 0.905, "step": 2272 }, { "epoch": 1.0299048482102402, "grad_norm": 1.3388353598642981, "learning_rate": 7.729593066510372e-05, "loss": 0.8934, "step": 2273 }, { "epoch": 1.0303579519710013, "grad_norm": 1.2683239174952823, "learning_rate": 7.729135534711689e-05, "loss": 0.9056, "step": 2274 }, { "epoch": 1.0308110557317627, "grad_norm": 0.7577843937824048, "learning_rate": 7.728677629729922e-05, "loss": 0.8868, "step": 2275 }, { "epoch": 1.0312641594925238, "grad_norm": 1.041822785308757, "learning_rate": 7.728219351610898e-05, "loss": 0.8897, "step": 2276 }, { "epoch": 1.031717263253285, "grad_norm": 1.046833275768975, "learning_rate": 7.727760700400477e-05, "loss": 0.909, "step": 2277 }, { "epoch": 1.0321703670140463, "grad_norm": 1.3931324191900119, "learning_rate": 7.727301676144559e-05, "loss": 0.8864, "step": 2278 }, { "epoch": 1.0326234707748074, "grad_norm": 1.1021109523733674, "learning_rate": 7.726842278889078e-05, "loss": 0.9273, "step": 2279 }, { "epoch": 1.0330765745355686, "grad_norm": 1.0078481289547385, "learning_rate": 7.726382508680007e-05, "loss": 0.912, "step": 2280 }, { "epoch": 1.03352967829633, "grad_norm": 1.0722602392082732, "learning_rate": 7.725922365563355e-05, "loss": 0.8799, "step": 2281 }, { "epoch": 1.033982782057091, "grad_norm": 0.9644637754791666, "learning_rate": 7.725461849585173e-05, "loss": 0.91, "step": 2282 }, { "epoch": 1.0344358858178522, "grad_norm": 1.3380630658342607, "learning_rate": 7.725000960791545e-05, "loss": 0.9043, "step": 2283 }, { "epoch": 1.0348889895786135, "grad_norm": 1.2102790171558482, "learning_rate": 7.72453969922859e-05, "loss": 0.9026, "step": 2284 }, { "epoch": 1.0353420933393747, "grad_norm": 1.2688006420601263, "learning_rate": 7.72407806494247e-05, "loss": 0.8871, "step": 2285 }, { "epoch": 1.035795197100136, "grad_norm": 0.6872260047655594, "learning_rate": 7.723616057979382e-05, "loss": 0.9006, "step": 2286 }, { "epoch": 1.0362483008608971, "grad_norm": 0.9788509154109835, "learning_rate": 7.72315367838556e-05, "loss": 0.9078, "step": 2287 }, { "epoch": 1.0367014046216583, "grad_norm": 1.3721978552139025, "learning_rate": 7.722690926207273e-05, "loss": 0.9128, "step": 2288 }, { "epoch": 1.0371545083824196, "grad_norm": 0.9198636942010002, "learning_rate": 7.722227801490834e-05, "loss": 0.9229, "step": 2289 }, { "epoch": 1.0376076121431808, "grad_norm": 1.6514515783622503, "learning_rate": 7.721764304282585e-05, "loss": 0.8954, "step": 2290 }, { "epoch": 1.038060715903942, "grad_norm": 0.9122287434651662, "learning_rate": 7.721300434628911e-05, "loss": 0.8981, "step": 2291 }, { "epoch": 1.0385138196647032, "grad_norm": 1.3615691446399756, "learning_rate": 7.720836192576232e-05, "loss": 0.9075, "step": 2292 }, { "epoch": 1.0389669234254644, "grad_norm": 0.848476676368999, "learning_rate": 7.720371578171007e-05, "loss": 0.889, "step": 2293 }, { "epoch": 1.0394200271862257, "grad_norm": 1.6347998311538379, "learning_rate": 7.71990659145973e-05, "loss": 0.9115, "step": 2294 }, { "epoch": 1.0398731309469869, "grad_norm": 0.8777191852087272, "learning_rate": 7.719441232488931e-05, "loss": 0.9012, "step": 2295 }, { "epoch": 1.040326234707748, "grad_norm": 1.491876828465684, "learning_rate": 7.718975501305184e-05, "loss": 0.9179, "step": 2296 }, { "epoch": 1.0407793384685093, "grad_norm": 1.0139876290726733, "learning_rate": 7.718509397955093e-05, "loss": 0.8866, "step": 2297 }, { "epoch": 1.0412324422292705, "grad_norm": 1.4613282365364357, "learning_rate": 7.718042922485304e-05, "loss": 0.9081, "step": 2298 }, { "epoch": 1.0416855459900318, "grad_norm": 1.3385060055991538, "learning_rate": 7.717576074942495e-05, "loss": 0.908, "step": 2299 }, { "epoch": 1.042138649750793, "grad_norm": 1.0529226448088949, "learning_rate": 7.717108855373387e-05, "loss": 0.9078, "step": 2300 }, { "epoch": 1.042591753511554, "grad_norm": 1.1080887984501142, "learning_rate": 7.716641263824736e-05, "loss": 0.8924, "step": 2301 }, { "epoch": 1.0430448572723154, "grad_norm": 1.2378093967406034, "learning_rate": 7.716173300343333e-05, "loss": 0.8994, "step": 2302 }, { "epoch": 1.0434979610330766, "grad_norm": 1.441003590271644, "learning_rate": 7.71570496497601e-05, "loss": 0.907, "step": 2303 }, { "epoch": 1.0439510647938377, "grad_norm": 1.8948085139000894, "learning_rate": 7.715236257769632e-05, "loss": 0.9183, "step": 2304 }, { "epoch": 1.044404168554599, "grad_norm": 1.1187478309814483, "learning_rate": 7.714767178771106e-05, "loss": 0.9, "step": 2305 }, { "epoch": 1.0448572723153602, "grad_norm": 2.1922913159416626, "learning_rate": 7.714297728027374e-05, "loss": 0.9144, "step": 2306 }, { "epoch": 1.0453103760761215, "grad_norm": 1.2619355191388462, "learning_rate": 7.713827905585414e-05, "loss": 0.9107, "step": 2307 }, { "epoch": 1.0457634798368827, "grad_norm": 2.819332986913738, "learning_rate": 7.71335771149224e-05, "loss": 0.9224, "step": 2308 }, { "epoch": 1.0462165835976438, "grad_norm": 2.2012220826250224, "learning_rate": 7.71288714579491e-05, "loss": 0.9046, "step": 2309 }, { "epoch": 1.0466696873584052, "grad_norm": 2.406307850354757, "learning_rate": 7.71241620854051e-05, "loss": 0.9184, "step": 2310 }, { "epoch": 1.0471227911191663, "grad_norm": 2.339238966823013, "learning_rate": 7.711944899776173e-05, "loss": 0.9063, "step": 2311 }, { "epoch": 1.0475758948799274, "grad_norm": 1.4920529731066259, "learning_rate": 7.711473219549059e-05, "loss": 0.9098, "step": 2312 }, { "epoch": 1.0480289986406888, "grad_norm": 1.1843478079828185, "learning_rate": 7.711001167906372e-05, "loss": 0.8861, "step": 2313 }, { "epoch": 1.04848210240145, "grad_norm": 2.395860989976041, "learning_rate": 7.710528744895352e-05, "loss": 0.9071, "step": 2314 }, { "epoch": 1.0489352061622113, "grad_norm": 1.761618078928907, "learning_rate": 7.710055950563275e-05, "loss": 0.9032, "step": 2315 }, { "epoch": 1.0493883099229724, "grad_norm": 2.487011651709097, "learning_rate": 7.709582784957454e-05, "loss": 0.9265, "step": 2316 }, { "epoch": 1.0498414136837335, "grad_norm": 2.243898857060106, "learning_rate": 7.70910924812524e-05, "loss": 0.9121, "step": 2317 }, { "epoch": 1.0502945174444949, "grad_norm": 1.896746384395067, "learning_rate": 7.708635340114021e-05, "loss": 0.9339, "step": 2318 }, { "epoch": 1.050747621205256, "grad_norm": 1.5994433224259752, "learning_rate": 7.708161060971221e-05, "loss": 0.8884, "step": 2319 }, { "epoch": 1.0512007249660171, "grad_norm": 2.152844791582675, "learning_rate": 7.707686410744303e-05, "loss": 0.9201, "step": 2320 }, { "epoch": 1.0516538287267785, "grad_norm": 1.6556215366532263, "learning_rate": 7.70721138948077e-05, "loss": 0.8784, "step": 2321 }, { "epoch": 1.0521069324875396, "grad_norm": 2.2479325093501923, "learning_rate": 7.706735997228151e-05, "loss": 0.9048, "step": 2322 }, { "epoch": 1.052560036248301, "grad_norm": 1.9965283697176717, "learning_rate": 7.706260234034027e-05, "loss": 0.9166, "step": 2323 }, { "epoch": 1.053013140009062, "grad_norm": 2.0522062776045966, "learning_rate": 7.705784099946003e-05, "loss": 0.9101, "step": 2324 }, { "epoch": 1.0534662437698232, "grad_norm": 1.850498251097121, "learning_rate": 7.705307595011731e-05, "loss": 0.9128, "step": 2325 }, { "epoch": 1.0539193475305846, "grad_norm": 2.018225665193534, "learning_rate": 7.704830719278892e-05, "loss": 0.9038, "step": 2326 }, { "epoch": 1.0543724512913457, "grad_norm": 1.76888299497361, "learning_rate": 7.704353472795212e-05, "loss": 0.9302, "step": 2327 }, { "epoch": 1.0548255550521068, "grad_norm": 1.8708824687059766, "learning_rate": 7.703875855608447e-05, "loss": 0.9221, "step": 2328 }, { "epoch": 1.0552786588128682, "grad_norm": 1.585877234211227, "learning_rate": 7.703397867766395e-05, "loss": 0.9033, "step": 2329 }, { "epoch": 1.0557317625736293, "grad_norm": 2.138280945607383, "learning_rate": 7.70291950931689e-05, "loss": 0.8916, "step": 2330 }, { "epoch": 1.0561848663343907, "grad_norm": 1.7559685286228246, "learning_rate": 7.7024407803078e-05, "loss": 0.8915, "step": 2331 }, { "epoch": 1.0566379700951518, "grad_norm": 2.032959038690613, "learning_rate": 7.701961680787034e-05, "loss": 0.902, "step": 2332 }, { "epoch": 1.057091073855913, "grad_norm": 1.7621771881575554, "learning_rate": 7.701482210802535e-05, "loss": 0.8887, "step": 2333 }, { "epoch": 1.0575441776166743, "grad_norm": 2.030560191806519, "learning_rate": 7.701002370402289e-05, "loss": 0.8994, "step": 2334 }, { "epoch": 1.0579972813774354, "grad_norm": 1.6179575021846726, "learning_rate": 7.70052215963431e-05, "loss": 0.896, "step": 2335 }, { "epoch": 1.0584503851381966, "grad_norm": 2.1139391989943905, "learning_rate": 7.700041578546655e-05, "loss": 0.8964, "step": 2336 }, { "epoch": 1.058903488898958, "grad_norm": 1.8971911473817038, "learning_rate": 7.699560627187418e-05, "loss": 0.9125, "step": 2337 }, { "epoch": 1.059356592659719, "grad_norm": 1.9152087078588966, "learning_rate": 7.699079305604728e-05, "loss": 0.8938, "step": 2338 }, { "epoch": 1.0598096964204804, "grad_norm": 1.6527225029950987, "learning_rate": 7.698597613846754e-05, "loss": 0.9076, "step": 2339 }, { "epoch": 1.0602628001812415, "grad_norm": 2.007216755524644, "learning_rate": 7.698115551961695e-05, "loss": 0.9027, "step": 2340 }, { "epoch": 1.0607159039420027, "grad_norm": 1.7086328322771362, "learning_rate": 7.697633119997797e-05, "loss": 0.9186, "step": 2341 }, { "epoch": 1.061169007702764, "grad_norm": 2.1227751412564713, "learning_rate": 7.697150318003337e-05, "loss": 0.9066, "step": 2342 }, { "epoch": 1.0616221114635251, "grad_norm": 1.9088659007891005, "learning_rate": 7.696667146026628e-05, "loss": 0.8892, "step": 2343 }, { "epoch": 1.0620752152242863, "grad_norm": 1.6991307788599834, "learning_rate": 7.696183604116024e-05, "loss": 0.9086, "step": 2344 }, { "epoch": 1.0625283189850476, "grad_norm": 1.44190037223379, "learning_rate": 7.695699692319913e-05, "loss": 0.8868, "step": 2345 }, { "epoch": 1.0629814227458088, "grad_norm": 2.177994242750637, "learning_rate": 7.695215410686723e-05, "loss": 0.9163, "step": 2346 }, { "epoch": 1.0634345265065699, "grad_norm": 1.8995537289230933, "learning_rate": 7.694730759264915e-05, "loss": 0.8951, "step": 2347 }, { "epoch": 1.0638876302673312, "grad_norm": 1.8897164905057826, "learning_rate": 7.694245738102991e-05, "loss": 0.9144, "step": 2348 }, { "epoch": 1.0643407340280924, "grad_norm": 1.7226439042781847, "learning_rate": 7.693760347249486e-05, "loss": 0.8997, "step": 2349 }, { "epoch": 1.0647938377888537, "grad_norm": 1.7589438039432912, "learning_rate": 7.693274586752977e-05, "loss": 0.8762, "step": 2350 }, { "epoch": 1.0652469415496149, "grad_norm": 1.4611336814135292, "learning_rate": 7.692788456662072e-05, "loss": 0.8912, "step": 2351 }, { "epoch": 1.065700045310376, "grad_norm": 2.0817293669646206, "learning_rate": 7.692301957025423e-05, "loss": 0.9086, "step": 2352 }, { "epoch": 1.0661531490711373, "grad_norm": 1.8408202846483848, "learning_rate": 7.69181508789171e-05, "loss": 0.898, "step": 2353 }, { "epoch": 1.0666062528318985, "grad_norm": 1.7866122237758368, "learning_rate": 7.691327849309661e-05, "loss": 0.9028, "step": 2354 }, { "epoch": 1.0670593565926598, "grad_norm": 2.0363501938405757, "learning_rate": 7.69084024132803e-05, "loss": 0.9442, "step": 2355 }, { "epoch": 1.067512460353421, "grad_norm": 1.1129545491478257, "learning_rate": 7.690352263995617e-05, "loss": 0.8952, "step": 2356 }, { "epoch": 1.067965564114182, "grad_norm": 0.755366074554094, "learning_rate": 7.689863917361251e-05, "loss": 0.8999, "step": 2357 }, { "epoch": 1.0684186678749434, "grad_norm": 1.2259462820618026, "learning_rate": 7.689375201473805e-05, "loss": 0.8891, "step": 2358 }, { "epoch": 1.0688717716357046, "grad_norm": 0.8940090404692054, "learning_rate": 7.688886116382185e-05, "loss": 0.8803, "step": 2359 }, { "epoch": 1.0693248753964657, "grad_norm": 1.6642733748589729, "learning_rate": 7.688396662135334e-05, "loss": 0.886, "step": 2360 }, { "epoch": 1.069777979157227, "grad_norm": 1.0149430130698969, "learning_rate": 7.687906838782235e-05, "loss": 0.9227, "step": 2361 }, { "epoch": 1.0702310829179882, "grad_norm": 2.228691567137153, "learning_rate": 7.687416646371904e-05, "loss": 0.9086, "step": 2362 }, { "epoch": 1.0706841866787493, "grad_norm": 1.5066054946680214, "learning_rate": 7.686926084953396e-05, "loss": 0.9079, "step": 2363 }, { "epoch": 1.0711372904395107, "grad_norm": 2.373308122565669, "learning_rate": 7.686435154575803e-05, "loss": 0.8947, "step": 2364 }, { "epoch": 1.0715903942002718, "grad_norm": 2.333086178676265, "learning_rate": 7.685943855288253e-05, "loss": 0.9195, "step": 2365 }, { "epoch": 1.0720434979610332, "grad_norm": 1.5405351606379427, "learning_rate": 7.685452187139912e-05, "loss": 0.8753, "step": 2366 }, { "epoch": 1.0724966017217943, "grad_norm": 1.6311383786272171, "learning_rate": 7.68496015017998e-05, "loss": 0.8943, "step": 2367 }, { "epoch": 1.0729497054825554, "grad_norm": 1.5025957913202772, "learning_rate": 7.6844677444577e-05, "loss": 0.8972, "step": 2368 }, { "epoch": 1.0734028092433168, "grad_norm": 1.661150018194262, "learning_rate": 7.683974970022345e-05, "loss": 0.9003, "step": 2369 }, { "epoch": 1.073855913004078, "grad_norm": 1.1731588222995863, "learning_rate": 7.683481826923231e-05, "loss": 0.9032, "step": 2370 }, { "epoch": 1.0743090167648393, "grad_norm": 2.0823041748646323, "learning_rate": 7.682988315209706e-05, "loss": 0.9017, "step": 2371 }, { "epoch": 1.0747621205256004, "grad_norm": 1.6150741390014307, "learning_rate": 7.682494434931157e-05, "loss": 0.886, "step": 2372 }, { "epoch": 1.0752152242863615, "grad_norm": 1.7368012250517249, "learning_rate": 7.682000186137007e-05, "loss": 0.8861, "step": 2373 }, { "epoch": 1.0756683280471229, "grad_norm": 1.3772671810407555, "learning_rate": 7.681505568876719e-05, "loss": 0.8975, "step": 2374 }, { "epoch": 1.076121431807884, "grad_norm": 1.9572568299877375, "learning_rate": 7.681010583199789e-05, "loss": 0.8949, "step": 2375 }, { "epoch": 1.0765745355686451, "grad_norm": 1.272451573151053, "learning_rate": 7.68051522915575e-05, "loss": 0.8964, "step": 2376 }, { "epoch": 1.0770276393294065, "grad_norm": 1.935788729955422, "learning_rate": 7.680019506794176e-05, "loss": 0.8797, "step": 2377 }, { "epoch": 1.0774807430901676, "grad_norm": 1.4963898493305494, "learning_rate": 7.679523416164672e-05, "loss": 0.8885, "step": 2378 }, { "epoch": 1.0779338468509287, "grad_norm": 2.0750712910452966, "learning_rate": 7.679026957316886e-05, "loss": 0.9086, "step": 2379 }, { "epoch": 1.07838695061169, "grad_norm": 1.5917108357629213, "learning_rate": 7.678530130300498e-05, "loss": 0.8973, "step": 2380 }, { "epoch": 1.0788400543724512, "grad_norm": 1.774428277993835, "learning_rate": 7.678032935165228e-05, "loss": 0.8786, "step": 2381 }, { "epoch": 1.0792931581332126, "grad_norm": 1.4886739278193832, "learning_rate": 7.67753537196083e-05, "loss": 0.8901, "step": 2382 }, { "epoch": 1.0797462618939737, "grad_norm": 1.7795684286726443, "learning_rate": 7.677037440737096e-05, "loss": 0.9144, "step": 2383 }, { "epoch": 1.0801993656547348, "grad_norm": 1.2331115591809787, "learning_rate": 7.676539141543854e-05, "loss": 0.8993, "step": 2384 }, { "epoch": 1.0806524694154962, "grad_norm": 1.75553522764519, "learning_rate": 7.676040474430974e-05, "loss": 0.8808, "step": 2385 }, { "epoch": 1.0811055731762573, "grad_norm": 1.2848274592246065, "learning_rate": 7.675541439448356e-05, "loss": 0.9138, "step": 2386 }, { "epoch": 1.0815586769370187, "grad_norm": 1.8592830123039485, "learning_rate": 7.675042036645941e-05, "loss": 0.8897, "step": 2387 }, { "epoch": 1.0820117806977798, "grad_norm": 1.4849700829166668, "learning_rate": 7.674542266073703e-05, "loss": 0.9002, "step": 2388 }, { "epoch": 1.082464884458541, "grad_norm": 1.6188363104168189, "learning_rate": 7.674042127781657e-05, "loss": 0.9108, "step": 2389 }, { "epoch": 1.0829179882193023, "grad_norm": 1.4724271591716016, "learning_rate": 7.673541621819853e-05, "loss": 0.9085, "step": 2390 }, { "epoch": 1.0833710919800634, "grad_norm": 1.1429917600462032, "learning_rate": 7.673040748238376e-05, "loss": 0.8922, "step": 2391 }, { "epoch": 1.0838241957408246, "grad_norm": 1.6242467413982407, "learning_rate": 7.672539507087353e-05, "loss": 0.8817, "step": 2392 }, { "epoch": 1.084277299501586, "grad_norm": 0.713790223943186, "learning_rate": 7.672037898416942e-05, "loss": 0.9021, "step": 2393 }, { "epoch": 1.084730403262347, "grad_norm": 1.705355937881628, "learning_rate": 7.671535922277338e-05, "loss": 0.9004, "step": 2394 }, { "epoch": 1.0851835070231084, "grad_norm": 0.8846464650076875, "learning_rate": 7.67103357871878e-05, "loss": 0.9088, "step": 2395 }, { "epoch": 1.0856366107838695, "grad_norm": 1.6339260696738667, "learning_rate": 7.670530867791535e-05, "loss": 0.8929, "step": 2396 }, { "epoch": 1.0860897145446307, "grad_norm": 1.1341528859085592, "learning_rate": 7.670027789545911e-05, "loss": 0.9297, "step": 2397 }, { "epoch": 1.086542818305392, "grad_norm": 1.7476013105708186, "learning_rate": 7.669524344032252e-05, "loss": 0.9116, "step": 2398 }, { "epoch": 1.0869959220661531, "grad_norm": 1.4105861614543955, "learning_rate": 7.669020531300939e-05, "loss": 0.919, "step": 2399 }, { "epoch": 1.0874490258269143, "grad_norm": 1.44669962705421, "learning_rate": 7.668516351402392e-05, "loss": 0.8987, "step": 2400 }, { "epoch": 1.0879021295876756, "grad_norm": 1.2256124785689628, "learning_rate": 7.668011804387064e-05, "loss": 0.9225, "step": 2401 }, { "epoch": 1.0883552333484368, "grad_norm": 1.5060844385494698, "learning_rate": 7.667506890305444e-05, "loss": 0.9037, "step": 2402 }, { "epoch": 1.088808337109198, "grad_norm": 1.0205484202505686, "learning_rate": 7.667001609208061e-05, "loss": 0.9019, "step": 2403 }, { "epoch": 1.0892614408699592, "grad_norm": 1.3135894170478073, "learning_rate": 7.666495961145483e-05, "loss": 0.8955, "step": 2404 }, { "epoch": 1.0897145446307204, "grad_norm": 0.9509995936224866, "learning_rate": 7.665989946168307e-05, "loss": 0.8895, "step": 2405 }, { "epoch": 1.0901676483914817, "grad_norm": 1.135268405558866, "learning_rate": 7.665483564327172e-05, "loss": 0.8986, "step": 2406 }, { "epoch": 1.0906207521522429, "grad_norm": 0.9092779145492391, "learning_rate": 7.664976815672755e-05, "loss": 0.8914, "step": 2407 }, { "epoch": 1.091073855913004, "grad_norm": 1.1419606101626985, "learning_rate": 7.664469700255766e-05, "loss": 0.9007, "step": 2408 }, { "epoch": 1.0915269596737653, "grad_norm": 0.9077313872318958, "learning_rate": 7.663962218126951e-05, "loss": 0.8907, "step": 2409 }, { "epoch": 1.0919800634345265, "grad_norm": 1.1828520921033319, "learning_rate": 7.6634543693371e-05, "loss": 0.9157, "step": 2410 }, { "epoch": 1.0924331671952878, "grad_norm": 1.1762775747810967, "learning_rate": 7.662946153937031e-05, "loss": 0.9155, "step": 2411 }, { "epoch": 1.092886270956049, "grad_norm": 1.0244867314394923, "learning_rate": 7.662437571977601e-05, "loss": 0.8899, "step": 2412 }, { "epoch": 1.09333937471681, "grad_norm": 0.9293735979799402, "learning_rate": 7.661928623509709e-05, "loss": 0.9047, "step": 2413 }, { "epoch": 1.0937924784775714, "grad_norm": 0.8368755863811849, "learning_rate": 7.661419308584283e-05, "loss": 0.8992, "step": 2414 }, { "epoch": 1.0942455822383326, "grad_norm": 1.0730111075583222, "learning_rate": 7.660909627252294e-05, "loss": 0.8833, "step": 2415 }, { "epoch": 1.0946986859990937, "grad_norm": 0.9696017140207275, "learning_rate": 7.660399579564745e-05, "loss": 0.9103, "step": 2416 }, { "epoch": 1.095151789759855, "grad_norm": 0.892447728070019, "learning_rate": 7.659889165572678e-05, "loss": 0.8959, "step": 2417 }, { "epoch": 1.0956048935206162, "grad_norm": 0.9586372759773943, "learning_rate": 7.659378385327174e-05, "loss": 0.9115, "step": 2418 }, { "epoch": 1.0960579972813775, "grad_norm": 1.2734900676642131, "learning_rate": 7.658867238879344e-05, "loss": 0.8915, "step": 2419 }, { "epoch": 1.0965111010421387, "grad_norm": 0.9111158304636244, "learning_rate": 7.65835572628034e-05, "loss": 0.9132, "step": 2420 }, { "epoch": 1.0969642048028998, "grad_norm": 0.6736847276256464, "learning_rate": 7.657843847581353e-05, "loss": 0.8806, "step": 2421 }, { "epoch": 1.0974173085636612, "grad_norm": 0.5688999998631088, "learning_rate": 7.657331602833607e-05, "loss": 0.8952, "step": 2422 }, { "epoch": 1.0978704123244223, "grad_norm": 0.6332401561165614, "learning_rate": 7.656818992088362e-05, "loss": 0.907, "step": 2423 }, { "epoch": 1.0983235160851834, "grad_norm": 0.7569678477673942, "learning_rate": 7.656306015396917e-05, "loss": 0.8968, "step": 2424 }, { "epoch": 1.0987766198459448, "grad_norm": 1.0260289360140897, "learning_rate": 7.655792672810606e-05, "loss": 0.8959, "step": 2425 }, { "epoch": 1.099229723606706, "grad_norm": 1.3619055619526208, "learning_rate": 7.655278964380803e-05, "loss": 0.9074, "step": 2426 }, { "epoch": 1.0996828273674673, "grad_norm": 0.8415265996638772, "learning_rate": 7.654764890158912e-05, "loss": 0.92, "step": 2427 }, { "epoch": 1.1001359311282284, "grad_norm": 0.6444217828595754, "learning_rate": 7.654250450196379e-05, "loss": 0.8918, "step": 2428 }, { "epoch": 1.1005890348889895, "grad_norm": 0.6008796600792148, "learning_rate": 7.653735644544689e-05, "loss": 0.9126, "step": 2429 }, { "epoch": 1.1010421386497509, "grad_norm": 0.7160683264888935, "learning_rate": 7.653220473255355e-05, "loss": 0.9048, "step": 2430 }, { "epoch": 1.101495242410512, "grad_norm": 0.779531152365368, "learning_rate": 7.652704936379932e-05, "loss": 0.8835, "step": 2431 }, { "epoch": 1.1019483461712731, "grad_norm": 1.042490891857465, "learning_rate": 7.652189033970013e-05, "loss": 0.8872, "step": 2432 }, { "epoch": 1.1024014499320345, "grad_norm": 1.5358968325911309, "learning_rate": 7.651672766077224e-05, "loss": 0.8912, "step": 2433 }, { "epoch": 1.1028545536927956, "grad_norm": 0.7218792918186365, "learning_rate": 7.65115613275323e-05, "loss": 0.8758, "step": 2434 }, { "epoch": 1.103307657453557, "grad_norm": 1.0895002310483632, "learning_rate": 7.65063913404973e-05, "loss": 0.9061, "step": 2435 }, { "epoch": 1.103760761214318, "grad_norm": 1.6845468886954433, "learning_rate": 7.650121770018463e-05, "loss": 0.9087, "step": 2436 }, { "epoch": 1.1042138649750792, "grad_norm": 0.7313781674203238, "learning_rate": 7.649604040711202e-05, "loss": 0.9024, "step": 2437 }, { "epoch": 1.1046669687358406, "grad_norm": 1.8713977493631124, "learning_rate": 7.649085946179758e-05, "loss": 0.9195, "step": 2438 }, { "epoch": 1.1051200724966017, "grad_norm": 0.9409174196710532, "learning_rate": 7.648567486475979e-05, "loss": 0.8906, "step": 2439 }, { "epoch": 1.1055731762573628, "grad_norm": 2.11913987354013, "learning_rate": 7.648048661651744e-05, "loss": 0.8979, "step": 2440 }, { "epoch": 1.1060262800181242, "grad_norm": 1.574367163821878, "learning_rate": 7.647529471758977e-05, "loss": 0.9237, "step": 2441 }, { "epoch": 1.1064793837788853, "grad_norm": 1.8326871903637578, "learning_rate": 7.647009916849634e-05, "loss": 0.9037, "step": 2442 }, { "epoch": 1.1069324875396467, "grad_norm": 1.666665124131595, "learning_rate": 7.646489996975708e-05, "loss": 0.9082, "step": 2443 }, { "epoch": 1.1073855913004078, "grad_norm": 1.5394579696188702, "learning_rate": 7.645969712189227e-05, "loss": 0.9048, "step": 2444 }, { "epoch": 1.107838695061169, "grad_norm": 1.2246743744788742, "learning_rate": 7.645449062542257e-05, "loss": 0.9089, "step": 2445 }, { "epoch": 1.1082917988219303, "grad_norm": 1.626467256568137, "learning_rate": 7.644928048086903e-05, "loss": 0.9023, "step": 2446 }, { "epoch": 1.1087449025826914, "grad_norm": 1.2350457460974182, "learning_rate": 7.644406668875302e-05, "loss": 0.8895, "step": 2447 }, { "epoch": 1.1091980063434526, "grad_norm": 1.7596256805109975, "learning_rate": 7.643884924959631e-05, "loss": 0.9244, "step": 2448 }, { "epoch": 1.109651110104214, "grad_norm": 1.4878736595329782, "learning_rate": 7.643362816392103e-05, "loss": 0.9182, "step": 2449 }, { "epoch": 1.110104213864975, "grad_norm": 1.4034848424633262, "learning_rate": 7.642840343224964e-05, "loss": 0.8952, "step": 2450 }, { "epoch": 1.1105573176257364, "grad_norm": 1.3164775631027015, "learning_rate": 7.642317505510498e-05, "loss": 0.9027, "step": 2451 }, { "epoch": 1.1110104213864975, "grad_norm": 1.2790467245664723, "learning_rate": 7.641794303301032e-05, "loss": 0.8799, "step": 2452 }, { "epoch": 1.1114635251472587, "grad_norm": 1.037729926816263, "learning_rate": 7.64127073664892e-05, "loss": 0.9031, "step": 2453 }, { "epoch": 1.11191662890802, "grad_norm": 1.4542817955365495, "learning_rate": 7.640746805606556e-05, "loss": 0.907, "step": 2454 }, { "epoch": 1.1123697326687811, "grad_norm": 1.1237390896834103, "learning_rate": 7.640222510226372e-05, "loss": 0.8915, "step": 2455 }, { "epoch": 1.1128228364295423, "grad_norm": 1.4820758432120709, "learning_rate": 7.639697850560836e-05, "loss": 0.9128, "step": 2456 }, { "epoch": 1.1132759401903036, "grad_norm": 1.337489187543574, "learning_rate": 7.639172826662453e-05, "loss": 0.9019, "step": 2457 }, { "epoch": 1.1137290439510648, "grad_norm": 1.322602848412767, "learning_rate": 7.63864743858376e-05, "loss": 0.9128, "step": 2458 }, { "epoch": 1.114182147711826, "grad_norm": 1.2286989088177818, "learning_rate": 7.638121686377337e-05, "loss": 0.9051, "step": 2459 }, { "epoch": 1.1146352514725872, "grad_norm": 1.2758019975775967, "learning_rate": 7.637595570095796e-05, "loss": 0.9066, "step": 2460 }, { "epoch": 1.1150883552333484, "grad_norm": 1.038483951543874, "learning_rate": 7.637069089791786e-05, "loss": 0.8913, "step": 2461 }, { "epoch": 1.1155414589941097, "grad_norm": 1.439346485574069, "learning_rate": 7.636542245517995e-05, "loss": 0.8998, "step": 2462 }, { "epoch": 1.1159945627548709, "grad_norm": 1.145976300511986, "learning_rate": 7.636015037327142e-05, "loss": 0.9053, "step": 2463 }, { "epoch": 1.116447666515632, "grad_norm": 1.4396051236007736, "learning_rate": 7.63548746527199e-05, "loss": 0.8943, "step": 2464 }, { "epoch": 1.1169007702763933, "grad_norm": 1.3141506706693753, "learning_rate": 7.63495952940533e-05, "loss": 0.9085, "step": 2465 }, { "epoch": 1.1173538740371545, "grad_norm": 1.1688843778338955, "learning_rate": 7.634431229779998e-05, "loss": 0.9004, "step": 2466 }, { "epoch": 1.1178069777979158, "grad_norm": 1.122522737490401, "learning_rate": 7.63390256644886e-05, "loss": 0.9107, "step": 2467 }, { "epoch": 1.118260081558677, "grad_norm": 1.4491411887900991, "learning_rate": 7.63337353946482e-05, "loss": 0.9177, "step": 2468 }, { "epoch": 1.118713185319438, "grad_norm": 1.0952439257744093, "learning_rate": 7.632844148880822e-05, "loss": 0.9052, "step": 2469 }, { "epoch": 1.1191662890801994, "grad_norm": 1.424288031160261, "learning_rate": 7.632314394749839e-05, "loss": 0.899, "step": 2470 }, { "epoch": 1.1196193928409606, "grad_norm": 1.1060200869786705, "learning_rate": 7.631784277124887e-05, "loss": 0.9011, "step": 2471 }, { "epoch": 1.1200724966017217, "grad_norm": 1.3329982562246196, "learning_rate": 7.631253796059016e-05, "loss": 0.9113, "step": 2472 }, { "epoch": 1.120525600362483, "grad_norm": 1.2524660992080876, "learning_rate": 7.630722951605314e-05, "loss": 0.9155, "step": 2473 }, { "epoch": 1.1209787041232442, "grad_norm": 1.1882835005654142, "learning_rate": 7.6301917438169e-05, "loss": 0.8968, "step": 2474 }, { "epoch": 1.1214318078840055, "grad_norm": 1.1035426739280592, "learning_rate": 7.629660172746937e-05, "loss": 0.8921, "step": 2475 }, { "epoch": 1.1218849116447667, "grad_norm": 1.4155141250629364, "learning_rate": 7.629128238448617e-05, "loss": 0.9141, "step": 2476 }, { "epoch": 1.1223380154055278, "grad_norm": 1.0370676414495088, "learning_rate": 7.628595940975175e-05, "loss": 0.9127, "step": 2477 }, { "epoch": 1.1227911191662892, "grad_norm": 1.4769374870376195, "learning_rate": 7.628063280379878e-05, "loss": 0.9028, "step": 2478 }, { "epoch": 1.1232442229270503, "grad_norm": 1.0783495886817116, "learning_rate": 7.627530256716029e-05, "loss": 0.9237, "step": 2479 }, { "epoch": 1.1236973266878114, "grad_norm": 1.0443559733420122, "learning_rate": 7.626996870036971e-05, "loss": 0.9057, "step": 2480 }, { "epoch": 1.1241504304485728, "grad_norm": 0.9625445280332027, "learning_rate": 7.626463120396082e-05, "loss": 0.8992, "step": 2481 }, { "epoch": 1.124603534209334, "grad_norm": 1.0682455840280343, "learning_rate": 7.625929007846773e-05, "loss": 0.9059, "step": 2482 }, { "epoch": 1.1250566379700953, "grad_norm": 1.1247629498456055, "learning_rate": 7.625394532442494e-05, "loss": 0.8982, "step": 2483 }, { "epoch": 1.1255097417308564, "grad_norm": 1.8164564680563333, "learning_rate": 7.624859694236734e-05, "loss": 0.9288, "step": 2484 }, { "epoch": 1.1259628454916175, "grad_norm": 1.095311846174531, "learning_rate": 7.624324493283013e-05, "loss": 0.8932, "step": 2485 }, { "epoch": 1.1264159492523789, "grad_norm": 2.2614403570745223, "learning_rate": 7.62378892963489e-05, "loss": 0.9159, "step": 2486 }, { "epoch": 1.12686905301314, "grad_norm": 1.7014928150780788, "learning_rate": 7.623253003345961e-05, "loss": 0.9101, "step": 2487 }, { "epoch": 1.1273221567739011, "grad_norm": 1.7248447993325555, "learning_rate": 7.622716714469857e-05, "loss": 0.9247, "step": 2488 }, { "epoch": 1.1277752605346625, "grad_norm": 1.173737942071054, "learning_rate": 7.622180063060245e-05, "loss": 0.8976, "step": 2489 }, { "epoch": 1.1282283642954236, "grad_norm": 1.2330322218708065, "learning_rate": 7.621643049170831e-05, "loss": 0.9222, "step": 2490 }, { "epoch": 1.1286814680561847, "grad_norm": 1.4308332190219062, "learning_rate": 7.62110567285535e-05, "loss": 0.8931, "step": 2491 }, { "epoch": 1.129134571816946, "grad_norm": 1.2204481358441428, "learning_rate": 7.620567934167586e-05, "loss": 0.9152, "step": 2492 }, { "epoch": 1.1295876755777072, "grad_norm": 0.8755811412041021, "learning_rate": 7.620029833161345e-05, "loss": 0.8861, "step": 2493 }, { "epoch": 1.1300407793384686, "grad_norm": 1.5231414061480701, "learning_rate": 7.619491369890482e-05, "loss": 0.9066, "step": 2494 }, { "epoch": 1.1304938830992297, "grad_norm": 0.946519582692234, "learning_rate": 7.618952544408876e-05, "loss": 0.8912, "step": 2495 }, { "epoch": 1.1309469868599908, "grad_norm": 0.804641926230491, "learning_rate": 7.618413356770452e-05, "loss": 0.8903, "step": 2496 }, { "epoch": 1.1314000906207522, "grad_norm": 1.1189569548835177, "learning_rate": 7.61787380702917e-05, "loss": 0.9015, "step": 2497 }, { "epoch": 1.1318531943815133, "grad_norm": 1.2361907181875176, "learning_rate": 7.617333895239018e-05, "loss": 0.9042, "step": 2498 }, { "epoch": 1.1323062981422747, "grad_norm": 0.8299217422127488, "learning_rate": 7.616793621454032e-05, "loss": 0.8987, "step": 2499 }, { "epoch": 1.1327594019030358, "grad_norm": 1.091918459558223, "learning_rate": 7.616252985728274e-05, "loss": 0.9028, "step": 2500 }, { "epoch": 1.133212505663797, "grad_norm": 1.3535189470012956, "learning_rate": 7.615711988115849e-05, "loss": 0.9009, "step": 2501 }, { "epoch": 1.1336656094245583, "grad_norm": 0.7371893855723126, "learning_rate": 7.615170628670896e-05, "loss": 0.8948, "step": 2502 }, { "epoch": 1.1341187131853194, "grad_norm": 0.9016663558515977, "learning_rate": 7.614628907447589e-05, "loss": 0.9168, "step": 2503 }, { "epoch": 1.1345718169460806, "grad_norm": 1.2026643783901327, "learning_rate": 7.614086824500141e-05, "loss": 0.9046, "step": 2504 }, { "epoch": 1.135024920706842, "grad_norm": 0.7178672481769706, "learning_rate": 7.613544379882796e-05, "loss": 0.9008, "step": 2505 }, { "epoch": 1.135478024467603, "grad_norm": 0.7639574356757772, "learning_rate": 7.613001573649841e-05, "loss": 0.9111, "step": 2506 }, { "epoch": 1.1359311282283642, "grad_norm": 1.1158116891631638, "learning_rate": 7.612458405855595e-05, "loss": 0.9035, "step": 2507 }, { "epoch": 1.1363842319891255, "grad_norm": 1.2593423370140862, "learning_rate": 7.611914876554415e-05, "loss": 0.912, "step": 2508 }, { "epoch": 1.1368373357498867, "grad_norm": 0.9141447738132088, "learning_rate": 7.611370985800688e-05, "loss": 0.905, "step": 2509 }, { "epoch": 1.137290439510648, "grad_norm": 0.992287262579865, "learning_rate": 7.61082673364885e-05, "loss": 0.8949, "step": 2510 }, { "epoch": 1.1377435432714091, "grad_norm": 1.0969258389700074, "learning_rate": 7.610282120153361e-05, "loss": 0.8945, "step": 2511 }, { "epoch": 1.1381966470321703, "grad_norm": 1.2349398063043657, "learning_rate": 7.609737145368723e-05, "loss": 0.9208, "step": 2512 }, { "epoch": 1.1386497507929316, "grad_norm": 1.3145136390992247, "learning_rate": 7.609191809349473e-05, "loss": 0.9034, "step": 2513 }, { "epoch": 1.1391028545536928, "grad_norm": 0.6596030950709092, "learning_rate": 7.608646112150183e-05, "loss": 0.8897, "step": 2514 }, { "epoch": 1.139555958314454, "grad_norm": 1.0275128001853653, "learning_rate": 7.608100053825463e-05, "loss": 0.882, "step": 2515 }, { "epoch": 1.1400090620752152, "grad_norm": 1.7017020352784604, "learning_rate": 7.607553634429956e-05, "loss": 0.904, "step": 2516 }, { "epoch": 1.1404621658359764, "grad_norm": 0.6224462410009706, "learning_rate": 7.607006854018347e-05, "loss": 0.9029, "step": 2517 }, { "epoch": 1.1409152695967377, "grad_norm": 1.6953177830520503, "learning_rate": 7.606459712645352e-05, "loss": 0.904, "step": 2518 }, { "epoch": 1.1413683733574989, "grad_norm": 0.79726000354287, "learning_rate": 7.605912210365725e-05, "loss": 0.8987, "step": 2519 }, { "epoch": 1.14182147711826, "grad_norm": 1.5089668829144671, "learning_rate": 7.605364347234255e-05, "loss": 0.9012, "step": 2520 }, { "epoch": 1.1422745808790213, "grad_norm": 0.8011331531108514, "learning_rate": 7.604816123305768e-05, "loss": 0.898, "step": 2521 }, { "epoch": 1.1427276846397825, "grad_norm": 1.211113934680831, "learning_rate": 7.604267538635127e-05, "loss": 0.9194, "step": 2522 }, { "epoch": 1.1431807884005436, "grad_norm": 1.260050413581243, "learning_rate": 7.603718593277229e-05, "loss": 0.8889, "step": 2523 }, { "epoch": 1.143633892161305, "grad_norm": 1.559198242687642, "learning_rate": 7.60316928728701e-05, "loss": 0.9018, "step": 2524 }, { "epoch": 1.144086995922066, "grad_norm": 0.8384271694568277, "learning_rate": 7.602619620719438e-05, "loss": 0.8957, "step": 2525 }, { "epoch": 1.1445400996828274, "grad_norm": 2.187653321180497, "learning_rate": 7.602069593629521e-05, "loss": 0.9052, "step": 2526 }, { "epoch": 1.1449932034435886, "grad_norm": 1.461875477174143, "learning_rate": 7.601519206072301e-05, "loss": 0.9089, "step": 2527 }, { "epoch": 1.1454463072043497, "grad_norm": 1.8288095872029724, "learning_rate": 7.600968458102855e-05, "loss": 0.9225, "step": 2528 }, { "epoch": 1.145899410965111, "grad_norm": 1.416393915984799, "learning_rate": 7.6004173497763e-05, "loss": 0.8998, "step": 2529 }, { "epoch": 1.1463525147258722, "grad_norm": 1.4598930116661744, "learning_rate": 7.599865881147786e-05, "loss": 0.8883, "step": 2530 }, { "epoch": 1.1468056184866335, "grad_norm": 1.2475518652641642, "learning_rate": 7.5993140522725e-05, "loss": 0.9194, "step": 2531 }, { "epoch": 1.1472587222473947, "grad_norm": 1.0945981442080077, "learning_rate": 7.598761863205662e-05, "loss": 0.891, "step": 2532 }, { "epoch": 1.1477118260081558, "grad_norm": 1.0386261267041468, "learning_rate": 7.598209314002535e-05, "loss": 0.8907, "step": 2533 }, { "epoch": 1.1481649297689172, "grad_norm": 0.9300780197000655, "learning_rate": 7.597656404718411e-05, "loss": 0.9093, "step": 2534 }, { "epoch": 1.1486180335296783, "grad_norm": 1.74677943855791, "learning_rate": 7.597103135408622e-05, "loss": 0.9007, "step": 2535 }, { "epoch": 1.1490711372904394, "grad_norm": 0.9240872756454068, "learning_rate": 7.596549506128534e-05, "loss": 0.9055, "step": 2536 }, { "epoch": 1.1495242410512008, "grad_norm": 1.6025583293557257, "learning_rate": 7.595995516933551e-05, "loss": 0.9034, "step": 2537 }, { "epoch": 1.149977344811962, "grad_norm": 1.307333943683373, "learning_rate": 7.595441167879111e-05, "loss": 0.8973, "step": 2538 }, { "epoch": 1.150430448572723, "grad_norm": 0.9343184808100908, "learning_rate": 7.594886459020691e-05, "loss": 0.891, "step": 2539 }, { "epoch": 1.1508835523334844, "grad_norm": 1.882536086841714, "learning_rate": 7.5943313904138e-05, "loss": 0.8904, "step": 2540 }, { "epoch": 1.1513366560942455, "grad_norm": 1.2330089654798388, "learning_rate": 7.593775962113985e-05, "loss": 0.893, "step": 2541 }, { "epoch": 1.1517897598550069, "grad_norm": 1.8370615605087475, "learning_rate": 7.59322017417683e-05, "loss": 0.9142, "step": 2542 }, { "epoch": 1.152242863615768, "grad_norm": 1.4892612228872482, "learning_rate": 7.592664026657953e-05, "loss": 0.9111, "step": 2543 }, { "epoch": 1.1526959673765291, "grad_norm": 1.5400524592711715, "learning_rate": 7.592107519613009e-05, "loss": 0.887, "step": 2544 }, { "epoch": 1.1531490711372905, "grad_norm": 1.601296305379991, "learning_rate": 7.59155065309769e-05, "loss": 0.9018, "step": 2545 }, { "epoch": 1.1536021748980516, "grad_norm": 1.451054083572372, "learning_rate": 7.590993427167723e-05, "loss": 0.9097, "step": 2546 }, { "epoch": 1.154055278658813, "grad_norm": 1.5702497512747484, "learning_rate": 7.590435841878869e-05, "loss": 0.8914, "step": 2547 }, { "epoch": 1.154508382419574, "grad_norm": 1.23388523400511, "learning_rate": 7.589877897286928e-05, "loss": 0.8859, "step": 2548 }, { "epoch": 1.1549614861803352, "grad_norm": 1.423171871949912, "learning_rate": 7.589319593447734e-05, "loss": 0.9299, "step": 2549 }, { "epoch": 1.1554145899410966, "grad_norm": 1.0949901136093583, "learning_rate": 7.58876093041716e-05, "loss": 0.8999, "step": 2550 }, { "epoch": 1.1558676937018577, "grad_norm": 1.6106840243933969, "learning_rate": 7.58820190825111e-05, "loss": 0.9127, "step": 2551 }, { "epoch": 1.1563207974626188, "grad_norm": 0.8558433374134498, "learning_rate": 7.587642527005527e-05, "loss": 0.9198, "step": 2552 }, { "epoch": 1.1567739012233802, "grad_norm": 1.8238249251945915, "learning_rate": 7.587082786736391e-05, "loss": 0.9084, "step": 2553 }, { "epoch": 1.1572270049841413, "grad_norm": 1.1889508227923724, "learning_rate": 7.586522687499718e-05, "loss": 0.9203, "step": 2554 }, { "epoch": 1.1576801087449025, "grad_norm": 1.5417256223447993, "learning_rate": 7.585962229351554e-05, "loss": 0.8952, "step": 2555 }, { "epoch": 1.1581332125056638, "grad_norm": 1.4855219588442565, "learning_rate": 7.585401412347986e-05, "loss": 0.9237, "step": 2556 }, { "epoch": 1.158586316266425, "grad_norm": 1.3402070771615322, "learning_rate": 7.58484023654514e-05, "loss": 0.8961, "step": 2557 }, { "epoch": 1.1590394200271863, "grad_norm": 1.162015055013641, "learning_rate": 7.584278701999172e-05, "loss": 0.886, "step": 2558 }, { "epoch": 1.1594925237879474, "grad_norm": 1.2345825618718071, "learning_rate": 7.583716808766275e-05, "loss": 0.8851, "step": 2559 }, { "epoch": 1.1599456275487086, "grad_norm": 0.9390574385012429, "learning_rate": 7.583154556902679e-05, "loss": 0.8703, "step": 2560 }, { "epoch": 1.16039873130947, "grad_norm": 0.9400066228207095, "learning_rate": 7.582591946464652e-05, "loss": 0.8898, "step": 2561 }, { "epoch": 1.160851835070231, "grad_norm": 0.7714973602379194, "learning_rate": 7.582028977508495e-05, "loss": 0.8839, "step": 2562 }, { "epoch": 1.1613049388309924, "grad_norm": 0.9811253459659807, "learning_rate": 7.581465650090544e-05, "loss": 0.8881, "step": 2563 }, { "epoch": 1.1617580425917535, "grad_norm": 0.7817938148503721, "learning_rate": 7.580901964267173e-05, "loss": 0.8913, "step": 2564 }, { "epoch": 1.1622111463525147, "grad_norm": 1.1102663805090036, "learning_rate": 7.580337920094792e-05, "loss": 0.9037, "step": 2565 }, { "epoch": 1.162664250113276, "grad_norm": 1.4536608404526383, "learning_rate": 7.579773517629847e-05, "loss": 0.9036, "step": 2566 }, { "epoch": 1.1631173538740371, "grad_norm": 0.6363676370070246, "learning_rate": 7.579208756928816e-05, "loss": 0.9154, "step": 2567 }, { "epoch": 1.1635704576347983, "grad_norm": 0.9280328517042782, "learning_rate": 7.578643638048219e-05, "loss": 0.8944, "step": 2568 }, { "epoch": 1.1640235613955596, "grad_norm": 1.3568346956500814, "learning_rate": 7.578078161044608e-05, "loss": 0.9023, "step": 2569 }, { "epoch": 1.1644766651563208, "grad_norm": 0.9623900182341611, "learning_rate": 7.577512325974571e-05, "loss": 0.8936, "step": 2570 }, { "epoch": 1.1649297689170819, "grad_norm": 1.0121668457287536, "learning_rate": 7.576946132894732e-05, "loss": 0.8803, "step": 2571 }, { "epoch": 1.1653828726778432, "grad_norm": 1.1394785070945326, "learning_rate": 7.576379581861752e-05, "loss": 0.8967, "step": 2572 }, { "epoch": 1.1658359764386044, "grad_norm": 1.3580381838880353, "learning_rate": 7.575812672932326e-05, "loss": 0.8858, "step": 2573 }, { "epoch": 1.1662890801993657, "grad_norm": 0.5866092653779166, "learning_rate": 7.575245406163188e-05, "loss": 0.8999, "step": 2574 }, { "epoch": 1.1667421839601269, "grad_norm": 1.1045787908620466, "learning_rate": 7.574677781611104e-05, "loss": 0.889, "step": 2575 }, { "epoch": 1.167195287720888, "grad_norm": 1.730479517707596, "learning_rate": 7.574109799332878e-05, "loss": 0.8947, "step": 2576 }, { "epoch": 1.1676483914816493, "grad_norm": 0.6098537418858359, "learning_rate": 7.573541459385349e-05, "loss": 0.904, "step": 2577 }, { "epoch": 1.1681014952424105, "grad_norm": 1.602014148868311, "learning_rate": 7.572972761825393e-05, "loss": 0.8866, "step": 2578 }, { "epoch": 1.1685545990031718, "grad_norm": 0.9403183505494258, "learning_rate": 7.572403706709918e-05, "loss": 0.9016, "step": 2579 }, { "epoch": 1.169007702763933, "grad_norm": 1.2863680116173193, "learning_rate": 7.571834294095875e-05, "loss": 0.9016, "step": 2580 }, { "epoch": 1.169460806524694, "grad_norm": 1.1176179010076406, "learning_rate": 7.571264524040242e-05, "loss": 0.8769, "step": 2581 }, { "epoch": 1.1699139102854554, "grad_norm": 0.7749990321354635, "learning_rate": 7.570694396600041e-05, "loss": 0.9328, "step": 2582 }, { "epoch": 1.1703670140462166, "grad_norm": 1.402772409027116, "learning_rate": 7.570123911832323e-05, "loss": 0.8938, "step": 2583 }, { "epoch": 1.170820117806978, "grad_norm": 0.9476792997364255, "learning_rate": 7.569553069794181e-05, "loss": 0.8988, "step": 2584 }, { "epoch": 1.171273221567739, "grad_norm": 1.1902819552966173, "learning_rate": 7.568981870542736e-05, "loss": 0.8953, "step": 2585 }, { "epoch": 1.1717263253285002, "grad_norm": 0.8949516531305024, "learning_rate": 7.568410314135153e-05, "loss": 0.9007, "step": 2586 }, { "epoch": 1.1721794290892615, "grad_norm": 0.9468324526670355, "learning_rate": 7.567838400628627e-05, "loss": 0.9019, "step": 2587 }, { "epoch": 1.1726325328500227, "grad_norm": 0.988651952115229, "learning_rate": 7.567266130080392e-05, "loss": 0.8902, "step": 2588 }, { "epoch": 1.1730856366107838, "grad_norm": 1.428944744289653, "learning_rate": 7.566693502547715e-05, "loss": 0.9048, "step": 2589 }, { "epoch": 1.1735387403715452, "grad_norm": 0.7240917774842935, "learning_rate": 7.566120518087902e-05, "loss": 0.8948, "step": 2590 }, { "epoch": 1.1739918441323063, "grad_norm": 1.1428720766309122, "learning_rate": 7.56554717675829e-05, "loss": 0.8955, "step": 2591 }, { "epoch": 1.1744449478930674, "grad_norm": 0.7126607402191453, "learning_rate": 7.564973478616257e-05, "loss": 0.8873, "step": 2592 }, { "epoch": 1.1748980516538288, "grad_norm": 1.1294726683027851, "learning_rate": 7.564399423719214e-05, "loss": 0.8881, "step": 2593 }, { "epoch": 1.17535115541459, "grad_norm": 0.8131919919355227, "learning_rate": 7.563825012124607e-05, "loss": 0.8949, "step": 2594 }, { "epoch": 1.1758042591753513, "grad_norm": 1.1737837637694364, "learning_rate": 7.56325024388992e-05, "loss": 0.896, "step": 2595 }, { "epoch": 1.1762573629361124, "grad_norm": 1.3881395988493723, "learning_rate": 7.56267511907267e-05, "loss": 0.9009, "step": 2596 }, { "epoch": 1.1767104666968735, "grad_norm": 0.500589922237516, "learning_rate": 7.56209963773041e-05, "loss": 0.8893, "step": 2597 }, { "epoch": 1.1771635704576349, "grad_norm": 0.8984531850115355, "learning_rate": 7.561523799920734e-05, "loss": 0.8931, "step": 2598 }, { "epoch": 1.177616674218396, "grad_norm": 0.938313171039311, "learning_rate": 7.560947605701265e-05, "loss": 0.9159, "step": 2599 }, { "epoch": 1.1780697779791574, "grad_norm": 1.3080469335730842, "learning_rate": 7.560371055129661e-05, "loss": 0.8835, "step": 2600 }, { "epoch": 1.1785228817399185, "grad_norm": 1.1723298510718534, "learning_rate": 7.559794148263624e-05, "loss": 0.8925, "step": 2601 }, { "epoch": 1.1789759855006796, "grad_norm": 1.026603520992225, "learning_rate": 7.559216885160883e-05, "loss": 0.894, "step": 2602 }, { "epoch": 1.179429089261441, "grad_norm": 0.993330805864796, "learning_rate": 7.558639265879207e-05, "loss": 0.8967, "step": 2603 }, { "epoch": 1.179882193022202, "grad_norm": 0.9676129468041501, "learning_rate": 7.5580612904764e-05, "loss": 0.9106, "step": 2604 }, { "epoch": 1.1803352967829632, "grad_norm": 1.2618648158902837, "learning_rate": 7.5574829590103e-05, "loss": 0.8955, "step": 2605 }, { "epoch": 1.1807884005437246, "grad_norm": 1.1388154042731478, "learning_rate": 7.556904271538783e-05, "loss": 0.8886, "step": 2606 }, { "epoch": 1.1812415043044857, "grad_norm": 1.0180459254704015, "learning_rate": 7.556325228119761e-05, "loss": 0.8783, "step": 2607 }, { "epoch": 1.1816946080652468, "grad_norm": 1.0229137757485443, "learning_rate": 7.555745828811177e-05, "loss": 0.9047, "step": 2608 }, { "epoch": 1.1821477118260082, "grad_norm": 1.0504190633241952, "learning_rate": 7.555166073671015e-05, "loss": 0.8919, "step": 2609 }, { "epoch": 1.1826008155867693, "grad_norm": 1.1536295632646234, "learning_rate": 7.554585962757292e-05, "loss": 0.8869, "step": 2610 }, { "epoch": 1.1830539193475307, "grad_norm": 1.0500773057320834, "learning_rate": 7.554005496128062e-05, "loss": 0.8967, "step": 2611 }, { "epoch": 1.1835070231082918, "grad_norm": 1.2241242202714555, "learning_rate": 7.553424673841411e-05, "loss": 0.8956, "step": 2612 }, { "epoch": 1.183960126869053, "grad_norm": 0.964010299556886, "learning_rate": 7.552843495955465e-05, "loss": 0.9037, "step": 2613 }, { "epoch": 1.1844132306298143, "grad_norm": 1.1799620590880335, "learning_rate": 7.552261962528383e-05, "loss": 0.8706, "step": 2614 }, { "epoch": 1.1848663343905754, "grad_norm": 0.9511683425325416, "learning_rate": 7.551680073618362e-05, "loss": 0.9052, "step": 2615 }, { "epoch": 1.1853194381513368, "grad_norm": 0.9784652867120691, "learning_rate": 7.551097829283631e-05, "loss": 0.9103, "step": 2616 }, { "epoch": 1.185772541912098, "grad_norm": 1.245085693705854, "learning_rate": 7.550515229582458e-05, "loss": 0.9183, "step": 2617 }, { "epoch": 1.186225645672859, "grad_norm": 1.097430764077981, "learning_rate": 7.549932274573144e-05, "loss": 0.9036, "step": 2618 }, { "epoch": 1.1866787494336204, "grad_norm": 1.1150973169041944, "learning_rate": 7.549348964314027e-05, "loss": 0.9089, "step": 2619 }, { "epoch": 1.1871318531943815, "grad_norm": 1.0076949611435342, "learning_rate": 7.548765298863482e-05, "loss": 0.8828, "step": 2620 }, { "epoch": 1.1875849569551427, "grad_norm": 1.1540709982687007, "learning_rate": 7.548181278279914e-05, "loss": 0.9113, "step": 2621 }, { "epoch": 1.188038060715904, "grad_norm": 1.0727345238766888, "learning_rate": 7.54759690262177e-05, "loss": 0.8824, "step": 2622 }, { "epoch": 1.1884911644766651, "grad_norm": 1.009485692550353, "learning_rate": 7.54701217194753e-05, "loss": 0.9045, "step": 2623 }, { "epoch": 1.1889442682374263, "grad_norm": 1.1560188997040328, "learning_rate": 7.546427086315707e-05, "loss": 0.8833, "step": 2624 }, { "epoch": 1.1893973719981876, "grad_norm": 0.9017316361891702, "learning_rate": 7.545841645784853e-05, "loss": 0.8923, "step": 2625 }, { "epoch": 1.1898504757589488, "grad_norm": 0.8555932268310321, "learning_rate": 7.545255850413554e-05, "loss": 0.9081, "step": 2626 }, { "epoch": 1.19030357951971, "grad_norm": 0.9329124423323369, "learning_rate": 7.544669700260434e-05, "loss": 0.9114, "step": 2627 }, { "epoch": 1.1907566832804712, "grad_norm": 1.244402429147307, "learning_rate": 7.54408319538415e-05, "loss": 0.907, "step": 2628 }, { "epoch": 1.1912097870412324, "grad_norm": 1.041939749318294, "learning_rate": 7.543496335843392e-05, "loss": 0.9021, "step": 2629 }, { "epoch": 1.1916628908019937, "grad_norm": 1.0356774730115184, "learning_rate": 7.54290912169689e-05, "loss": 0.894, "step": 2630 }, { "epoch": 1.1921159945627549, "grad_norm": 1.0663838377647523, "learning_rate": 7.542321553003409e-05, "loss": 0.8969, "step": 2631 }, { "epoch": 1.1925690983235162, "grad_norm": 1.2393622575714915, "learning_rate": 7.541733629821745e-05, "loss": 0.932, "step": 2632 }, { "epoch": 1.1930222020842773, "grad_norm": 0.9545852754302522, "learning_rate": 7.541145352210737e-05, "loss": 0.8765, "step": 2633 }, { "epoch": 1.1934753058450385, "grad_norm": 0.961343390126277, "learning_rate": 7.540556720229253e-05, "loss": 0.896, "step": 2634 }, { "epoch": 1.1939284096057998, "grad_norm": 1.025392062878274, "learning_rate": 7.539967733936201e-05, "loss": 0.8917, "step": 2635 }, { "epoch": 1.194381513366561, "grad_norm": 1.215275626372242, "learning_rate": 7.539378393390518e-05, "loss": 0.9008, "step": 2636 }, { "epoch": 1.194834617127322, "grad_norm": 1.0251670290229935, "learning_rate": 7.538788698651183e-05, "loss": 0.9162, "step": 2637 }, { "epoch": 1.1952877208880834, "grad_norm": 1.157484175323903, "learning_rate": 7.538198649777208e-05, "loss": 0.9025, "step": 2638 }, { "epoch": 1.1957408246488446, "grad_norm": 0.914576322574953, "learning_rate": 7.537608246827642e-05, "loss": 0.898, "step": 2639 }, { "epoch": 1.1961939284096057, "grad_norm": 1.1162012908435799, "learning_rate": 7.537017489861566e-05, "loss": 0.8919, "step": 2640 }, { "epoch": 1.196647032170367, "grad_norm": 1.0321516978452137, "learning_rate": 7.536426378938098e-05, "loss": 0.8974, "step": 2641 }, { "epoch": 1.1971001359311282, "grad_norm": 1.0578491645657844, "learning_rate": 7.535834914116396e-05, "loss": 0.8916, "step": 2642 }, { "epoch": 1.1975532396918895, "grad_norm": 1.2634518677218085, "learning_rate": 7.535243095455643e-05, "loss": 0.8898, "step": 2643 }, { "epoch": 1.1980063434526507, "grad_norm": 0.8377510502720152, "learning_rate": 7.534650923015066e-05, "loss": 0.8674, "step": 2644 }, { "epoch": 1.1984594472134118, "grad_norm": 0.9205901271081269, "learning_rate": 7.534058396853928e-05, "loss": 0.9067, "step": 2645 }, { "epoch": 1.1989125509741732, "grad_norm": 0.9260454215538035, "learning_rate": 7.533465517031521e-05, "loss": 0.8972, "step": 2646 }, { "epoch": 1.1993656547349343, "grad_norm": 1.0856258796439837, "learning_rate": 7.532872283607177e-05, "loss": 0.904, "step": 2647 }, { "epoch": 1.1998187584956956, "grad_norm": 1.3252329194419734, "learning_rate": 7.532278696640263e-05, "loss": 0.9176, "step": 2648 }, { "epoch": 1.2002718622564568, "grad_norm": 0.8334466059881674, "learning_rate": 7.531684756190179e-05, "loss": 0.8909, "step": 2649 }, { "epoch": 1.200724966017218, "grad_norm": 0.898449907790359, "learning_rate": 7.531090462316363e-05, "loss": 0.8845, "step": 2650 }, { "epoch": 1.2011780697779793, "grad_norm": 1.0594474076744054, "learning_rate": 7.530495815078287e-05, "loss": 0.8949, "step": 2651 }, { "epoch": 1.2016311735387404, "grad_norm": 1.060599976817127, "learning_rate": 7.52990081453546e-05, "loss": 0.904, "step": 2652 }, { "epoch": 1.2020842772995015, "grad_norm": 0.9828867378034244, "learning_rate": 7.529305460747422e-05, "loss": 0.8845, "step": 2653 }, { "epoch": 1.2025373810602629, "grad_norm": 1.3032321312924475, "learning_rate": 7.528709753773754e-05, "loss": 0.9334, "step": 2654 }, { "epoch": 1.202990484821024, "grad_norm": 0.8354572095966699, "learning_rate": 7.528113693674069e-05, "loss": 0.91, "step": 2655 }, { "epoch": 1.2034435885817851, "grad_norm": 1.1054774195593418, "learning_rate": 7.527517280508017e-05, "loss": 0.8966, "step": 2656 }, { "epoch": 1.2038966923425465, "grad_norm": 1.2040498781513145, "learning_rate": 7.52692051433528e-05, "loss": 0.8996, "step": 2657 }, { "epoch": 1.2043497961033076, "grad_norm": 0.8492742273686542, "learning_rate": 7.52632339521558e-05, "loss": 0.9003, "step": 2658 }, { "epoch": 1.204802899864069, "grad_norm": 1.095815326492595, "learning_rate": 7.525725923208673e-05, "loss": 0.9053, "step": 2659 }, { "epoch": 1.20525600362483, "grad_norm": 1.3468651478710427, "learning_rate": 7.525128098374345e-05, "loss": 0.9179, "step": 2660 }, { "epoch": 1.2057091073855912, "grad_norm": 0.7148425543244216, "learning_rate": 7.524529920772426e-05, "loss": 0.8725, "step": 2661 }, { "epoch": 1.2061622111463526, "grad_norm": 0.7360519667803848, "learning_rate": 7.523931390462774e-05, "loss": 0.873, "step": 2662 }, { "epoch": 1.2066153149071137, "grad_norm": 0.8262564055377528, "learning_rate": 7.52333250750529e-05, "loss": 0.9481, "step": 2663 }, { "epoch": 1.207068418667875, "grad_norm": 0.7698788008394439, "learning_rate": 7.5227332719599e-05, "loss": 0.888, "step": 2664 }, { "epoch": 1.2075215224286362, "grad_norm": 1.4276808629814008, "learning_rate": 7.522133683886574e-05, "loss": 0.9471, "step": 2665 }, { "epoch": 1.2079746261893973, "grad_norm": 0.7332493684131751, "learning_rate": 7.52153374334531e-05, "loss": 0.9167, "step": 2666 }, { "epoch": 1.2084277299501587, "grad_norm": 0.9912571614057348, "learning_rate": 7.520933450396154e-05, "loss": 0.8919, "step": 2667 }, { "epoch": 1.2088808337109198, "grad_norm": 0.9575658061278504, "learning_rate": 7.52033280509917e-05, "loss": 0.8893, "step": 2668 }, { "epoch": 1.209333937471681, "grad_norm": 0.9792520247983353, "learning_rate": 7.51973180751447e-05, "loss": 0.892, "step": 2669 }, { "epoch": 1.2097870412324423, "grad_norm": 1.2129490106719516, "learning_rate": 7.519130457702197e-05, "loss": 0.9174, "step": 2670 }, { "epoch": 1.2102401449932034, "grad_norm": 0.7587932553482725, "learning_rate": 7.518528755722528e-05, "loss": 0.8543, "step": 2671 }, { "epoch": 1.2106932487539646, "grad_norm": 1.0738476479390238, "learning_rate": 7.517926701635679e-05, "loss": 0.8972, "step": 2672 }, { "epoch": 1.211146352514726, "grad_norm": 1.2534326745419153, "learning_rate": 7.517324295501897e-05, "loss": 0.9072, "step": 2673 }, { "epoch": 1.211599456275487, "grad_norm": 1.0360649412846143, "learning_rate": 7.516721537381467e-05, "loss": 0.9053, "step": 2674 }, { "epoch": 1.2120525600362484, "grad_norm": 1.258471368907704, "learning_rate": 7.516118427334708e-05, "loss": 0.9052, "step": 2675 }, { "epoch": 1.2125056637970095, "grad_norm": 0.948364627115508, "learning_rate": 7.515514965421976e-05, "loss": 0.9112, "step": 2676 }, { "epoch": 1.2129587675577707, "grad_norm": 1.116992647076451, "learning_rate": 7.514911151703657e-05, "loss": 0.8935, "step": 2677 }, { "epoch": 1.213411871318532, "grad_norm": 0.6951435853682353, "learning_rate": 7.51430698624018e-05, "loss": 0.9096, "step": 2678 }, { "epoch": 1.2138649750792931, "grad_norm": 1.1136748235791167, "learning_rate": 7.513702469092004e-05, "loss": 0.9015, "step": 2679 }, { "epoch": 1.2143180788400545, "grad_norm": 0.8112043656175747, "learning_rate": 7.513097600319624e-05, "loss": 0.8924, "step": 2680 }, { "epoch": 1.2147711826008156, "grad_norm": 0.5499568285526372, "learning_rate": 7.512492379983571e-05, "loss": 0.9013, "step": 2681 }, { "epoch": 1.2152242863615768, "grad_norm": 0.8635699989842661, "learning_rate": 7.51188680814441e-05, "loss": 0.8952, "step": 2682 }, { "epoch": 1.215677390122338, "grad_norm": 1.1353007079255393, "learning_rate": 7.511280884862742e-05, "loss": 0.8896, "step": 2683 }, { "epoch": 1.2161304938830992, "grad_norm": 1.2672451411574739, "learning_rate": 7.510674610199203e-05, "loss": 0.887, "step": 2684 }, { "epoch": 1.2165835976438604, "grad_norm": 1.0784291796679393, "learning_rate": 7.510067984214467e-05, "loss": 0.9115, "step": 2685 }, { "epoch": 1.2170367014046217, "grad_norm": 0.9510986755134807, "learning_rate": 7.509461006969238e-05, "loss": 0.9099, "step": 2686 }, { "epoch": 1.2174898051653829, "grad_norm": 1.177688351134757, "learning_rate": 7.508853678524257e-05, "loss": 0.9202, "step": 2687 }, { "epoch": 1.217942908926144, "grad_norm": 0.9834568488833887, "learning_rate": 7.508245998940302e-05, "loss": 0.8845, "step": 2688 }, { "epoch": 1.2183960126869053, "grad_norm": 1.2259121337614265, "learning_rate": 7.507637968278185e-05, "loss": 0.8981, "step": 2689 }, { "epoch": 1.2188491164476665, "grad_norm": 0.8283571597399331, "learning_rate": 7.507029586598752e-05, "loss": 0.8848, "step": 2690 }, { "epoch": 1.2193022202084278, "grad_norm": 1.1599115552338077, "learning_rate": 7.506420853962887e-05, "loss": 0.9002, "step": 2691 }, { "epoch": 1.219755323969189, "grad_norm": 1.3130781105277742, "learning_rate": 7.505811770431504e-05, "loss": 0.8914, "step": 2692 }, { "epoch": 1.22020842772995, "grad_norm": 0.900937896747201, "learning_rate": 7.505202336065559e-05, "loss": 0.8794, "step": 2693 }, { "epoch": 1.2206615314907114, "grad_norm": 0.8907045873496086, "learning_rate": 7.504592550926037e-05, "loss": 0.9157, "step": 2694 }, { "epoch": 1.2211146352514726, "grad_norm": 0.9807938945506497, "learning_rate": 7.503982415073961e-05, "loss": 0.9082, "step": 2695 }, { "epoch": 1.221567739012234, "grad_norm": 1.2985063369398138, "learning_rate": 7.50337192857039e-05, "loss": 0.8865, "step": 2696 }, { "epoch": 1.222020842772995, "grad_norm": 0.7991796387600667, "learning_rate": 7.502761091476416e-05, "loss": 0.9008, "step": 2697 }, { "epoch": 1.2224739465337562, "grad_norm": 0.9703395295022555, "learning_rate": 7.502149903853166e-05, "loss": 0.9296, "step": 2698 }, { "epoch": 1.2229270502945175, "grad_norm": 1.1889324297914257, "learning_rate": 7.501538365761804e-05, "loss": 0.9, "step": 2699 }, { "epoch": 1.2233801540552787, "grad_norm": 1.2871061274414, "learning_rate": 7.500926477263527e-05, "loss": 0.919, "step": 2700 }, { "epoch": 1.2238332578160398, "grad_norm": 0.7589083420267362, "learning_rate": 7.50031423841957e-05, "loss": 0.9097, "step": 2701 }, { "epoch": 1.2242863615768012, "grad_norm": 1.2778798090212538, "learning_rate": 7.499701649291199e-05, "loss": 0.9172, "step": 2702 }, { "epoch": 1.2247394653375623, "grad_norm": 0.7589904921983173, "learning_rate": 7.49908870993972e-05, "loss": 0.9123, "step": 2703 }, { "epoch": 1.2251925690983234, "grad_norm": 0.785619635368983, "learning_rate": 7.498475420426467e-05, "loss": 0.9186, "step": 2704 }, { "epoch": 1.2256456728590848, "grad_norm": 0.9595196451068231, "learning_rate": 7.497861780812816e-05, "loss": 0.8738, "step": 2705 }, { "epoch": 1.226098776619846, "grad_norm": 0.9281939415934357, "learning_rate": 7.497247791160174e-05, "loss": 0.8996, "step": 2706 }, { "epoch": 1.2265518803806073, "grad_norm": 1.3030781120031165, "learning_rate": 7.496633451529986e-05, "loss": 0.9079, "step": 2707 }, { "epoch": 1.2270049841413684, "grad_norm": 1.3348807286950177, "learning_rate": 7.496018761983731e-05, "loss": 0.9125, "step": 2708 }, { "epoch": 1.2274580879021295, "grad_norm": 0.6446126093459635, "learning_rate": 7.49540372258292e-05, "loss": 0.9085, "step": 2709 }, { "epoch": 1.2279111916628909, "grad_norm": 1.119124182117548, "learning_rate": 7.494788333389103e-05, "loss": 0.8937, "step": 2710 }, { "epoch": 1.228364295423652, "grad_norm": 1.736021332537093, "learning_rate": 7.494172594463863e-05, "loss": 0.9069, "step": 2711 }, { "epoch": 1.2288173991844134, "grad_norm": 0.6441501741752539, "learning_rate": 7.493556505868817e-05, "loss": 0.9198, "step": 2712 }, { "epoch": 1.2292705029451745, "grad_norm": 1.7783539861001059, "learning_rate": 7.492940067665619e-05, "loss": 0.9102, "step": 2713 }, { "epoch": 1.2297236067059356, "grad_norm": 0.9213258118939345, "learning_rate": 7.49232327991596e-05, "loss": 0.9046, "step": 2714 }, { "epoch": 1.230176710466697, "grad_norm": 1.5530932851766912, "learning_rate": 7.491706142681561e-05, "loss": 0.9054, "step": 2715 }, { "epoch": 1.230629814227458, "grad_norm": 0.9505812087788601, "learning_rate": 7.491088656024179e-05, "loss": 0.9135, "step": 2716 }, { "epoch": 1.2310829179882192, "grad_norm": 1.8539756071871878, "learning_rate": 7.49047082000561e-05, "loss": 0.8876, "step": 2717 }, { "epoch": 1.2315360217489806, "grad_norm": 1.592240713083235, "learning_rate": 7.48985263468768e-05, "loss": 0.9202, "step": 2718 }, { "epoch": 1.2319891255097417, "grad_norm": 1.2231912912405551, "learning_rate": 7.489234100132254e-05, "loss": 0.9113, "step": 2719 }, { "epoch": 1.2324422292705028, "grad_norm": 1.444768619341265, "learning_rate": 7.488615216401231e-05, "loss": 0.9076, "step": 2720 }, { "epoch": 1.2328953330312642, "grad_norm": 0.7341874210140708, "learning_rate": 7.487995983556541e-05, "loss": 0.8834, "step": 2721 }, { "epoch": 1.2333484367920253, "grad_norm": 1.3284334634456862, "learning_rate": 7.487376401660154e-05, "loss": 0.9126, "step": 2722 }, { "epoch": 1.2338015405527867, "grad_norm": 1.1607167714347377, "learning_rate": 7.486756470774073e-05, "loss": 0.8936, "step": 2723 }, { "epoch": 1.2342546443135478, "grad_norm": 0.9954895012211927, "learning_rate": 7.486136190960336e-05, "loss": 0.9193, "step": 2724 }, { "epoch": 1.234707748074309, "grad_norm": 0.939778134764094, "learning_rate": 7.485515562281014e-05, "loss": 0.9017, "step": 2725 }, { "epoch": 1.2351608518350703, "grad_norm": 0.9918400385792572, "learning_rate": 7.484894584798218e-05, "loss": 0.9145, "step": 2726 }, { "epoch": 1.2356139555958314, "grad_norm": 0.9092555925353281, "learning_rate": 7.484273258574088e-05, "loss": 0.9009, "step": 2727 }, { "epoch": 1.2360670593565928, "grad_norm": 1.1893822444588231, "learning_rate": 7.483651583670801e-05, "loss": 0.8969, "step": 2728 }, { "epoch": 1.236520163117354, "grad_norm": 0.9230721941754281, "learning_rate": 7.483029560150572e-05, "loss": 0.8909, "step": 2729 }, { "epoch": 1.236973266878115, "grad_norm": 1.261213703153279, "learning_rate": 7.482407188075646e-05, "loss": 0.9181, "step": 2730 }, { "epoch": 1.2374263706388764, "grad_norm": 1.1265916864143506, "learning_rate": 7.481784467508308e-05, "loss": 0.8942, "step": 2731 }, { "epoch": 1.2378794743996375, "grad_norm": 0.9890460877270388, "learning_rate": 7.481161398510873e-05, "loss": 0.8836, "step": 2732 }, { "epoch": 1.2383325781603987, "grad_norm": 0.9294555742745886, "learning_rate": 7.480537981145692e-05, "loss": 0.896, "step": 2733 }, { "epoch": 1.23878568192116, "grad_norm": 0.987980367586934, "learning_rate": 7.479914215475153e-05, "loss": 0.9039, "step": 2734 }, { "epoch": 1.2392387856819211, "grad_norm": 1.3604602941340913, "learning_rate": 7.479290101561678e-05, "loss": 0.9018, "step": 2735 }, { "epoch": 1.2396918894426823, "grad_norm": 0.896861855422012, "learning_rate": 7.478665639467722e-05, "loss": 0.8842, "step": 2736 }, { "epoch": 1.2401449932034436, "grad_norm": 0.9721347563429887, "learning_rate": 7.478040829255779e-05, "loss": 0.8933, "step": 2737 }, { "epoch": 1.2405980969642048, "grad_norm": 0.8777693192110709, "learning_rate": 7.477415670988373e-05, "loss": 0.8637, "step": 2738 }, { "epoch": 1.241051200724966, "grad_norm": 0.7352893704254828, "learning_rate": 7.476790164728065e-05, "loss": 0.8952, "step": 2739 }, { "epoch": 1.2415043044857272, "grad_norm": 0.8305861346044877, "learning_rate": 7.476164310537453e-05, "loss": 0.9221, "step": 2740 }, { "epoch": 1.2419574082464884, "grad_norm": 0.9546090156645911, "learning_rate": 7.475538108479165e-05, "loss": 0.9014, "step": 2741 }, { "epoch": 1.2424105120072497, "grad_norm": 1.314599005851199, "learning_rate": 7.474911558615869e-05, "loss": 0.8943, "step": 2742 }, { "epoch": 1.2428636157680109, "grad_norm": 0.7643453323367481, "learning_rate": 7.474284661010262e-05, "loss": 0.9027, "step": 2743 }, { "epoch": 1.2433167195287722, "grad_norm": 0.9991487975499138, "learning_rate": 7.473657415725083e-05, "loss": 0.8891, "step": 2744 }, { "epoch": 1.2437698232895333, "grad_norm": 1.372412728614557, "learning_rate": 7.473029822823098e-05, "loss": 0.8918, "step": 2745 }, { "epoch": 1.2442229270502945, "grad_norm": 0.7054051108474844, "learning_rate": 7.472401882367114e-05, "loss": 0.9023, "step": 2746 }, { "epoch": 1.2446760308110558, "grad_norm": 0.8893049890897539, "learning_rate": 7.47177359441997e-05, "loss": 0.8963, "step": 2747 }, { "epoch": 1.245129134571817, "grad_norm": 1.2746509661636305, "learning_rate": 7.47114495904454e-05, "loss": 0.8995, "step": 2748 }, { "epoch": 1.245582238332578, "grad_norm": 0.7133832836905928, "learning_rate": 7.470515976303732e-05, "loss": 0.8925, "step": 2749 }, { "epoch": 1.2460353420933394, "grad_norm": 0.9455986283564668, "learning_rate": 7.469886646260493e-05, "loss": 0.8964, "step": 2750 }, { "epoch": 1.2464884458541006, "grad_norm": 1.3028266868460572, "learning_rate": 7.469256968977797e-05, "loss": 0.8857, "step": 2751 }, { "epoch": 1.2469415496148617, "grad_norm": 0.955069970858139, "learning_rate": 7.468626944518662e-05, "loss": 0.886, "step": 2752 }, { "epoch": 1.247394653375623, "grad_norm": 1.6331230712974583, "learning_rate": 7.467996572946132e-05, "loss": 0.8888, "step": 2753 }, { "epoch": 1.2478477571363842, "grad_norm": 0.6564622839119899, "learning_rate": 7.467365854323293e-05, "loss": 0.8906, "step": 2754 }, { "epoch": 1.2483008608971455, "grad_norm": 1.8115673448130487, "learning_rate": 7.466734788713259e-05, "loss": 0.8986, "step": 2755 }, { "epoch": 1.2487539646579067, "grad_norm": 1.0222768229758057, "learning_rate": 7.466103376179183e-05, "loss": 0.8992, "step": 2756 }, { "epoch": 1.2492070684186678, "grad_norm": 1.7856547297479446, "learning_rate": 7.465471616784254e-05, "loss": 0.8874, "step": 2757 }, { "epoch": 1.2496601721794292, "grad_norm": 1.630128897199766, "learning_rate": 7.464839510591693e-05, "loss": 0.9184, "step": 2758 }, { "epoch": 1.2501132759401903, "grad_norm": 1.263693353500577, "learning_rate": 7.464207057664755e-05, "loss": 0.8973, "step": 2759 }, { "epoch": 1.2505663797009516, "grad_norm": 1.2575111847553153, "learning_rate": 7.46357425806673e-05, "loss": 0.9084, "step": 2760 }, { "epoch": 1.2510194834617128, "grad_norm": 1.0729263084109408, "learning_rate": 7.462941111860949e-05, "loss": 0.8874, "step": 2761 }, { "epoch": 1.251472587222474, "grad_norm": 1.3927867493621031, "learning_rate": 7.462307619110766e-05, "loss": 0.9168, "step": 2762 }, { "epoch": 1.251925690983235, "grad_norm": 0.9402415849749838, "learning_rate": 7.461673779879579e-05, "loss": 0.8856, "step": 2763 }, { "epoch": 1.2523787947439964, "grad_norm": 1.6254749347982396, "learning_rate": 7.461039594230817e-05, "loss": 0.8923, "step": 2764 }, { "epoch": 1.2528318985047575, "grad_norm": 1.0685564765449251, "learning_rate": 7.460405062227947e-05, "loss": 0.8944, "step": 2765 }, { "epoch": 1.2532850022655189, "grad_norm": 1.5982682442280496, "learning_rate": 7.459770183934465e-05, "loss": 0.9213, "step": 2766 }, { "epoch": 1.25373810602628, "grad_norm": 1.3999754961402473, "learning_rate": 7.459134959413906e-05, "loss": 0.9054, "step": 2767 }, { "epoch": 1.2541912097870411, "grad_norm": 1.3039624789049613, "learning_rate": 7.458499388729837e-05, "loss": 0.8902, "step": 2768 }, { "epoch": 1.2546443135478025, "grad_norm": 1.2464540094928533, "learning_rate": 7.457863471945864e-05, "loss": 0.8694, "step": 2769 }, { "epoch": 1.2550974173085636, "grad_norm": 1.150895214697517, "learning_rate": 7.45722720912562e-05, "loss": 0.9117, "step": 2770 }, { "epoch": 1.255550521069325, "grad_norm": 1.06457185505176, "learning_rate": 7.456590600332782e-05, "loss": 0.9116, "step": 2771 }, { "epoch": 1.256003624830086, "grad_norm": 1.414189146946511, "learning_rate": 7.455953645631054e-05, "loss": 0.9104, "step": 2772 }, { "epoch": 1.2564567285908472, "grad_norm": 0.9263322762975202, "learning_rate": 7.455316345084179e-05, "loss": 0.9009, "step": 2773 }, { "epoch": 1.2569098323516086, "grad_norm": 1.4079787814023081, "learning_rate": 7.454678698755932e-05, "loss": 0.9183, "step": 2774 }, { "epoch": 1.2573629361123697, "grad_norm": 1.0848268118277054, "learning_rate": 7.454040706710125e-05, "loss": 0.9106, "step": 2775 }, { "epoch": 1.257816039873131, "grad_norm": 1.1952197487931095, "learning_rate": 7.453402369010602e-05, "loss": 0.8779, "step": 2776 }, { "epoch": 1.2582691436338922, "grad_norm": 1.0684119924926216, "learning_rate": 7.452763685721243e-05, "loss": 0.9211, "step": 2777 }, { "epoch": 1.2587222473946533, "grad_norm": 1.2202010328990744, "learning_rate": 7.452124656905963e-05, "loss": 0.9105, "step": 2778 }, { "epoch": 1.2591753511554145, "grad_norm": 0.9038399786972255, "learning_rate": 7.451485282628713e-05, "loss": 0.9088, "step": 2779 }, { "epoch": 1.2596284549161758, "grad_norm": 1.187100691619804, "learning_rate": 7.450845562953472e-05, "loss": 0.887, "step": 2780 }, { "epoch": 1.260081558676937, "grad_norm": 0.928512011223461, "learning_rate": 7.450205497944262e-05, "loss": 0.8922, "step": 2781 }, { "epoch": 1.2605346624376983, "grad_norm": 1.1669860554358946, "learning_rate": 7.449565087665135e-05, "loss": 0.9001, "step": 2782 }, { "epoch": 1.2609877661984594, "grad_norm": 0.8608925550486591, "learning_rate": 7.448924332180178e-05, "loss": 0.8904, "step": 2783 }, { "epoch": 1.2614408699592206, "grad_norm": 1.1789154351598634, "learning_rate": 7.44828323155351e-05, "loss": 0.9019, "step": 2784 }, { "epoch": 1.261893973719982, "grad_norm": 0.8002528734441843, "learning_rate": 7.447641785849293e-05, "loss": 0.8954, "step": 2785 }, { "epoch": 1.262347077480743, "grad_norm": 1.09299621615279, "learning_rate": 7.446999995131715e-05, "loss": 0.9041, "step": 2786 }, { "epoch": 1.2628001812415044, "grad_norm": 1.085639803346195, "learning_rate": 7.446357859465e-05, "loss": 0.9037, "step": 2787 }, { "epoch": 1.2632532850022655, "grad_norm": 1.3764007486975436, "learning_rate": 7.445715378913412e-05, "loss": 0.9118, "step": 2788 }, { "epoch": 1.2637063887630267, "grad_norm": 0.8971631065684876, "learning_rate": 7.445072553541241e-05, "loss": 0.9089, "step": 2789 }, { "epoch": 1.264159492523788, "grad_norm": 1.1519094451163583, "learning_rate": 7.444429383412818e-05, "loss": 0.9023, "step": 2790 }, { "epoch": 1.2646125962845491, "grad_norm": 1.1474667073885245, "learning_rate": 7.443785868592508e-05, "loss": 0.894, "step": 2791 }, { "epoch": 1.2650657000453105, "grad_norm": 1.0120854254916067, "learning_rate": 7.443142009144707e-05, "loss": 0.9187, "step": 2792 }, { "epoch": 1.2655188038060716, "grad_norm": 1.650340769700587, "learning_rate": 7.442497805133848e-05, "loss": 0.8917, "step": 2793 }, { "epoch": 1.2659719075668328, "grad_norm": 0.6818300762866837, "learning_rate": 7.441853256624397e-05, "loss": 0.897, "step": 2794 }, { "epoch": 1.2664250113275939, "grad_norm": 1.3176350115276794, "learning_rate": 7.441208363680858e-05, "loss": 0.8737, "step": 2795 }, { "epoch": 1.2668781150883552, "grad_norm": 1.4387437403327685, "learning_rate": 7.440563126367765e-05, "loss": 0.9135, "step": 2796 }, { "epoch": 1.2673312188491164, "grad_norm": 0.7277330231195401, "learning_rate": 7.439917544749687e-05, "loss": 0.9069, "step": 2797 }, { "epoch": 1.2677843226098777, "grad_norm": 0.8556751325714931, "learning_rate": 7.439271618891232e-05, "loss": 0.9114, "step": 2798 }, { "epoch": 1.2682374263706389, "grad_norm": 1.296029754725755, "learning_rate": 7.438625348857037e-05, "loss": 0.9138, "step": 2799 }, { "epoch": 1.2686905301314, "grad_norm": 1.1452223710411087, "learning_rate": 7.437978734711776e-05, "loss": 0.9005, "step": 2800 }, { "epoch": 1.2691436338921613, "grad_norm": 1.6719960899389414, "learning_rate": 7.437331776520159e-05, "loss": 0.8871, "step": 2801 }, { "epoch": 1.2695967376529225, "grad_norm": 0.8257862967859431, "learning_rate": 7.436684474346928e-05, "loss": 0.9057, "step": 2802 }, { "epoch": 1.2700498414136838, "grad_norm": 1.934387609553433, "learning_rate": 7.436036828256857e-05, "loss": 0.9042, "step": 2803 }, { "epoch": 1.270502945174445, "grad_norm": 0.8547274149088285, "learning_rate": 7.43538883831476e-05, "loss": 0.906, "step": 2804 }, { "epoch": 1.270956048935206, "grad_norm": 2.6266724922064175, "learning_rate": 7.434740504585483e-05, "loss": 0.9277, "step": 2805 }, { "epoch": 1.2714091526959674, "grad_norm": 1.8496793852781783, "learning_rate": 7.434091827133906e-05, "loss": 0.9156, "step": 2806 }, { "epoch": 1.2718622564567286, "grad_norm": 2.30395241558666, "learning_rate": 7.433442806024943e-05, "loss": 0.9126, "step": 2807 }, { "epoch": 1.27231536021749, "grad_norm": 1.8281900877433663, "learning_rate": 7.432793441323544e-05, "loss": 0.9005, "step": 2808 }, { "epoch": 1.272768463978251, "grad_norm": 1.7510294576154821, "learning_rate": 7.432143733094691e-05, "loss": 0.91, "step": 2809 }, { "epoch": 1.2732215677390122, "grad_norm": 1.4248514444834415, "learning_rate": 7.431493681403403e-05, "loss": 0.8908, "step": 2810 }, { "epoch": 1.2736746714997733, "grad_norm": 1.7045334923815558, "learning_rate": 7.430843286314732e-05, "loss": 0.9011, "step": 2811 }, { "epoch": 1.2741277752605347, "grad_norm": 1.3202183520369086, "learning_rate": 7.430192547893763e-05, "loss": 0.903, "step": 2812 }, { "epoch": 1.2745808790212958, "grad_norm": 1.6538147794941738, "learning_rate": 7.429541466205619e-05, "loss": 0.9298, "step": 2813 }, { "epoch": 1.2750339827820572, "grad_norm": 1.3598476259565437, "learning_rate": 7.428890041315454e-05, "loss": 0.9117, "step": 2814 }, { "epoch": 1.2754870865428183, "grad_norm": 1.33948485761685, "learning_rate": 7.428238273288457e-05, "loss": 0.8917, "step": 2815 }, { "epoch": 1.2759401903035794, "grad_norm": 1.3952255866378915, "learning_rate": 7.427586162189853e-05, "loss": 0.9102, "step": 2816 }, { "epoch": 1.2763932940643408, "grad_norm": 1.1201308206460328, "learning_rate": 7.426933708084902e-05, "loss": 0.8881, "step": 2817 }, { "epoch": 1.276846397825102, "grad_norm": 1.3470674002183967, "learning_rate": 7.426280911038893e-05, "loss": 0.8997, "step": 2818 }, { "epoch": 1.2772995015858633, "grad_norm": 1.040552911328274, "learning_rate": 7.425627771117154e-05, "loss": 0.9017, "step": 2819 }, { "epoch": 1.2777526053466244, "grad_norm": 1.929775781505513, "learning_rate": 7.424974288385048e-05, "loss": 0.8894, "step": 2820 }, { "epoch": 1.2782057091073855, "grad_norm": 1.320677084938492, "learning_rate": 7.424320462907969e-05, "loss": 0.9275, "step": 2821 }, { "epoch": 1.2786588128681469, "grad_norm": 2.314765122178669, "learning_rate": 7.423666294751345e-05, "loss": 0.8995, "step": 2822 }, { "epoch": 1.279111916628908, "grad_norm": 2.0437339330716853, "learning_rate": 7.423011783980645e-05, "loss": 0.9054, "step": 2823 }, { "epoch": 1.2795650203896693, "grad_norm": 1.6819565966342593, "learning_rate": 7.422356930661362e-05, "loss": 0.9179, "step": 2824 }, { "epoch": 1.2800181241504305, "grad_norm": 1.4005839015278838, "learning_rate": 7.421701734859032e-05, "loss": 0.8966, "step": 2825 }, { "epoch": 1.2804712279111916, "grad_norm": 1.7708578402593242, "learning_rate": 7.421046196639222e-05, "loss": 0.8911, "step": 2826 }, { "epoch": 1.2809243316719527, "grad_norm": 1.2364486560211816, "learning_rate": 7.420390316067531e-05, "loss": 0.9017, "step": 2827 }, { "epoch": 1.281377435432714, "grad_norm": 1.8295403622493658, "learning_rate": 7.419734093209597e-05, "loss": 0.9074, "step": 2828 }, { "epoch": 1.2818305391934752, "grad_norm": 1.337792808163628, "learning_rate": 7.419077528131088e-05, "loss": 0.9011, "step": 2829 }, { "epoch": 1.2822836429542366, "grad_norm": 1.7622631273256562, "learning_rate": 7.418420620897708e-05, "loss": 0.8924, "step": 2830 }, { "epoch": 1.2827367467149977, "grad_norm": 1.5392603881482327, "learning_rate": 7.417763371575196e-05, "loss": 0.8975, "step": 2831 }, { "epoch": 1.2831898504757588, "grad_norm": 1.7644014313257077, "learning_rate": 7.417105780229323e-05, "loss": 0.8977, "step": 2832 }, { "epoch": 1.2836429542365202, "grad_norm": 1.671773352957547, "learning_rate": 7.416447846925897e-05, "loss": 0.9005, "step": 2833 }, { "epoch": 1.2840960579972813, "grad_norm": 1.413465810037402, "learning_rate": 7.415789571730759e-05, "loss": 0.9042, "step": 2834 }, { "epoch": 1.2845491617580427, "grad_norm": 1.3903686759111844, "learning_rate": 7.415130954709784e-05, "loss": 0.9252, "step": 2835 }, { "epoch": 1.2850022655188038, "grad_norm": 1.6638950174501392, "learning_rate": 7.41447199592888e-05, "loss": 0.899, "step": 2836 }, { "epoch": 1.285455369279565, "grad_norm": 1.2242768552543553, "learning_rate": 7.413812695453993e-05, "loss": 0.8917, "step": 2837 }, { "epoch": 1.2859084730403263, "grad_norm": 1.8201674323903123, "learning_rate": 7.413153053351097e-05, "loss": 0.9025, "step": 2838 }, { "epoch": 1.2863615768010874, "grad_norm": 1.4429804205531447, "learning_rate": 7.412493069686209e-05, "loss": 0.8752, "step": 2839 }, { "epoch": 1.2868146805618488, "grad_norm": 1.6500629929044897, "learning_rate": 7.41183274452537e-05, "loss": 0.9016, "step": 2840 }, { "epoch": 1.28726778432261, "grad_norm": 1.5612926125743445, "learning_rate": 7.411172077934663e-05, "loss": 0.8959, "step": 2841 }, { "epoch": 1.287720888083371, "grad_norm": 1.4189294519506617, "learning_rate": 7.410511069980201e-05, "loss": 0.8745, "step": 2842 }, { "epoch": 1.2881739918441322, "grad_norm": 1.3516055080399616, "learning_rate": 7.409849720728135e-05, "loss": 0.9002, "step": 2843 }, { "epoch": 1.2886270956048935, "grad_norm": 1.3259648246341615, "learning_rate": 7.409188030244645e-05, "loss": 0.8908, "step": 2844 }, { "epoch": 1.2890801993656547, "grad_norm": 0.9717814298136696, "learning_rate": 7.408525998595948e-05, "loss": 0.8985, "step": 2845 }, { "epoch": 1.289533303126416, "grad_norm": 1.8418602676951226, "learning_rate": 7.407863625848298e-05, "loss": 0.9144, "step": 2846 }, { "epoch": 1.2899864068871771, "grad_norm": 1.4177411094183694, "learning_rate": 7.407200912067978e-05, "loss": 0.8773, "step": 2847 }, { "epoch": 1.2904395106479383, "grad_norm": 1.735009818486593, "learning_rate": 7.406537857321306e-05, "loss": 0.9256, "step": 2848 }, { "epoch": 1.2908926144086996, "grad_norm": 1.6385777048207486, "learning_rate": 7.405874461674638e-05, "loss": 0.883, "step": 2849 }, { "epoch": 1.2913457181694608, "grad_norm": 1.1813510174826436, "learning_rate": 7.40521072519436e-05, "loss": 0.894, "step": 2850 }, { "epoch": 1.291798821930222, "grad_norm": 1.66010933594611, "learning_rate": 7.404546647946893e-05, "loss": 0.9032, "step": 2851 }, { "epoch": 1.2922519256909832, "grad_norm": 0.763743251187519, "learning_rate": 7.403882229998694e-05, "loss": 0.8911, "step": 2852 }, { "epoch": 1.2927050294517444, "grad_norm": 1.8718286940873308, "learning_rate": 7.403217471416254e-05, "loss": 0.8986, "step": 2853 }, { "epoch": 1.2931581332125057, "grad_norm": 0.8952219396039006, "learning_rate": 7.402552372266094e-05, "loss": 0.8932, "step": 2854 }, { "epoch": 1.2936112369732669, "grad_norm": 1.7857501608629882, "learning_rate": 7.401886932614774e-05, "loss": 0.8926, "step": 2855 }, { "epoch": 1.2940643407340282, "grad_norm": 1.2062385145472099, "learning_rate": 7.401221152528884e-05, "loss": 0.8796, "step": 2856 }, { "epoch": 1.2945174444947893, "grad_norm": 1.2601487242531113, "learning_rate": 7.400555032075054e-05, "loss": 0.8898, "step": 2857 }, { "epoch": 1.2949705482555505, "grad_norm": 1.5534080009542934, "learning_rate": 7.399888571319939e-05, "loss": 0.898, "step": 2858 }, { "epoch": 1.2954236520163116, "grad_norm": 0.969431755349281, "learning_rate": 7.399221770330239e-05, "loss": 0.8968, "step": 2859 }, { "epoch": 1.295876755777073, "grad_norm": 1.6844950106040792, "learning_rate": 7.398554629172676e-05, "loss": 0.8995, "step": 2860 }, { "epoch": 1.296329859537834, "grad_norm": 1.4424595515304304, "learning_rate": 7.397887147914017e-05, "loss": 0.8489, "step": 2861 }, { "epoch": 1.2967829632985954, "grad_norm": 1.3120253923396068, "learning_rate": 7.397219326621057e-05, "loss": 0.9084, "step": 2862 }, { "epoch": 1.2972360670593566, "grad_norm": 1.7352031638408605, "learning_rate": 7.396551165360627e-05, "loss": 0.9011, "step": 2863 }, { "epoch": 1.2976891708201177, "grad_norm": 1.1659399455082946, "learning_rate": 7.39588266419959e-05, "loss": 0.9057, "step": 2864 }, { "epoch": 1.298142274580879, "grad_norm": 1.932328318515867, "learning_rate": 7.395213823204845e-05, "loss": 0.8962, "step": 2865 }, { "epoch": 1.2985953783416402, "grad_norm": 1.5174842542934934, "learning_rate": 7.394544642443325e-05, "loss": 0.8952, "step": 2866 }, { "epoch": 1.2990484821024015, "grad_norm": 1.7032910464933917, "learning_rate": 7.393875121981997e-05, "loss": 0.8947, "step": 2867 }, { "epoch": 1.2995015858631627, "grad_norm": 1.1304664271222997, "learning_rate": 7.393205261887861e-05, "loss": 0.8905, "step": 2868 }, { "epoch": 1.2999546896239238, "grad_norm": 1.64895031219908, "learning_rate": 7.39253506222795e-05, "loss": 0.897, "step": 2869 }, { "epoch": 1.3004077933846852, "grad_norm": 1.019923128787382, "learning_rate": 7.391864523069333e-05, "loss": 0.8904, "step": 2870 }, { "epoch": 1.3008608971454463, "grad_norm": 1.741830156736573, "learning_rate": 7.391193644479113e-05, "loss": 0.9043, "step": 2871 }, { "epoch": 1.3013140009062076, "grad_norm": 1.2836709436312372, "learning_rate": 7.390522426524427e-05, "loss": 0.8966, "step": 2872 }, { "epoch": 1.3017671046669688, "grad_norm": 1.6379003592424446, "learning_rate": 7.389850869272442e-05, "loss": 0.9017, "step": 2873 }, { "epoch": 1.30222020842773, "grad_norm": 1.4312717619501043, "learning_rate": 7.389178972790365e-05, "loss": 0.9236, "step": 2874 }, { "epoch": 1.302673312188491, "grad_norm": 1.4367657975500048, "learning_rate": 7.388506737145436e-05, "loss": 0.9098, "step": 2875 }, { "epoch": 1.3031264159492524, "grad_norm": 1.3457087255782663, "learning_rate": 7.387834162404924e-05, "loss": 0.9003, "step": 2876 }, { "epoch": 1.3035795197100135, "grad_norm": 1.3020181040429502, "learning_rate": 7.387161248636136e-05, "loss": 0.9178, "step": 2877 }, { "epoch": 1.3040326234707749, "grad_norm": 1.2337402559033623, "learning_rate": 7.386487995906413e-05, "loss": 0.8635, "step": 2878 }, { "epoch": 1.304485727231536, "grad_norm": 1.0991276436961337, "learning_rate": 7.385814404283128e-05, "loss": 0.9089, "step": 2879 }, { "epoch": 1.3049388309922971, "grad_norm": 1.072487087898268, "learning_rate": 7.385140473833688e-05, "loss": 0.8967, "step": 2880 }, { "epoch": 1.3053919347530585, "grad_norm": 1.1590215547255454, "learning_rate": 7.384466204625536e-05, "loss": 0.905, "step": 2881 }, { "epoch": 1.3058450385138196, "grad_norm": 0.8539110175946345, "learning_rate": 7.383791596726146e-05, "loss": 0.8932, "step": 2882 }, { "epoch": 1.306298142274581, "grad_norm": 1.13268929723165, "learning_rate": 7.383116650203031e-05, "loss": 0.889, "step": 2883 }, { "epoch": 1.306751246035342, "grad_norm": 0.8920590409144281, "learning_rate": 7.382441365123733e-05, "loss": 0.8916, "step": 2884 }, { "epoch": 1.3072043497961032, "grad_norm": 1.0220253493133573, "learning_rate": 7.381765741555827e-05, "loss": 0.9003, "step": 2885 }, { "epoch": 1.3076574535568646, "grad_norm": 0.8431507003761101, "learning_rate": 7.381089779566924e-05, "loss": 0.9096, "step": 2886 }, { "epoch": 1.3081105573176257, "grad_norm": 0.9600688403802633, "learning_rate": 7.380413479224674e-05, "loss": 0.9066, "step": 2887 }, { "epoch": 1.308563661078387, "grad_norm": 0.8744776191211828, "learning_rate": 7.379736840596751e-05, "loss": 0.8992, "step": 2888 }, { "epoch": 1.3090167648391482, "grad_norm": 0.7609193796698226, "learning_rate": 7.37905986375087e-05, "loss": 0.9184, "step": 2889 }, { "epoch": 1.3094698685999093, "grad_norm": 1.0485864730338044, "learning_rate": 7.378382548754778e-05, "loss": 0.9016, "step": 2890 }, { "epoch": 1.3099229723606705, "grad_norm": 1.2865562078710469, "learning_rate": 7.377704895676253e-05, "loss": 0.8759, "step": 2891 }, { "epoch": 1.3103760761214318, "grad_norm": 0.7736576108673885, "learning_rate": 7.377026904583113e-05, "loss": 0.8847, "step": 2892 }, { "epoch": 1.310829179882193, "grad_norm": 0.748744539330596, "learning_rate": 7.376348575543203e-05, "loss": 0.9063, "step": 2893 }, { "epoch": 1.3112822836429543, "grad_norm": 0.563848347251493, "learning_rate": 7.375669908624406e-05, "loss": 0.8882, "step": 2894 }, { "epoch": 1.3117353874037154, "grad_norm": 0.9039841699368594, "learning_rate": 7.374990903894636e-05, "loss": 0.8972, "step": 2895 }, { "epoch": 1.3121884911644766, "grad_norm": 1.2259124133539625, "learning_rate": 7.374311561421846e-05, "loss": 0.8833, "step": 2896 }, { "epoch": 1.312641594925238, "grad_norm": 1.2740892059138489, "learning_rate": 7.373631881274018e-05, "loss": 0.8815, "step": 2897 }, { "epoch": 1.313094698685999, "grad_norm": 0.5969077302383474, "learning_rate": 7.372951863519169e-05, "loss": 0.8915, "step": 2898 }, { "epoch": 1.3135478024467604, "grad_norm": 1.0211031817808265, "learning_rate": 7.372271508225348e-05, "loss": 0.91, "step": 2899 }, { "epoch": 1.3140009062075215, "grad_norm": 1.7494358005749386, "learning_rate": 7.371590815460642e-05, "loss": 0.903, "step": 2900 }, { "epoch": 1.3144540099682827, "grad_norm": 0.7155341507028938, "learning_rate": 7.370909785293169e-05, "loss": 0.9131, "step": 2901 }, { "epoch": 1.314907113729044, "grad_norm": 2.0255538977345, "learning_rate": 7.370228417791081e-05, "loss": 0.8901, "step": 2902 }, { "epoch": 1.3153602174898051, "grad_norm": 1.0981836227467303, "learning_rate": 7.369546713022565e-05, "loss": 0.912, "step": 2903 }, { "epoch": 1.3158133212505665, "grad_norm": 2.3890274869626604, "learning_rate": 7.368864671055837e-05, "loss": 0.9, "step": 2904 }, { "epoch": 1.3162664250113276, "grad_norm": 2.084836608897643, "learning_rate": 7.368182291959155e-05, "loss": 0.9094, "step": 2905 }, { "epoch": 1.3167195287720888, "grad_norm": 1.56518750413508, "learning_rate": 7.367499575800806e-05, "loss": 0.9185, "step": 2906 }, { "epoch": 1.3171726325328499, "grad_norm": 1.5737938447937234, "learning_rate": 7.366816522649107e-05, "loss": 0.9066, "step": 2907 }, { "epoch": 1.3176257362936112, "grad_norm": 1.2823618960559549, "learning_rate": 7.366133132572415e-05, "loss": 0.9002, "step": 2908 }, { "epoch": 1.3180788400543724, "grad_norm": 1.3901770452030575, "learning_rate": 7.365449405639121e-05, "loss": 0.8891, "step": 2909 }, { "epoch": 1.3185319438151337, "grad_norm": 0.8838304952699748, "learning_rate": 7.364765341917644e-05, "loss": 0.9037, "step": 2910 }, { "epoch": 1.3189850475758949, "grad_norm": 1.1290367330686364, "learning_rate": 7.36408094147644e-05, "loss": 0.8921, "step": 2911 }, { "epoch": 1.319438151336656, "grad_norm": 1.49483828267372, "learning_rate": 7.363396204383998e-05, "loss": 0.888, "step": 2912 }, { "epoch": 1.3198912550974173, "grad_norm": 0.8816180721893832, "learning_rate": 7.362711130708844e-05, "loss": 0.8922, "step": 2913 }, { "epoch": 1.3203443588581785, "grad_norm": 1.4854507735592117, "learning_rate": 7.362025720519532e-05, "loss": 0.8942, "step": 2914 }, { "epoch": 1.3207974626189398, "grad_norm": 1.5132885805076377, "learning_rate": 7.361339973884654e-05, "loss": 0.904, "step": 2915 }, { "epoch": 1.321250566379701, "grad_norm": 0.9265115129186478, "learning_rate": 7.360653890872834e-05, "loss": 0.9059, "step": 2916 }, { "epoch": 1.321703670140462, "grad_norm": 2.218611589416932, "learning_rate": 7.359967471552729e-05, "loss": 0.8838, "step": 2917 }, { "epoch": 1.3221567739012234, "grad_norm": 1.4423714979481335, "learning_rate": 7.359280715993032e-05, "loss": 0.8985, "step": 2918 }, { "epoch": 1.3226098776619846, "grad_norm": 2.755420358552311, "learning_rate": 7.358593624262468e-05, "loss": 0.9256, "step": 2919 }, { "epoch": 1.323062981422746, "grad_norm": 2.707212538170887, "learning_rate": 7.357906196429795e-05, "loss": 0.9136, "step": 2920 }, { "epoch": 1.323516085183507, "grad_norm": 1.110094811227806, "learning_rate": 7.357218432563805e-05, "loss": 0.8972, "step": 2921 }, { "epoch": 1.3239691889442682, "grad_norm": 1.761958063239864, "learning_rate": 7.356530332733325e-05, "loss": 0.9133, "step": 2922 }, { "epoch": 1.3244222927050293, "grad_norm": 1.411002128832353, "learning_rate": 7.355841897007215e-05, "loss": 0.9004, "step": 2923 }, { "epoch": 1.3248753964657907, "grad_norm": 1.25058066837867, "learning_rate": 7.35515312545437e-05, "loss": 0.8851, "step": 2924 }, { "epoch": 1.3253285002265518, "grad_norm": 1.2105127901816781, "learning_rate": 7.354464018143712e-05, "loss": 0.8893, "step": 2925 }, { "epoch": 1.3257816039873132, "grad_norm": 1.2835718380090841, "learning_rate": 7.353774575144204e-05, "loss": 0.8933, "step": 2926 }, { "epoch": 1.3262347077480743, "grad_norm": 0.896121563182739, "learning_rate": 7.353084796524842e-05, "loss": 0.8999, "step": 2927 }, { "epoch": 1.3266878115088354, "grad_norm": 1.4078293766708925, "learning_rate": 7.352394682354653e-05, "loss": 0.886, "step": 2928 }, { "epoch": 1.3271409152695968, "grad_norm": 1.0084788388928256, "learning_rate": 7.351704232702697e-05, "loss": 0.893, "step": 2929 }, { "epoch": 1.327594019030358, "grad_norm": 1.427365011211372, "learning_rate": 7.351013447638068e-05, "loss": 0.8895, "step": 2930 }, { "epoch": 1.3280471227911192, "grad_norm": 1.059082183120484, "learning_rate": 7.350322327229897e-05, "loss": 0.8819, "step": 2931 }, { "epoch": 1.3285002265518804, "grad_norm": 1.3970847766546068, "learning_rate": 7.349630871547345e-05, "loss": 0.8924, "step": 2932 }, { "epoch": 1.3289533303126415, "grad_norm": 1.186939437022317, "learning_rate": 7.348939080659607e-05, "loss": 0.9012, "step": 2933 }, { "epoch": 1.3294064340734029, "grad_norm": 1.2126153889570066, "learning_rate": 7.348246954635912e-05, "loss": 0.9158, "step": 2934 }, { "epoch": 1.329859537834164, "grad_norm": 1.4818084713661128, "learning_rate": 7.347554493545525e-05, "loss": 0.8889, "step": 2935 }, { "epoch": 1.3303126415949253, "grad_norm": 0.9882104785850169, "learning_rate": 7.34686169745774e-05, "loss": 0.8914, "step": 2936 }, { "epoch": 1.3307657453556865, "grad_norm": 1.5885441829766171, "learning_rate": 7.346168566441884e-05, "loss": 0.9105, "step": 2937 }, { "epoch": 1.3312188491164476, "grad_norm": 1.2253777889744204, "learning_rate": 7.345475100567326e-05, "loss": 0.8839, "step": 2938 }, { "epoch": 1.3316719528772087, "grad_norm": 1.4627775965795589, "learning_rate": 7.344781299903458e-05, "loss": 0.9061, "step": 2939 }, { "epoch": 1.33212505663797, "grad_norm": 1.1555509526240433, "learning_rate": 7.344087164519714e-05, "loss": 0.8862, "step": 2940 }, { "epoch": 1.3325781603987312, "grad_norm": 1.316904708319868, "learning_rate": 7.343392694485555e-05, "loss": 0.8979, "step": 2941 }, { "epoch": 1.3330312641594926, "grad_norm": 0.9281588725905686, "learning_rate": 7.342697889870478e-05, "loss": 0.9054, "step": 2942 }, { "epoch": 1.3334843679202537, "grad_norm": 1.0764093078364707, "learning_rate": 7.342002750744016e-05, "loss": 0.8845, "step": 2943 }, { "epoch": 1.3339374716810148, "grad_norm": 1.140365089793188, "learning_rate": 7.341307277175731e-05, "loss": 0.8976, "step": 2944 }, { "epoch": 1.3343905754417762, "grad_norm": 0.8474296618097549, "learning_rate": 7.340611469235222e-05, "loss": 0.8651, "step": 2945 }, { "epoch": 1.3348436792025373, "grad_norm": 0.9205771374681293, "learning_rate": 7.339915326992119e-05, "loss": 0.8658, "step": 2946 }, { "epoch": 1.3352967829632987, "grad_norm": 1.1423494263770682, "learning_rate": 7.339218850516087e-05, "loss": 0.8986, "step": 2947 }, { "epoch": 1.3357498867240598, "grad_norm": 0.689534760344968, "learning_rate": 7.338522039876825e-05, "loss": 0.8943, "step": 2948 }, { "epoch": 1.336202990484821, "grad_norm": 1.1007819025654677, "learning_rate": 7.337824895144063e-05, "loss": 0.9074, "step": 2949 }, { "epoch": 1.3366560942455823, "grad_norm": 0.872327230606282, "learning_rate": 7.337127416387566e-05, "loss": 0.892, "step": 2950 }, { "epoch": 1.3371091980063434, "grad_norm": 0.5119740713041531, "learning_rate": 7.336429603677133e-05, "loss": 0.9076, "step": 2951 }, { "epoch": 1.3375623017671048, "grad_norm": 0.7637340953133165, "learning_rate": 7.335731457082594e-05, "loss": 0.8989, "step": 2952 }, { "epoch": 1.338015405527866, "grad_norm": 0.6671630725276808, "learning_rate": 7.335032976673818e-05, "loss": 0.9042, "step": 2953 }, { "epoch": 1.338468509288627, "grad_norm": 0.6513133566003382, "learning_rate": 7.334334162520698e-05, "loss": 0.8997, "step": 2954 }, { "epoch": 1.3389216130493884, "grad_norm": 0.6281756995744282, "learning_rate": 7.33363501469317e-05, "loss": 0.8898, "step": 2955 }, { "epoch": 1.3393747168101495, "grad_norm": 0.7144026153458901, "learning_rate": 7.3329355332612e-05, "loss": 0.8851, "step": 2956 }, { "epoch": 1.3398278205709109, "grad_norm": 0.8331605153723007, "learning_rate": 7.332235718294784e-05, "loss": 0.8977, "step": 2957 }, { "epoch": 1.340280924331672, "grad_norm": 1.053051171706433, "learning_rate": 7.331535569863953e-05, "loss": 0.8831, "step": 2958 }, { "epoch": 1.3407340280924331, "grad_norm": 1.240280916993373, "learning_rate": 7.330835088038776e-05, "loss": 0.8698, "step": 2959 }, { "epoch": 1.3411871318531943, "grad_norm": 0.7208867924956088, "learning_rate": 7.33013427288935e-05, "loss": 0.9035, "step": 2960 }, { "epoch": 1.3416402356139556, "grad_norm": 0.5556551522787903, "learning_rate": 7.329433124485807e-05, "loss": 0.9033, "step": 2961 }, { "epoch": 1.3420933393747168, "grad_norm": 0.4702639873333179, "learning_rate": 7.328731642898315e-05, "loss": 0.8913, "step": 2962 }, { "epoch": 1.342546443135478, "grad_norm": 0.5342738615026693, "learning_rate": 7.328029828197067e-05, "loss": 0.89, "step": 2963 }, { "epoch": 1.3429995468962392, "grad_norm": 0.7929768946824208, "learning_rate": 7.327327680452302e-05, "loss": 0.8999, "step": 2964 }, { "epoch": 1.3434526506570004, "grad_norm": 1.2416303646203575, "learning_rate": 7.32662519973428e-05, "loss": 0.9058, "step": 2965 }, { "epoch": 1.3439057544177617, "grad_norm": 1.0496917740814937, "learning_rate": 7.325922386113305e-05, "loss": 0.9169, "step": 2966 }, { "epoch": 1.3443588581785229, "grad_norm": 0.9383716098978258, "learning_rate": 7.325219239659706e-05, "loss": 0.8804, "step": 2967 }, { "epoch": 1.3448119619392842, "grad_norm": 0.8781434189045074, "learning_rate": 7.324515760443849e-05, "loss": 0.9043, "step": 2968 }, { "epoch": 1.3452650657000453, "grad_norm": 0.8541771845285312, "learning_rate": 7.323811948536131e-05, "loss": 0.8802, "step": 2969 }, { "epoch": 1.3457181694608065, "grad_norm": 0.9724584919983391, "learning_rate": 7.323107804006988e-05, "loss": 0.906, "step": 2970 }, { "epoch": 1.3461712732215678, "grad_norm": 1.4586288887478989, "learning_rate": 7.322403326926883e-05, "loss": 0.8897, "step": 2971 }, { "epoch": 1.346624376982329, "grad_norm": 0.5576911844039866, "learning_rate": 7.321698517366313e-05, "loss": 0.8906, "step": 2972 }, { "epoch": 1.3470774807430903, "grad_norm": 0.9101729146520612, "learning_rate": 7.320993375395813e-05, "loss": 0.8959, "step": 2973 }, { "epoch": 1.3475305845038514, "grad_norm": 1.8191271492603074, "learning_rate": 7.320287901085946e-05, "loss": 0.8927, "step": 2974 }, { "epoch": 1.3479836882646126, "grad_norm": 0.7152812301123123, "learning_rate": 7.319582094507312e-05, "loss": 0.8983, "step": 2975 }, { "epoch": 1.3484367920253737, "grad_norm": 2.326221928706702, "learning_rate": 7.31887595573054e-05, "loss": 0.9194, "step": 2976 }, { "epoch": 1.348889895786135, "grad_norm": 1.4143430356438869, "learning_rate": 7.3181694848263e-05, "loss": 0.901, "step": 2977 }, { "epoch": 1.3493429995468962, "grad_norm": 2.459517989631613, "learning_rate": 7.317462681865284e-05, "loss": 0.8884, "step": 2978 }, { "epoch": 1.3497961033076575, "grad_norm": 2.226721436878511, "learning_rate": 7.316755546918227e-05, "loss": 0.9188, "step": 2979 }, { "epoch": 1.3502492070684187, "grad_norm": 1.6238011615684036, "learning_rate": 7.316048080055895e-05, "loss": 0.9028, "step": 2980 }, { "epoch": 1.3507023108291798, "grad_norm": 1.6990640413168567, "learning_rate": 7.315340281349082e-05, "loss": 0.9027, "step": 2981 }, { "epoch": 1.3511554145899412, "grad_norm": 1.715432165053432, "learning_rate": 7.31463215086862e-05, "loss": 0.8768, "step": 2982 }, { "epoch": 1.3516085183507023, "grad_norm": 1.1812278779829697, "learning_rate": 7.313923688685375e-05, "loss": 0.907, "step": 2983 }, { "epoch": 1.3520616221114636, "grad_norm": 1.6991782741428554, "learning_rate": 7.313214894870244e-05, "loss": 0.893, "step": 2984 }, { "epoch": 1.3525147258722248, "grad_norm": 1.1358335078488078, "learning_rate": 7.312505769494157e-05, "loss": 0.8933, "step": 2985 }, { "epoch": 1.352967829632986, "grad_norm": 1.6285923536251687, "learning_rate": 7.311796312628078e-05, "loss": 0.9002, "step": 2986 }, { "epoch": 1.3534209333937472, "grad_norm": 1.344755900324435, "learning_rate": 7.311086524343003e-05, "loss": 0.911, "step": 2987 }, { "epoch": 1.3538740371545084, "grad_norm": 1.2641250681751806, "learning_rate": 7.310376404709963e-05, "loss": 0.8996, "step": 2988 }, { "epoch": 1.3543271409152697, "grad_norm": 1.6250610612449743, "learning_rate": 7.309665953800023e-05, "loss": 0.893, "step": 2989 }, { "epoch": 1.3547802446760309, "grad_norm": 0.9373814242286282, "learning_rate": 7.308955171684275e-05, "loss": 0.9088, "step": 2990 }, { "epoch": 1.355233348436792, "grad_norm": 2.161972683621953, "learning_rate": 7.308244058433853e-05, "loss": 0.8855, "step": 2991 }, { "epoch": 1.3556864521975531, "grad_norm": 1.5307936146114913, "learning_rate": 7.307532614119919e-05, "loss": 0.8854, "step": 2992 }, { "epoch": 1.3561395559583145, "grad_norm": 2.167513439752408, "learning_rate": 7.306820838813666e-05, "loss": 0.9029, "step": 2993 }, { "epoch": 1.3565926597190756, "grad_norm": 1.902732505451618, "learning_rate": 7.306108732586325e-05, "loss": 0.9096, "step": 2994 }, { "epoch": 1.357045763479837, "grad_norm": 1.5514229475051222, "learning_rate": 7.305396295509157e-05, "loss": 0.8922, "step": 2995 }, { "epoch": 1.357498867240598, "grad_norm": 1.3018700682239968, "learning_rate": 7.30468352765346e-05, "loss": 0.889, "step": 2996 }, { "epoch": 1.3579519710013592, "grad_norm": 1.9717013022597423, "learning_rate": 7.303970429090559e-05, "loss": 0.9203, "step": 2997 }, { "epoch": 1.3584050747621206, "grad_norm": 1.552893585444873, "learning_rate": 7.303256999891818e-05, "loss": 0.8843, "step": 2998 }, { "epoch": 1.3588581785228817, "grad_norm": 1.9250181324081301, "learning_rate": 7.30254324012863e-05, "loss": 0.9007, "step": 2999 }, { "epoch": 1.359311282283643, "grad_norm": 1.5543467928353778, "learning_rate": 7.301829149872422e-05, "loss": 0.895, "step": 3000 }, { "epoch": 1.3597643860444042, "grad_norm": 1.8564894566239487, "learning_rate": 7.301114729194655e-05, "loss": 0.9116, "step": 3001 }, { "epoch": 1.3602174898051653, "grad_norm": 1.6293301640174618, "learning_rate": 7.300399978166825e-05, "loss": 0.8908, "step": 3002 }, { "epoch": 1.3606705935659267, "grad_norm": 1.7962577311083863, "learning_rate": 7.299684896860457e-05, "loss": 0.9201, "step": 3003 }, { "epoch": 1.3611236973266878, "grad_norm": 1.4798422126170692, "learning_rate": 7.29896948534711e-05, "loss": 0.8895, "step": 3004 }, { "epoch": 1.3615768010874492, "grad_norm": 1.7077586884078555, "learning_rate": 7.298253743698376e-05, "loss": 0.8891, "step": 3005 }, { "epoch": 1.3620299048482103, "grad_norm": 1.312300695572604, "learning_rate": 7.297537671985885e-05, "loss": 0.8956, "step": 3006 }, { "epoch": 1.3624830086089714, "grad_norm": 1.8919789365703539, "learning_rate": 7.296821270281292e-05, "loss": 0.9033, "step": 3007 }, { "epoch": 1.3629361123697326, "grad_norm": 1.5218930833346986, "learning_rate": 7.296104538656291e-05, "loss": 0.8917, "step": 3008 }, { "epoch": 1.363389216130494, "grad_norm": 1.964150738829793, "learning_rate": 7.295387477182607e-05, "loss": 0.8936, "step": 3009 }, { "epoch": 1.363842319891255, "grad_norm": 1.6385579435399447, "learning_rate": 7.294670085931996e-05, "loss": 0.9339, "step": 3010 }, { "epoch": 1.3642954236520164, "grad_norm": 1.7239343822989668, "learning_rate": 7.293952364976251e-05, "loss": 0.8978, "step": 3011 }, { "epoch": 1.3647485274127775, "grad_norm": 1.417860528327187, "learning_rate": 7.293234314387196e-05, "loss": 0.8939, "step": 3012 }, { "epoch": 1.3652016311735387, "grad_norm": 1.8822337266501554, "learning_rate": 7.292515934236687e-05, "loss": 0.9067, "step": 3013 }, { "epoch": 1.3656547349343, "grad_norm": 1.4074824262854382, "learning_rate": 7.291797224596613e-05, "loss": 0.8908, "step": 3014 }, { "epoch": 1.3661078386950611, "grad_norm": 1.9430378006482019, "learning_rate": 7.291078185538899e-05, "loss": 0.8725, "step": 3015 }, { "epoch": 1.3665609424558225, "grad_norm": 1.781872057473471, "learning_rate": 7.2903588171355e-05, "loss": 0.9109, "step": 3016 }, { "epoch": 1.3670140462165836, "grad_norm": 1.5950769298696341, "learning_rate": 7.289639119458404e-05, "loss": 0.8834, "step": 3017 }, { "epoch": 1.3674671499773448, "grad_norm": 1.391597407154388, "learning_rate": 7.288919092579634e-05, "loss": 0.8987, "step": 3018 }, { "epoch": 1.367920253738106, "grad_norm": 1.7906220046768029, "learning_rate": 7.288198736571245e-05, "loss": 0.8873, "step": 3019 }, { "epoch": 1.3683733574988672, "grad_norm": 1.389412937632218, "learning_rate": 7.287478051505323e-05, "loss": 0.9099, "step": 3020 }, { "epoch": 1.3688264612596286, "grad_norm": 1.9512353583979807, "learning_rate": 7.286757037453991e-05, "loss": 0.8934, "step": 3021 }, { "epoch": 1.3692795650203897, "grad_norm": 1.6833075516367841, "learning_rate": 7.286035694489399e-05, "loss": 0.8989, "step": 3022 }, { "epoch": 1.3697326687811509, "grad_norm": 1.7049594075527825, "learning_rate": 7.285314022683737e-05, "loss": 0.9054, "step": 3023 }, { "epoch": 1.370185772541912, "grad_norm": 1.5903897253627421, "learning_rate": 7.284592022109222e-05, "loss": 0.8923, "step": 3024 }, { "epoch": 1.3706388763026733, "grad_norm": 1.6179858429523992, "learning_rate": 7.283869692838108e-05, "loss": 0.9034, "step": 3025 }, { "epoch": 1.3710919800634345, "grad_norm": 1.3780852023266108, "learning_rate": 7.283147034942679e-05, "loss": 0.8922, "step": 3026 }, { "epoch": 1.3715450838241958, "grad_norm": 1.6849980833300857, "learning_rate": 7.282424048495251e-05, "loss": 0.8913, "step": 3027 }, { "epoch": 1.371998187584957, "grad_norm": 1.3327383746493806, "learning_rate": 7.28170073356818e-05, "loss": 0.9, "step": 3028 }, { "epoch": 1.372451291345718, "grad_norm": 1.8194678606949366, "learning_rate": 7.280977090233845e-05, "loss": 0.9128, "step": 3029 }, { "epoch": 1.3729043951064794, "grad_norm": 1.4764366941839873, "learning_rate": 7.280253118564667e-05, "loss": 0.9167, "step": 3030 }, { "epoch": 1.3733574988672406, "grad_norm": 1.7597496285034784, "learning_rate": 7.279528818633092e-05, "loss": 0.8931, "step": 3031 }, { "epoch": 1.373810602628002, "grad_norm": 1.5093955466911184, "learning_rate": 7.278804190511602e-05, "loss": 0.9093, "step": 3032 }, { "epoch": 1.374263706388763, "grad_norm": 1.5879365829786507, "learning_rate": 7.278079234272716e-05, "loss": 0.9013, "step": 3033 }, { "epoch": 1.3747168101495242, "grad_norm": 1.3357549429864564, "learning_rate": 7.277353949988978e-05, "loss": 0.8965, "step": 3034 }, { "epoch": 1.3751699139102855, "grad_norm": 1.564090940767059, "learning_rate": 7.27662833773297e-05, "loss": 0.8921, "step": 3035 }, { "epoch": 1.3756230176710467, "grad_norm": 1.1974484121669449, "learning_rate": 7.275902397577307e-05, "loss": 0.8921, "step": 3036 }, { "epoch": 1.376076121431808, "grad_norm": 1.6363442002241104, "learning_rate": 7.275176129594635e-05, "loss": 0.8976, "step": 3037 }, { "epoch": 1.3765292251925692, "grad_norm": 1.2749991095131088, "learning_rate": 7.274449533857632e-05, "loss": 0.8789, "step": 3038 }, { "epoch": 1.3769823289533303, "grad_norm": 1.6176738907607573, "learning_rate": 7.273722610439011e-05, "loss": 0.8677, "step": 3039 }, { "epoch": 1.3774354327140914, "grad_norm": 1.421649671528802, "learning_rate": 7.272995359411519e-05, "loss": 0.9268, "step": 3040 }, { "epoch": 1.3778885364748528, "grad_norm": 1.5499206579900833, "learning_rate": 7.272267780847928e-05, "loss": 0.8963, "step": 3041 }, { "epoch": 1.378341640235614, "grad_norm": 1.3454901554642438, "learning_rate": 7.271539874821054e-05, "loss": 0.8979, "step": 3042 }, { "epoch": 1.3787947439963752, "grad_norm": 1.4515499219741739, "learning_rate": 7.270811641403738e-05, "loss": 0.9016, "step": 3043 }, { "epoch": 1.3792478477571364, "grad_norm": 1.271667428950454, "learning_rate": 7.270083080668855e-05, "loss": 0.9041, "step": 3044 }, { "epoch": 1.3797009515178975, "grad_norm": 1.6481119633904622, "learning_rate": 7.269354192689316e-05, "loss": 0.8826, "step": 3045 }, { "epoch": 1.3801540552786589, "grad_norm": 1.461732094166383, "learning_rate": 7.268624977538062e-05, "loss": 0.9085, "step": 3046 }, { "epoch": 1.38060715903942, "grad_norm": 1.243935285630102, "learning_rate": 7.267895435288066e-05, "loss": 0.8927, "step": 3047 }, { "epoch": 1.3810602628001813, "grad_norm": 1.263658221628216, "learning_rate": 7.267165566012334e-05, "loss": 0.9056, "step": 3048 }, { "epoch": 1.3815133665609425, "grad_norm": 1.2320997784061283, "learning_rate": 7.266435369783909e-05, "loss": 0.8909, "step": 3049 }, { "epoch": 1.3819664703217036, "grad_norm": 1.0641382505207446, "learning_rate": 7.265704846675862e-05, "loss": 0.9231, "step": 3050 }, { "epoch": 1.382419574082465, "grad_norm": 1.3467456951809773, "learning_rate": 7.264973996761297e-05, "loss": 0.8866, "step": 3051 }, { "epoch": 1.382872677843226, "grad_norm": 0.955135723518989, "learning_rate": 7.264242820113353e-05, "loss": 0.8858, "step": 3052 }, { "epoch": 1.3833257816039874, "grad_norm": 1.5366152556003942, "learning_rate": 7.2635113168052e-05, "loss": 0.909, "step": 3053 }, { "epoch": 1.3837788853647486, "grad_norm": 1.2100297760071754, "learning_rate": 7.26277948691004e-05, "loss": 0.8921, "step": 3054 }, { "epoch": 1.3842319891255097, "grad_norm": 1.4794822136136636, "learning_rate": 7.26204733050111e-05, "loss": 0.8745, "step": 3055 }, { "epoch": 1.3846850928862708, "grad_norm": 1.4365166663475009, "learning_rate": 7.261314847651681e-05, "loss": 0.893, "step": 3056 }, { "epoch": 1.3851381966470322, "grad_norm": 1.1977721710938167, "learning_rate": 7.260582038435051e-05, "loss": 0.8779, "step": 3057 }, { "epoch": 1.3855913004077933, "grad_norm": 1.0311137081871877, "learning_rate": 7.259848902924554e-05, "loss": 0.8905, "step": 3058 }, { "epoch": 1.3860444041685547, "grad_norm": 1.2791328882929027, "learning_rate": 7.259115441193557e-05, "loss": 0.8908, "step": 3059 }, { "epoch": 1.3864975079293158, "grad_norm": 1.0080854662366838, "learning_rate": 7.258381653315461e-05, "loss": 0.8905, "step": 3060 }, { "epoch": 1.386950611690077, "grad_norm": 1.4009243651118646, "learning_rate": 7.257647539363696e-05, "loss": 0.8969, "step": 3061 }, { "epoch": 1.3874037154508383, "grad_norm": 1.1866774183206732, "learning_rate": 7.256913099411726e-05, "loss": 0.9049, "step": 3062 }, { "epoch": 1.3878568192115994, "grad_norm": 1.3151636869669838, "learning_rate": 7.256178333533049e-05, "loss": 0.8973, "step": 3063 }, { "epoch": 1.3883099229723608, "grad_norm": 1.2183485045329157, "learning_rate": 7.255443241801196e-05, "loss": 0.8801, "step": 3064 }, { "epoch": 1.388763026733122, "grad_norm": 1.170052661443549, "learning_rate": 7.254707824289726e-05, "loss": 0.8752, "step": 3065 }, { "epoch": 1.389216130493883, "grad_norm": 1.0024357480809898, "learning_rate": 7.253972081072237e-05, "loss": 0.8731, "step": 3066 }, { "epoch": 1.3896692342546444, "grad_norm": 1.0590323101166321, "learning_rate": 7.253236012222355e-05, "loss": 0.9022, "step": 3067 }, { "epoch": 1.3901223380154055, "grad_norm": 0.9272665697811131, "learning_rate": 7.25249961781374e-05, "loss": 0.8957, "step": 3068 }, { "epoch": 1.3905754417761669, "grad_norm": 1.3074194096285803, "learning_rate": 7.251762897920086e-05, "loss": 0.88, "step": 3069 }, { "epoch": 1.391028545536928, "grad_norm": 0.8692585071147092, "learning_rate": 7.251025852615117e-05, "loss": 0.894, "step": 3070 }, { "epoch": 1.3914816492976891, "grad_norm": 1.3876148057434723, "learning_rate": 7.25028848197259e-05, "loss": 0.8868, "step": 3071 }, { "epoch": 1.3919347530584503, "grad_norm": 1.0241838980879623, "learning_rate": 7.249550786066298e-05, "loss": 0.8956, "step": 3072 }, { "epoch": 1.3923878568192116, "grad_norm": 1.273468363786029, "learning_rate": 7.24881276497006e-05, "loss": 0.8835, "step": 3073 }, { "epoch": 1.3928409605799728, "grad_norm": 1.0811124445254106, "learning_rate": 7.248074418757736e-05, "loss": 0.8875, "step": 3074 }, { "epoch": 1.393294064340734, "grad_norm": 1.0333017936709878, "learning_rate": 7.247335747503211e-05, "loss": 0.8943, "step": 3075 }, { "epoch": 1.3937471681014952, "grad_norm": 0.7922548478530567, "learning_rate": 7.246596751280406e-05, "loss": 0.9009, "step": 3076 }, { "epoch": 1.3942002718622564, "grad_norm": 1.0018740309996577, "learning_rate": 7.245857430163276e-05, "loss": 0.8743, "step": 3077 }, { "epoch": 1.3946533756230177, "grad_norm": 0.8218658564662644, "learning_rate": 7.245117784225804e-05, "loss": 0.9047, "step": 3078 }, { "epoch": 1.3951064793837789, "grad_norm": 0.8383025903443085, "learning_rate": 7.244377813542008e-05, "loss": 0.8823, "step": 3079 }, { "epoch": 1.3955595831445402, "grad_norm": 0.760699907997529, "learning_rate": 7.24363751818594e-05, "loss": 0.9195, "step": 3080 }, { "epoch": 1.3960126869053013, "grad_norm": 0.7492491521348111, "learning_rate": 7.242896898231682e-05, "loss": 0.8789, "step": 3081 }, { "epoch": 1.3964657906660625, "grad_norm": 0.8434051171971592, "learning_rate": 7.242155953753351e-05, "loss": 0.8862, "step": 3082 }, { "epoch": 1.3969188944268238, "grad_norm": 0.845860334539331, "learning_rate": 7.241414684825094e-05, "loss": 0.9012, "step": 3083 }, { "epoch": 1.397371998187585, "grad_norm": 0.7780917243374864, "learning_rate": 7.240673091521091e-05, "loss": 0.8942, "step": 3084 }, { "epoch": 1.3978251019483463, "grad_norm": 0.6426500504946769, "learning_rate": 7.239931173915556e-05, "loss": 0.8887, "step": 3085 }, { "epoch": 1.3982782057091074, "grad_norm": 0.6955091212956528, "learning_rate": 7.239188932082734e-05, "loss": 0.9136, "step": 3086 }, { "epoch": 1.3987313094698686, "grad_norm": 0.6574060467198353, "learning_rate": 7.238446366096901e-05, "loss": 0.8846, "step": 3087 }, { "epoch": 1.3991844132306297, "grad_norm": 0.6926949717836107, "learning_rate": 7.23770347603237e-05, "loss": 0.9011, "step": 3088 }, { "epoch": 1.399637516991391, "grad_norm": 0.7189291951064826, "learning_rate": 7.236960261963483e-05, "loss": 0.8921, "step": 3089 }, { "epoch": 1.4000906207521522, "grad_norm": 0.6338786750440548, "learning_rate": 7.236216723964615e-05, "loss": 0.9063, "step": 3090 }, { "epoch": 1.4005437245129135, "grad_norm": 0.7779182043030637, "learning_rate": 7.235472862110172e-05, "loss": 0.8671, "step": 3091 }, { "epoch": 1.4009968282736747, "grad_norm": 0.9751759778998415, "learning_rate": 7.234728676474596e-05, "loss": 0.8722, "step": 3092 }, { "epoch": 1.4014499320344358, "grad_norm": 4.010298935006607, "learning_rate": 7.233984167132359e-05, "loss": 0.9256, "step": 3093 }, { "epoch": 1.4019030357951971, "grad_norm": 1.8590781162353696, "learning_rate": 7.233239334157966e-05, "loss": 0.9017, "step": 3094 }, { "epoch": 1.4023561395559583, "grad_norm": 1.1694173849816538, "learning_rate": 7.232494177625951e-05, "loss": 0.9071, "step": 3095 }, { "epoch": 1.4028092433167196, "grad_norm": 1.0203443890279902, "learning_rate": 7.231748697610887e-05, "loss": 0.9151, "step": 3096 }, { "epoch": 1.4032623470774808, "grad_norm": 0.5693665172244927, "learning_rate": 7.231002894187375e-05, "loss": 0.9082, "step": 3097 }, { "epoch": 1.403715450838242, "grad_norm": 0.65642191207587, "learning_rate": 7.23025676743005e-05, "loss": 0.8958, "step": 3098 }, { "epoch": 1.4041685545990032, "grad_norm": 0.7477896464615974, "learning_rate": 7.229510317413578e-05, "loss": 0.9214, "step": 3099 }, { "epoch": 1.4046216583597644, "grad_norm": 0.9079131812234458, "learning_rate": 7.228763544212656e-05, "loss": 0.904, "step": 3100 }, { "epoch": 1.4050747621205257, "grad_norm": 1.0117312998969956, "learning_rate": 7.228016447902017e-05, "loss": 0.9322, "step": 3101 }, { "epoch": 1.4055278658812869, "grad_norm": 1.170474930487897, "learning_rate": 7.227269028556427e-05, "loss": 0.9032, "step": 3102 }, { "epoch": 1.405980969642048, "grad_norm": 0.8160601494453376, "learning_rate": 7.226521286250677e-05, "loss": 0.8993, "step": 3103 }, { "epoch": 1.4064340734028091, "grad_norm": 0.7845208136232551, "learning_rate": 7.225773221059598e-05, "loss": 0.8868, "step": 3104 }, { "epoch": 1.4068871771635705, "grad_norm": 0.7541403455965089, "learning_rate": 7.225024833058051e-05, "loss": 0.899, "step": 3105 }, { "epoch": 1.4073402809243316, "grad_norm": 0.8180574875544928, "learning_rate": 7.22427612232093e-05, "loss": 0.8969, "step": 3106 }, { "epoch": 1.407793384685093, "grad_norm": 0.8544042888468079, "learning_rate": 7.223527088923154e-05, "loss": 0.9054, "step": 3107 }, { "epoch": 1.408246488445854, "grad_norm": 0.6306702919016515, "learning_rate": 7.22277773293969e-05, "loss": 0.8899, "step": 3108 }, { "epoch": 1.4086995922066152, "grad_norm": 0.5286391844366651, "learning_rate": 7.222028054445519e-05, "loss": 0.9079, "step": 3109 }, { "epoch": 1.4091526959673766, "grad_norm": 0.718810040175335, "learning_rate": 7.221278053515668e-05, "loss": 0.9041, "step": 3110 }, { "epoch": 1.4096057997281377, "grad_norm": 0.6534314507585051, "learning_rate": 7.220527730225191e-05, "loss": 0.9057, "step": 3111 }, { "epoch": 1.410058903488899, "grad_norm": 0.6570812168836848, "learning_rate": 7.219777084649173e-05, "loss": 0.9018, "step": 3112 }, { "epoch": 1.4105120072496602, "grad_norm": 0.767114422934399, "learning_rate": 7.219026116862733e-05, "loss": 0.9107, "step": 3113 }, { "epoch": 1.4109651110104213, "grad_norm": 1.055452227266792, "learning_rate": 7.218274826941023e-05, "loss": 0.9192, "step": 3114 }, { "epoch": 1.4114182147711827, "grad_norm": 1.218382094049941, "learning_rate": 7.217523214959225e-05, "loss": 0.902, "step": 3115 }, { "epoch": 1.4118713185319438, "grad_norm": 0.7850983913836461, "learning_rate": 7.216771280992556e-05, "loss": 0.882, "step": 3116 }, { "epoch": 1.4123244222927052, "grad_norm": 0.8977076782452105, "learning_rate": 7.216019025116263e-05, "loss": 0.8977, "step": 3117 }, { "epoch": 1.4127775260534663, "grad_norm": 1.0291727506568726, "learning_rate": 7.215266447405626e-05, "loss": 0.8889, "step": 3118 }, { "epoch": 1.4132306298142274, "grad_norm": 0.6358595809938545, "learning_rate": 7.214513547935956e-05, "loss": 0.8975, "step": 3119 }, { "epoch": 1.4136837335749886, "grad_norm": 0.979538638869614, "learning_rate": 7.213760326782601e-05, "loss": 0.9206, "step": 3120 }, { "epoch": 1.41413683733575, "grad_norm": 0.46320788677478886, "learning_rate": 7.213006784020932e-05, "loss": 0.8744, "step": 3121 }, { "epoch": 1.414589941096511, "grad_norm": 0.5489815558727639, "learning_rate": 7.212252919726363e-05, "loss": 0.8804, "step": 3122 }, { "epoch": 1.4150430448572724, "grad_norm": 0.8395365794597586, "learning_rate": 7.211498733974331e-05, "loss": 0.9208, "step": 3123 }, { "epoch": 1.4154961486180335, "grad_norm": 1.1047227216748001, "learning_rate": 7.210744226840313e-05, "loss": 0.8827, "step": 3124 }, { "epoch": 1.4159492523787947, "grad_norm": 1.1238893558750251, "learning_rate": 7.20998939839981e-05, "loss": 0.8811, "step": 3125 }, { "epoch": 1.416402356139556, "grad_norm": 1.0292228512328134, "learning_rate": 7.209234248728364e-05, "loss": 0.8933, "step": 3126 }, { "epoch": 1.4168554599003171, "grad_norm": 1.15398240717686, "learning_rate": 7.20847877790154e-05, "loss": 0.8959, "step": 3127 }, { "epoch": 1.4173085636610785, "grad_norm": 1.0085525136101379, "learning_rate": 7.207722985994941e-05, "loss": 0.9045, "step": 3128 }, { "epoch": 1.4177616674218396, "grad_norm": 1.0553555675916233, "learning_rate": 7.206966873084204e-05, "loss": 0.9007, "step": 3129 }, { "epoch": 1.4182147711826008, "grad_norm": 0.8146787483143054, "learning_rate": 7.206210439244991e-05, "loss": 0.898, "step": 3130 }, { "epoch": 1.418667874943362, "grad_norm": 0.9197987477572535, "learning_rate": 7.205453684553001e-05, "loss": 0.9047, "step": 3131 }, { "epoch": 1.4191209787041232, "grad_norm": 0.9449017038140756, "learning_rate": 7.204696609083966e-05, "loss": 0.8967, "step": 3132 }, { "epoch": 1.4195740824648846, "grad_norm": 0.943641772824657, "learning_rate": 7.203939212913646e-05, "loss": 0.8973, "step": 3133 }, { "epoch": 1.4200271862256457, "grad_norm": 1.0501391819638601, "learning_rate": 7.203181496117836e-05, "loss": 0.9003, "step": 3134 }, { "epoch": 1.4204802899864069, "grad_norm": 1.2212811101634262, "learning_rate": 7.202423458772364e-05, "loss": 0.8942, "step": 3135 }, { "epoch": 1.420933393747168, "grad_norm": 0.6289715919294007, "learning_rate": 7.201665100953085e-05, "loss": 0.9179, "step": 3136 }, { "epoch": 1.4213864975079293, "grad_norm": 0.4079464514266739, "learning_rate": 7.200906422735895e-05, "loss": 0.8935, "step": 3137 }, { "epoch": 1.4218396012686905, "grad_norm": 0.5718925855661433, "learning_rate": 7.200147424196711e-05, "loss": 0.9095, "step": 3138 }, { "epoch": 1.4222927050294518, "grad_norm": 0.811216542461963, "learning_rate": 7.19938810541149e-05, "loss": 0.8879, "step": 3139 }, { "epoch": 1.422745808790213, "grad_norm": 1.1263431883689676, "learning_rate": 7.198628466456221e-05, "loss": 0.9116, "step": 3140 }, { "epoch": 1.423198912550974, "grad_norm": 1.0823493219610782, "learning_rate": 7.19786850740692e-05, "loss": 0.8939, "step": 3141 }, { "epoch": 1.4236520163117354, "grad_norm": 0.982184204528922, "learning_rate": 7.197108228339638e-05, "loss": 0.8912, "step": 3142 }, { "epoch": 1.4241051200724966, "grad_norm": 1.12354889434874, "learning_rate": 7.19634762933046e-05, "loss": 0.8961, "step": 3143 }, { "epoch": 1.424558223833258, "grad_norm": 0.9222215585796397, "learning_rate": 7.195586710455497e-05, "loss": 0.897, "step": 3144 }, { "epoch": 1.425011327594019, "grad_norm": 0.7203434890417191, "learning_rate": 7.1948254717909e-05, "loss": 0.9048, "step": 3145 }, { "epoch": 1.4254644313547802, "grad_norm": 0.49479276100350905, "learning_rate": 7.194063913412845e-05, "loss": 0.8797, "step": 3146 }, { "epoch": 1.4259175351155415, "grad_norm": 0.6028971978322192, "learning_rate": 7.193302035397545e-05, "loss": 0.893, "step": 3147 }, { "epoch": 1.4263706388763027, "grad_norm": 0.7567441610640943, "learning_rate": 7.192539837821242e-05, "loss": 0.8958, "step": 3148 }, { "epoch": 1.426823742637064, "grad_norm": 0.8261794882604402, "learning_rate": 7.19177732076021e-05, "loss": 0.8913, "step": 3149 }, { "epoch": 1.4272768463978251, "grad_norm": 0.9209590171566016, "learning_rate": 7.191014484290756e-05, "loss": 0.907, "step": 3150 }, { "epoch": 1.4277299501585863, "grad_norm": 1.121041331176206, "learning_rate": 7.19025132848922e-05, "loss": 0.9048, "step": 3151 }, { "epoch": 1.4281830539193474, "grad_norm": 1.0666133834175908, "learning_rate": 7.189487853431973e-05, "loss": 0.893, "step": 3152 }, { "epoch": 1.4286361576801088, "grad_norm": 0.9238929034982115, "learning_rate": 7.188724059195415e-05, "loss": 0.8916, "step": 3153 }, { "epoch": 1.42908926144087, "grad_norm": 0.8817970875211582, "learning_rate": 7.187959945855984e-05, "loss": 0.9041, "step": 3154 }, { "epoch": 1.4295423652016312, "grad_norm": 0.7963334673729507, "learning_rate": 7.187195513490143e-05, "loss": 0.8893, "step": 3155 }, { "epoch": 1.4299954689623924, "grad_norm": 0.7930378513788792, "learning_rate": 7.186430762174393e-05, "loss": 0.8828, "step": 3156 }, { "epoch": 1.4304485727231535, "grad_norm": 0.9273132162171612, "learning_rate": 7.185665691985265e-05, "loss": 0.8755, "step": 3157 }, { "epoch": 1.4309016764839149, "grad_norm": 1.105691615136974, "learning_rate": 7.184900302999319e-05, "loss": 0.8943, "step": 3158 }, { "epoch": 1.431354780244676, "grad_norm": 0.8850546167690825, "learning_rate": 7.184134595293151e-05, "loss": 0.8913, "step": 3159 }, { "epoch": 1.4318078840054373, "grad_norm": 0.887678044465657, "learning_rate": 7.183368568943386e-05, "loss": 0.9095, "step": 3160 }, { "epoch": 1.4322609877661985, "grad_norm": 1.0856018875938813, "learning_rate": 7.182602224026682e-05, "loss": 0.8897, "step": 3161 }, { "epoch": 1.4327140915269596, "grad_norm": 0.9627250609863394, "learning_rate": 7.18183556061973e-05, "loss": 0.9034, "step": 3162 }, { "epoch": 1.433167195287721, "grad_norm": 0.8918663739382396, "learning_rate": 7.181068578799252e-05, "loss": 0.9228, "step": 3163 }, { "epoch": 1.433620299048482, "grad_norm": 0.9056306903937612, "learning_rate": 7.180301278641998e-05, "loss": 0.8935, "step": 3164 }, { "epoch": 1.4340734028092434, "grad_norm": 0.9924260286238505, "learning_rate": 7.17953366022476e-05, "loss": 0.8777, "step": 3165 }, { "epoch": 1.4345265065700046, "grad_norm": 1.2757033628467846, "learning_rate": 7.178765723624349e-05, "loss": 0.8798, "step": 3166 }, { "epoch": 1.4349796103307657, "grad_norm": 0.6267232460665861, "learning_rate": 7.177997468917616e-05, "loss": 0.8768, "step": 3167 }, { "epoch": 1.4354327140915268, "grad_norm": 0.49726323162592884, "learning_rate": 7.177228896181444e-05, "loss": 0.8548, "step": 3168 }, { "epoch": 1.4358858178522882, "grad_norm": 0.9397780938393971, "learning_rate": 7.176460005492744e-05, "loss": 0.8895, "step": 3169 }, { "epoch": 1.4363389216130493, "grad_norm": 1.3820731419142613, "learning_rate": 7.175690796928462e-05, "loss": 0.9018, "step": 3170 }, { "epoch": 1.4367920253738107, "grad_norm": 0.5901320492736939, "learning_rate": 7.174921270565572e-05, "loss": 0.9068, "step": 3171 }, { "epoch": 1.4372451291345718, "grad_norm": 0.7666402103198962, "learning_rate": 7.174151426481086e-05, "loss": 0.9084, "step": 3172 }, { "epoch": 1.437698232895333, "grad_norm": 0.9196376704369599, "learning_rate": 7.17338126475204e-05, "loss": 0.9042, "step": 3173 }, { "epoch": 1.4381513366560943, "grad_norm": 1.0657609414851783, "learning_rate": 7.172610785455508e-05, "loss": 0.8936, "step": 3174 }, { "epoch": 1.4386044404168554, "grad_norm": 1.1418841804599777, "learning_rate": 7.171839988668595e-05, "loss": 0.9087, "step": 3175 }, { "epoch": 1.4390575441776168, "grad_norm": 1.0495924056693589, "learning_rate": 7.171068874468432e-05, "loss": 0.8919, "step": 3176 }, { "epoch": 1.439510647938378, "grad_norm": 0.9275485776934391, "learning_rate": 7.170297442932191e-05, "loss": 0.8951, "step": 3177 }, { "epoch": 1.439963751699139, "grad_norm": 0.7662438357450909, "learning_rate": 7.169525694137068e-05, "loss": 0.8848, "step": 3178 }, { "epoch": 1.4404168554599004, "grad_norm": 0.6566239831689977, "learning_rate": 7.168753628160295e-05, "loss": 0.8903, "step": 3179 }, { "epoch": 1.4408699592206615, "grad_norm": 0.5619480896415378, "learning_rate": 7.167981245079132e-05, "loss": 0.9098, "step": 3180 }, { "epoch": 1.4413230629814229, "grad_norm": 0.45451578958007793, "learning_rate": 7.167208544970878e-05, "loss": 0.8972, "step": 3181 }, { "epoch": 1.441776166742184, "grad_norm": 0.4638780752069402, "learning_rate": 7.166435527912853e-05, "loss": 0.9096, "step": 3182 }, { "epoch": 1.4422292705029451, "grad_norm": 0.6381056244416821, "learning_rate": 7.165662193982419e-05, "loss": 0.9054, "step": 3183 }, { "epoch": 1.4426823742637063, "grad_norm": 0.8532133286526447, "learning_rate": 7.164888543256963e-05, "loss": 0.8898, "step": 3184 }, { "epoch": 1.4431354780244676, "grad_norm": 1.0598979710988954, "learning_rate": 7.164114575813907e-05, "loss": 0.9016, "step": 3185 }, { "epoch": 1.4435885817852288, "grad_norm": 1.21256791619585, "learning_rate": 7.163340291730705e-05, "loss": 0.8856, "step": 3186 }, { "epoch": 1.44404168554599, "grad_norm": 0.8267028210305167, "learning_rate": 7.162565691084837e-05, "loss": 0.9057, "step": 3187 }, { "epoch": 1.4444947893067512, "grad_norm": 0.7658545686476594, "learning_rate": 7.161790773953823e-05, "loss": 0.8831, "step": 3188 }, { "epoch": 1.4449478930675124, "grad_norm": 0.8252424305894196, "learning_rate": 7.161015540415209e-05, "loss": 0.8893, "step": 3189 }, { "epoch": 1.4454009968282737, "grad_norm": 0.945264467570463, "learning_rate": 7.160239990546575e-05, "loss": 0.9143, "step": 3190 }, { "epoch": 1.4458541005890349, "grad_norm": 1.097998292935251, "learning_rate": 7.159464124425533e-05, "loss": 0.9041, "step": 3191 }, { "epoch": 1.4463072043497962, "grad_norm": 1.0693926900528434, "learning_rate": 7.158687942129725e-05, "loss": 0.8972, "step": 3192 }, { "epoch": 1.4467603081105573, "grad_norm": 0.9536762051754049, "learning_rate": 7.157911443736822e-05, "loss": 0.914, "step": 3193 }, { "epoch": 1.4472134118713185, "grad_norm": 0.931617731836376, "learning_rate": 7.157134629324535e-05, "loss": 0.8859, "step": 3194 }, { "epoch": 1.4476665156320798, "grad_norm": 1.0276826747853613, "learning_rate": 7.1563574989706e-05, "loss": 0.8978, "step": 3195 }, { "epoch": 1.448119619392841, "grad_norm": 0.9478590743088988, "learning_rate": 7.155580052752785e-05, "loss": 0.906, "step": 3196 }, { "epoch": 1.4485727231536023, "grad_norm": 0.925752876200872, "learning_rate": 7.154802290748892e-05, "loss": 0.8846, "step": 3197 }, { "epoch": 1.4490258269143634, "grad_norm": 0.9718676336795493, "learning_rate": 7.154024213036752e-05, "loss": 0.8942, "step": 3198 }, { "epoch": 1.4494789306751246, "grad_norm": 1.1496226805971372, "learning_rate": 7.153245819694231e-05, "loss": 0.8912, "step": 3199 }, { "epoch": 1.4499320344358857, "grad_norm": 0.8730025021426028, "learning_rate": 7.152467110799222e-05, "loss": 0.8881, "step": 3200 }, { "epoch": 1.450385138196647, "grad_norm": 0.555604861877808, "learning_rate": 7.151688086429655e-05, "loss": 0.8909, "step": 3201 }, { "epoch": 1.4508382419574082, "grad_norm": 0.5526684444257617, "learning_rate": 7.150908746663487e-05, "loss": 0.9045, "step": 3202 }, { "epoch": 1.4512913457181695, "grad_norm": 0.5615684728333997, "learning_rate": 7.15012909157871e-05, "loss": 0.9045, "step": 3203 }, { "epoch": 1.4517444494789307, "grad_norm": 0.4746699319404592, "learning_rate": 7.149349121253343e-05, "loss": 0.887, "step": 3204 }, { "epoch": 1.4521975532396918, "grad_norm": 0.5266742376872605, "learning_rate": 7.14856883576544e-05, "loss": 0.9088, "step": 3205 }, { "epoch": 1.4526506570004531, "grad_norm": 0.6808368161188605, "learning_rate": 7.147788235193089e-05, "loss": 0.9017, "step": 3206 }, { "epoch": 1.4531037607612143, "grad_norm": 0.653538873283881, "learning_rate": 7.147007319614406e-05, "loss": 0.8969, "step": 3207 }, { "epoch": 1.4535568645219756, "grad_norm": 0.4784095693141157, "learning_rate": 7.146226089107533e-05, "loss": 0.9023, "step": 3208 }, { "epoch": 1.4540099682827368, "grad_norm": 0.48119591613941237, "learning_rate": 7.145444543750657e-05, "loss": 0.9013, "step": 3209 }, { "epoch": 1.454463072043498, "grad_norm": 0.6858012523635783, "learning_rate": 7.144662683621986e-05, "loss": 0.8794, "step": 3210 }, { "epoch": 1.4549161758042592, "grad_norm": 0.9333078230477828, "learning_rate": 7.143880508799761e-05, "loss": 0.8864, "step": 3211 }, { "epoch": 1.4553692795650204, "grad_norm": 1.2207114506006336, "learning_rate": 7.143098019362258e-05, "loss": 0.8889, "step": 3212 }, { "epoch": 1.4558223833257817, "grad_norm": 0.9078628245752651, "learning_rate": 7.142315215387783e-05, "loss": 0.8793, "step": 3213 }, { "epoch": 1.4562754870865429, "grad_norm": 0.8786453684132461, "learning_rate": 7.141532096954671e-05, "loss": 0.8728, "step": 3214 }, { "epoch": 1.456728590847304, "grad_norm": 0.9147055185371431, "learning_rate": 7.140748664141292e-05, "loss": 0.893, "step": 3215 }, { "epoch": 1.4571816946080651, "grad_norm": 1.011842272993526, "learning_rate": 7.139964917026046e-05, "loss": 0.8882, "step": 3216 }, { "epoch": 1.4576347983688265, "grad_norm": 1.193824000018561, "learning_rate": 7.139180855687362e-05, "loss": 0.8814, "step": 3217 }, { "epoch": 1.4580879021295876, "grad_norm": 0.7335286283049168, "learning_rate": 7.138396480203705e-05, "loss": 0.8842, "step": 3218 }, { "epoch": 1.458541005890349, "grad_norm": 0.5587578029289476, "learning_rate": 7.137611790653569e-05, "loss": 0.8881, "step": 3219 }, { "epoch": 1.45899410965111, "grad_norm": 0.5157619125942362, "learning_rate": 7.136826787115481e-05, "loss": 0.8716, "step": 3220 }, { "epoch": 1.4594472134118712, "grad_norm": 0.49475804083277075, "learning_rate": 7.136041469667993e-05, "loss": 0.8851, "step": 3221 }, { "epoch": 1.4599003171726326, "grad_norm": 0.567298139052923, "learning_rate": 7.1352558383897e-05, "loss": 0.9008, "step": 3222 }, { "epoch": 1.4603534209333937, "grad_norm": 0.7343108014928421, "learning_rate": 7.134469893359218e-05, "loss": 0.9123, "step": 3223 }, { "epoch": 1.460806524694155, "grad_norm": 1.0454235235619374, "learning_rate": 7.133683634655199e-05, "loss": 0.9017, "step": 3224 }, { "epoch": 1.4612596284549162, "grad_norm": 1.3167249985418767, "learning_rate": 7.132897062356326e-05, "loss": 0.9016, "step": 3225 }, { "epoch": 1.4617127322156773, "grad_norm": 0.6942469352718307, "learning_rate": 7.132110176541313e-05, "loss": 0.8972, "step": 3226 }, { "epoch": 1.4621658359764387, "grad_norm": 0.5924370660277701, "learning_rate": 7.131322977288906e-05, "loss": 0.8862, "step": 3227 }, { "epoch": 1.4626189397371998, "grad_norm": 0.8790733410713917, "learning_rate": 7.13053546467788e-05, "loss": 0.9184, "step": 3228 }, { "epoch": 1.4630720434979612, "grad_norm": 1.3720375677574894, "learning_rate": 7.129747638787045e-05, "loss": 0.9225, "step": 3229 }, { "epoch": 1.4635251472587223, "grad_norm": 0.8196786454734566, "learning_rate": 7.12895949969524e-05, "loss": 0.8834, "step": 3230 }, { "epoch": 1.4639782510194834, "grad_norm": 0.9399100739589467, "learning_rate": 7.128171047481336e-05, "loss": 0.8922, "step": 3231 }, { "epoch": 1.4644313547802446, "grad_norm": 0.9465100098213162, "learning_rate": 7.127382282224235e-05, "loss": 0.8904, "step": 3232 }, { "epoch": 1.464884458541006, "grad_norm": 0.7313595759994519, "learning_rate": 7.126593204002872e-05, "loss": 0.8738, "step": 3233 }, { "epoch": 1.465337562301767, "grad_norm": 0.7225716979953196, "learning_rate": 7.125803812896209e-05, "loss": 0.875, "step": 3234 }, { "epoch": 1.4657906660625284, "grad_norm": 0.9657041229881717, "learning_rate": 7.125014108983243e-05, "loss": 0.8868, "step": 3235 }, { "epoch": 1.4662437698232895, "grad_norm": 1.3316667122285835, "learning_rate": 7.124224092343003e-05, "loss": 0.9129, "step": 3236 }, { "epoch": 1.4666968735840507, "grad_norm": 0.7387247075611779, "learning_rate": 7.123433763054547e-05, "loss": 0.8898, "step": 3237 }, { "epoch": 1.467149977344812, "grad_norm": 0.788663271695546, "learning_rate": 7.122643121196964e-05, "loss": 0.9155, "step": 3238 }, { "epoch": 1.4676030811055731, "grad_norm": 0.8325410349479185, "learning_rate": 7.121852166849378e-05, "loss": 0.9073, "step": 3239 }, { "epoch": 1.4680561848663345, "grad_norm": 0.987624913781269, "learning_rate": 7.121060900090938e-05, "loss": 0.9199, "step": 3240 }, { "epoch": 1.4685092886270956, "grad_norm": 1.2606906940610898, "learning_rate": 7.12026932100083e-05, "loss": 0.8798, "step": 3241 }, { "epoch": 1.4689623923878568, "grad_norm": 0.6130191976214382, "learning_rate": 7.11947742965827e-05, "loss": 0.8855, "step": 3242 }, { "epoch": 1.469415496148618, "grad_norm": 1.0171383745626021, "learning_rate": 7.118685226142501e-05, "loss": 0.9019, "step": 3243 }, { "epoch": 1.4698685999093792, "grad_norm": 1.1926488993154885, "learning_rate": 7.117892710532806e-05, "loss": 0.9121, "step": 3244 }, { "epoch": 1.4703217036701406, "grad_norm": 0.444666012283508, "learning_rate": 7.117099882908488e-05, "loss": 0.8917, "step": 3245 }, { "epoch": 1.4707748074309017, "grad_norm": 0.8694316611708524, "learning_rate": 7.116306743348891e-05, "loss": 0.8958, "step": 3246 }, { "epoch": 1.4712279111916629, "grad_norm": 1.1844824138834495, "learning_rate": 7.115513291933384e-05, "loss": 0.8932, "step": 3247 }, { "epoch": 1.471681014952424, "grad_norm": 0.6058771733451676, "learning_rate": 7.114719528741371e-05, "loss": 0.9109, "step": 3248 }, { "epoch": 1.4721341187131853, "grad_norm": 1.0936178914214496, "learning_rate": 7.113925453852287e-05, "loss": 0.8853, "step": 3249 }, { "epoch": 1.4725872224739465, "grad_norm": 1.0753330553711777, "learning_rate": 7.113131067345592e-05, "loss": 0.8989, "step": 3250 }, { "epoch": 1.4730403262347078, "grad_norm": 0.8075052714418732, "learning_rate": 7.112336369300786e-05, "loss": 0.9008, "step": 3251 }, { "epoch": 1.473493429995469, "grad_norm": 1.1099069611084937, "learning_rate": 7.111541359797397e-05, "loss": 0.8745, "step": 3252 }, { "epoch": 1.47394653375623, "grad_norm": 0.8219924225526095, "learning_rate": 7.110746038914979e-05, "loss": 0.8941, "step": 3253 }, { "epoch": 1.4743996375169914, "grad_norm": 1.0348660782354944, "learning_rate": 7.109950406733127e-05, "loss": 0.8882, "step": 3254 }, { "epoch": 1.4748527412777526, "grad_norm": 1.076271976504931, "learning_rate": 7.109154463331458e-05, "loss": 0.8906, "step": 3255 }, { "epoch": 1.475305845038514, "grad_norm": 0.8063982807002872, "learning_rate": 7.108358208789625e-05, "loss": 0.8963, "step": 3256 }, { "epoch": 1.475758948799275, "grad_norm": 1.148193778692857, "learning_rate": 7.10756164318731e-05, "loss": 0.8836, "step": 3257 }, { "epoch": 1.4762120525600362, "grad_norm": 0.5484570229210824, "learning_rate": 7.106764766604229e-05, "loss": 0.9029, "step": 3258 }, { "epoch": 1.4766651563207975, "grad_norm": 0.9141901872805501, "learning_rate": 7.105967579120125e-05, "loss": 0.904, "step": 3259 }, { "epoch": 1.4771182600815587, "grad_norm": 1.1051935783566713, "learning_rate": 7.105170080814777e-05, "loss": 0.8819, "step": 3260 }, { "epoch": 1.47757136384232, "grad_norm": 0.670190362796194, "learning_rate": 7.104372271767991e-05, "loss": 0.8808, "step": 3261 }, { "epoch": 1.4780244676030811, "grad_norm": 2.1747645704926795, "learning_rate": 7.103574152059607e-05, "loss": 0.9152, "step": 3262 }, { "epoch": 1.4784775713638423, "grad_norm": 0.9516623279713216, "learning_rate": 7.102775721769491e-05, "loss": 0.8946, "step": 3263 }, { "epoch": 1.4789306751246034, "grad_norm": 1.6880349616985777, "learning_rate": 7.101976980977547e-05, "loss": 0.9026, "step": 3264 }, { "epoch": 1.4793837788853648, "grad_norm": 1.118490280325732, "learning_rate": 7.101177929763707e-05, "loss": 0.9035, "step": 3265 }, { "epoch": 1.479836882646126, "grad_norm": 1.4269089793143837, "learning_rate": 7.100378568207932e-05, "loss": 0.9001, "step": 3266 }, { "epoch": 1.4802899864068872, "grad_norm": 1.3140434940629675, "learning_rate": 7.099578896390217e-05, "loss": 0.9149, "step": 3267 }, { "epoch": 1.4807430901676484, "grad_norm": 1.1762638805942591, "learning_rate": 7.098778914390585e-05, "loss": 0.893, "step": 3268 }, { "epoch": 1.4811961939284095, "grad_norm": 0.8613123808372599, "learning_rate": 7.097978622289096e-05, "loss": 0.8835, "step": 3269 }, { "epoch": 1.4816492976891709, "grad_norm": 1.0939359845773386, "learning_rate": 7.097178020165833e-05, "loss": 0.8872, "step": 3270 }, { "epoch": 1.482102401449932, "grad_norm": 0.9990502005663385, "learning_rate": 7.096377108100917e-05, "loss": 0.8956, "step": 3271 }, { "epoch": 1.4825555052106933, "grad_norm": 1.3514229583865887, "learning_rate": 7.095575886174497e-05, "loss": 0.9095, "step": 3272 }, { "epoch": 1.4830086089714545, "grad_norm": 0.9031975459364059, "learning_rate": 7.09477435446675e-05, "loss": 0.901, "step": 3273 }, { "epoch": 1.4834617127322156, "grad_norm": 1.0770121702699855, "learning_rate": 7.09397251305789e-05, "loss": 0.9074, "step": 3274 }, { "epoch": 1.483914816492977, "grad_norm": 1.230693883820848, "learning_rate": 7.093170362028158e-05, "loss": 0.9076, "step": 3275 }, { "epoch": 1.484367920253738, "grad_norm": 1.0275786761979953, "learning_rate": 7.092367901457827e-05, "loss": 0.8839, "step": 3276 }, { "epoch": 1.4848210240144994, "grad_norm": 0.8858737938485299, "learning_rate": 7.091565131427201e-05, "loss": 0.8862, "step": 3277 }, { "epoch": 1.4852741277752606, "grad_norm": 0.7244728468182233, "learning_rate": 7.090762052016617e-05, "loss": 0.9062, "step": 3278 }, { "epoch": 1.4857272315360217, "grad_norm": 0.7803785397479307, "learning_rate": 7.089958663306439e-05, "loss": 0.8973, "step": 3279 }, { "epoch": 1.4861803352967828, "grad_norm": 0.743831644610428, "learning_rate": 7.089154965377064e-05, "loss": 0.8935, "step": 3280 }, { "epoch": 1.4866334390575442, "grad_norm": 0.8214092294363566, "learning_rate": 7.08835095830892e-05, "loss": 0.889, "step": 3281 }, { "epoch": 1.4870865428183053, "grad_norm": 0.820437634927794, "learning_rate": 7.087546642182465e-05, "loss": 0.8989, "step": 3282 }, { "epoch": 1.4875396465790667, "grad_norm": 0.8650091195248676, "learning_rate": 7.086742017078191e-05, "loss": 0.8734, "step": 3283 }, { "epoch": 1.4879927503398278, "grad_norm": 1.1786197943230825, "learning_rate": 7.085937083076618e-05, "loss": 0.8989, "step": 3284 }, { "epoch": 1.488445854100589, "grad_norm": 1.171552419864845, "learning_rate": 7.085131840258298e-05, "loss": 0.9109, "step": 3285 }, { "epoch": 1.4888989578613503, "grad_norm": 0.7878204199824722, "learning_rate": 7.084326288703809e-05, "loss": 0.8947, "step": 3286 }, { "epoch": 1.4893520616221114, "grad_norm": 0.6911085842050984, "learning_rate": 7.08352042849377e-05, "loss": 0.889, "step": 3287 }, { "epoch": 1.4898051653828728, "grad_norm": 0.5151115635339854, "learning_rate": 7.082714259708824e-05, "loss": 0.9084, "step": 3288 }, { "epoch": 1.490258269143634, "grad_norm": 0.5728612503165279, "learning_rate": 7.081907782429643e-05, "loss": 0.8965, "step": 3289 }, { "epoch": 1.490711372904395, "grad_norm": 0.8784402302469341, "learning_rate": 7.081100996736937e-05, "loss": 0.883, "step": 3290 }, { "epoch": 1.4911644766651564, "grad_norm": 1.2596827084003892, "learning_rate": 7.080293902711441e-05, "loss": 0.8786, "step": 3291 }, { "epoch": 1.4916175804259175, "grad_norm": 0.8754748517268356, "learning_rate": 7.079486500433922e-05, "loss": 0.8784, "step": 3292 }, { "epoch": 1.4920706841866789, "grad_norm": 0.728053527623023, "learning_rate": 7.07867878998518e-05, "loss": 0.8947, "step": 3293 }, { "epoch": 1.49252378794744, "grad_norm": 0.6645583785037039, "learning_rate": 7.077870771446045e-05, "loss": 0.8993, "step": 3294 }, { "epoch": 1.4929768917082011, "grad_norm": 0.6064938125150995, "learning_rate": 7.077062444897375e-05, "loss": 0.899, "step": 3295 }, { "epoch": 1.4934299954689623, "grad_norm": 0.7504976340062879, "learning_rate": 7.076253810420063e-05, "loss": 0.879, "step": 3296 }, { "epoch": 1.4938830992297236, "grad_norm": 0.792088350206204, "learning_rate": 7.07544486809503e-05, "loss": 0.9107, "step": 3297 }, { "epoch": 1.4943362029904848, "grad_norm": 0.8350290227612993, "learning_rate": 7.074635618003229e-05, "loss": 0.8719, "step": 3298 }, { "epoch": 1.494789306751246, "grad_norm": 1.058804122021197, "learning_rate": 7.073826060225642e-05, "loss": 0.8988, "step": 3299 }, { "epoch": 1.4952424105120072, "grad_norm": 1.2555340105075643, "learning_rate": 7.073016194843285e-05, "loss": 0.8908, "step": 3300 }, { "epoch": 1.4956955142727684, "grad_norm": 0.6408325923358602, "learning_rate": 7.072206021937203e-05, "loss": 0.905, "step": 3301 }, { "epoch": 1.4961486180335297, "grad_norm": 0.5647774767846386, "learning_rate": 7.071395541588472e-05, "loss": 0.8772, "step": 3302 }, { "epoch": 1.4966017217942909, "grad_norm": 1.0042327020983237, "learning_rate": 7.070584753878198e-05, "loss": 0.8934, "step": 3303 }, { "epoch": 1.4970548255550522, "grad_norm": 1.444094047799636, "learning_rate": 7.069773658887519e-05, "loss": 0.9155, "step": 3304 }, { "epoch": 1.4975079293158133, "grad_norm": 0.613619279480082, "learning_rate": 7.068962256697601e-05, "loss": 0.9156, "step": 3305 }, { "epoch": 1.4979610330765745, "grad_norm": 1.1916892834839503, "learning_rate": 7.068150547389646e-05, "loss": 0.8884, "step": 3306 }, { "epoch": 1.4984141368373358, "grad_norm": 1.1479657870176756, "learning_rate": 7.067338531044882e-05, "loss": 0.8908, "step": 3307 }, { "epoch": 1.498867240598097, "grad_norm": 0.8482037381145376, "learning_rate": 7.06652620774457e-05, "loss": 0.8986, "step": 3308 }, { "epoch": 1.4993203443588583, "grad_norm": 1.2246513877557412, "learning_rate": 7.065713577570001e-05, "loss": 0.9151, "step": 3309 }, { "epoch": 1.4997734481196194, "grad_norm": 1.1210284257003345, "learning_rate": 7.064900640602494e-05, "loss": 0.8934, "step": 3310 }, { "epoch": 1.5002265518803806, "grad_norm": 0.7449750589331005, "learning_rate": 7.064087396923403e-05, "loss": 0.898, "step": 3311 }, { "epoch": 1.5006796556411417, "grad_norm": 0.8664187341282109, "learning_rate": 7.063273846614115e-05, "loss": 0.9002, "step": 3312 }, { "epoch": 1.501132759401903, "grad_norm": 1.3257591805242932, "learning_rate": 7.062459989756039e-05, "loss": 0.8899, "step": 3313 }, { "epoch": 1.5015858631626644, "grad_norm": 0.9322895985298765, "learning_rate": 7.061645826430621e-05, "loss": 0.9225, "step": 3314 }, { "epoch": 1.5020389669234255, "grad_norm": 1.0463533081773368, "learning_rate": 7.060831356719338e-05, "loss": 0.9083, "step": 3315 }, { "epoch": 1.5024920706841867, "grad_norm": 0.9314330152345638, "learning_rate": 7.060016580703693e-05, "loss": 0.8815, "step": 3316 }, { "epoch": 1.5029451744449478, "grad_norm": 1.4152090239919997, "learning_rate": 7.059201498465224e-05, "loss": 0.9032, "step": 3317 }, { "epoch": 1.5033982782057091, "grad_norm": 0.6639683665300411, "learning_rate": 7.058386110085497e-05, "loss": 0.8785, "step": 3318 }, { "epoch": 1.5038513819664703, "grad_norm": 0.9742179968295097, "learning_rate": 7.057570415646111e-05, "loss": 0.8954, "step": 3319 }, { "epoch": 1.5043044857272316, "grad_norm": 1.0857313714048253, "learning_rate": 7.056754415228693e-05, "loss": 0.9232, "step": 3320 }, { "epoch": 1.5047575894879928, "grad_norm": 0.8098353417358428, "learning_rate": 7.055938108914905e-05, "loss": 0.897, "step": 3321 }, { "epoch": 1.505210693248754, "grad_norm": 0.7766638518440637, "learning_rate": 7.055121496786433e-05, "loss": 0.9114, "step": 3322 }, { "epoch": 1.505663797009515, "grad_norm": 0.8678227556053785, "learning_rate": 7.054304578925e-05, "loss": 0.8985, "step": 3323 }, { "epoch": 1.5061169007702764, "grad_norm": 0.7519595325275809, "learning_rate": 7.053487355412355e-05, "loss": 0.8953, "step": 3324 }, { "epoch": 1.5065700045310377, "grad_norm": 0.5537511765706019, "learning_rate": 7.05266982633028e-05, "loss": 0.8903, "step": 3325 }, { "epoch": 1.5070231082917989, "grad_norm": 0.705675420390086, "learning_rate": 7.051851991760588e-05, "loss": 0.9059, "step": 3326 }, { "epoch": 1.50747621205256, "grad_norm": 0.7948135184117949, "learning_rate": 7.05103385178512e-05, "loss": 0.901, "step": 3327 }, { "epoch": 1.5079293158133211, "grad_norm": 0.800780216370637, "learning_rate": 7.050215406485751e-05, "loss": 0.8866, "step": 3328 }, { "epoch": 1.5083824195740825, "grad_norm": 0.8215462140183432, "learning_rate": 7.049396655944381e-05, "loss": 0.8999, "step": 3329 }, { "epoch": 1.5088355233348438, "grad_norm": 0.9065038672324, "learning_rate": 7.04857760024295e-05, "loss": 0.8883, "step": 3330 }, { "epoch": 1.509288627095605, "grad_norm": 1.0409901725920876, "learning_rate": 7.047758239463417e-05, "loss": 0.8965, "step": 3331 }, { "epoch": 1.509741730856366, "grad_norm": 1.1333047581956177, "learning_rate": 7.046938573687782e-05, "loss": 0.9021, "step": 3332 }, { "epoch": 1.5101948346171272, "grad_norm": 1.0268059413961985, "learning_rate": 7.046118602998066e-05, "loss": 0.8924, "step": 3333 }, { "epoch": 1.5106479383778886, "grad_norm": 0.9064830460652765, "learning_rate": 7.04529832747633e-05, "loss": 0.8892, "step": 3334 }, { "epoch": 1.5111010421386497, "grad_norm": 0.8994100948002577, "learning_rate": 7.044477747204656e-05, "loss": 0.8977, "step": 3335 }, { "epoch": 1.511554145899411, "grad_norm": 0.7568664760267921, "learning_rate": 7.043656862265166e-05, "loss": 0.8764, "step": 3336 }, { "epoch": 1.5120072496601722, "grad_norm": 0.6625800713977102, "learning_rate": 7.042835672740007e-05, "loss": 0.9119, "step": 3337 }, { "epoch": 1.5124603534209333, "grad_norm": 0.7834337924858503, "learning_rate": 7.042014178711355e-05, "loss": 0.8944, "step": 3338 }, { "epoch": 1.5129134571816945, "grad_norm": 0.959871500950362, "learning_rate": 7.041192380261421e-05, "loss": 0.8948, "step": 3339 }, { "epoch": 1.5133665609424558, "grad_norm": 1.1882843209824308, "learning_rate": 7.040370277472444e-05, "loss": 0.9229, "step": 3340 }, { "epoch": 1.5138196647032172, "grad_norm": 0.8542070406875875, "learning_rate": 7.03954787042669e-05, "loss": 0.8892, "step": 3341 }, { "epoch": 1.5142727684639783, "grad_norm": 0.7798657755241186, "learning_rate": 7.038725159206463e-05, "loss": 0.9313, "step": 3342 }, { "epoch": 1.5147258722247394, "grad_norm": 0.674790323881245, "learning_rate": 7.037902143894094e-05, "loss": 0.8919, "step": 3343 }, { "epoch": 1.5151789759855006, "grad_norm": 0.6667876871557108, "learning_rate": 7.037078824571942e-05, "loss": 0.8956, "step": 3344 }, { "epoch": 1.515632079746262, "grad_norm": 0.6968933245555669, "learning_rate": 7.0362552013224e-05, "loss": 0.8913, "step": 3345 }, { "epoch": 1.5160851835070233, "grad_norm": 0.7261260591361013, "learning_rate": 7.035431274227888e-05, "loss": 0.9005, "step": 3346 }, { "epoch": 1.5165382872677844, "grad_norm": 0.9153926819339302, "learning_rate": 7.034607043370858e-05, "loss": 0.899, "step": 3347 }, { "epoch": 1.5169913910285455, "grad_norm": 1.1326228585192257, "learning_rate": 7.033782508833797e-05, "loss": 0.8804, "step": 3348 }, { "epoch": 1.5174444947893067, "grad_norm": 0.9351187179213047, "learning_rate": 7.032957670699212e-05, "loss": 0.8786, "step": 3349 }, { "epoch": 1.517897598550068, "grad_norm": 0.923756194162044, "learning_rate": 7.032132529049652e-05, "loss": 0.8981, "step": 3350 }, { "epoch": 1.5183507023108291, "grad_norm": 1.038191887109609, "learning_rate": 7.031307083967687e-05, "loss": 0.8996, "step": 3351 }, { "epoch": 1.5188038060715905, "grad_norm": 0.9823487809272572, "learning_rate": 7.030481335535923e-05, "loss": 0.9168, "step": 3352 }, { "epoch": 1.5192569098323516, "grad_norm": 1.1548923193471061, "learning_rate": 7.029655283836992e-05, "loss": 0.8899, "step": 3353 }, { "epoch": 1.5197100135931128, "grad_norm": 0.8655278071183692, "learning_rate": 7.028828928953562e-05, "loss": 0.8959, "step": 3354 }, { "epoch": 1.5201631173538739, "grad_norm": 0.6868179459690534, "learning_rate": 7.028002270968328e-05, "loss": 0.8933, "step": 3355 }, { "epoch": 1.5206162211146352, "grad_norm": 0.6365125449103037, "learning_rate": 7.027175309964014e-05, "loss": 0.8964, "step": 3356 }, { "epoch": 1.5210693248753966, "grad_norm": 0.7430691618022832, "learning_rate": 7.026348046023376e-05, "loss": 0.9077, "step": 3357 }, { "epoch": 1.5215224286361577, "grad_norm": 0.9805392168043618, "learning_rate": 7.025520479229201e-05, "loss": 0.8995, "step": 3358 }, { "epoch": 1.5219755323969189, "grad_norm": 1.5808780403267162, "learning_rate": 7.024692609664303e-05, "loss": 0.9089, "step": 3359 }, { "epoch": 1.52242863615768, "grad_norm": 0.5588372738740941, "learning_rate": 7.023864437411533e-05, "loss": 0.8885, "step": 3360 }, { "epoch": 1.5228817399184413, "grad_norm": 1.683949254774955, "learning_rate": 7.023035962553765e-05, "loss": 0.8969, "step": 3361 }, { "epoch": 1.5233348436792027, "grad_norm": 0.5799660770742319, "learning_rate": 7.022207185173908e-05, "loss": 0.8941, "step": 3362 }, { "epoch": 1.5237879474399638, "grad_norm": 1.2443652741482045, "learning_rate": 7.021378105354899e-05, "loss": 0.8941, "step": 3363 }, { "epoch": 1.524241051200725, "grad_norm": 0.8957739435906127, "learning_rate": 7.020548723179707e-05, "loss": 0.9088, "step": 3364 }, { "epoch": 1.524694154961486, "grad_norm": 1.1315196883367824, "learning_rate": 7.019719038731326e-05, "loss": 0.9081, "step": 3365 }, { "epoch": 1.5251472587222474, "grad_norm": 1.0066229754535483, "learning_rate": 7.01888905209279e-05, "loss": 0.8886, "step": 3366 }, { "epoch": 1.5256003624830086, "grad_norm": 1.2878024710262157, "learning_rate": 7.018058763347155e-05, "loss": 0.8824, "step": 3367 }, { "epoch": 1.52605346624377, "grad_norm": 0.6444926802968852, "learning_rate": 7.017228172577508e-05, "loss": 0.9091, "step": 3368 }, { "epoch": 1.526506570004531, "grad_norm": 1.152350337767983, "learning_rate": 7.016397279866973e-05, "loss": 0.8898, "step": 3369 }, { "epoch": 1.5269596737652922, "grad_norm": 1.3487349416661307, "learning_rate": 7.015566085298693e-05, "loss": 0.885, "step": 3370 }, { "epoch": 1.5274127775260533, "grad_norm": 0.7094390083015316, "learning_rate": 7.014734588955852e-05, "loss": 0.8892, "step": 3371 }, { "epoch": 1.5278658812868147, "grad_norm": 1.0740136163600404, "learning_rate": 7.01390279092166e-05, "loss": 0.8865, "step": 3372 }, { "epoch": 1.528318985047576, "grad_norm": 1.2629751581187985, "learning_rate": 7.013070691279356e-05, "loss": 0.9042, "step": 3373 }, { "epoch": 1.5287720888083371, "grad_norm": 0.9638594415495758, "learning_rate": 7.012238290112208e-05, "loss": 0.8855, "step": 3374 }, { "epoch": 1.5292251925690983, "grad_norm": 0.8186592939867153, "learning_rate": 7.011405587503519e-05, "loss": 0.8818, "step": 3375 }, { "epoch": 1.5296782963298594, "grad_norm": 0.6599152314697402, "learning_rate": 7.010572583536617e-05, "loss": 0.9005, "step": 3376 }, { "epoch": 1.5301314000906208, "grad_norm": 0.6401743396203744, "learning_rate": 7.009739278294865e-05, "loss": 0.9053, "step": 3377 }, { "epoch": 1.5305845038513821, "grad_norm": 0.6681954857660707, "learning_rate": 7.008905671861654e-05, "loss": 0.8849, "step": 3378 }, { "epoch": 1.5310376076121432, "grad_norm": 0.8651980044459626, "learning_rate": 7.008071764320401e-05, "loss": 0.8786, "step": 3379 }, { "epoch": 1.5314907113729044, "grad_norm": 1.070626381142338, "learning_rate": 7.007237555754561e-05, "loss": 0.8818, "step": 3380 }, { "epoch": 1.5319438151336655, "grad_norm": 1.021149446136332, "learning_rate": 7.006403046247615e-05, "loss": 0.8832, "step": 3381 }, { "epoch": 1.5323969188944269, "grad_norm": 1.1423733770518736, "learning_rate": 7.005568235883073e-05, "loss": 0.9148, "step": 3382 }, { "epoch": 1.532850022655188, "grad_norm": 0.8768610798811475, "learning_rate": 7.004733124744475e-05, "loss": 0.8889, "step": 3383 }, { "epoch": 1.5333031264159493, "grad_norm": 0.7385390492522523, "learning_rate": 7.003897712915394e-05, "loss": 0.9147, "step": 3384 }, { "epoch": 1.5337562301767105, "grad_norm": 0.6385477873933806, "learning_rate": 7.003062000479432e-05, "loss": 0.9091, "step": 3385 }, { "epoch": 1.5342093339374716, "grad_norm": 0.6449008798623395, "learning_rate": 7.00222598752022e-05, "loss": 0.8991, "step": 3386 }, { "epoch": 1.5346624376982327, "grad_norm": 0.6336749507518524, "learning_rate": 7.00138967412142e-05, "loss": 0.9221, "step": 3387 }, { "epoch": 1.535115541458994, "grad_norm": 0.5907933378887937, "learning_rate": 7.000553060366724e-05, "loss": 0.9034, "step": 3388 }, { "epoch": 1.5355686452197554, "grad_norm": 0.6400381015284416, "learning_rate": 6.999716146339853e-05, "loss": 0.9045, "step": 3389 }, { "epoch": 1.5360217489805166, "grad_norm": 0.8127779082572151, "learning_rate": 6.998878932124561e-05, "loss": 0.9062, "step": 3390 }, { "epoch": 1.5364748527412777, "grad_norm": 1.0876426079627992, "learning_rate": 6.998041417804626e-05, "loss": 0.892, "step": 3391 }, { "epoch": 1.5369279565020388, "grad_norm": 0.9844373469530302, "learning_rate": 6.997203603463864e-05, "loss": 0.9076, "step": 3392 }, { "epoch": 1.5373810602628002, "grad_norm": 1.034878625553893, "learning_rate": 6.996365489186114e-05, "loss": 0.8834, "step": 3393 }, { "epoch": 1.5378341640235615, "grad_norm": 1.0625568169353223, "learning_rate": 6.99552707505525e-05, "loss": 0.8825, "step": 3394 }, { "epoch": 1.5382872677843227, "grad_norm": 0.7878569053111423, "learning_rate": 6.994688361155174e-05, "loss": 0.8951, "step": 3395 }, { "epoch": 1.5387403715450838, "grad_norm": 0.5848994697660591, "learning_rate": 6.993849347569816e-05, "loss": 0.8923, "step": 3396 }, { "epoch": 1.539193475305845, "grad_norm": 0.5412674199622064, "learning_rate": 6.99301003438314e-05, "loss": 0.8979, "step": 3397 }, { "epoch": 1.5396465790666063, "grad_norm": 0.5211369764832144, "learning_rate": 6.992170421679138e-05, "loss": 0.8803, "step": 3398 }, { "epoch": 1.5400996828273674, "grad_norm": 0.4969659134068957, "learning_rate": 6.99133050954183e-05, "loss": 0.9004, "step": 3399 }, { "epoch": 1.5405527865881288, "grad_norm": 0.48644232656788017, "learning_rate": 6.990490298055269e-05, "loss": 0.9135, "step": 3400 }, { "epoch": 1.54100589034889, "grad_norm": 0.49676532069544377, "learning_rate": 6.989649787303538e-05, "loss": 0.8721, "step": 3401 }, { "epoch": 1.541458994109651, "grad_norm": 0.6797331875066164, "learning_rate": 6.988808977370748e-05, "loss": 0.9205, "step": 3402 }, { "epoch": 1.5419120978704122, "grad_norm": 0.8871181509517331, "learning_rate": 6.98796786834104e-05, "loss": 0.8964, "step": 3403 }, { "epoch": 1.5423652016311735, "grad_norm": 1.142305881442712, "learning_rate": 6.987126460298587e-05, "loss": 0.8845, "step": 3404 }, { "epoch": 1.5428183053919349, "grad_norm": 1.0202077048892852, "learning_rate": 6.986284753327589e-05, "loss": 0.9084, "step": 3405 }, { "epoch": 1.543271409152696, "grad_norm": 1.0720051032881408, "learning_rate": 6.985442747512279e-05, "loss": 0.9026, "step": 3406 }, { "epoch": 1.5437245129134571, "grad_norm": 0.8756801500720297, "learning_rate": 6.984600442936918e-05, "loss": 0.89, "step": 3407 }, { "epoch": 1.5441776166742183, "grad_norm": 0.837931114187693, "learning_rate": 6.983757839685797e-05, "loss": 0.8993, "step": 3408 }, { "epoch": 1.5446307204349796, "grad_norm": 0.843367234024653, "learning_rate": 6.982914937843238e-05, "loss": 0.9007, "step": 3409 }, { "epoch": 1.545083824195741, "grad_norm": 0.9454478098482529, "learning_rate": 6.982071737493591e-05, "loss": 0.8873, "step": 3410 }, { "epoch": 1.545536927956502, "grad_norm": 1.104707467295669, "learning_rate": 6.981228238721237e-05, "loss": 0.897, "step": 3411 }, { "epoch": 1.5459900317172632, "grad_norm": 0.9548191527744149, "learning_rate": 6.980384441610588e-05, "loss": 0.892, "step": 3412 }, { "epoch": 1.5464431354780244, "grad_norm": 0.8508676151393817, "learning_rate": 6.979540346246085e-05, "loss": 0.9109, "step": 3413 }, { "epoch": 1.5468962392387857, "grad_norm": 0.9258737125679659, "learning_rate": 6.978695952712197e-05, "loss": 0.9028, "step": 3414 }, { "epoch": 1.5473493429995469, "grad_norm": 1.0046450333261552, "learning_rate": 6.977851261093425e-05, "loss": 0.8926, "step": 3415 }, { "epoch": 1.5478024467603082, "grad_norm": 1.1166004875285322, "learning_rate": 6.977006271474299e-05, "loss": 0.8938, "step": 3416 }, { "epoch": 1.5482555505210693, "grad_norm": 0.7239478092124956, "learning_rate": 6.97616098393938e-05, "loss": 0.9013, "step": 3417 }, { "epoch": 1.5487086542818305, "grad_norm": 0.4765514860377708, "learning_rate": 6.975315398573257e-05, "loss": 0.8792, "step": 3418 }, { "epoch": 1.5491617580425916, "grad_norm": 0.3926069811597738, "learning_rate": 6.974469515460548e-05, "loss": 0.9016, "step": 3419 }, { "epoch": 1.549614861803353, "grad_norm": 0.47417626680109703, "learning_rate": 6.973623334685906e-05, "loss": 0.8974, "step": 3420 }, { "epoch": 1.5500679655641143, "grad_norm": 0.643947318052294, "learning_rate": 6.972776856334008e-05, "loss": 0.8934, "step": 3421 }, { "epoch": 1.5505210693248754, "grad_norm": 0.9829160506398852, "learning_rate": 6.971930080489564e-05, "loss": 0.9014, "step": 3422 }, { "epoch": 1.5509741730856366, "grad_norm": 1.3894211583463807, "learning_rate": 6.971083007237312e-05, "loss": 0.8817, "step": 3423 }, { "epoch": 1.5514272768463977, "grad_norm": 0.4073444393395162, "learning_rate": 6.97023563666202e-05, "loss": 0.8871, "step": 3424 }, { "epoch": 1.551880380607159, "grad_norm": 0.9114691363329962, "learning_rate": 6.969387968848487e-05, "loss": 0.9153, "step": 3425 }, { "epoch": 1.5523334843679204, "grad_norm": 1.6984880921405183, "learning_rate": 6.968540003881542e-05, "loss": 0.915, "step": 3426 }, { "epoch": 1.5527865881286815, "grad_norm": 0.5704242632332651, "learning_rate": 6.967691741846043e-05, "loss": 0.8895, "step": 3427 }, { "epoch": 1.5532396918894427, "grad_norm": 1.9488529022470913, "learning_rate": 6.966843182826875e-05, "loss": 0.9188, "step": 3428 }, { "epoch": 1.5536927956502038, "grad_norm": 0.8432058617002056, "learning_rate": 6.965994326908956e-05, "loss": 0.9197, "step": 3429 }, { "epoch": 1.5541458994109651, "grad_norm": 2.2998411423902727, "learning_rate": 6.965145174177234e-05, "loss": 0.9062, "step": 3430 }, { "epoch": 1.5545990031717263, "grad_norm": 1.88434741471113, "learning_rate": 6.964295724716688e-05, "loss": 0.8849, "step": 3431 }, { "epoch": 1.5550521069324876, "grad_norm": 1.5523432929825207, "learning_rate": 6.963445978612319e-05, "loss": 0.9043, "step": 3432 }, { "epoch": 1.5555052106932488, "grad_norm": 1.5725943583603217, "learning_rate": 6.962595935949165e-05, "loss": 0.9082, "step": 3433 }, { "epoch": 1.55595831445401, "grad_norm": 1.2280352660112122, "learning_rate": 6.961745596812294e-05, "loss": 0.9077, "step": 3434 }, { "epoch": 1.556411418214771, "grad_norm": 1.3955277272816764, "learning_rate": 6.9608949612868e-05, "loss": 0.893, "step": 3435 }, { "epoch": 1.5568645219755324, "grad_norm": 0.8212467554933703, "learning_rate": 6.960044029457807e-05, "loss": 0.898, "step": 3436 }, { "epoch": 1.5573176257362937, "grad_norm": 1.5339425562642202, "learning_rate": 6.95919280141047e-05, "loss": 0.8891, "step": 3437 }, { "epoch": 1.5577707294970549, "grad_norm": 0.8685092025411969, "learning_rate": 6.958341277229974e-05, "loss": 0.8955, "step": 3438 }, { "epoch": 1.558223833257816, "grad_norm": 1.5134503755467779, "learning_rate": 6.957489457001534e-05, "loss": 0.9342, "step": 3439 }, { "epoch": 1.5586769370185771, "grad_norm": 1.1624816327068455, "learning_rate": 6.956637340810391e-05, "loss": 0.9034, "step": 3440 }, { "epoch": 1.5591300407793385, "grad_norm": 1.278890369852667, "learning_rate": 6.95578492874182e-05, "loss": 0.8962, "step": 3441 }, { "epoch": 1.5595831445400998, "grad_norm": 1.3217402282754989, "learning_rate": 6.954932220881122e-05, "loss": 0.8964, "step": 3442 }, { "epoch": 1.560036248300861, "grad_norm": 0.8326349774758334, "learning_rate": 6.954079217313632e-05, "loss": 0.9111, "step": 3443 }, { "epoch": 1.560489352061622, "grad_norm": 1.3198657974654977, "learning_rate": 6.953225918124711e-05, "loss": 0.8793, "step": 3444 }, { "epoch": 1.5609424558223832, "grad_norm": 0.72368358700592, "learning_rate": 6.95237232339975e-05, "loss": 0.8874, "step": 3445 }, { "epoch": 1.5613955595831446, "grad_norm": 1.0670845715658315, "learning_rate": 6.95151843322417e-05, "loss": 0.9141, "step": 3446 }, { "epoch": 1.5618486633439057, "grad_norm": 0.9631451251891258, "learning_rate": 6.950664247683423e-05, "loss": 0.8893, "step": 3447 }, { "epoch": 1.562301767104667, "grad_norm": 0.865359357313757, "learning_rate": 6.949809766862989e-05, "loss": 0.8915, "step": 3448 }, { "epoch": 1.5627548708654282, "grad_norm": 0.9558652147211348, "learning_rate": 6.948954990848376e-05, "loss": 0.8949, "step": 3449 }, { "epoch": 1.5632079746261893, "grad_norm": 0.8188606329120686, "learning_rate": 6.948099919725126e-05, "loss": 0.904, "step": 3450 }, { "epoch": 1.5636610783869505, "grad_norm": 0.7599140743685551, "learning_rate": 6.947244553578806e-05, "loss": 0.9157, "step": 3451 }, { "epoch": 1.5641141821477118, "grad_norm": 0.7744475968400465, "learning_rate": 6.946388892495016e-05, "loss": 0.8902, "step": 3452 }, { "epoch": 1.5645672859084732, "grad_norm": 0.8163701410930587, "learning_rate": 6.945532936559383e-05, "loss": 0.8907, "step": 3453 }, { "epoch": 1.5650203896692343, "grad_norm": 0.7590393577079584, "learning_rate": 6.944676685857564e-05, "loss": 0.8902, "step": 3454 }, { "epoch": 1.5654734934299954, "grad_norm": 0.7665672020024272, "learning_rate": 6.943820140475246e-05, "loss": 0.8969, "step": 3455 }, { "epoch": 1.5659265971907566, "grad_norm": 0.8573685262071652, "learning_rate": 6.942963300498147e-05, "loss": 0.8886, "step": 3456 }, { "epoch": 1.566379700951518, "grad_norm": 0.901658997491469, "learning_rate": 6.942106166012011e-05, "loss": 0.8958, "step": 3457 }, { "epoch": 1.5668328047122793, "grad_norm": 0.9116514560390151, "learning_rate": 6.941248737102614e-05, "loss": 0.8889, "step": 3458 }, { "epoch": 1.5672859084730404, "grad_norm": 0.8478943905065546, "learning_rate": 6.940391013855762e-05, "loss": 0.8998, "step": 3459 }, { "epoch": 1.5677390122338015, "grad_norm": 0.8463719947592555, "learning_rate": 6.939532996357289e-05, "loss": 0.8948, "step": 3460 }, { "epoch": 1.5681921159945627, "grad_norm": 0.9474632009289496, "learning_rate": 6.938674684693057e-05, "loss": 0.8956, "step": 3461 }, { "epoch": 1.568645219755324, "grad_norm": 1.092453679549176, "learning_rate": 6.93781607894896e-05, "loss": 0.8961, "step": 3462 }, { "epoch": 1.5690983235160851, "grad_norm": 0.9039432947961877, "learning_rate": 6.936957179210922e-05, "loss": 0.8916, "step": 3463 }, { "epoch": 1.5695514272768465, "grad_norm": 0.71190847326076, "learning_rate": 6.936097985564894e-05, "loss": 0.9047, "step": 3464 }, { "epoch": 1.5700045310376076, "grad_norm": 0.5529472268759954, "learning_rate": 6.935238498096858e-05, "loss": 0.9022, "step": 3465 }, { "epoch": 1.5704576347983688, "grad_norm": 0.5639704228024494, "learning_rate": 6.934378716892824e-05, "loss": 0.9027, "step": 3466 }, { "epoch": 1.5709107385591299, "grad_norm": 0.6161843870701832, "learning_rate": 6.933518642038832e-05, "loss": 0.8831, "step": 3467 }, { "epoch": 1.5713638423198912, "grad_norm": 0.5729904625052932, "learning_rate": 6.932658273620954e-05, "loss": 0.9008, "step": 3468 }, { "epoch": 1.5718169460806526, "grad_norm": 0.4474830975472187, "learning_rate": 6.931797611725285e-05, "loss": 0.8824, "step": 3469 }, { "epoch": 1.5722700498414137, "grad_norm": 0.4427376642219476, "learning_rate": 6.930936656437958e-05, "loss": 0.9091, "step": 3470 }, { "epoch": 1.5727231536021749, "grad_norm": 0.5305147363399897, "learning_rate": 6.930075407845127e-05, "loss": 0.9138, "step": 3471 }, { "epoch": 1.573176257362936, "grad_norm": 0.49933511241470413, "learning_rate": 6.929213866032981e-05, "loss": 0.8849, "step": 3472 }, { "epoch": 1.5736293611236973, "grad_norm": 0.6838552818300394, "learning_rate": 6.928352031087737e-05, "loss": 0.8929, "step": 3473 }, { "epoch": 1.5740824648844587, "grad_norm": 0.9023895829294115, "learning_rate": 6.92748990309564e-05, "loss": 0.9044, "step": 3474 }, { "epoch": 1.5745355686452198, "grad_norm": 1.1316413339614748, "learning_rate": 6.926627482142965e-05, "loss": 0.901, "step": 3475 }, { "epoch": 1.574988672405981, "grad_norm": 1.139050122132923, "learning_rate": 6.925764768316015e-05, "loss": 0.9232, "step": 3476 }, { "epoch": 1.575441776166742, "grad_norm": 0.958271583129145, "learning_rate": 6.924901761701124e-05, "loss": 0.8821, "step": 3477 }, { "epoch": 1.5758948799275034, "grad_norm": 0.8500088989702472, "learning_rate": 6.924038462384659e-05, "loss": 0.8694, "step": 3478 }, { "epoch": 1.5763479836882646, "grad_norm": 0.6947853822984605, "learning_rate": 6.923174870453008e-05, "loss": 0.8797, "step": 3479 }, { "epoch": 1.576801087449026, "grad_norm": 0.5365081678777683, "learning_rate": 6.922310985992594e-05, "loss": 0.8939, "step": 3480 }, { "epoch": 1.577254191209787, "grad_norm": 0.521498614493454, "learning_rate": 6.921446809089869e-05, "loss": 0.9036, "step": 3481 }, { "epoch": 1.5777072949705482, "grad_norm": 0.5596461128922123, "learning_rate": 6.920582339831311e-05, "loss": 0.8961, "step": 3482 }, { "epoch": 1.5781603987313093, "grad_norm": 0.6497969021487037, "learning_rate": 6.919717578303432e-05, "loss": 0.8853, "step": 3483 }, { "epoch": 1.5786135024920707, "grad_norm": 0.8373076641823668, "learning_rate": 6.918852524592769e-05, "loss": 0.8946, "step": 3484 }, { "epoch": 1.579066606252832, "grad_norm": 0.9869931761245858, "learning_rate": 6.917987178785889e-05, "loss": 0.9037, "step": 3485 }, { "epoch": 1.5795197100135931, "grad_norm": 1.0481755816570983, "learning_rate": 6.91712154096939e-05, "loss": 0.8751, "step": 3486 }, { "epoch": 1.5799728137743543, "grad_norm": 1.1235890661480339, "learning_rate": 6.916255611229899e-05, "loss": 0.885, "step": 3487 }, { "epoch": 1.5804259175351154, "grad_norm": 0.9507095696785681, "learning_rate": 6.915389389654072e-05, "loss": 0.8961, "step": 3488 }, { "epoch": 1.5808790212958768, "grad_norm": 0.9036453334147057, "learning_rate": 6.914522876328592e-05, "loss": 0.9085, "step": 3489 }, { "epoch": 1.5813321250566381, "grad_norm": 0.8025541433438265, "learning_rate": 6.913656071340175e-05, "loss": 0.9117, "step": 3490 }, { "epoch": 1.5817852288173992, "grad_norm": 0.6747561396305085, "learning_rate": 6.91278897477556e-05, "loss": 0.8865, "step": 3491 }, { "epoch": 1.5822383325781604, "grad_norm": 0.5541102589866065, "learning_rate": 6.911921586721526e-05, "loss": 0.8769, "step": 3492 }, { "epoch": 1.5826914363389215, "grad_norm": 0.7255727057419836, "learning_rate": 6.91105390726487e-05, "loss": 0.8976, "step": 3493 }, { "epoch": 1.5831445400996829, "grad_norm": 0.45838868155041274, "learning_rate": 6.910185936492422e-05, "loss": 0.8854, "step": 3494 }, { "epoch": 1.583597643860444, "grad_norm": 0.5655378479896495, "learning_rate": 6.909317674491044e-05, "loss": 0.889, "step": 3495 }, { "epoch": 1.5840507476212053, "grad_norm": 0.971882974420061, "learning_rate": 6.908449121347624e-05, "loss": 0.8986, "step": 3496 }, { "epoch": 1.5845038513819665, "grad_norm": 0.9161781019356826, "learning_rate": 6.90758027714908e-05, "loss": 0.8855, "step": 3497 }, { "epoch": 1.5849569551427276, "grad_norm": 1.1431722731550675, "learning_rate": 6.906711141982361e-05, "loss": 0.9199, "step": 3498 }, { "epoch": 1.5854100589034887, "grad_norm": 1.0739887213450006, "learning_rate": 6.905841715934441e-05, "loss": 0.8982, "step": 3499 }, { "epoch": 1.58586316266425, "grad_norm": 0.9036917844088225, "learning_rate": 6.904971999092327e-05, "loss": 0.887, "step": 3500 }, { "epoch": 1.5863162664250114, "grad_norm": 0.8730515751658392, "learning_rate": 6.904101991543052e-05, "loss": 0.9159, "step": 3501 }, { "epoch": 1.5867693701857726, "grad_norm": 0.8427841997016255, "learning_rate": 6.903231693373681e-05, "loss": 0.9083, "step": 3502 }, { "epoch": 1.5872224739465337, "grad_norm": 0.8021829840259671, "learning_rate": 6.902361104671306e-05, "loss": 0.8994, "step": 3503 }, { "epoch": 1.5876755777072948, "grad_norm": 0.8340825190655512, "learning_rate": 6.901490225523049e-05, "loss": 0.8866, "step": 3504 }, { "epoch": 1.5881286814680562, "grad_norm": 0.9598194957519738, "learning_rate": 6.900619056016061e-05, "loss": 0.9139, "step": 3505 }, { "epoch": 1.5885817852288175, "grad_norm": 1.218473600962941, "learning_rate": 6.899747596237522e-05, "loss": 0.9003, "step": 3506 }, { "epoch": 1.5890348889895787, "grad_norm": 0.8380171075792581, "learning_rate": 6.898875846274638e-05, "loss": 0.904, "step": 3507 }, { "epoch": 1.5894879927503398, "grad_norm": 0.6341773131930805, "learning_rate": 6.898003806214654e-05, "loss": 0.8839, "step": 3508 }, { "epoch": 1.589941096511101, "grad_norm": 0.5578458168613567, "learning_rate": 6.897131476144831e-05, "loss": 0.8939, "step": 3509 }, { "epoch": 1.5903942002718623, "grad_norm": 0.6204494133375175, "learning_rate": 6.896258856152467e-05, "loss": 0.8912, "step": 3510 }, { "epoch": 1.5908473040326234, "grad_norm": 0.9304657578831995, "learning_rate": 6.895385946324886e-05, "loss": 0.9082, "step": 3511 }, { "epoch": 1.5913004077933848, "grad_norm": 1.1428767759252194, "learning_rate": 6.894512746749444e-05, "loss": 0.8896, "step": 3512 }, { "epoch": 1.591753511554146, "grad_norm": 0.9871817667097879, "learning_rate": 6.893639257513523e-05, "loss": 0.9096, "step": 3513 }, { "epoch": 1.592206615314907, "grad_norm": 1.0668081683968846, "learning_rate": 6.892765478704536e-05, "loss": 0.9026, "step": 3514 }, { "epoch": 1.5926597190756682, "grad_norm": 0.8681328797491151, "learning_rate": 6.891891410409921e-05, "loss": 0.9156, "step": 3515 }, { "epoch": 1.5931128228364295, "grad_norm": 0.7976442211400667, "learning_rate": 6.891017052717151e-05, "loss": 0.883, "step": 3516 }, { "epoch": 1.5935659265971909, "grad_norm": 0.6652167843194743, "learning_rate": 6.890142405713725e-05, "loss": 0.9104, "step": 3517 }, { "epoch": 1.594019030357952, "grad_norm": 0.627402310552794, "learning_rate": 6.88926746948717e-05, "loss": 0.9081, "step": 3518 }, { "epoch": 1.5944721341187131, "grad_norm": 0.6342134015584443, "learning_rate": 6.888392244125042e-05, "loss": 0.9079, "step": 3519 }, { "epoch": 1.5949252378794743, "grad_norm": 0.6746797338848383, "learning_rate": 6.887516729714929e-05, "loss": 0.9018, "step": 3520 }, { "epoch": 1.5953783416402356, "grad_norm": 0.7152932831626311, "learning_rate": 6.886640926344442e-05, "loss": 0.8965, "step": 3521 }, { "epoch": 1.595831445400997, "grad_norm": 0.7790404659542389, "learning_rate": 6.88576483410123e-05, "loss": 0.9032, "step": 3522 }, { "epoch": 1.596284549161758, "grad_norm": 0.8798370611525538, "learning_rate": 6.88488845307296e-05, "loss": 0.8885, "step": 3523 }, { "epoch": 1.5967376529225192, "grad_norm": 0.8894618107931617, "learning_rate": 6.884011783347336e-05, "loss": 0.8921, "step": 3524 }, { "epoch": 1.5971907566832804, "grad_norm": 1.016065366307064, "learning_rate": 6.883134825012091e-05, "loss": 0.8948, "step": 3525 }, { "epoch": 1.5976438604440417, "grad_norm": 1.1980474183002199, "learning_rate": 6.88225757815498e-05, "loss": 0.8796, "step": 3526 }, { "epoch": 1.5980969642048029, "grad_norm": 0.8426250483844328, "learning_rate": 6.881380042863793e-05, "loss": 0.9096, "step": 3527 }, { "epoch": 1.5985500679655642, "grad_norm": 0.7701702954872676, "learning_rate": 6.880502219226346e-05, "loss": 0.8978, "step": 3528 }, { "epoch": 1.5990031717263253, "grad_norm": 0.6001760131979587, "learning_rate": 6.879624107330486e-05, "loss": 0.9131, "step": 3529 }, { "epoch": 1.5994562754870865, "grad_norm": 0.48947163747711353, "learning_rate": 6.878745707264088e-05, "loss": 0.8908, "step": 3530 }, { "epoch": 1.5999093792478476, "grad_norm": 0.5334486477479601, "learning_rate": 6.877867019115054e-05, "loss": 0.9095, "step": 3531 }, { "epoch": 1.600362483008609, "grad_norm": 0.6854231654444036, "learning_rate": 6.876988042971315e-05, "loss": 0.9023, "step": 3532 }, { "epoch": 1.6008155867693703, "grad_norm": 0.781893193767191, "learning_rate": 6.876108778920836e-05, "loss": 0.9085, "step": 3533 }, { "epoch": 1.6012686905301314, "grad_norm": 1.0228806802101238, "learning_rate": 6.875229227051604e-05, "loss": 0.9096, "step": 3534 }, { "epoch": 1.6017217942908926, "grad_norm": 1.3582271984736785, "learning_rate": 6.87434938745164e-05, "loss": 0.8986, "step": 3535 }, { "epoch": 1.6021748980516537, "grad_norm": 0.49794113607579954, "learning_rate": 6.87346926020899e-05, "loss": 0.9051, "step": 3536 }, { "epoch": 1.602628001812415, "grad_norm": 0.7164821695170884, "learning_rate": 6.87258884541173e-05, "loss": 0.8941, "step": 3537 }, { "epoch": 1.6030811055731764, "grad_norm": 1.2999463690460151, "learning_rate": 6.871708143147966e-05, "loss": 0.8926, "step": 3538 }, { "epoch": 1.6035342093339375, "grad_norm": 0.9839622551505298, "learning_rate": 6.870827153505831e-05, "loss": 0.8988, "step": 3539 }, { "epoch": 1.6039873130946987, "grad_norm": 1.1177491636871069, "learning_rate": 6.869945876573489e-05, "loss": 0.8998, "step": 3540 }, { "epoch": 1.6044404168554598, "grad_norm": 0.8710628625443587, "learning_rate": 6.869064312439132e-05, "loss": 0.9046, "step": 3541 }, { "epoch": 1.6048935206162211, "grad_norm": 0.892246980602926, "learning_rate": 6.868182461190976e-05, "loss": 0.8779, "step": 3542 }, { "epoch": 1.6053466243769823, "grad_norm": 0.9917305853531345, "learning_rate": 6.867300322917275e-05, "loss": 0.9223, "step": 3543 }, { "epoch": 1.6057997281377436, "grad_norm": 0.8888335686059172, "learning_rate": 6.866417897706303e-05, "loss": 0.8986, "step": 3544 }, { "epoch": 1.6062528318985048, "grad_norm": 0.6352883010366969, "learning_rate": 6.86553518564637e-05, "loss": 0.8967, "step": 3545 }, { "epoch": 1.606705935659266, "grad_norm": 0.5460047792338215, "learning_rate": 6.864652186825806e-05, "loss": 0.9124, "step": 3546 }, { "epoch": 1.607159039420027, "grad_norm": 0.694870812819432, "learning_rate": 6.863768901332978e-05, "loss": 0.9097, "step": 3547 }, { "epoch": 1.6076121431807884, "grad_norm": 0.790493788537813, "learning_rate": 6.862885329256277e-05, "loss": 0.8765, "step": 3548 }, { "epoch": 1.6080652469415497, "grad_norm": 0.8848465820116366, "learning_rate": 6.862001470684124e-05, "loss": 0.8905, "step": 3549 }, { "epoch": 1.6085183507023109, "grad_norm": 0.9645180473888139, "learning_rate": 6.861117325704971e-05, "loss": 0.9096, "step": 3550 }, { "epoch": 1.608971454463072, "grad_norm": 1.0044136446485379, "learning_rate": 6.860232894407295e-05, "loss": 0.904, "step": 3551 }, { "epoch": 1.6094245582238331, "grad_norm": 1.099069234191287, "learning_rate": 6.859348176879603e-05, "loss": 0.8839, "step": 3552 }, { "epoch": 1.6098776619845945, "grad_norm": 1.0278941010168525, "learning_rate": 6.85846317321043e-05, "loss": 0.8756, "step": 3553 }, { "epoch": 1.6103307657453558, "grad_norm": 0.999925875008792, "learning_rate": 6.857577883488341e-05, "loss": 0.9007, "step": 3554 }, { "epoch": 1.610783869506117, "grad_norm": 1.101769930819662, "learning_rate": 6.856692307801929e-05, "loss": 0.892, "step": 3555 }, { "epoch": 1.611236973266878, "grad_norm": 0.7496685053533017, "learning_rate": 6.855806446239815e-05, "loss": 0.8925, "step": 3556 }, { "epoch": 1.6116900770276392, "grad_norm": 0.4482992933219128, "learning_rate": 6.854920298890651e-05, "loss": 0.8707, "step": 3557 }, { "epoch": 1.6121431807884006, "grad_norm": 0.3355468921082313, "learning_rate": 6.854033865843112e-05, "loss": 0.8815, "step": 3558 }, { "epoch": 1.6125962845491617, "grad_norm": 0.3785159014710561, "learning_rate": 6.85314714718591e-05, "loss": 0.8746, "step": 3559 }, { "epoch": 1.613049388309923, "grad_norm": 0.542733781048537, "learning_rate": 6.852260143007778e-05, "loss": 0.9009, "step": 3560 }, { "epoch": 1.6135024920706842, "grad_norm": 0.5242572348085032, "learning_rate": 6.851372853397483e-05, "loss": 0.878, "step": 3561 }, { "epoch": 1.6139555958314453, "grad_norm": 1.5222681960648803, "learning_rate": 6.850485278443814e-05, "loss": 0.8842, "step": 3562 }, { "epoch": 1.6144086995922065, "grad_norm": 0.3427583944749791, "learning_rate": 6.849597418235596e-05, "loss": 0.879, "step": 3563 }, { "epoch": 1.6148618033529678, "grad_norm": 6.837089215792265, "learning_rate": 6.848709272861677e-05, "loss": 0.9005, "step": 3564 }, { "epoch": 1.6153149071137292, "grad_norm": 2.2649678865199983, "learning_rate": 6.847820842410938e-05, "loss": 0.9265, "step": 3565 }, { "epoch": 1.6157680108744903, "grad_norm": 0.5149790549629617, "learning_rate": 6.846932126972286e-05, "loss": 0.9034, "step": 3566 }, { "epoch": 1.6162211146352514, "grad_norm": 0.7894379282688089, "learning_rate": 6.846043126634655e-05, "loss": 0.9049, "step": 3567 }, { "epoch": 1.6166742183960126, "grad_norm": 0.9381317854000399, "learning_rate": 6.84515384148701e-05, "loss": 0.8957, "step": 3568 }, { "epoch": 1.617127322156774, "grad_norm": 0.9071542981047086, "learning_rate": 6.844264271618345e-05, "loss": 0.8985, "step": 3569 }, { "epoch": 1.6175804259175353, "grad_norm": 1.2908800304728802, "learning_rate": 6.84337441711768e-05, "loss": 0.9033, "step": 3570 }, { "epoch": 1.6180335296782964, "grad_norm": 0.8770628332564391, "learning_rate": 6.842484278074064e-05, "loss": 0.9007, "step": 3571 }, { "epoch": 1.6184866334390575, "grad_norm": 0.8382461247463775, "learning_rate": 6.841593854576579e-05, "loss": 0.9073, "step": 3572 }, { "epoch": 1.6189397371998187, "grad_norm": 0.903901849542989, "learning_rate": 6.840703146714326e-05, "loss": 0.9062, "step": 3573 }, { "epoch": 1.61939284096058, "grad_norm": 1.0940834678277995, "learning_rate": 6.839812154576445e-05, "loss": 0.8967, "step": 3574 }, { "epoch": 1.6198459447213411, "grad_norm": 1.015359327247719, "learning_rate": 6.838920878252097e-05, "loss": 0.8884, "step": 3575 }, { "epoch": 1.6202990484821025, "grad_norm": 0.9534414316316906, "learning_rate": 6.838029317830476e-05, "loss": 0.897, "step": 3576 }, { "epoch": 1.6207521522428636, "grad_norm": 1.074421650763258, "learning_rate": 6.8371374734008e-05, "loss": 0.8998, "step": 3577 }, { "epoch": 1.6212052560036248, "grad_norm": 0.8489646981670734, "learning_rate": 6.83624534505232e-05, "loss": 0.8985, "step": 3578 }, { "epoch": 1.6216583597643859, "grad_norm": 0.7713345617193064, "learning_rate": 6.835352932874312e-05, "loss": 0.8898, "step": 3579 }, { "epoch": 1.6221114635251472, "grad_norm": 1.0247577064991407, "learning_rate": 6.834460236956084e-05, "loss": 0.9021, "step": 3580 }, { "epoch": 1.6225645672859086, "grad_norm": 1.1231797460194115, "learning_rate": 6.833567257386967e-05, "loss": 0.8919, "step": 3581 }, { "epoch": 1.6230176710466697, "grad_norm": 0.8860782803110929, "learning_rate": 6.832673994256326e-05, "loss": 0.9061, "step": 3582 }, { "epoch": 1.6234707748074308, "grad_norm": 1.0520433800759648, "learning_rate": 6.83178044765355e-05, "loss": 0.8886, "step": 3583 }, { "epoch": 1.623923878568192, "grad_norm": 0.777074985775903, "learning_rate": 6.830886617668058e-05, "loss": 0.9058, "step": 3584 }, { "epoch": 1.6243769823289533, "grad_norm": 0.6002996322919721, "learning_rate": 6.8299925043893e-05, "loss": 0.8755, "step": 3585 }, { "epoch": 1.6248300860897147, "grad_norm": 0.7198384996028465, "learning_rate": 6.82909810790675e-05, "loss": 0.9248, "step": 3586 }, { "epoch": 1.6252831898504758, "grad_norm": 0.610680964246901, "learning_rate": 6.828203428309913e-05, "loss": 0.8918, "step": 3587 }, { "epoch": 1.625736293611237, "grad_norm": 0.5253746637381921, "learning_rate": 6.82730846568832e-05, "loss": 0.8794, "step": 3588 }, { "epoch": 1.626189397371998, "grad_norm": 0.6591734013343589, "learning_rate": 6.826413220131536e-05, "loss": 0.8845, "step": 3589 }, { "epoch": 1.6266425011327594, "grad_norm": 0.709744886469526, "learning_rate": 6.825517691729148e-05, "loss": 0.9031, "step": 3590 }, { "epoch": 1.6270956048935206, "grad_norm": 0.755160041471946, "learning_rate": 6.824621880570771e-05, "loss": 0.888, "step": 3591 }, { "epoch": 1.627548708654282, "grad_norm": 0.909743611624254, "learning_rate": 6.823725786746053e-05, "loss": 0.903, "step": 3592 }, { "epoch": 1.628001812415043, "grad_norm": 1.22553493488482, "learning_rate": 6.82282941034467e-05, "loss": 0.901, "step": 3593 }, { "epoch": 1.6284549161758042, "grad_norm": 0.6652817255446694, "learning_rate": 6.821932751456322e-05, "loss": 0.8863, "step": 3594 }, { "epoch": 1.6289080199365653, "grad_norm": 0.6599327862550087, "learning_rate": 6.82103581017074e-05, "loss": 0.8836, "step": 3595 }, { "epoch": 1.6293611236973267, "grad_norm": 0.6801269833319625, "learning_rate": 6.820138586577685e-05, "loss": 0.9068, "step": 3596 }, { "epoch": 1.629814227458088, "grad_norm": 0.6687941601694739, "learning_rate": 6.81924108076694e-05, "loss": 0.8961, "step": 3597 }, { "epoch": 1.6302673312188491, "grad_norm": 0.9003264788716343, "learning_rate": 6.818343292828325e-05, "loss": 0.897, "step": 3598 }, { "epoch": 1.6307204349796103, "grad_norm": 1.2435023006526045, "learning_rate": 6.817445222851682e-05, "loss": 0.8835, "step": 3599 }, { "epoch": 1.6311735387403714, "grad_norm": 0.8412285011923561, "learning_rate": 6.816546870926881e-05, "loss": 0.9055, "step": 3600 }, { "epoch": 1.6316266425011328, "grad_norm": 0.7229915903813585, "learning_rate": 6.815648237143825e-05, "loss": 0.8749, "step": 3601 }, { "epoch": 1.6320797462618941, "grad_norm": 0.6519520626984064, "learning_rate": 6.81474932159244e-05, "loss": 0.8894, "step": 3602 }, { "epoch": 1.6325328500226552, "grad_norm": 2.0914820690221663, "learning_rate": 6.813850124362684e-05, "loss": 0.9355, "step": 3603 }, { "epoch": 1.6329859537834164, "grad_norm": 0.8470764397923084, "learning_rate": 6.812950645544541e-05, "loss": 0.9103, "step": 3604 }, { "epoch": 1.6334390575441775, "grad_norm": 1.8559156385093085, "learning_rate": 6.812050885228024e-05, "loss": 0.9135, "step": 3605 }, { "epoch": 1.6338921613049389, "grad_norm": 1.3147390238311905, "learning_rate": 6.811150843503175e-05, "loss": 0.9265, "step": 3606 }, { "epoch": 1.6343452650657, "grad_norm": 1.630008918563957, "learning_rate": 6.810250520460063e-05, "loss": 0.9384, "step": 3607 }, { "epoch": 1.6347983688264613, "grad_norm": 1.408928655167699, "learning_rate": 6.809349916188784e-05, "loss": 0.9055, "step": 3608 }, { "epoch": 1.6352514725872225, "grad_norm": 2.298421252906979, "learning_rate": 6.808449030779464e-05, "loss": 0.9077, "step": 3609 }, { "epoch": 1.6357045763479836, "grad_norm": 1.5074518511418336, "learning_rate": 6.807547864322257e-05, "loss": 0.9048, "step": 3610 }, { "epoch": 1.6361576801087447, "grad_norm": 1.489573900846232, "learning_rate": 6.806646416907345e-05, "loss": 0.9076, "step": 3611 }, { "epoch": 1.636610783869506, "grad_norm": 1.7719351395883731, "learning_rate": 6.805744688624938e-05, "loss": 0.9139, "step": 3612 }, { "epoch": 1.6370638876302674, "grad_norm": 1.0630128778725443, "learning_rate": 6.804842679565273e-05, "loss": 0.9386, "step": 3613 }, { "epoch": 1.6375169913910286, "grad_norm": 2.474119863077503, "learning_rate": 6.803940389818618e-05, "loss": 0.9398, "step": 3614 }, { "epoch": 1.6379700951517897, "grad_norm": 1.5194383002246954, "learning_rate": 6.803037819475265e-05, "loss": 0.9452, "step": 3615 }, { "epoch": 1.6384231989125508, "grad_norm": 2.283164883524124, "learning_rate": 6.802134968625538e-05, "loss": 0.9284, "step": 3616 }, { "epoch": 1.6388763026733122, "grad_norm": 1.6955639769240638, "learning_rate": 6.801231837359786e-05, "loss": 0.9388, "step": 3617 }, { "epoch": 1.6393294064340735, "grad_norm": 2.4299932508417097, "learning_rate": 6.800328425768388e-05, "loss": 0.9377, "step": 3618 }, { "epoch": 1.6397825101948347, "grad_norm": 1.3535089223743224, "learning_rate": 6.79942473394175e-05, "loss": 0.9648, "step": 3619 }, { "epoch": 1.6402356139555958, "grad_norm": 1.6565674717850298, "learning_rate": 6.798520761970308e-05, "loss": 0.939, "step": 3620 }, { "epoch": 1.640688717716357, "grad_norm": 1.1653717645645847, "learning_rate": 6.797616509944524e-05, "loss": 0.9213, "step": 3621 }, { "epoch": 1.6411418214771183, "grad_norm": 1.4142774550294392, "learning_rate": 6.796711977954886e-05, "loss": 0.9042, "step": 3622 }, { "epoch": 1.6415949252378794, "grad_norm": 1.1340365254039464, "learning_rate": 6.795807166091916e-05, "loss": 0.9189, "step": 3623 }, { "epoch": 1.6420480289986408, "grad_norm": 1.5239794691734156, "learning_rate": 6.794902074446159e-05, "loss": 0.9221, "step": 3624 }, { "epoch": 1.642501132759402, "grad_norm": 1.329609692395993, "learning_rate": 6.79399670310819e-05, "loss": 0.9344, "step": 3625 }, { "epoch": 1.642954236520163, "grad_norm": 1.6285805283666823, "learning_rate": 6.79309105216861e-05, "loss": 0.9214, "step": 3626 }, { "epoch": 1.6434073402809242, "grad_norm": 1.2499710445493024, "learning_rate": 6.792185121718053e-05, "loss": 0.9256, "step": 3627 }, { "epoch": 1.6438604440416855, "grad_norm": 1.3877555524757288, "learning_rate": 6.791278911847172e-05, "loss": 0.9163, "step": 3628 }, { "epoch": 1.6443135478024469, "grad_norm": 0.9791445593868057, "learning_rate": 6.790372422646662e-05, "loss": 0.927, "step": 3629 }, { "epoch": 1.644766651563208, "grad_norm": 1.5471959061234666, "learning_rate": 6.789465654207228e-05, "loss": 0.9368, "step": 3630 }, { "epoch": 1.6452197553239691, "grad_norm": 1.2245786648384118, "learning_rate": 6.788558606619619e-05, "loss": 0.9194, "step": 3631 }, { "epoch": 1.6456728590847303, "grad_norm": 1.532382517244527, "learning_rate": 6.787651279974602e-05, "loss": 0.9039, "step": 3632 }, { "epoch": 1.6461259628454916, "grad_norm": 1.3268881173378513, "learning_rate": 6.786743674362977e-05, "loss": 0.8964, "step": 3633 }, { "epoch": 1.646579066606253, "grad_norm": 1.2086517290457222, "learning_rate": 6.785835789875569e-05, "loss": 0.9075, "step": 3634 }, { "epoch": 1.647032170367014, "grad_norm": 1.1631330709215852, "learning_rate": 6.784927626603232e-05, "loss": 0.9178, "step": 3635 }, { "epoch": 1.6474852741277752, "grad_norm": 1.2273518295711585, "learning_rate": 6.784019184636849e-05, "loss": 0.9006, "step": 3636 }, { "epoch": 1.6479383778885364, "grad_norm": 0.872408358654909, "learning_rate": 6.783110464067329e-05, "loss": 0.9228, "step": 3637 }, { "epoch": 1.6483914816492977, "grad_norm": 1.6565528457815846, "learning_rate": 6.78220146498561e-05, "loss": 0.9308, "step": 3638 }, { "epoch": 1.6488445854100588, "grad_norm": 0.9738195897997454, "learning_rate": 6.781292187482658e-05, "loss": 0.9035, "step": 3639 }, { "epoch": 1.6492976891708202, "grad_norm": 1.610580617176628, "learning_rate": 6.780382631649467e-05, "loss": 0.9448, "step": 3640 }, { "epoch": 1.6497507929315813, "grad_norm": 1.2138218446283366, "learning_rate": 6.779472797577056e-05, "loss": 0.9255, "step": 3641 }, { "epoch": 1.6502038966923425, "grad_norm": 1.360342783123942, "learning_rate": 6.778562685356477e-05, "loss": 0.9079, "step": 3642 }, { "epoch": 1.6506570004531036, "grad_norm": 1.1977860931184685, "learning_rate": 6.777652295078804e-05, "loss": 0.8894, "step": 3643 }, { "epoch": 1.651110104213865, "grad_norm": 1.2508614051491864, "learning_rate": 6.776741626835144e-05, "loss": 0.8923, "step": 3644 }, { "epoch": 1.6515632079746263, "grad_norm": 1.175119636816322, "learning_rate": 6.77583068071663e-05, "loss": 0.9303, "step": 3645 }, { "epoch": 1.6520163117353874, "grad_norm": 1.1359978100617494, "learning_rate": 6.774919456814421e-05, "loss": 0.8893, "step": 3646 }, { "epoch": 1.6524694154961486, "grad_norm": 1.0086536656579619, "learning_rate": 6.774007955219706e-05, "loss": 0.9082, "step": 3647 }, { "epoch": 1.6529225192569097, "grad_norm": 1.1119846931724393, "learning_rate": 6.7730961760237e-05, "loss": 0.8877, "step": 3648 }, { "epoch": 1.653375623017671, "grad_norm": 1.1161429536592316, "learning_rate": 6.772184119317649e-05, "loss": 0.9053, "step": 3649 }, { "epoch": 1.6538287267784324, "grad_norm": 0.9718077839130882, "learning_rate": 6.771271785192821e-05, "loss": 0.8861, "step": 3650 }, { "epoch": 1.6542818305391935, "grad_norm": 0.9188291800564393, "learning_rate": 6.770359173740521e-05, "loss": 0.9059, "step": 3651 }, { "epoch": 1.6547349342999547, "grad_norm": 0.8733832198860321, "learning_rate": 6.76944628505207e-05, "loss": 0.9022, "step": 3652 }, { "epoch": 1.6551880380607158, "grad_norm": 0.9717519834638781, "learning_rate": 6.768533119218827e-05, "loss": 0.9107, "step": 3653 }, { "epoch": 1.6556411418214771, "grad_norm": 0.8161647106758889, "learning_rate": 6.767619676332173e-05, "loss": 0.9083, "step": 3654 }, { "epoch": 1.6560942455822383, "grad_norm": 0.927032563587589, "learning_rate": 6.766705956483518e-05, "loss": 0.9107, "step": 3655 }, { "epoch": 1.6565473493429996, "grad_norm": 0.9258531838075749, "learning_rate": 6.7657919597643e-05, "loss": 0.9046, "step": 3656 }, { "epoch": 1.6570004531037608, "grad_norm": 0.6489864054391326, "learning_rate": 6.764877686265985e-05, "loss": 0.9121, "step": 3657 }, { "epoch": 1.657453556864522, "grad_norm": 0.5883596347649203, "learning_rate": 6.763963136080067e-05, "loss": 0.8911, "step": 3658 }, { "epoch": 1.657906660625283, "grad_norm": 0.6511927971359812, "learning_rate": 6.763048309298069e-05, "loss": 0.9124, "step": 3659 }, { "epoch": 1.6583597643860444, "grad_norm": 0.5387016394439597, "learning_rate": 6.762133206011535e-05, "loss": 0.9102, "step": 3660 }, { "epoch": 1.6588128681468057, "grad_norm": 0.636318827293877, "learning_rate": 6.761217826312044e-05, "loss": 0.8886, "step": 3661 }, { "epoch": 1.6592659719075669, "grad_norm": 0.8390279701965127, "learning_rate": 6.760302170291202e-05, "loss": 0.8846, "step": 3662 }, { "epoch": 1.659719075668328, "grad_norm": 0.8493078486891416, "learning_rate": 6.759386238040637e-05, "loss": 0.8972, "step": 3663 }, { "epoch": 1.6601721794290891, "grad_norm": 0.8688248145041134, "learning_rate": 6.758470029652011e-05, "loss": 0.9075, "step": 3664 }, { "epoch": 1.6606252831898505, "grad_norm": 0.9588256990306583, "learning_rate": 6.757553545217011e-05, "loss": 0.9002, "step": 3665 }, { "epoch": 1.6610783869506118, "grad_norm": 1.0662540439237864, "learning_rate": 6.756636784827351e-05, "loss": 0.8787, "step": 3666 }, { "epoch": 1.661531490711373, "grad_norm": 0.7945229055229595, "learning_rate": 6.755719748574773e-05, "loss": 0.8909, "step": 3667 }, { "epoch": 1.661984594472134, "grad_norm": 0.569857223492817, "learning_rate": 6.754802436551046e-05, "loss": 0.8873, "step": 3668 }, { "epoch": 1.6624376982328952, "grad_norm": 0.3830709235486501, "learning_rate": 6.753884848847972e-05, "loss": 0.8996, "step": 3669 }, { "epoch": 1.6628908019936566, "grad_norm": 0.3566035751444805, "learning_rate": 6.75296698555737e-05, "loss": 0.8935, "step": 3670 }, { "epoch": 1.6633439057544177, "grad_norm": 0.3618261089785541, "learning_rate": 6.752048846771097e-05, "loss": 0.8958, "step": 3671 }, { "epoch": 1.663797009515179, "grad_norm": 0.40453355799900087, "learning_rate": 6.75113043258103e-05, "loss": 0.8937, "step": 3672 }, { "epoch": 1.6642501132759402, "grad_norm": 0.5146990987696083, "learning_rate": 6.75021174307908e-05, "loss": 0.8893, "step": 3673 }, { "epoch": 1.6647032170367013, "grad_norm": 0.6294467636284425, "learning_rate": 6.749292778357182e-05, "loss": 0.8824, "step": 3674 }, { "epoch": 1.6651563207974625, "grad_norm": 0.6496953455551749, "learning_rate": 6.748373538507298e-05, "loss": 0.8937, "step": 3675 }, { "epoch": 1.6656094245582238, "grad_norm": 0.7814624824546768, "learning_rate": 6.747454023621416e-05, "loss": 0.8889, "step": 3676 }, { "epoch": 1.6660625283189852, "grad_norm": 0.9281623986515087, "learning_rate": 6.746534233791558e-05, "loss": 0.8726, "step": 3677 }, { "epoch": 1.6665156320797463, "grad_norm": 1.0956274064989726, "learning_rate": 6.745614169109767e-05, "loss": 0.8795, "step": 3678 }, { "epoch": 1.6669687358405074, "grad_norm": 0.9147679211522559, "learning_rate": 6.744693829668116e-05, "loss": 0.8884, "step": 3679 }, { "epoch": 1.6674218396012686, "grad_norm": 0.7391417694286008, "learning_rate": 6.743773215558706e-05, "loss": 0.8978, "step": 3680 }, { "epoch": 1.66787494336203, "grad_norm": 0.5615406724840001, "learning_rate": 6.742852326873667e-05, "loss": 0.9074, "step": 3681 }, { "epoch": 1.6683280471227913, "grad_norm": 0.4164513187528811, "learning_rate": 6.74193116370515e-05, "loss": 0.8951, "step": 3682 }, { "epoch": 1.6687811508835524, "grad_norm": 0.3642373441606687, "learning_rate": 6.741009726145343e-05, "loss": 0.883, "step": 3683 }, { "epoch": 1.6692342546443135, "grad_norm": 0.30320592481153597, "learning_rate": 6.740088014286451e-05, "loss": 0.9027, "step": 3684 }, { "epoch": 1.6696873584050747, "grad_norm": 0.34229839957073627, "learning_rate": 6.739166028220716e-05, "loss": 0.9003, "step": 3685 }, { "epoch": 1.670140462165836, "grad_norm": 0.4818963368632765, "learning_rate": 6.738243768040402e-05, "loss": 0.8989, "step": 3686 }, { "epoch": 1.6705935659265974, "grad_norm": 0.5929370555499773, "learning_rate": 6.737321233837802e-05, "loss": 0.9196, "step": 3687 }, { "epoch": 1.6710466696873585, "grad_norm": 0.6459485193147658, "learning_rate": 6.736398425705235e-05, "loss": 0.8941, "step": 3688 }, { "epoch": 1.6714997734481196, "grad_norm": 0.5897058032283434, "learning_rate": 6.735475343735049e-05, "loss": 0.8994, "step": 3689 }, { "epoch": 1.6719528772088808, "grad_norm": 0.4697297198890535, "learning_rate": 6.734551988019618e-05, "loss": 0.8937, "step": 3690 }, { "epoch": 1.672405980969642, "grad_norm": 0.5203767925006226, "learning_rate": 6.733628358651346e-05, "loss": 0.865, "step": 3691 }, { "epoch": 1.6728590847304032, "grad_norm": 0.7794660950827464, "learning_rate": 6.732704455722662e-05, "loss": 0.8932, "step": 3692 }, { "epoch": 1.6733121884911646, "grad_norm": 1.1147699887573925, "learning_rate": 6.731780279326023e-05, "loss": 0.9023, "step": 3693 }, { "epoch": 1.6737652922519257, "grad_norm": 1.0247051354440957, "learning_rate": 6.730855829553914e-05, "loss": 0.9025, "step": 3694 }, { "epoch": 1.6742183960126868, "grad_norm": 1.0034147469017147, "learning_rate": 6.729931106498846e-05, "loss": 0.8993, "step": 3695 }, { "epoch": 1.674671499773448, "grad_norm": 1.0310774192850078, "learning_rate": 6.729006110253358e-05, "loss": 0.8832, "step": 3696 }, { "epoch": 1.6751246035342093, "grad_norm": 0.9479340871328026, "learning_rate": 6.728080840910018e-05, "loss": 0.9049, "step": 3697 }, { "epoch": 1.6755777072949707, "grad_norm": 0.8762602732704422, "learning_rate": 6.727155298561418e-05, "loss": 0.8886, "step": 3698 }, { "epoch": 1.6760308110557318, "grad_norm": 0.8982189655975866, "learning_rate": 6.72622948330018e-05, "loss": 0.9078, "step": 3699 }, { "epoch": 1.676483914816493, "grad_norm": 0.9492228807275845, "learning_rate": 6.72530339521895e-05, "loss": 0.9018, "step": 3700 }, { "epoch": 1.676937018577254, "grad_norm": 0.9967410397203389, "learning_rate": 6.724377034410409e-05, "loss": 0.8756, "step": 3701 }, { "epoch": 1.6773901223380154, "grad_norm": 0.9613916873835391, "learning_rate": 6.723450400967257e-05, "loss": 0.9233, "step": 3702 }, { "epoch": 1.6778432260987768, "grad_norm": 0.9036998904727348, "learning_rate": 6.722523494982223e-05, "loss": 0.8871, "step": 3703 }, { "epoch": 1.678296329859538, "grad_norm": 0.8164320290484371, "learning_rate": 6.721596316548067e-05, "loss": 0.8962, "step": 3704 }, { "epoch": 1.678749433620299, "grad_norm": 0.6992872808951331, "learning_rate": 6.720668865757571e-05, "loss": 0.8799, "step": 3705 }, { "epoch": 1.6792025373810602, "grad_norm": 0.515871745253487, "learning_rate": 6.71974114270355e-05, "loss": 0.8907, "step": 3706 }, { "epoch": 1.6796556411418215, "grad_norm": 0.3744053030236735, "learning_rate": 6.718813147478843e-05, "loss": 0.8929, "step": 3707 }, { "epoch": 1.6801087449025827, "grad_norm": 0.27789016103090985, "learning_rate": 6.717884880176316e-05, "loss": 0.8835, "step": 3708 }, { "epoch": 1.680561848663344, "grad_norm": 0.27791365073430974, "learning_rate": 6.716956340888861e-05, "loss": 0.9035, "step": 3709 }, { "epoch": 1.6810149524241051, "grad_norm": 0.3054792323999052, "learning_rate": 6.716027529709403e-05, "loss": 0.8932, "step": 3710 }, { "epoch": 1.6814680561848663, "grad_norm": 0.3610729495527018, "learning_rate": 6.715098446730885e-05, "loss": 0.8907, "step": 3711 }, { "epoch": 1.6819211599456274, "grad_norm": 0.39160934583355533, "learning_rate": 6.714169092046288e-05, "loss": 0.8956, "step": 3712 }, { "epoch": 1.6823742637063888, "grad_norm": 0.4774711733596979, "learning_rate": 6.713239465748611e-05, "loss": 0.8964, "step": 3713 }, { "epoch": 1.6828273674671501, "grad_norm": 0.543623465700534, "learning_rate": 6.712309567930886e-05, "loss": 0.8855, "step": 3714 }, { "epoch": 1.6832804712279112, "grad_norm": 0.7040422233741136, "learning_rate": 6.711379398686168e-05, "loss": 0.8824, "step": 3715 }, { "epoch": 1.6837335749886724, "grad_norm": 0.9237623781911403, "learning_rate": 6.710448958107541e-05, "loss": 0.8797, "step": 3716 }, { "epoch": 1.6841866787494335, "grad_norm": 1.220977672759251, "learning_rate": 6.70951824628812e-05, "loss": 0.8964, "step": 3717 }, { "epoch": 1.6846397825101949, "grad_norm": 0.6892171662907455, "learning_rate": 6.708587263321039e-05, "loss": 0.8749, "step": 3718 }, { "epoch": 1.6850928862709562, "grad_norm": 0.3689468429434547, "learning_rate": 6.707656009299465e-05, "loss": 0.88, "step": 3719 }, { "epoch": 1.6855459900317173, "grad_norm": 0.5124303425811731, "learning_rate": 6.706724484316592e-05, "loss": 0.8885, "step": 3720 }, { "epoch": 1.6859990937924785, "grad_norm": 1.0281456793858479, "learning_rate": 6.70579268846564e-05, "loss": 0.8759, "step": 3721 }, { "epoch": 1.6864521975532396, "grad_norm": 1.4253455909657111, "learning_rate": 6.704860621839853e-05, "loss": 0.8878, "step": 3722 }, { "epoch": 1.686905301314001, "grad_norm": 0.3238116593448224, "learning_rate": 6.703928284532508e-05, "loss": 0.9012, "step": 3723 }, { "epoch": 1.687358405074762, "grad_norm": 1.4148595373165507, "learning_rate": 6.702995676636904e-05, "loss": 0.888, "step": 3724 }, { "epoch": 1.6878115088355234, "grad_norm": 1.0109323991685, "learning_rate": 6.702062798246371e-05, "loss": 0.9036, "step": 3725 }, { "epoch": 1.6882646125962846, "grad_norm": 0.9609270776222236, "learning_rate": 6.701129649454262e-05, "loss": 0.8785, "step": 3726 }, { "epoch": 1.6887177163570457, "grad_norm": 1.200998100485842, "learning_rate": 6.700196230353962e-05, "loss": 0.869, "step": 3727 }, { "epoch": 1.6891708201178068, "grad_norm": 0.9448319728578055, "learning_rate": 6.699262541038879e-05, "loss": 0.8842, "step": 3728 }, { "epoch": 1.6896239238785682, "grad_norm": 0.7450816107307349, "learning_rate": 6.69832858160245e-05, "loss": 0.8761, "step": 3729 }, { "epoch": 1.6900770276393295, "grad_norm": 0.5487477298193583, "learning_rate": 6.697394352138137e-05, "loss": 0.8992, "step": 3730 }, { "epoch": 1.6905301314000907, "grad_norm": 0.5558924745788719, "learning_rate": 6.696459852739432e-05, "loss": 0.8967, "step": 3731 }, { "epoch": 1.6909832351608518, "grad_norm": 0.5719952406813689, "learning_rate": 6.695525083499852e-05, "loss": 0.8791, "step": 3732 }, { "epoch": 1.691436338921613, "grad_norm": 0.7061984142110167, "learning_rate": 6.694590044512943e-05, "loss": 0.8959, "step": 3733 }, { "epoch": 1.6918894426823743, "grad_norm": 0.9883295012597704, "learning_rate": 6.693654735872273e-05, "loss": 0.8841, "step": 3734 }, { "epoch": 1.6923425464431356, "grad_norm": 1.125699569007161, "learning_rate": 6.692719157671443e-05, "loss": 0.8909, "step": 3735 }, { "epoch": 1.6927956502038968, "grad_norm": 0.9425936244336002, "learning_rate": 6.691783310004077e-05, "loss": 0.9016, "step": 3736 }, { "epoch": 1.693248753964658, "grad_norm": 1.0237805298051093, "learning_rate": 6.69084719296383e-05, "loss": 0.8909, "step": 3737 }, { "epoch": 1.693701857725419, "grad_norm": 1.131055933049995, "learning_rate": 6.689910806644378e-05, "loss": 0.878, "step": 3738 }, { "epoch": 1.6941549614861804, "grad_norm": 0.9342381106704366, "learning_rate": 6.68897415113943e-05, "loss": 0.913, "step": 3739 }, { "epoch": 1.6946080652469415, "grad_norm": 0.8924856527189611, "learning_rate": 6.688037226542717e-05, "loss": 0.8706, "step": 3740 }, { "epoch": 1.6950611690077029, "grad_norm": 0.7153580464625944, "learning_rate": 6.687100032948002e-05, "loss": 0.8953, "step": 3741 }, { "epoch": 1.695514272768464, "grad_norm": 0.4820480326264549, "learning_rate": 6.686162570449068e-05, "loss": 0.8851, "step": 3742 }, { "epoch": 1.6959673765292251, "grad_norm": 0.521901559788853, "learning_rate": 6.685224839139734e-05, "loss": 0.9057, "step": 3743 }, { "epoch": 1.6964204802899863, "grad_norm": 0.5625445260757328, "learning_rate": 6.684286839113837e-05, "loss": 0.8711, "step": 3744 }, { "epoch": 1.6968735840507476, "grad_norm": 0.41835545516506434, "learning_rate": 6.683348570465247e-05, "loss": 0.868, "step": 3745 }, { "epoch": 1.697326687811509, "grad_norm": 0.4776166497930051, "learning_rate": 6.682410033287857e-05, "loss": 0.8713, "step": 3746 }, { "epoch": 1.69777979157227, "grad_norm": 0.5910704815593527, "learning_rate": 6.681471227675589e-05, "loss": 0.8983, "step": 3747 }, { "epoch": 1.6982328953330312, "grad_norm": 0.5470499922918121, "learning_rate": 6.680532153722392e-05, "loss": 0.88, "step": 3748 }, { "epoch": 1.6986859990937924, "grad_norm": 0.5716685516480372, "learning_rate": 6.679592811522243e-05, "loss": 0.9113, "step": 3749 }, { "epoch": 1.6991391028545537, "grad_norm": 0.7989344406429992, "learning_rate": 6.67865320116914e-05, "loss": 0.8841, "step": 3750 }, { "epoch": 1.699592206615315, "grad_norm": 0.9973414238861055, "learning_rate": 6.677713322757115e-05, "loss": 0.8664, "step": 3751 }, { "epoch": 1.7000453103760762, "grad_norm": 1.2782942449127095, "learning_rate": 6.676773176380222e-05, "loss": 0.9104, "step": 3752 }, { "epoch": 1.7004984141368373, "grad_norm": 0.5804385861485809, "learning_rate": 6.675832762132545e-05, "loss": 0.8848, "step": 3753 }, { "epoch": 1.7009515178975985, "grad_norm": 0.6195755668375709, "learning_rate": 6.674892080108193e-05, "loss": 0.887, "step": 3754 }, { "epoch": 1.7014046216583598, "grad_norm": 0.985964508725399, "learning_rate": 6.673951130401303e-05, "loss": 0.8931, "step": 3755 }, { "epoch": 1.701857725419121, "grad_norm": 1.211292261292287, "learning_rate": 6.673009913106035e-05, "loss": 0.8982, "step": 3756 }, { "epoch": 1.7023108291798823, "grad_norm": 0.8268137002478885, "learning_rate": 6.672068428316583e-05, "loss": 0.8858, "step": 3757 }, { "epoch": 1.7027639329406434, "grad_norm": 0.8967757568365443, "learning_rate": 6.671126676127159e-05, "loss": 0.8967, "step": 3758 }, { "epoch": 1.7032170367014046, "grad_norm": 0.8737737674061984, "learning_rate": 6.670184656632011e-05, "loss": 0.8859, "step": 3759 }, { "epoch": 1.7036701404621657, "grad_norm": 0.6566583262069591, "learning_rate": 6.669242369925406e-05, "loss": 0.8911, "step": 3760 }, { "epoch": 1.704123244222927, "grad_norm": 0.4435097344342695, "learning_rate": 6.668299816101639e-05, "loss": 0.9098, "step": 3761 }, { "epoch": 1.7045763479836884, "grad_norm": 0.5689808648182075, "learning_rate": 6.667356995255038e-05, "loss": 0.8915, "step": 3762 }, { "epoch": 1.7050294517444495, "grad_norm": 0.7784635972054635, "learning_rate": 6.666413907479951e-05, "loss": 0.9017, "step": 3763 }, { "epoch": 1.7054825555052107, "grad_norm": 0.7109940131107474, "learning_rate": 6.665470552870755e-05, "loss": 0.8843, "step": 3764 }, { "epoch": 1.7059356592659718, "grad_norm": 0.6246232323646222, "learning_rate": 6.664526931521855e-05, "loss": 0.8827, "step": 3765 }, { "epoch": 1.7063887630267331, "grad_norm": 0.7334511400651871, "learning_rate": 6.663583043527679e-05, "loss": 0.8982, "step": 3766 }, { "epoch": 1.7068418667874945, "grad_norm": 0.9328950911382233, "learning_rate": 6.662638888982684e-05, "loss": 0.8977, "step": 3767 }, { "epoch": 1.7072949705482556, "grad_norm": 1.1623641754694987, "learning_rate": 6.661694467981356e-05, "loss": 0.8942, "step": 3768 }, { "epoch": 1.7077480743090168, "grad_norm": 0.7301716479174466, "learning_rate": 6.660749780618204e-05, "loss": 0.8963, "step": 3769 }, { "epoch": 1.708201178069778, "grad_norm": 0.629159023540157, "learning_rate": 6.659804826987765e-05, "loss": 0.88, "step": 3770 }, { "epoch": 1.7086542818305392, "grad_norm": 0.7552058222183419, "learning_rate": 6.658859607184604e-05, "loss": 0.8933, "step": 3771 }, { "epoch": 1.7091073855913004, "grad_norm": 0.6567212779175771, "learning_rate": 6.657914121303309e-05, "loss": 0.902, "step": 3772 }, { "epoch": 1.7095604893520617, "grad_norm": 0.43646185523546877, "learning_rate": 6.656968369438498e-05, "loss": 0.8986, "step": 3773 }, { "epoch": 1.7100135931128229, "grad_norm": 0.41396214473739545, "learning_rate": 6.656022351684815e-05, "loss": 0.9033, "step": 3774 }, { "epoch": 1.710466696873584, "grad_norm": 0.4838559011545612, "learning_rate": 6.655076068136929e-05, "loss": 0.87, "step": 3775 }, { "epoch": 1.7109198006343451, "grad_norm": 0.5840048493079041, "learning_rate": 6.65412951888954e-05, "loss": 0.9009, "step": 3776 }, { "epoch": 1.7113729043951065, "grad_norm": 0.682121857953136, "learning_rate": 6.653182704037365e-05, "loss": 0.9099, "step": 3777 }, { "epoch": 1.7118260081558678, "grad_norm": 0.6596935165395246, "learning_rate": 6.652235623675159e-05, "loss": 0.8953, "step": 3778 }, { "epoch": 1.712279111916629, "grad_norm": 0.506796685527015, "learning_rate": 6.651288277897698e-05, "loss": 0.8933, "step": 3779 }, { "epoch": 1.71273221567739, "grad_norm": 0.35747550785972765, "learning_rate": 6.650340666799784e-05, "loss": 0.8899, "step": 3780 }, { "epoch": 1.7131853194381512, "grad_norm": 0.5456680742751104, "learning_rate": 6.649392790476247e-05, "loss": 0.886, "step": 3781 }, { "epoch": 1.7136384231989126, "grad_norm": 0.7804407616699013, "learning_rate": 6.648444649021944e-05, "loss": 0.8816, "step": 3782 }, { "epoch": 1.714091526959674, "grad_norm": 0.8093752611888069, "learning_rate": 6.647496242531756e-05, "loss": 0.8969, "step": 3783 }, { "epoch": 1.714544630720435, "grad_norm": 0.6053886678706283, "learning_rate": 6.646547571100591e-05, "loss": 0.8814, "step": 3784 }, { "epoch": 1.7149977344811962, "grad_norm": 0.4409191897347701, "learning_rate": 6.645598634823388e-05, "loss": 0.8899, "step": 3785 }, { "epoch": 1.7154508382419573, "grad_norm": 0.4616614770447869, "learning_rate": 6.644649433795109e-05, "loss": 0.9028, "step": 3786 }, { "epoch": 1.7159039420027187, "grad_norm": 0.6256614975699648, "learning_rate": 6.64369996811074e-05, "loss": 0.8643, "step": 3787 }, { "epoch": 1.7163570457634798, "grad_norm": 0.6759293731490351, "learning_rate": 6.6427502378653e-05, "loss": 0.8847, "step": 3788 }, { "epoch": 1.7168101495242412, "grad_norm": 0.7291236173989774, "learning_rate": 6.641800243153825e-05, "loss": 0.8932, "step": 3789 }, { "epoch": 1.7172632532850023, "grad_norm": 0.9779886618876941, "learning_rate": 6.640849984071388e-05, "loss": 0.8925, "step": 3790 }, { "epoch": 1.7177163570457634, "grad_norm": 1.3762552852617154, "learning_rate": 6.639899460713082e-05, "loss": 0.8899, "step": 3791 }, { "epoch": 1.7181694608065246, "grad_norm": 0.5736737228682732, "learning_rate": 6.638948673174028e-05, "loss": 0.9057, "step": 3792 }, { "epoch": 1.718622564567286, "grad_norm": 0.9025840988197611, "learning_rate": 6.637997621549375e-05, "loss": 0.9159, "step": 3793 }, { "epoch": 1.7190756683280473, "grad_norm": 1.4511239025660596, "learning_rate": 6.637046305934295e-05, "loss": 0.8809, "step": 3794 }, { "epoch": 1.7195287720888084, "grad_norm": 0.8713320320463533, "learning_rate": 6.636094726423988e-05, "loss": 0.8797, "step": 3795 }, { "epoch": 1.7199818758495695, "grad_norm": 1.4198249051455833, "learning_rate": 6.635142883113681e-05, "loss": 0.8947, "step": 3796 }, { "epoch": 1.7204349796103307, "grad_norm": 0.9256137519789993, "learning_rate": 6.634190776098629e-05, "loss": 0.8986, "step": 3797 }, { "epoch": 1.720888083371092, "grad_norm": 1.5649167411550908, "learning_rate": 6.63323840547411e-05, "loss": 0.8861, "step": 3798 }, { "epoch": 1.7213411871318534, "grad_norm": 0.6418914712416719, "learning_rate": 6.63228577133543e-05, "loss": 0.8917, "step": 3799 }, { "epoch": 1.7217942908926145, "grad_norm": 1.6611385304685053, "learning_rate": 6.63133287377792e-05, "loss": 0.8999, "step": 3800 }, { "epoch": 1.7222473946533756, "grad_norm": 0.936208772406714, "learning_rate": 6.630379712896942e-05, "loss": 0.8829, "step": 3801 }, { "epoch": 1.7227004984141367, "grad_norm": 1.7511277921745592, "learning_rate": 6.629426288787878e-05, "loss": 0.8939, "step": 3802 }, { "epoch": 1.723153602174898, "grad_norm": 1.455199171944526, "learning_rate": 6.62847260154614e-05, "loss": 0.8827, "step": 3803 }, { "epoch": 1.7236067059356592, "grad_norm": 1.2943748067134242, "learning_rate": 6.627518651267167e-05, "loss": 0.8848, "step": 3804 }, { "epoch": 1.7240598096964206, "grad_norm": 1.2152442673754467, "learning_rate": 6.62656443804642e-05, "loss": 0.8976, "step": 3805 }, { "epoch": 1.7245129134571817, "grad_norm": 1.2745722606381467, "learning_rate": 6.625609961979394e-05, "loss": 0.9161, "step": 3806 }, { "epoch": 1.7249660172179428, "grad_norm": 0.9695616123289494, "learning_rate": 6.624655223161602e-05, "loss": 0.8961, "step": 3807 }, { "epoch": 1.725419120978704, "grad_norm": 1.627946293655513, "learning_rate": 6.623700221688585e-05, "loss": 0.8907, "step": 3808 }, { "epoch": 1.7258722247394653, "grad_norm": 1.3067588862570307, "learning_rate": 6.622744957655918e-05, "loss": 0.8813, "step": 3809 }, { "epoch": 1.7263253285002267, "grad_norm": 1.30171674831825, "learning_rate": 6.621789431159192e-05, "loss": 0.9049, "step": 3810 }, { "epoch": 1.7267784322609878, "grad_norm": 1.2551789098701924, "learning_rate": 6.620833642294031e-05, "loss": 0.8729, "step": 3811 }, { "epoch": 1.727231536021749, "grad_norm": 0.9890658682152803, "learning_rate": 6.619877591156081e-05, "loss": 0.9243, "step": 3812 }, { "epoch": 1.72768463978251, "grad_norm": 1.1225131292158539, "learning_rate": 6.618921277841017e-05, "loss": 0.8968, "step": 3813 }, { "epoch": 1.7281377435432714, "grad_norm": 1.0830452878660253, "learning_rate": 6.61796470244454e-05, "loss": 0.8862, "step": 3814 }, { "epoch": 1.7285908473040328, "grad_norm": 0.8647797587307924, "learning_rate": 6.617007865062377e-05, "loss": 0.9001, "step": 3815 }, { "epoch": 1.729043951064794, "grad_norm": 0.8913914537621603, "learning_rate": 6.616050765790279e-05, "loss": 0.9019, "step": 3816 }, { "epoch": 1.729497054825555, "grad_norm": 0.8175617254239462, "learning_rate": 6.615093404724028e-05, "loss": 0.887, "step": 3817 }, { "epoch": 1.7299501585863162, "grad_norm": 0.8441391445476116, "learning_rate": 6.614135781959426e-05, "loss": 0.8823, "step": 3818 }, { "epoch": 1.7304032623470775, "grad_norm": 0.7342294596732132, "learning_rate": 6.613177897592307e-05, "loss": 0.9002, "step": 3819 }, { "epoch": 1.7308563661078387, "grad_norm": 0.6844537141714987, "learning_rate": 6.612219751718527e-05, "loss": 0.9014, "step": 3820 }, { "epoch": 1.7313094698686, "grad_norm": 0.5606095808177308, "learning_rate": 6.611261344433971e-05, "loss": 0.8958, "step": 3821 }, { "epoch": 1.7317625736293611, "grad_norm": 0.5970334739227013, "learning_rate": 6.610302675834549e-05, "loss": 0.8942, "step": 3822 }, { "epoch": 1.7322156773901223, "grad_norm": 0.5476971319325165, "learning_rate": 6.609343746016196e-05, "loss": 0.8943, "step": 3823 }, { "epoch": 1.7326687811508834, "grad_norm": 0.6914523227993391, "learning_rate": 6.608384555074875e-05, "loss": 0.8842, "step": 3824 }, { "epoch": 1.7331218849116448, "grad_norm": 0.841024534726979, "learning_rate": 6.607425103106576e-05, "loss": 0.8913, "step": 3825 }, { "epoch": 1.7335749886724061, "grad_norm": 0.9465816816261521, "learning_rate": 6.606465390207311e-05, "loss": 0.8805, "step": 3826 }, { "epoch": 1.7340280924331672, "grad_norm": 0.9167287922741649, "learning_rate": 6.60550541647312e-05, "loss": 0.8789, "step": 3827 }, { "epoch": 1.7344811961939284, "grad_norm": 1.0537703419516535, "learning_rate": 6.604545182000075e-05, "loss": 0.8799, "step": 3828 }, { "epoch": 1.7349342999546895, "grad_norm": 1.0539043145498461, "learning_rate": 6.603584686884262e-05, "loss": 0.9016, "step": 3829 }, { "epoch": 1.7353874037154509, "grad_norm": 0.8046252508940924, "learning_rate": 6.602623931221805e-05, "loss": 0.8866, "step": 3830 }, { "epoch": 1.7358405074762122, "grad_norm": 0.705024052459053, "learning_rate": 6.601662915108847e-05, "loss": 0.8873, "step": 3831 }, { "epoch": 1.7362936112369733, "grad_norm": 0.5549921570746292, "learning_rate": 6.600701638641559e-05, "loss": 0.9199, "step": 3832 }, { "epoch": 1.7367467149977345, "grad_norm": 0.4761392961961818, "learning_rate": 6.599740101916138e-05, "loss": 0.8875, "step": 3833 }, { "epoch": 1.7371998187584956, "grad_norm": 0.44474778194219433, "learning_rate": 6.59877830502881e-05, "loss": 0.8865, "step": 3834 }, { "epoch": 1.737652922519257, "grad_norm": 0.46462382198379704, "learning_rate": 6.597816248075819e-05, "loss": 0.8791, "step": 3835 }, { "epoch": 1.738106026280018, "grad_norm": 0.415570287878265, "learning_rate": 6.596853931153444e-05, "loss": 0.8912, "step": 3836 }, { "epoch": 1.7385591300407794, "grad_norm": 0.40505374760650986, "learning_rate": 6.595891354357985e-05, "loss": 0.8916, "step": 3837 }, { "epoch": 1.7390122338015406, "grad_norm": 0.39798717759113766, "learning_rate": 6.59492851778577e-05, "loss": 0.8857, "step": 3838 }, { "epoch": 1.7394653375623017, "grad_norm": 0.4644452557355205, "learning_rate": 6.59396542153315e-05, "loss": 0.8928, "step": 3839 }, { "epoch": 1.7399184413230628, "grad_norm": 0.6012059091949454, "learning_rate": 6.593002065696508e-05, "loss": 0.8808, "step": 3840 }, { "epoch": 1.7403715450838242, "grad_norm": 0.4708483845214585, "learning_rate": 6.592038450372247e-05, "loss": 0.8965, "step": 3841 }, { "epoch": 1.7408246488445855, "grad_norm": 0.35526778957715394, "learning_rate": 6.591074575656799e-05, "loss": 0.8821, "step": 3842 }, { "epoch": 1.7412777526053467, "grad_norm": 0.488706228727804, "learning_rate": 6.590110441646619e-05, "loss": 0.8669, "step": 3843 }, { "epoch": 1.7417308563661078, "grad_norm": 0.5762219405256507, "learning_rate": 6.589146048438192e-05, "loss": 0.8887, "step": 3844 }, { "epoch": 1.742183960126869, "grad_norm": 0.7020633826519185, "learning_rate": 6.588181396128027e-05, "loss": 0.8961, "step": 3845 }, { "epoch": 1.7426370638876303, "grad_norm": 0.9054055193085659, "learning_rate": 6.587216484812657e-05, "loss": 0.8883, "step": 3846 }, { "epoch": 1.7430901676483916, "grad_norm": 1.2539331639599252, "learning_rate": 6.586251314588646e-05, "loss": 0.9052, "step": 3847 }, { "epoch": 1.7435432714091528, "grad_norm": 0.8340882466121821, "learning_rate": 6.585285885552581e-05, "loss": 0.8907, "step": 3848 }, { "epoch": 1.743996375169914, "grad_norm": 0.5539793794141492, "learning_rate": 6.58432019780107e-05, "loss": 0.8979, "step": 3849 }, { "epoch": 1.744449478930675, "grad_norm": 0.4502275922690575, "learning_rate": 6.583354251430754e-05, "loss": 0.8887, "step": 3850 }, { "epoch": 1.7449025826914364, "grad_norm": 0.6059149555627876, "learning_rate": 6.582388046538301e-05, "loss": 0.8947, "step": 3851 }, { "epoch": 1.7453556864521975, "grad_norm": 0.7290662392220372, "learning_rate": 6.581421583220396e-05, "loss": 0.9133, "step": 3852 }, { "epoch": 1.7458087902129589, "grad_norm": 0.9535337723031823, "learning_rate": 6.580454861573758e-05, "loss": 0.9013, "step": 3853 }, { "epoch": 1.74626189397372, "grad_norm": 1.3138690751964717, "learning_rate": 6.579487881695129e-05, "loss": 0.8831, "step": 3854 }, { "epoch": 1.7467149977344811, "grad_norm": 0.6424424812021458, "learning_rate": 6.578520643681275e-05, "loss": 0.8918, "step": 3855 }, { "epoch": 1.7471681014952423, "grad_norm": 0.5895589485661706, "learning_rate": 6.577553147628993e-05, "loss": 0.899, "step": 3856 }, { "epoch": 1.7476212052560036, "grad_norm": 1.0425445575150558, "learning_rate": 6.5765853936351e-05, "loss": 0.8791, "step": 3857 }, { "epoch": 1.748074309016765, "grad_norm": 1.4704526828069788, "learning_rate": 6.575617381796441e-05, "loss": 0.8756, "step": 3858 }, { "epoch": 1.748527412777526, "grad_norm": 0.39287177928053685, "learning_rate": 6.574649112209888e-05, "loss": 0.8891, "step": 3859 }, { "epoch": 1.7489805165382872, "grad_norm": 1.3499067688282627, "learning_rate": 6.573680584972339e-05, "loss": 0.8767, "step": 3860 }, { "epoch": 1.7494336202990484, "grad_norm": 1.0403381012170116, "learning_rate": 6.572711800180715e-05, "loss": 0.8902, "step": 3861 }, { "epoch": 1.7498867240598097, "grad_norm": 0.7273594231937359, "learning_rate": 6.571742757931966e-05, "loss": 0.8906, "step": 3862 }, { "epoch": 1.750339827820571, "grad_norm": 0.5049630165576735, "learning_rate": 6.570773458323066e-05, "loss": 0.8819, "step": 3863 }, { "epoch": 1.7507929315813322, "grad_norm": 0.4893614664835116, "learning_rate": 6.569803901451014e-05, "loss": 0.895, "step": 3864 }, { "epoch": 1.7512460353420933, "grad_norm": 0.672574699176374, "learning_rate": 6.568834087412836e-05, "loss": 0.8989, "step": 3865 }, { "epoch": 1.7516991391028545, "grad_norm": 1.039679879346113, "learning_rate": 6.567864016305583e-05, "loss": 0.9045, "step": 3866 }, { "epoch": 1.7521522428636158, "grad_norm": 1.2260761458706921, "learning_rate": 6.566893688226333e-05, "loss": 0.8883, "step": 3867 }, { "epoch": 1.752605346624377, "grad_norm": 0.5972597153239148, "learning_rate": 6.565923103272191e-05, "loss": 0.8943, "step": 3868 }, { "epoch": 1.7530584503851383, "grad_norm": 0.6368320104665945, "learning_rate": 6.564952261540283e-05, "loss": 0.8865, "step": 3869 }, { "epoch": 1.7535115541458994, "grad_norm": 1.0814153278968996, "learning_rate": 6.563981163127762e-05, "loss": 0.871, "step": 3870 }, { "epoch": 1.7539646579066606, "grad_norm": 1.1779695751225012, "learning_rate": 6.563009808131811e-05, "loss": 0.8722, "step": 3871 }, { "epoch": 1.7544177616674217, "grad_norm": 0.7539460004269893, "learning_rate": 6.562038196649635e-05, "loss": 0.8884, "step": 3872 }, { "epoch": 1.754870865428183, "grad_norm": 0.7101824775462835, "learning_rate": 6.561066328778467e-05, "loss": 0.8918, "step": 3873 }, { "epoch": 1.7553239691889444, "grad_norm": 0.6422036662867449, "learning_rate": 6.56009420461556e-05, "loss": 0.8947, "step": 3874 }, { "epoch": 1.7557770729497055, "grad_norm": 0.573203784214368, "learning_rate": 6.559121824258198e-05, "loss": 0.889, "step": 3875 }, { "epoch": 1.7562301767104667, "grad_norm": 0.7106829647649101, "learning_rate": 6.558149187803691e-05, "loss": 0.9022, "step": 3876 }, { "epoch": 1.7566832804712278, "grad_norm": 0.8879550242807123, "learning_rate": 6.557176295349372e-05, "loss": 0.9001, "step": 3877 }, { "epoch": 1.7571363842319891, "grad_norm": 0.9146815473041416, "learning_rate": 6.556203146992601e-05, "loss": 0.9077, "step": 3878 }, { "epoch": 1.7575894879927505, "grad_norm": 1.174832369605837, "learning_rate": 6.555229742830762e-05, "loss": 0.8981, "step": 3879 }, { "epoch": 1.7580425917535116, "grad_norm": 0.8587371245613193, "learning_rate": 6.554256082961268e-05, "loss": 0.8849, "step": 3880 }, { "epoch": 1.7584956955142728, "grad_norm": 0.5860867820129156, "learning_rate": 6.553282167481552e-05, "loss": 0.8818, "step": 3881 }, { "epoch": 1.758948799275034, "grad_norm": 0.6066476977153399, "learning_rate": 6.552307996489079e-05, "loss": 0.8982, "step": 3882 }, { "epoch": 1.7594019030357952, "grad_norm": 0.8697408405097001, "learning_rate": 6.551333570081335e-05, "loss": 0.8969, "step": 3883 }, { "epoch": 1.7598550067965564, "grad_norm": 0.9326029434078977, "learning_rate": 6.550358888355833e-05, "loss": 0.8855, "step": 3884 }, { "epoch": 1.7603081105573177, "grad_norm": 0.8372970259233278, "learning_rate": 6.549383951410113e-05, "loss": 0.8776, "step": 3885 }, { "epoch": 1.7607612143180789, "grad_norm": 0.7489276605660564, "learning_rate": 6.548408759341737e-05, "loss": 0.9038, "step": 3886 }, { "epoch": 1.76121431807884, "grad_norm": 0.7820402440457934, "learning_rate": 6.547433312248297e-05, "loss": 0.8995, "step": 3887 }, { "epoch": 1.7616674218396011, "grad_norm": 0.8031938732182399, "learning_rate": 6.546457610227408e-05, "loss": 0.896, "step": 3888 }, { "epoch": 1.7621205256003625, "grad_norm": 0.8255959170292123, "learning_rate": 6.545481653376707e-05, "loss": 0.9077, "step": 3889 }, { "epoch": 1.7625736293611238, "grad_norm": 0.7767694745374341, "learning_rate": 6.544505441793866e-05, "loss": 0.8844, "step": 3890 }, { "epoch": 1.763026733121885, "grad_norm": 0.7031756723586808, "learning_rate": 6.54352897557657e-05, "loss": 0.886, "step": 3891 }, { "epoch": 1.763479836882646, "grad_norm": 0.4896774519599837, "learning_rate": 6.542552254822542e-05, "loss": 0.877, "step": 3892 }, { "epoch": 1.7639329406434072, "grad_norm": 0.5811999026349352, "learning_rate": 6.541575279629524e-05, "loss": 0.8985, "step": 3893 }, { "epoch": 1.7643860444041686, "grad_norm": 0.6113735775129413, "learning_rate": 6.540598050095281e-05, "loss": 0.8788, "step": 3894 }, { "epoch": 1.76483914816493, "grad_norm": 0.5620958205676141, "learning_rate": 6.539620566317609e-05, "loss": 0.8965, "step": 3895 }, { "epoch": 1.765292251925691, "grad_norm": 0.5262901783853808, "learning_rate": 6.538642828394325e-05, "loss": 0.88, "step": 3896 }, { "epoch": 1.7657453556864522, "grad_norm": 0.5989765464946278, "learning_rate": 6.537664836423276e-05, "loss": 0.905, "step": 3897 }, { "epoch": 1.7661984594472133, "grad_norm": 0.7031006866159374, "learning_rate": 6.536686590502331e-05, "loss": 0.9008, "step": 3898 }, { "epoch": 1.7666515632079747, "grad_norm": 0.9245501899804779, "learning_rate": 6.535708090729385e-05, "loss": 0.9029, "step": 3899 }, { "epoch": 1.7671046669687358, "grad_norm": 1.1467975893083133, "learning_rate": 6.534729337202358e-05, "loss": 0.9135, "step": 3900 }, { "epoch": 1.7675577707294972, "grad_norm": 0.8405436705633793, "learning_rate": 6.533750330019197e-05, "loss": 0.8935, "step": 3901 }, { "epoch": 1.7680108744902583, "grad_norm": 0.7871902639197557, "learning_rate": 6.532771069277874e-05, "loss": 0.8828, "step": 3902 }, { "epoch": 1.7684639782510194, "grad_norm": 0.7122113605379804, "learning_rate": 6.531791555076384e-05, "loss": 0.9095, "step": 3903 }, { "epoch": 1.7689170820117806, "grad_norm": 0.5519878688793342, "learning_rate": 6.530811787512751e-05, "loss": 0.8888, "step": 3904 }, { "epoch": 1.769370185772542, "grad_norm": 0.5327045355671639, "learning_rate": 6.529831766685022e-05, "loss": 0.9014, "step": 3905 }, { "epoch": 1.7698232895333033, "grad_norm": 0.6061088499633142, "learning_rate": 6.528851492691269e-05, "loss": 0.8951, "step": 3906 }, { "epoch": 1.7702763932940644, "grad_norm": 0.518929404381037, "learning_rate": 6.527870965629591e-05, "loss": 0.9301, "step": 3907 }, { "epoch": 1.7707294970548255, "grad_norm": 0.4497939158537952, "learning_rate": 6.526890185598113e-05, "loss": 0.8865, "step": 3908 }, { "epoch": 1.7711826008155866, "grad_norm": 0.5140383559995867, "learning_rate": 6.525909152694982e-05, "loss": 0.8841, "step": 3909 }, { "epoch": 1.771635704576348, "grad_norm": 0.5365857704424637, "learning_rate": 6.524927867018373e-05, "loss": 0.8731, "step": 3910 }, { "epoch": 1.7720888083371094, "grad_norm": 0.5211155154117972, "learning_rate": 6.523946328666484e-05, "loss": 0.9061, "step": 3911 }, { "epoch": 1.7725419120978705, "grad_norm": 0.5259175612095026, "learning_rate": 6.52296453773754e-05, "loss": 0.8783, "step": 3912 }, { "epoch": 1.7729950158586316, "grad_norm": 0.6292576291532613, "learning_rate": 6.521982494329794e-05, "loss": 0.8694, "step": 3913 }, { "epoch": 1.7734481196193927, "grad_norm": 0.765385995223778, "learning_rate": 6.521000198541519e-05, "loss": 0.8852, "step": 3914 }, { "epoch": 1.773901223380154, "grad_norm": 0.9141032836833193, "learning_rate": 6.520017650471018e-05, "loss": 0.8915, "step": 3915 }, { "epoch": 1.7743543271409152, "grad_norm": 1.1006666897046533, "learning_rate": 6.519034850216612e-05, "loss": 0.8949, "step": 3916 }, { "epoch": 1.7748074309016766, "grad_norm": 0.9254762528254277, "learning_rate": 6.518051797876656e-05, "loss": 0.8801, "step": 3917 }, { "epoch": 1.7752605346624377, "grad_norm": 0.8382070944081289, "learning_rate": 6.517068493549523e-05, "loss": 0.8848, "step": 3918 }, { "epoch": 1.7757136384231988, "grad_norm": 0.7237545843767166, "learning_rate": 6.516084937333618e-05, "loss": 0.9053, "step": 3919 }, { "epoch": 1.77616674218396, "grad_norm": 0.6729274389638767, "learning_rate": 6.515101129327369e-05, "loss": 0.9109, "step": 3920 }, { "epoch": 1.7766198459447213, "grad_norm": 0.6301939021942204, "learning_rate": 6.514117069629222e-05, "loss": 0.8805, "step": 3921 }, { "epoch": 1.7770729497054827, "grad_norm": 0.5838480018985684, "learning_rate": 6.513132758337657e-05, "loss": 0.8756, "step": 3922 }, { "epoch": 1.7775260534662438, "grad_norm": 0.6391488317452831, "learning_rate": 6.512148195551178e-05, "loss": 0.8998, "step": 3923 }, { "epoch": 1.777979157227005, "grad_norm": 0.7477889727143019, "learning_rate": 6.511163381368312e-05, "loss": 0.885, "step": 3924 }, { "epoch": 1.778432260987766, "grad_norm": 0.8387695338756241, "learning_rate": 6.510178315887607e-05, "loss": 0.8939, "step": 3925 }, { "epoch": 1.7788853647485274, "grad_norm": 0.8638243538584158, "learning_rate": 6.509192999207649e-05, "loss": 0.8786, "step": 3926 }, { "epoch": 1.7793384685092888, "grad_norm": 0.8607047363222857, "learning_rate": 6.508207431427033e-05, "loss": 0.88, "step": 3927 }, { "epoch": 1.77979157227005, "grad_norm": 0.9895332371010727, "learning_rate": 6.507221612644392e-05, "loss": 0.9112, "step": 3928 }, { "epoch": 1.780244676030811, "grad_norm": 1.1397188694499183, "learning_rate": 6.506235542958376e-05, "loss": 0.9187, "step": 3929 }, { "epoch": 1.7806977797915722, "grad_norm": 0.8745501562151013, "learning_rate": 6.505249222467665e-05, "loss": 0.8884, "step": 3930 }, { "epoch": 1.7811508835523335, "grad_norm": 0.7450640336266414, "learning_rate": 6.504262651270962e-05, "loss": 0.8681, "step": 3931 }, { "epoch": 1.7816039873130947, "grad_norm": 0.6045324208022943, "learning_rate": 6.503275829466996e-05, "loss": 0.9067, "step": 3932 }, { "epoch": 1.782057091073856, "grad_norm": 0.4782067539874632, "learning_rate": 6.502288757154519e-05, "loss": 0.8939, "step": 3933 }, { "epoch": 1.7825101948346171, "grad_norm": 0.4856837161667929, "learning_rate": 6.501301434432312e-05, "loss": 0.9055, "step": 3934 }, { "epoch": 1.7829632985953783, "grad_norm": 0.30866827126589735, "learning_rate": 6.500313861399175e-05, "loss": 0.8832, "step": 3935 }, { "epoch": 1.7834164023561394, "grad_norm": 0.425118485126124, "learning_rate": 6.499326038153942e-05, "loss": 0.9022, "step": 3936 }, { "epoch": 1.7838695061169008, "grad_norm": 0.46338767152274274, "learning_rate": 6.498337964795459e-05, "loss": 0.9066, "step": 3937 }, { "epoch": 1.7843226098776621, "grad_norm": 0.4462347036759082, "learning_rate": 6.497349641422612e-05, "loss": 0.8964, "step": 3938 }, { "epoch": 1.7847757136384232, "grad_norm": 0.5118311746070867, "learning_rate": 6.496361068134301e-05, "loss": 0.8736, "step": 3939 }, { "epoch": 1.7852288173991844, "grad_norm": 0.5557144920168263, "learning_rate": 6.495372245029457e-05, "loss": 0.8854, "step": 3940 }, { "epoch": 1.7856819211599455, "grad_norm": 0.5810031445938219, "learning_rate": 6.494383172207032e-05, "loss": 0.911, "step": 3941 }, { "epoch": 1.7861350249207069, "grad_norm": 0.6634055217666525, "learning_rate": 6.493393849766005e-05, "loss": 0.9249, "step": 3942 }, { "epoch": 1.7865881286814682, "grad_norm": 0.9435770190238307, "learning_rate": 6.492404277805382e-05, "loss": 0.8955, "step": 3943 }, { "epoch": 1.7870412324422293, "grad_norm": 1.1772291098774312, "learning_rate": 6.491414456424188e-05, "loss": 0.8843, "step": 3944 }, { "epoch": 1.7874943362029905, "grad_norm": 0.7499157657940416, "learning_rate": 6.490424385721479e-05, "loss": 0.8973, "step": 3945 }, { "epoch": 1.7879474399637516, "grad_norm": 0.5474183384836888, "learning_rate": 6.489434065796332e-05, "loss": 0.8962, "step": 3946 }, { "epoch": 1.788400543724513, "grad_norm": 0.44999344299349203, "learning_rate": 6.488443496747852e-05, "loss": 0.8796, "step": 3947 }, { "epoch": 1.788853647485274, "grad_norm": 0.4893024346799743, "learning_rate": 6.487452678675169e-05, "loss": 0.9022, "step": 3948 }, { "epoch": 1.7893067512460354, "grad_norm": 0.4423348276853175, "learning_rate": 6.486461611677433e-05, "loss": 0.8914, "step": 3949 }, { "epoch": 1.7897598550067966, "grad_norm": 0.477728890278391, "learning_rate": 6.485470295853825e-05, "loss": 0.8745, "step": 3950 }, { "epoch": 1.7902129587675577, "grad_norm": 0.6202643139369237, "learning_rate": 6.484478731303545e-05, "loss": 0.891, "step": 3951 }, { "epoch": 1.7906660625283188, "grad_norm": 0.7966617528075871, "learning_rate": 6.483486918125825e-05, "loss": 0.8864, "step": 3952 }, { "epoch": 1.7911191662890802, "grad_norm": 0.8789404626861369, "learning_rate": 6.482494856419915e-05, "loss": 0.8949, "step": 3953 }, { "epoch": 1.7915722700498415, "grad_norm": 0.9711227802131811, "learning_rate": 6.481502546285097e-05, "loss": 0.897, "step": 3954 }, { "epoch": 1.7920253738106027, "grad_norm": 1.1948296678869224, "learning_rate": 6.480509987820668e-05, "loss": 0.8841, "step": 3955 }, { "epoch": 1.7924784775713638, "grad_norm": 0.7654433877599729, "learning_rate": 6.47951718112596e-05, "loss": 0.9024, "step": 3956 }, { "epoch": 1.792931581332125, "grad_norm": 0.5568992179036155, "learning_rate": 6.478524126300324e-05, "loss": 0.8839, "step": 3957 }, { "epoch": 1.7933846850928863, "grad_norm": 0.4556741874499601, "learning_rate": 6.477530823443138e-05, "loss": 0.8778, "step": 3958 }, { "epoch": 1.7938377888536476, "grad_norm": 0.5233912267020912, "learning_rate": 6.476537272653801e-05, "loss": 0.894, "step": 3959 }, { "epoch": 1.7942908926144088, "grad_norm": 0.7312473470801998, "learning_rate": 6.475543474031742e-05, "loss": 0.8812, "step": 3960 }, { "epoch": 1.79474399637517, "grad_norm": 0.8282234220854647, "learning_rate": 6.474549427676415e-05, "loss": 0.8845, "step": 3961 }, { "epoch": 1.795197100135931, "grad_norm": 0.776303217730237, "learning_rate": 6.473555133687292e-05, "loss": 0.8951, "step": 3962 }, { "epoch": 1.7956502038966924, "grad_norm": 0.7721292275293749, "learning_rate": 6.472560592163879e-05, "loss": 0.9142, "step": 3963 }, { "epoch": 1.7961033076574535, "grad_norm": 0.9189781548810402, "learning_rate": 6.471565803205698e-05, "loss": 0.8947, "step": 3964 }, { "epoch": 1.7965564114182149, "grad_norm": 0.9902793168749593, "learning_rate": 6.470570766912302e-05, "loss": 0.884, "step": 3965 }, { "epoch": 1.797009515178976, "grad_norm": 1.0540467745499176, "learning_rate": 6.469575483383265e-05, "loss": 0.9175, "step": 3966 }, { "epoch": 1.7974626189397371, "grad_norm": 1.046277140118661, "learning_rate": 6.46857995271819e-05, "loss": 0.8946, "step": 3967 }, { "epoch": 1.7979157227004983, "grad_norm": 0.9136160016875691, "learning_rate": 6.467584175016699e-05, "loss": 0.8897, "step": 3968 }, { "epoch": 1.7983688264612596, "grad_norm": 0.7968440967317088, "learning_rate": 6.466588150378443e-05, "loss": 0.9049, "step": 3969 }, { "epoch": 1.798821930222021, "grad_norm": 0.6759729370994306, "learning_rate": 6.465591878903096e-05, "loss": 0.8793, "step": 3970 }, { "epoch": 1.799275033982782, "grad_norm": 0.5256069217048146, "learning_rate": 6.46459536069036e-05, "loss": 0.8933, "step": 3971 }, { "epoch": 1.7997281377435432, "grad_norm": 0.4799154302174843, "learning_rate": 6.463598595839953e-05, "loss": 0.8689, "step": 3972 }, { "epoch": 1.8001812415043044, "grad_norm": 0.6635615800654341, "learning_rate": 6.46260158445163e-05, "loss": 0.889, "step": 3973 }, { "epoch": 1.8006343452650657, "grad_norm": 0.7123785925576692, "learning_rate": 6.46160432662516e-05, "loss": 0.9059, "step": 3974 }, { "epoch": 1.801087449025827, "grad_norm": 0.7091360182418022, "learning_rate": 6.460606822460341e-05, "loss": 0.8883, "step": 3975 }, { "epoch": 1.8015405527865882, "grad_norm": 0.7900574827570422, "learning_rate": 6.459609072056998e-05, "loss": 0.8787, "step": 3976 }, { "epoch": 1.8019936565473493, "grad_norm": 0.8848944758673141, "learning_rate": 6.458611075514977e-05, "loss": 0.9006, "step": 3977 }, { "epoch": 1.8024467603081105, "grad_norm": 0.9992906070069143, "learning_rate": 6.457612832934148e-05, "loss": 0.8931, "step": 3978 }, { "epoch": 1.8028998640688718, "grad_norm": 1.1502388864602975, "learning_rate": 6.456614344414409e-05, "loss": 0.8918, "step": 3979 }, { "epoch": 1.803352967829633, "grad_norm": 0.7112982202476814, "learning_rate": 6.455615610055681e-05, "loss": 0.9053, "step": 3980 }, { "epoch": 1.8038060715903943, "grad_norm": 0.4130347033656845, "learning_rate": 6.45461662995791e-05, "loss": 0.8991, "step": 3981 }, { "epoch": 1.8042591753511554, "grad_norm": 0.37818749439700033, "learning_rate": 6.453617404221066e-05, "loss": 0.8987, "step": 3982 }, { "epoch": 1.8047122791119166, "grad_norm": 0.47435403131567244, "learning_rate": 6.452617932945143e-05, "loss": 0.9174, "step": 3983 }, { "epoch": 1.8051653828726777, "grad_norm": 0.6893498603984071, "learning_rate": 6.451618216230162e-05, "loss": 0.8738, "step": 3984 }, { "epoch": 1.805618486633439, "grad_norm": 0.8913923811840104, "learning_rate": 6.450618254176165e-05, "loss": 0.8789, "step": 3985 }, { "epoch": 1.8060715903942004, "grad_norm": 1.062799636097053, "learning_rate": 6.449618046883222e-05, "loss": 0.8969, "step": 3986 }, { "epoch": 1.8065246941549615, "grad_norm": 0.9846833344664645, "learning_rate": 6.448617594451424e-05, "loss": 0.8925, "step": 3987 }, { "epoch": 1.8069777979157227, "grad_norm": 0.9998697048793649, "learning_rate": 6.44761689698089e-05, "loss": 0.8876, "step": 3988 }, { "epoch": 1.8074309016764838, "grad_norm": 0.9805134756613603, "learning_rate": 6.446615954571764e-05, "loss": 0.8893, "step": 3989 }, { "epoch": 1.8078840054372451, "grad_norm": 0.8409928365471854, "learning_rate": 6.445614767324208e-05, "loss": 0.8862, "step": 3990 }, { "epoch": 1.8083371091980065, "grad_norm": 0.7110985143320421, "learning_rate": 6.444613335338416e-05, "loss": 0.887, "step": 3991 }, { "epoch": 1.8087902129587676, "grad_norm": 0.674167820625282, "learning_rate": 6.443611658714604e-05, "loss": 0.8778, "step": 3992 }, { "epoch": 1.8092433167195288, "grad_norm": 0.553797774485727, "learning_rate": 6.442609737553012e-05, "loss": 0.8817, "step": 3993 }, { "epoch": 1.80969642048029, "grad_norm": 0.47826728096196475, "learning_rate": 6.441607571953902e-05, "loss": 0.8932, "step": 3994 }, { "epoch": 1.8101495242410512, "grad_norm": 0.388436269894752, "learning_rate": 6.440605162017566e-05, "loss": 0.8804, "step": 3995 }, { "epoch": 1.8106026280018124, "grad_norm": 0.3820220397633483, "learning_rate": 6.439602507844316e-05, "loss": 0.8966, "step": 3996 }, { "epoch": 1.8110557317625737, "grad_norm": 0.4973383594425029, "learning_rate": 6.438599609534491e-05, "loss": 0.8889, "step": 3997 }, { "epoch": 1.8115088355233349, "grad_norm": 0.599778598287365, "learning_rate": 6.437596467188451e-05, "loss": 0.8929, "step": 3998 }, { "epoch": 1.811961939284096, "grad_norm": 0.763216803671346, "learning_rate": 6.436593080906585e-05, "loss": 0.8916, "step": 3999 }, { "epoch": 1.8124150430448571, "grad_norm": 0.8195368365692092, "learning_rate": 6.435589450789301e-05, "loss": 0.8817, "step": 4000 }, { "epoch": 1.8128681468056185, "grad_norm": 0.84810902023434, "learning_rate": 6.434585576937041e-05, "loss": 0.9033, "step": 4001 }, { "epoch": 1.8133212505663798, "grad_norm": 0.9761356962495836, "learning_rate": 6.433581459450258e-05, "loss": 0.914, "step": 4002 }, { "epoch": 1.813774354327141, "grad_norm": 1.169670355301037, "learning_rate": 6.432577098429441e-05, "loss": 0.8768, "step": 4003 }, { "epoch": 1.814227458087902, "grad_norm": 0.7829666554734951, "learning_rate": 6.431572493975095e-05, "loss": 0.8805, "step": 4004 }, { "epoch": 1.8146805618486632, "grad_norm": 0.49235270961359434, "learning_rate": 6.430567646187755e-05, "loss": 0.8767, "step": 4005 }, { "epoch": 1.8151336656094246, "grad_norm": 0.5149740363425871, "learning_rate": 6.429562555167979e-05, "loss": 0.8821, "step": 4006 }, { "epoch": 1.815586769370186, "grad_norm": 0.8073294550697957, "learning_rate": 6.428557221016349e-05, "loss": 0.879, "step": 4007 }, { "epoch": 1.816039873130947, "grad_norm": 1.044988282700362, "learning_rate": 6.427551643833469e-05, "loss": 0.8776, "step": 4008 }, { "epoch": 1.8164929768917082, "grad_norm": 1.007362219015377, "learning_rate": 6.42654582371997e-05, "loss": 0.9041, "step": 4009 }, { "epoch": 1.8169460806524693, "grad_norm": 1.0309543876572433, "learning_rate": 6.425539760776507e-05, "loss": 0.8869, "step": 4010 }, { "epoch": 1.8173991844132307, "grad_norm": 0.9779331839129113, "learning_rate": 6.42453345510376e-05, "loss": 0.8955, "step": 4011 }, { "epoch": 1.8178522881739918, "grad_norm": 0.8367386151415731, "learning_rate": 6.423526906802431e-05, "loss": 0.8803, "step": 4012 }, { "epoch": 1.8183053919347532, "grad_norm": 0.633730580813127, "learning_rate": 6.42252011597325e-05, "loss": 0.8913, "step": 4013 }, { "epoch": 1.8187584956955143, "grad_norm": 0.5287081008514404, "learning_rate": 6.421513082716965e-05, "loss": 0.907, "step": 4014 }, { "epoch": 1.8192115994562754, "grad_norm": 0.41258571766558283, "learning_rate": 6.420505807134355e-05, "loss": 0.8873, "step": 4015 }, { "epoch": 1.8196647032170365, "grad_norm": 0.45070735567390663, "learning_rate": 6.419498289326218e-05, "loss": 0.8931, "step": 4016 }, { "epoch": 1.820117806977798, "grad_norm": 0.4812827117827195, "learning_rate": 6.418490529393382e-05, "loss": 0.8743, "step": 4017 }, { "epoch": 1.8205709107385593, "grad_norm": 0.39514587854080063, "learning_rate": 6.417482527436692e-05, "loss": 0.8944, "step": 4018 }, { "epoch": 1.8210240144993204, "grad_norm": 0.35342867928184957, "learning_rate": 6.416474283557024e-05, "loss": 0.8917, "step": 4019 }, { "epoch": 1.8214771182600815, "grad_norm": 0.3807533836289167, "learning_rate": 6.415465797855274e-05, "loss": 0.8798, "step": 4020 }, { "epoch": 1.8219302220208426, "grad_norm": 0.35016429418537653, "learning_rate": 6.414457070432363e-05, "loss": 0.876, "step": 4021 }, { "epoch": 1.822383325781604, "grad_norm": 0.364112792074443, "learning_rate": 6.413448101389239e-05, "loss": 0.9045, "step": 4022 }, { "epoch": 1.8228364295423654, "grad_norm": 0.5404796394406632, "learning_rate": 6.412438890826869e-05, "loss": 0.8955, "step": 4023 }, { "epoch": 1.8232895333031265, "grad_norm": 0.547786051767178, "learning_rate": 6.411429438846248e-05, "loss": 0.8976, "step": 4024 }, { "epoch": 1.8237426370638876, "grad_norm": 0.5423966356158272, "learning_rate": 6.410419745548394e-05, "loss": 0.8701, "step": 4025 }, { "epoch": 1.8241957408246487, "grad_norm": 0.6898464114803704, "learning_rate": 6.40940981103435e-05, "loss": 0.8904, "step": 4026 }, { "epoch": 1.82464884458541, "grad_norm": 0.8723326966665359, "learning_rate": 6.408399635405182e-05, "loss": 0.8758, "step": 4027 }, { "epoch": 1.8251019483461712, "grad_norm": 1.0295220283632407, "learning_rate": 6.407389218761981e-05, "loss": 0.8883, "step": 4028 }, { "epoch": 1.8255550521069326, "grad_norm": 1.0108668886528653, "learning_rate": 6.40637856120586e-05, "loss": 0.883, "step": 4029 }, { "epoch": 1.8260081558676937, "grad_norm": 0.9248059121286857, "learning_rate": 6.40536766283796e-05, "loss": 0.8903, "step": 4030 }, { "epoch": 1.8264612596284548, "grad_norm": 0.8793445067668431, "learning_rate": 6.404356523759444e-05, "loss": 0.9098, "step": 4031 }, { "epoch": 1.826914363389216, "grad_norm": 0.7536054229668194, "learning_rate": 6.403345144071496e-05, "loss": 0.902, "step": 4032 }, { "epoch": 1.8273674671499773, "grad_norm": 0.6084407114246054, "learning_rate": 6.402333523875331e-05, "loss": 0.8641, "step": 4033 }, { "epoch": 1.8278205709107387, "grad_norm": 0.45983399750774323, "learning_rate": 6.40132166327218e-05, "loss": 0.8979, "step": 4034 }, { "epoch": 1.8282736746714998, "grad_norm": 0.4361712998864458, "learning_rate": 6.400309562363305e-05, "loss": 0.8796, "step": 4035 }, { "epoch": 1.828726778432261, "grad_norm": 0.45589515067585723, "learning_rate": 6.39929722124999e-05, "loss": 0.8934, "step": 4036 }, { "epoch": 1.829179882193022, "grad_norm": 0.4681597632371948, "learning_rate": 6.39828464003354e-05, "loss": 0.8845, "step": 4037 }, { "epoch": 1.8296329859537834, "grad_norm": 0.6946047961076076, "learning_rate": 6.397271818815287e-05, "loss": 0.9081, "step": 4038 }, { "epoch": 1.8300860897145448, "grad_norm": 0.93256162950957, "learning_rate": 6.396258757696586e-05, "loss": 0.9051, "step": 4039 }, { "epoch": 1.830539193475306, "grad_norm": 1.073002713918133, "learning_rate": 6.395245456778818e-05, "loss": 0.8668, "step": 4040 }, { "epoch": 1.830992297236067, "grad_norm": 0.9016719151675711, "learning_rate": 6.394231916163385e-05, "loss": 0.8949, "step": 4041 }, { "epoch": 1.8314454009968282, "grad_norm": 0.8268016060025021, "learning_rate": 6.393218135951714e-05, "loss": 0.8981, "step": 4042 }, { "epoch": 1.8318985047575895, "grad_norm": 0.7828248483110718, "learning_rate": 6.392204116245257e-05, "loss": 0.8794, "step": 4043 }, { "epoch": 1.8323516085183507, "grad_norm": 0.728801533744088, "learning_rate": 6.391189857145489e-05, "loss": 0.8797, "step": 4044 }, { "epoch": 1.832804712279112, "grad_norm": 0.6071002829501524, "learning_rate": 6.390175358753909e-05, "loss": 0.8979, "step": 4045 }, { "epoch": 1.8332578160398731, "grad_norm": 0.5261640720852429, "learning_rate": 6.38916062117204e-05, "loss": 0.8969, "step": 4046 }, { "epoch": 1.8337109198006343, "grad_norm": 0.4334333412252468, "learning_rate": 6.388145644501432e-05, "loss": 0.9, "step": 4047 }, { "epoch": 1.8341640235613954, "grad_norm": 0.3097760608452073, "learning_rate": 6.387130428843651e-05, "loss": 0.9095, "step": 4048 }, { "epoch": 1.8346171273221568, "grad_norm": 0.2885676889785043, "learning_rate": 6.386114974300295e-05, "loss": 0.8794, "step": 4049 }, { "epoch": 1.8350702310829181, "grad_norm": 0.3400508864622852, "learning_rate": 6.385099280972983e-05, "loss": 0.881, "step": 4050 }, { "epoch": 1.8355233348436792, "grad_norm": 0.3759287804854425, "learning_rate": 6.384083348963357e-05, "loss": 0.8858, "step": 4051 }, { "epoch": 1.8359764386044404, "grad_norm": 0.45144845947821527, "learning_rate": 6.383067178373083e-05, "loss": 0.8667, "step": 4052 }, { "epoch": 1.8364295423652015, "grad_norm": 0.5142262462969249, "learning_rate": 6.382050769303852e-05, "loss": 0.8954, "step": 4053 }, { "epoch": 1.8368826461259629, "grad_norm": 0.6055098489086699, "learning_rate": 6.381034121857381e-05, "loss": 0.8642, "step": 4054 }, { "epoch": 1.8373357498867242, "grad_norm": 0.7464133002391914, "learning_rate": 6.380017236135404e-05, "loss": 0.8892, "step": 4055 }, { "epoch": 1.8377888536474853, "grad_norm": 0.9665104739013085, "learning_rate": 6.379000112239684e-05, "loss": 0.8887, "step": 4056 }, { "epoch": 1.8382419574082465, "grad_norm": 1.2308380092211626, "learning_rate": 6.37798275027201e-05, "loss": 0.8926, "step": 4057 }, { "epoch": 1.8386950611690076, "grad_norm": 0.6119806896001694, "learning_rate": 6.376965150334188e-05, "loss": 0.9071, "step": 4058 }, { "epoch": 1.839148164929769, "grad_norm": 0.4012300197644531, "learning_rate": 6.375947312528054e-05, "loss": 0.8933, "step": 4059 }, { "epoch": 1.83960126869053, "grad_norm": 0.6053068215972193, "learning_rate": 6.374929236955463e-05, "loss": 0.8802, "step": 4060 }, { "epoch": 1.8400543724512914, "grad_norm": 0.852190065888709, "learning_rate": 6.373910923718299e-05, "loss": 0.8766, "step": 4061 }, { "epoch": 1.8405074762120526, "grad_norm": 1.1593758161245673, "learning_rate": 6.372892372918465e-05, "loss": 0.9056, "step": 4062 }, { "epoch": 1.8409605799728137, "grad_norm": 0.8791177333018141, "learning_rate": 6.37187358465789e-05, "loss": 0.884, "step": 4063 }, { "epoch": 1.8414136837335748, "grad_norm": 0.7286339231940276, "learning_rate": 6.370854559038526e-05, "loss": 0.8936, "step": 4064 }, { "epoch": 1.8418667874943362, "grad_norm": 0.6112913707800957, "learning_rate": 6.369835296162351e-05, "loss": 0.9022, "step": 4065 }, { "epoch": 1.8423198912550975, "grad_norm": 0.6397258255669686, "learning_rate": 6.368815796131363e-05, "loss": 0.8687, "step": 4066 }, { "epoch": 1.8427729950158587, "grad_norm": 0.6223549185215336, "learning_rate": 6.367796059047586e-05, "loss": 0.8875, "step": 4067 }, { "epoch": 1.8432260987766198, "grad_norm": 0.5735593690966659, "learning_rate": 6.366776085013069e-05, "loss": 0.8945, "step": 4068 }, { "epoch": 1.843679202537381, "grad_norm": 0.6349995118341686, "learning_rate": 6.365755874129882e-05, "loss": 0.8832, "step": 4069 }, { "epoch": 1.8441323062981423, "grad_norm": 0.7053114887885775, "learning_rate": 6.364735426500119e-05, "loss": 0.8845, "step": 4070 }, { "epoch": 1.8445854100589036, "grad_norm": 0.80119105996365, "learning_rate": 6.3637147422259e-05, "loss": 0.8919, "step": 4071 }, { "epoch": 1.8450385138196648, "grad_norm": 0.8708149453792321, "learning_rate": 6.362693821409366e-05, "loss": 0.8815, "step": 4072 }, { "epoch": 1.845491617580426, "grad_norm": 0.9836672706727184, "learning_rate": 6.361672664152684e-05, "loss": 0.8968, "step": 4073 }, { "epoch": 1.845944721341187, "grad_norm": 1.017521483350627, "learning_rate": 6.360651270558042e-05, "loss": 0.8986, "step": 4074 }, { "epoch": 1.8463978251019484, "grad_norm": 0.9258662049039521, "learning_rate": 6.359629640727655e-05, "loss": 0.8968, "step": 4075 }, { "epoch": 1.8468509288627095, "grad_norm": 0.8306366844932499, "learning_rate": 6.358607774763759e-05, "loss": 0.8729, "step": 4076 }, { "epoch": 1.8473040326234709, "grad_norm": 0.7515631937597963, "learning_rate": 6.357585672768615e-05, "loss": 0.9039, "step": 4077 }, { "epoch": 1.847757136384232, "grad_norm": 0.6377060560312852, "learning_rate": 6.356563334844507e-05, "loss": 0.8902, "step": 4078 }, { "epoch": 1.8482102401449931, "grad_norm": 0.6582152374595273, "learning_rate": 6.355540761093739e-05, "loss": 0.8801, "step": 4079 }, { "epoch": 1.8486633439057543, "grad_norm": 0.6520424552544216, "learning_rate": 6.354517951618648e-05, "loss": 0.8904, "step": 4080 }, { "epoch": 1.8491164476665156, "grad_norm": 0.5579075633229209, "learning_rate": 6.353494906521585e-05, "loss": 0.898, "step": 4081 }, { "epoch": 1.849569551427277, "grad_norm": 0.4466349787342145, "learning_rate": 6.352471625904932e-05, "loss": 0.8988, "step": 4082 }, { "epoch": 1.850022655188038, "grad_norm": 0.4927718405067608, "learning_rate": 6.351448109871087e-05, "loss": 0.8791, "step": 4083 }, { "epoch": 1.8504757589487992, "grad_norm": 0.5312445145537483, "learning_rate": 6.350424358522478e-05, "loss": 0.916, "step": 4084 }, { "epoch": 1.8509288627095604, "grad_norm": 0.543909973753519, "learning_rate": 6.349400371961553e-05, "loss": 0.8908, "step": 4085 }, { "epoch": 1.8513819664703217, "grad_norm": 0.5017673878837148, "learning_rate": 6.348376150290785e-05, "loss": 0.8706, "step": 4086 }, { "epoch": 1.851835070231083, "grad_norm": 0.4853575247801631, "learning_rate": 6.347351693612671e-05, "loss": 0.8992, "step": 4087 }, { "epoch": 1.8522881739918442, "grad_norm": 0.4978096003720829, "learning_rate": 6.34632700202973e-05, "loss": 0.8977, "step": 4088 }, { "epoch": 1.8527412777526053, "grad_norm": 0.5766330127080789, "learning_rate": 6.345302075644504e-05, "loss": 0.8826, "step": 4089 }, { "epoch": 1.8531943815133665, "grad_norm": 0.6941053139160145, "learning_rate": 6.344276914559564e-05, "loss": 0.8999, "step": 4090 }, { "epoch": 1.8536474852741278, "grad_norm": 0.8120307152901352, "learning_rate": 6.343251518877495e-05, "loss": 0.8821, "step": 4091 }, { "epoch": 1.854100589034889, "grad_norm": 1.0069403764583376, "learning_rate": 6.342225888700912e-05, "loss": 0.9019, "step": 4092 }, { "epoch": 1.8545536927956503, "grad_norm": 1.1183453083009778, "learning_rate": 6.341200024132455e-05, "loss": 0.8911, "step": 4093 }, { "epoch": 1.8550067965564114, "grad_norm": 0.7379430868148158, "learning_rate": 6.340173925274781e-05, "loss": 0.8955, "step": 4094 }, { "epoch": 1.8554599003171726, "grad_norm": 0.49267785403198183, "learning_rate": 6.339147592230577e-05, "loss": 0.8898, "step": 4095 }, { "epoch": 1.8559130040779337, "grad_norm": 0.39646939733119274, "learning_rate": 6.338121025102549e-05, "loss": 0.8823, "step": 4096 }, { "epoch": 1.856366107838695, "grad_norm": 0.40341045100610134, "learning_rate": 6.337094223993427e-05, "loss": 0.8938, "step": 4097 }, { "epoch": 1.8568192115994564, "grad_norm": 0.5349712138219576, "learning_rate": 6.336067189005969e-05, "loss": 0.883, "step": 4098 }, { "epoch": 1.8572723153602175, "grad_norm": 0.7014431399336083, "learning_rate": 6.335039920242948e-05, "loss": 0.9082, "step": 4099 }, { "epoch": 1.8577254191209787, "grad_norm": 0.7707790331881305, "learning_rate": 6.334012417807166e-05, "loss": 0.8825, "step": 4100 }, { "epoch": 1.8581785228817398, "grad_norm": 0.9312304438336148, "learning_rate": 6.332984681801452e-05, "loss": 0.8916, "step": 4101 }, { "epoch": 1.8586316266425011, "grad_norm": 1.166666340409403, "learning_rate": 6.331956712328649e-05, "loss": 0.8865, "step": 4102 }, { "epoch": 1.8590847304032625, "grad_norm": 0.7234517015252323, "learning_rate": 6.33092850949163e-05, "loss": 0.8729, "step": 4103 }, { "epoch": 1.8595378341640236, "grad_norm": 0.5648034878377308, "learning_rate": 6.329900073393288e-05, "loss": 0.8802, "step": 4104 }, { "epoch": 1.8599909379247848, "grad_norm": 0.7022461981518934, "learning_rate": 6.328871404136544e-05, "loss": 0.8907, "step": 4105 }, { "epoch": 1.860444041685546, "grad_norm": 0.8794437761977942, "learning_rate": 6.327842501824339e-05, "loss": 0.8922, "step": 4106 }, { "epoch": 1.8608971454463072, "grad_norm": 1.153330995147307, "learning_rate": 6.326813366559634e-05, "loss": 0.9118, "step": 4107 }, { "epoch": 1.8613502492070684, "grad_norm": 0.8931675505600226, "learning_rate": 6.32578399844542e-05, "loss": 0.8921, "step": 4108 }, { "epoch": 1.8618033529678297, "grad_norm": 0.6767160627852744, "learning_rate": 6.324754397584709e-05, "loss": 0.8811, "step": 4109 }, { "epoch": 1.8622564567285909, "grad_norm": 0.5942121989012964, "learning_rate": 6.323724564080532e-05, "loss": 0.8949, "step": 4110 }, { "epoch": 1.862709560489352, "grad_norm": 0.6620477251026435, "learning_rate": 6.322694498035949e-05, "loss": 0.8885, "step": 4111 }, { "epoch": 1.8631626642501131, "grad_norm": 0.6598924482885759, "learning_rate": 6.321664199554041e-05, "loss": 0.8717, "step": 4112 }, { "epoch": 1.8636157680108745, "grad_norm": 0.6075742299505759, "learning_rate": 6.320633668737911e-05, "loss": 0.8718, "step": 4113 }, { "epoch": 1.8640688717716358, "grad_norm": 0.5159759327186287, "learning_rate": 6.319602905690689e-05, "loss": 0.9146, "step": 4114 }, { "epoch": 1.864521975532397, "grad_norm": 0.41803437075840916, "learning_rate": 6.318571910515525e-05, "loss": 0.8669, "step": 4115 }, { "epoch": 1.864975079293158, "grad_norm": 0.3807401521642518, "learning_rate": 6.317540683315593e-05, "loss": 0.8958, "step": 4116 }, { "epoch": 1.8654281830539192, "grad_norm": 0.36587283677619875, "learning_rate": 6.316509224194087e-05, "loss": 0.9032, "step": 4117 }, { "epoch": 1.8658812868146806, "grad_norm": 0.4234699132298564, "learning_rate": 6.315477533254232e-05, "loss": 0.8783, "step": 4118 }, { "epoch": 1.866334390575442, "grad_norm": 0.49705518431737555, "learning_rate": 6.314445610599269e-05, "loss": 0.8798, "step": 4119 }, { "epoch": 1.866787494336203, "grad_norm": 0.44319579074547527, "learning_rate": 6.313413456332466e-05, "loss": 0.8878, "step": 4120 }, { "epoch": 1.8672405980969642, "grad_norm": 0.550685461786317, "learning_rate": 6.312381070557113e-05, "loss": 0.9211, "step": 4121 }, { "epoch": 1.8676937018577253, "grad_norm": 0.6586263042123565, "learning_rate": 6.311348453376523e-05, "loss": 0.8797, "step": 4122 }, { "epoch": 1.8681468056184867, "grad_norm": 0.7206584333224713, "learning_rate": 6.310315604894031e-05, "loss": 0.8943, "step": 4123 }, { "epoch": 1.8685999093792478, "grad_norm": 0.9419979200029934, "learning_rate": 6.309282525212999e-05, "loss": 0.8993, "step": 4124 }, { "epoch": 1.8690530131400092, "grad_norm": 1.1692373645226488, "learning_rate": 6.308249214436808e-05, "loss": 0.879, "step": 4125 }, { "epoch": 1.8695061169007703, "grad_norm": 0.761099496456244, "learning_rate": 6.307215672668866e-05, "loss": 0.8847, "step": 4126 }, { "epoch": 1.8699592206615314, "grad_norm": 0.7020363780368568, "learning_rate": 6.306181900012598e-05, "loss": 0.9009, "step": 4127 }, { "epoch": 1.8704123244222925, "grad_norm": 0.5480213579092129, "learning_rate": 6.305147896571457e-05, "loss": 0.8916, "step": 4128 }, { "epoch": 1.870865428183054, "grad_norm": 0.5124035083139558, "learning_rate": 6.304113662448922e-05, "loss": 0.8998, "step": 4129 }, { "epoch": 1.8713185319438153, "grad_norm": 0.5498056854317357, "learning_rate": 6.303079197748487e-05, "loss": 0.893, "step": 4130 }, { "epoch": 1.8717716357045764, "grad_norm": 0.5967465044725703, "learning_rate": 6.302044502573674e-05, "loss": 0.8742, "step": 4131 }, { "epoch": 1.8722247394653375, "grad_norm": 0.6677607027026501, "learning_rate": 6.301009577028028e-05, "loss": 0.8829, "step": 4132 }, { "epoch": 1.8726778432260986, "grad_norm": 0.7389316858874354, "learning_rate": 6.299974421215116e-05, "loss": 0.8825, "step": 4133 }, { "epoch": 1.87313094698686, "grad_norm": 0.7843322192641047, "learning_rate": 6.298939035238529e-05, "loss": 0.901, "step": 4134 }, { "epoch": 1.8735840507476214, "grad_norm": 0.8008715064311792, "learning_rate": 6.297903419201881e-05, "loss": 0.8944, "step": 4135 }, { "epoch": 1.8740371545083825, "grad_norm": 0.8514851359724431, "learning_rate": 6.296867573208807e-05, "loss": 0.9105, "step": 4136 }, { "epoch": 1.8744902582691436, "grad_norm": 0.9404983878547252, "learning_rate": 6.295831497362965e-05, "loss": 0.8863, "step": 4137 }, { "epoch": 1.8749433620299047, "grad_norm": 1.1365053921984947, "learning_rate": 6.29479519176804e-05, "loss": 0.9014, "step": 4138 }, { "epoch": 1.875396465790666, "grad_norm": 0.8300407635656564, "learning_rate": 6.293758656527739e-05, "loss": 0.8839, "step": 4139 }, { "epoch": 1.8758495695514272, "grad_norm": 0.5628474241759052, "learning_rate": 6.292721891745785e-05, "loss": 0.9069, "step": 4140 }, { "epoch": 1.8763026733121886, "grad_norm": 0.42946817779355734, "learning_rate": 6.291684897525935e-05, "loss": 0.9052, "step": 4141 }, { "epoch": 1.8767557770729497, "grad_norm": 0.5313551220885061, "learning_rate": 6.290647673971961e-05, "loss": 0.892, "step": 4142 }, { "epoch": 1.8772088808337108, "grad_norm": 0.6642247934643714, "learning_rate": 6.289610221187659e-05, "loss": 0.8881, "step": 4143 }, { "epoch": 1.877661984594472, "grad_norm": 0.7998839887949597, "learning_rate": 6.288572539276854e-05, "loss": 0.9131, "step": 4144 }, { "epoch": 1.8781150883552333, "grad_norm": 1.0695838296881972, "learning_rate": 6.287534628343382e-05, "loss": 0.8981, "step": 4145 }, { "epoch": 1.8785681921159947, "grad_norm": 1.0546479641283084, "learning_rate": 6.286496488491115e-05, "loss": 0.8975, "step": 4146 }, { "epoch": 1.8790212958767558, "grad_norm": 0.8767995628412749, "learning_rate": 6.285458119823939e-05, "loss": 0.889, "step": 4147 }, { "epoch": 1.879474399637517, "grad_norm": 0.7401084494811594, "learning_rate": 6.284419522445766e-05, "loss": 0.9049, "step": 4148 }, { "epoch": 1.879927503398278, "grad_norm": 0.6067220776911135, "learning_rate": 6.283380696460534e-05, "loss": 0.8844, "step": 4149 }, { "epoch": 1.8803806071590394, "grad_norm": 0.47064549086273216, "learning_rate": 6.282341641972198e-05, "loss": 0.8678, "step": 4150 }, { "epoch": 1.8808337109198008, "grad_norm": 0.38668863957370203, "learning_rate": 6.281302359084737e-05, "loss": 0.8896, "step": 4151 }, { "epoch": 1.881286814680562, "grad_norm": 0.3505820368703267, "learning_rate": 6.280262847902158e-05, "loss": 0.8872, "step": 4152 }, { "epoch": 1.881739918441323, "grad_norm": 0.3482694321123255, "learning_rate": 6.279223108528485e-05, "loss": 0.8864, "step": 4153 }, { "epoch": 1.8821930222020842, "grad_norm": 0.42664311760802254, "learning_rate": 6.278183141067768e-05, "loss": 0.8981, "step": 4154 }, { "epoch": 1.8826461259628455, "grad_norm": 0.4873462378894363, "learning_rate": 6.277142945624079e-05, "loss": 0.8884, "step": 4155 }, { "epoch": 1.8830992297236067, "grad_norm": 0.4940634428765285, "learning_rate": 6.276102522301511e-05, "loss": 0.8727, "step": 4156 }, { "epoch": 1.883552333484368, "grad_norm": 0.5170148624075921, "learning_rate": 6.275061871204183e-05, "loss": 0.8993, "step": 4157 }, { "epoch": 1.8840054372451291, "grad_norm": 0.4616028849632208, "learning_rate": 6.274020992436236e-05, "loss": 0.9165, "step": 4158 }, { "epoch": 1.8844585410058903, "grad_norm": 0.5057038821009546, "learning_rate": 6.272979886101831e-05, "loss": 0.8748, "step": 4159 }, { "epoch": 1.8849116447666514, "grad_norm": 0.6599986340221943, "learning_rate": 6.271938552305157e-05, "loss": 0.897, "step": 4160 }, { "epoch": 1.8853647485274128, "grad_norm": 0.80271392696244, "learning_rate": 6.270896991150419e-05, "loss": 0.8859, "step": 4161 }, { "epoch": 1.8858178522881741, "grad_norm": 0.9710003910430843, "learning_rate": 6.269855202741852e-05, "loss": 0.8937, "step": 4162 }, { "epoch": 1.8862709560489352, "grad_norm": 1.0557718432882353, "learning_rate": 6.268813187183707e-05, "loss": 0.8872, "step": 4163 }, { "epoch": 1.8867240598096964, "grad_norm": 0.9618296455989324, "learning_rate": 6.267770944580262e-05, "loss": 0.8766, "step": 4164 }, { "epoch": 1.8871771635704575, "grad_norm": 0.9966668681141326, "learning_rate": 6.266728475035818e-05, "loss": 0.8936, "step": 4165 }, { "epoch": 1.8876302673312189, "grad_norm": 1.079188611151995, "learning_rate": 6.265685778654694e-05, "loss": 0.8924, "step": 4166 }, { "epoch": 1.8880833710919802, "grad_norm": 0.8845010458784155, "learning_rate": 6.264642855541239e-05, "loss": 0.8948, "step": 4167 }, { "epoch": 1.8885364748527413, "grad_norm": 0.5720598653146858, "learning_rate": 6.263599705799819e-05, "loss": 0.8782, "step": 4168 }, { "epoch": 1.8889895786135025, "grad_norm": 0.46316344309135415, "learning_rate": 6.262556329534822e-05, "loss": 0.9, "step": 4169 }, { "epoch": 1.8894426823742636, "grad_norm": 0.6249059014444537, "learning_rate": 6.261512726850666e-05, "loss": 0.8839, "step": 4170 }, { "epoch": 1.889895786135025, "grad_norm": 0.7846941758138251, "learning_rate": 6.26046889785178e-05, "loss": 0.8913, "step": 4171 }, { "epoch": 1.890348889895786, "grad_norm": 0.9885530754774693, "learning_rate": 6.25942484264263e-05, "loss": 0.8982, "step": 4172 }, { "epoch": 1.8908019936565474, "grad_norm": 1.069186751798196, "learning_rate": 6.258380561327691e-05, "loss": 0.8749, "step": 4173 }, { "epoch": 1.8912550974173086, "grad_norm": 0.826280481896329, "learning_rate": 6.25733605401147e-05, "loss": 0.8758, "step": 4174 }, { "epoch": 1.8917082011780697, "grad_norm": 0.6237229237298871, "learning_rate": 6.256291320798491e-05, "loss": 0.8798, "step": 4175 }, { "epoch": 1.8921613049388308, "grad_norm": 0.47775433583642385, "learning_rate": 6.255246361793305e-05, "loss": 0.8925, "step": 4176 }, { "epoch": 1.8926144086995922, "grad_norm": 0.45579048808374667, "learning_rate": 6.254201177100482e-05, "loss": 0.8884, "step": 4177 }, { "epoch": 1.8930675124603535, "grad_norm": 0.3446351466174614, "learning_rate": 6.253155766824616e-05, "loss": 0.8988, "step": 4178 }, { "epoch": 1.8935206162211147, "grad_norm": 0.4368264259593028, "learning_rate": 6.252110131070323e-05, "loss": 0.8939, "step": 4179 }, { "epoch": 1.8939737199818758, "grad_norm": 0.6070569033947285, "learning_rate": 6.251064269942245e-05, "loss": 0.8933, "step": 4180 }, { "epoch": 1.894426823742637, "grad_norm": 0.5737717647495172, "learning_rate": 6.25001818354504e-05, "loss": 0.8945, "step": 4181 }, { "epoch": 1.8948799275033983, "grad_norm": 0.5808720459560045, "learning_rate": 6.248971871983395e-05, "loss": 0.9165, "step": 4182 }, { "epoch": 1.8953330312641596, "grad_norm": 0.6828678529734642, "learning_rate": 6.247925335362014e-05, "loss": 0.898, "step": 4183 }, { "epoch": 1.8957861350249208, "grad_norm": 0.7421289464880222, "learning_rate": 6.24687857378563e-05, "loss": 0.8951, "step": 4184 }, { "epoch": 1.896239238785682, "grad_norm": 0.8746623779281493, "learning_rate": 6.245831587358992e-05, "loss": 0.898, "step": 4185 }, { "epoch": 1.896692342546443, "grad_norm": 1.0676412747159538, "learning_rate": 6.244784376186876e-05, "loss": 0.9033, "step": 4186 }, { "epoch": 1.8971454463072044, "grad_norm": 1.0506519214201364, "learning_rate": 6.243736940374078e-05, "loss": 0.8855, "step": 4187 }, { "epoch": 1.8975985500679655, "grad_norm": 0.8682949651300984, "learning_rate": 6.242689280025415e-05, "loss": 0.9053, "step": 4188 }, { "epoch": 1.8980516538287269, "grad_norm": 0.7385662592439821, "learning_rate": 6.241641395245733e-05, "loss": 0.8716, "step": 4189 }, { "epoch": 1.898504757589488, "grad_norm": 0.6990398934188211, "learning_rate": 6.240593286139895e-05, "loss": 0.8811, "step": 4190 }, { "epoch": 1.8989578613502491, "grad_norm": 0.8314799277255526, "learning_rate": 6.239544952812786e-05, "loss": 0.8913, "step": 4191 }, { "epoch": 1.8994109651110103, "grad_norm": 0.9641339927240961, "learning_rate": 6.238496395369316e-05, "loss": 0.8917, "step": 4192 }, { "epoch": 1.8998640688717716, "grad_norm": 1.0703834223215623, "learning_rate": 6.237447613914415e-05, "loss": 0.9001, "step": 4193 }, { "epoch": 1.900317172632533, "grad_norm": 0.8472973241558929, "learning_rate": 6.236398608553041e-05, "loss": 0.8676, "step": 4194 }, { "epoch": 1.900770276393294, "grad_norm": 0.7429818485130683, "learning_rate": 6.235349379390166e-05, "loss": 0.8793, "step": 4195 }, { "epoch": 1.9012233801540552, "grad_norm": 0.5938090461155467, "learning_rate": 6.234299926530792e-05, "loss": 0.8868, "step": 4196 }, { "epoch": 1.9016764839148164, "grad_norm": 0.4310366040868412, "learning_rate": 6.233250250079938e-05, "loss": 0.8858, "step": 4197 }, { "epoch": 1.9021295876755777, "grad_norm": 0.4456871826916923, "learning_rate": 6.232200350142649e-05, "loss": 0.8851, "step": 4198 }, { "epoch": 1.902582691436339, "grad_norm": 0.5033630064135427, "learning_rate": 6.231150226823989e-05, "loss": 0.8606, "step": 4199 }, { "epoch": 1.9030357951971002, "grad_norm": 0.6692326594709057, "learning_rate": 6.230099880229046e-05, "loss": 0.8801, "step": 4200 }, { "epoch": 1.9034888989578613, "grad_norm": 0.5198130472081913, "learning_rate": 6.229049310462936e-05, "loss": 0.9039, "step": 4201 }, { "epoch": 1.9039420027186225, "grad_norm": 0.45238833551626406, "learning_rate": 6.227998517630785e-05, "loss": 0.8776, "step": 4202 }, { "epoch": 1.9043951064793838, "grad_norm": 0.42572299618065274, "learning_rate": 6.226947501837752e-05, "loss": 0.8611, "step": 4203 }, { "epoch": 1.904848210240145, "grad_norm": 0.4806723503455454, "learning_rate": 6.225896263189013e-05, "loss": 0.8846, "step": 4204 }, { "epoch": 1.9053013140009063, "grad_norm": 0.49832162557158455, "learning_rate": 6.224844801789769e-05, "loss": 0.9099, "step": 4205 }, { "epoch": 1.9057544177616674, "grad_norm": 0.5507192061963861, "learning_rate": 6.22379311774524e-05, "loss": 0.8814, "step": 4206 }, { "epoch": 1.9062075215224286, "grad_norm": 0.5891232556600382, "learning_rate": 6.222741211160675e-05, "loss": 0.9002, "step": 4207 }, { "epoch": 1.9066606252831897, "grad_norm": 0.5246603502459889, "learning_rate": 6.221689082141335e-05, "loss": 0.8976, "step": 4208 }, { "epoch": 1.907113729043951, "grad_norm": 0.5663706337786437, "learning_rate": 6.220636730792512e-05, "loss": 0.8942, "step": 4209 }, { "epoch": 1.9075668328047124, "grad_norm": 0.6269822622787063, "learning_rate": 6.219584157219517e-05, "loss": 0.9073, "step": 4210 }, { "epoch": 1.9080199365654735, "grad_norm": 0.6709657499334317, "learning_rate": 6.218531361527686e-05, "loss": 0.8853, "step": 4211 }, { "epoch": 1.9084730403262347, "grad_norm": 0.7198788615598909, "learning_rate": 6.217478343822368e-05, "loss": 0.8945, "step": 4212 }, { "epoch": 1.9089261440869958, "grad_norm": 0.7329015346924111, "learning_rate": 6.216425104208947e-05, "loss": 0.9049, "step": 4213 }, { "epoch": 1.9093792478477571, "grad_norm": 0.7668761937333932, "learning_rate": 6.21537164279282e-05, "loss": 0.8924, "step": 4214 }, { "epoch": 1.9098323516085185, "grad_norm": 0.8303269672812594, "learning_rate": 6.21431795967941e-05, "loss": 0.8995, "step": 4215 }, { "epoch": 1.9102854553692796, "grad_norm": 0.9206797315961693, "learning_rate": 6.213264054974163e-05, "loss": 0.8992, "step": 4216 }, { "epoch": 1.9107385591300408, "grad_norm": 0.9756832219138135, "learning_rate": 6.212209928782544e-05, "loss": 0.9093, "step": 4217 }, { "epoch": 1.911191662890802, "grad_norm": 0.9312612915160894, "learning_rate": 6.211155581210041e-05, "loss": 0.8994, "step": 4218 }, { "epoch": 1.9116447666515632, "grad_norm": 0.778618124116339, "learning_rate": 6.210101012362166e-05, "loss": 0.8927, "step": 4219 }, { "epoch": 1.9120978704123244, "grad_norm": 0.6935075691698102, "learning_rate": 6.209046222344453e-05, "loss": 0.888, "step": 4220 }, { "epoch": 1.9125509741730857, "grad_norm": 0.7531360801265733, "learning_rate": 6.207991211262456e-05, "loss": 0.8918, "step": 4221 }, { "epoch": 1.9130040779338469, "grad_norm": 0.7567952857952743, "learning_rate": 6.206935979221752e-05, "loss": 0.8787, "step": 4222 }, { "epoch": 1.913457181694608, "grad_norm": 0.5790944275436405, "learning_rate": 6.205880526327942e-05, "loss": 0.8795, "step": 4223 }, { "epoch": 1.9139102854553691, "grad_norm": 0.39493001352377455, "learning_rate": 6.204824852686647e-05, "loss": 0.901, "step": 4224 }, { "epoch": 1.9143633892161305, "grad_norm": 0.4352796500371513, "learning_rate": 6.203768958403511e-05, "loss": 0.8756, "step": 4225 }, { "epoch": 1.9148164929768918, "grad_norm": 0.5581540336902452, "learning_rate": 6.202712843584198e-05, "loss": 0.897, "step": 4226 }, { "epoch": 1.915269596737653, "grad_norm": 0.5364957811154327, "learning_rate": 6.201656508334398e-05, "loss": 0.9009, "step": 4227 }, { "epoch": 1.915722700498414, "grad_norm": 0.6190297990345587, "learning_rate": 6.200599952759821e-05, "loss": 0.9023, "step": 4228 }, { "epoch": 1.9161758042591752, "grad_norm": 0.7455941011023854, "learning_rate": 6.199543176966195e-05, "loss": 0.8986, "step": 4229 }, { "epoch": 1.9166289080199366, "grad_norm": 0.8185808805042696, "learning_rate": 6.198486181059279e-05, "loss": 0.8807, "step": 4230 }, { "epoch": 1.917082011780698, "grad_norm": 0.9127523512666963, "learning_rate": 6.197428965144846e-05, "loss": 0.899, "step": 4231 }, { "epoch": 1.917535115541459, "grad_norm": 1.0388030516440003, "learning_rate": 6.196371529328694e-05, "loss": 0.9122, "step": 4232 }, { "epoch": 1.9179882193022202, "grad_norm": 1.024639430431278, "learning_rate": 6.195313873716644e-05, "loss": 0.8864, "step": 4233 }, { "epoch": 1.9184413230629813, "grad_norm": 0.9825307168028632, "learning_rate": 6.19425599841454e-05, "loss": 0.8854, "step": 4234 }, { "epoch": 1.9188944268237427, "grad_norm": 0.9639298273811195, "learning_rate": 6.193197903528241e-05, "loss": 0.9017, "step": 4235 }, { "epoch": 1.9193475305845038, "grad_norm": 0.9790209833212845, "learning_rate": 6.192139589163636e-05, "loss": 0.8907, "step": 4236 }, { "epoch": 1.9198006343452652, "grad_norm": 0.8544397773178959, "learning_rate": 6.191081055426633e-05, "loss": 0.9062, "step": 4237 }, { "epoch": 1.9202537381060263, "grad_norm": 0.6123572021967714, "learning_rate": 6.190022302423161e-05, "loss": 0.8928, "step": 4238 }, { "epoch": 1.9207068418667874, "grad_norm": 0.5004528241688517, "learning_rate": 6.188963330259174e-05, "loss": 0.8965, "step": 4239 }, { "epoch": 1.9211599456275485, "grad_norm": 0.434192640520721, "learning_rate": 6.187904139040644e-05, "loss": 0.8933, "step": 4240 }, { "epoch": 1.92161304938831, "grad_norm": 0.4019942855453845, "learning_rate": 6.186844728873566e-05, "loss": 0.8902, "step": 4241 }, { "epoch": 1.9220661531490713, "grad_norm": 0.4754882998650924, "learning_rate": 6.185785099863959e-05, "loss": 0.8878, "step": 4242 }, { "epoch": 1.9225192569098324, "grad_norm": 0.5913419223721693, "learning_rate": 6.18472525211786e-05, "loss": 0.8994, "step": 4243 }, { "epoch": 1.9229723606705935, "grad_norm": 0.7416705865767846, "learning_rate": 6.183665185741334e-05, "loss": 0.878, "step": 4244 }, { "epoch": 1.9234254644313546, "grad_norm": 0.8748792608513716, "learning_rate": 6.182604900840462e-05, "loss": 0.9025, "step": 4245 }, { "epoch": 1.923878568192116, "grad_norm": 0.9647779682912607, "learning_rate": 6.181544397521349e-05, "loss": 0.8639, "step": 4246 }, { "epoch": 1.9243316719528774, "grad_norm": 0.9931017433200647, "learning_rate": 6.180483675890124e-05, "loss": 0.8801, "step": 4247 }, { "epoch": 1.9247847757136385, "grad_norm": 0.9199462373820798, "learning_rate": 6.179422736052934e-05, "loss": 0.8846, "step": 4248 }, { "epoch": 1.9252378794743996, "grad_norm": 0.8187747331482743, "learning_rate": 6.178361578115949e-05, "loss": 0.9081, "step": 4249 }, { "epoch": 1.9256909832351607, "grad_norm": 0.6539154771067041, "learning_rate": 6.177300202185362e-05, "loss": 0.9063, "step": 4250 }, { "epoch": 1.926144086995922, "grad_norm": 0.4843752320748223, "learning_rate": 6.176238608367391e-05, "loss": 0.8734, "step": 4251 }, { "epoch": 1.9265971907566832, "grad_norm": 0.41804137473834985, "learning_rate": 6.175176796768266e-05, "loss": 0.8695, "step": 4252 }, { "epoch": 1.9270502945174446, "grad_norm": 0.4155193751669997, "learning_rate": 6.174114767494251e-05, "loss": 0.8773, "step": 4253 }, { "epoch": 1.9275033982782057, "grad_norm": 0.5848429856956747, "learning_rate": 6.17305252065162e-05, "loss": 0.8652, "step": 4254 }, { "epoch": 1.9279565020389668, "grad_norm": 0.7291339474733973, "learning_rate": 6.171990056346679e-05, "loss": 0.9029, "step": 4255 }, { "epoch": 1.928409605799728, "grad_norm": 0.7941706972238208, "learning_rate": 6.170927374685747e-05, "loss": 0.8929, "step": 4256 }, { "epoch": 1.9288627095604893, "grad_norm": 0.8431917435911509, "learning_rate": 6.169864475775172e-05, "loss": 0.89, "step": 4257 }, { "epoch": 1.9293158133212507, "grad_norm": 0.8629134340110758, "learning_rate": 6.168801359721321e-05, "loss": 0.8861, "step": 4258 }, { "epoch": 1.9297689170820118, "grad_norm": 0.8606808705176173, "learning_rate": 6.16773802663058e-05, "loss": 0.8876, "step": 4259 }, { "epoch": 1.930222020842773, "grad_norm": 0.7193934039360419, "learning_rate": 6.166674476609362e-05, "loss": 0.8862, "step": 4260 }, { "epoch": 1.930675124603534, "grad_norm": 0.4991506117681391, "learning_rate": 6.165610709764096e-05, "loss": 0.8861, "step": 4261 }, { "epoch": 1.9311282283642954, "grad_norm": 0.31056830235606053, "learning_rate": 6.164546726201238e-05, "loss": 0.9071, "step": 4262 }, { "epoch": 1.9315813321250568, "grad_norm": 0.3653231268558209, "learning_rate": 6.163482526027262e-05, "loss": 0.9094, "step": 4263 }, { "epoch": 1.932034435885818, "grad_norm": 0.5576922222608863, "learning_rate": 6.162418109348665e-05, "loss": 0.8831, "step": 4264 }, { "epoch": 1.932487539646579, "grad_norm": 0.6837962710524934, "learning_rate": 6.161353476271967e-05, "loss": 0.8918, "step": 4265 }, { "epoch": 1.9329406434073402, "grad_norm": 0.727782815801632, "learning_rate": 6.160288626903705e-05, "loss": 0.8727, "step": 4266 }, { "epoch": 1.9333937471681015, "grad_norm": 0.7802953739269209, "learning_rate": 6.159223561350444e-05, "loss": 0.8828, "step": 4267 }, { "epoch": 1.9338468509288627, "grad_norm": 0.7066551749336324, "learning_rate": 6.158158279718766e-05, "loss": 0.8885, "step": 4268 }, { "epoch": 1.934299954689624, "grad_norm": 0.6675831774938069, "learning_rate": 6.157092782115276e-05, "loss": 0.893, "step": 4269 }, { "epoch": 1.9347530584503851, "grad_norm": 0.6407558929747971, "learning_rate": 6.156027068646604e-05, "loss": 0.8792, "step": 4270 }, { "epoch": 1.9352061622111463, "grad_norm": 0.5921391777125348, "learning_rate": 6.154961139419394e-05, "loss": 0.873, "step": 4271 }, { "epoch": 1.9356592659719074, "grad_norm": 0.4043738031614325, "learning_rate": 6.153894994540319e-05, "loss": 0.8863, "step": 4272 }, { "epoch": 1.9361123697326688, "grad_norm": 0.34320171093268304, "learning_rate": 6.152828634116067e-05, "loss": 0.9007, "step": 4273 }, { "epoch": 1.9365654734934301, "grad_norm": 0.4044429984651648, "learning_rate": 6.151762058253355e-05, "loss": 0.8935, "step": 4274 }, { "epoch": 1.9370185772541912, "grad_norm": 0.38563027889860735, "learning_rate": 6.150695267058916e-05, "loss": 0.8946, "step": 4275 }, { "epoch": 1.9374716810149524, "grad_norm": 0.35040722444700134, "learning_rate": 6.149628260639505e-05, "loss": 0.9021, "step": 4276 }, { "epoch": 1.9379247847757135, "grad_norm": 0.38567000588424816, "learning_rate": 6.148561039101903e-05, "loss": 0.9011, "step": 4277 }, { "epoch": 1.9383778885364749, "grad_norm": 0.39719978763941777, "learning_rate": 6.147493602552906e-05, "loss": 0.9117, "step": 4278 }, { "epoch": 1.9388309922972362, "grad_norm": 0.49940318367975073, "learning_rate": 6.146425951099338e-05, "loss": 0.9042, "step": 4279 }, { "epoch": 1.9392840960579973, "grad_norm": 0.5841106797881711, "learning_rate": 6.145358084848038e-05, "loss": 0.8853, "step": 4280 }, { "epoch": 1.9397371998187585, "grad_norm": 0.7190260302268386, "learning_rate": 6.14429000390587e-05, "loss": 0.8683, "step": 4281 }, { "epoch": 1.9401903035795196, "grad_norm": 0.7649958350846261, "learning_rate": 6.143221708379722e-05, "loss": 0.8767, "step": 4282 }, { "epoch": 1.940643407340281, "grad_norm": 0.7638487442716168, "learning_rate": 6.1421531983765e-05, "loss": 0.884, "step": 4283 }, { "epoch": 1.941096511101042, "grad_norm": 0.8829671225400902, "learning_rate": 6.14108447400313e-05, "loss": 0.8796, "step": 4284 }, { "epoch": 1.9415496148618034, "grad_norm": 0.7943894223916276, "learning_rate": 6.140015535366564e-05, "loss": 0.887, "step": 4285 }, { "epoch": 1.9420027186225646, "grad_norm": 0.8242622465539066, "learning_rate": 6.138946382573772e-05, "loss": 0.9115, "step": 4286 }, { "epoch": 1.9424558223833257, "grad_norm": 0.8956708579738244, "learning_rate": 6.137877015731749e-05, "loss": 0.8927, "step": 4287 }, { "epoch": 1.9429089261440868, "grad_norm": 1.0039129739953678, "learning_rate": 6.136807434947505e-05, "loss": 0.8904, "step": 4288 }, { "epoch": 1.9433620299048482, "grad_norm": 1.0664760981918773, "learning_rate": 6.135737640328078e-05, "loss": 0.8968, "step": 4289 }, { "epoch": 1.9438151336656095, "grad_norm": 0.772238706417757, "learning_rate": 6.134667631980523e-05, "loss": 0.8812, "step": 4290 }, { "epoch": 1.9442682374263707, "grad_norm": 0.5695840990168402, "learning_rate": 6.13359741001192e-05, "loss": 0.8906, "step": 4291 }, { "epoch": 1.9447213411871318, "grad_norm": 0.5182911714630163, "learning_rate": 6.132526974529367e-05, "loss": 0.9014, "step": 4292 }, { "epoch": 1.945174444947893, "grad_norm": 0.580893253241591, "learning_rate": 6.131456325639987e-05, "loss": 0.875, "step": 4293 }, { "epoch": 1.9456275487086543, "grad_norm": 0.6025677245350964, "learning_rate": 6.130385463450919e-05, "loss": 0.8869, "step": 4294 }, { "epoch": 1.9460806524694156, "grad_norm": 0.5762870203872986, "learning_rate": 6.129314388069331e-05, "loss": 0.8906, "step": 4295 }, { "epoch": 1.9465337562301768, "grad_norm": 0.5838253990088913, "learning_rate": 6.128243099602404e-05, "loss": 0.8876, "step": 4296 }, { "epoch": 1.946986859990938, "grad_norm": 0.6879404092614619, "learning_rate": 6.127171598157347e-05, "loss": 0.9007, "step": 4297 }, { "epoch": 1.947439963751699, "grad_norm": 0.7647722973141506, "learning_rate": 6.126099883841384e-05, "loss": 0.8929, "step": 4298 }, { "epoch": 1.9478930675124604, "grad_norm": 0.8157631010404578, "learning_rate": 6.125027956761768e-05, "loss": 0.9163, "step": 4299 }, { "epoch": 1.9483461712732215, "grad_norm": 0.6376771656794586, "learning_rate": 6.123955817025768e-05, "loss": 0.8962, "step": 4300 }, { "epoch": 1.9487992750339829, "grad_norm": 0.5896322509237735, "learning_rate": 6.122883464740673e-05, "loss": 0.8814, "step": 4301 }, { "epoch": 1.949252378794744, "grad_norm": 0.543611623119012, "learning_rate": 6.1218109000138e-05, "loss": 0.9039, "step": 4302 }, { "epoch": 1.9497054825555051, "grad_norm": 0.5200309964341281, "learning_rate": 6.12073812295248e-05, "loss": 0.8885, "step": 4303 }, { "epoch": 1.9501585863162663, "grad_norm": 0.5072154879087917, "learning_rate": 6.119665133664068e-05, "loss": 0.8996, "step": 4304 }, { "epoch": 1.9506116900770276, "grad_norm": 0.3769133249863421, "learning_rate": 6.118591932255942e-05, "loss": 0.8799, "step": 4305 }, { "epoch": 1.951064793837789, "grad_norm": 0.37349727868199023, "learning_rate": 6.117518518835499e-05, "loss": 0.8911, "step": 4306 }, { "epoch": 1.95151789759855, "grad_norm": 0.47636267947755107, "learning_rate": 6.116444893510159e-05, "loss": 0.8801, "step": 4307 }, { "epoch": 1.9519710013593112, "grad_norm": 0.4610856476465857, "learning_rate": 6.115371056387358e-05, "loss": 0.8831, "step": 4308 }, { "epoch": 1.9524241051200724, "grad_norm": 0.47945642096017044, "learning_rate": 6.114297007574564e-05, "loss": 0.8733, "step": 4309 }, { "epoch": 1.9528772088808337, "grad_norm": 0.5598278711976397, "learning_rate": 6.113222747179254e-05, "loss": 0.8748, "step": 4310 }, { "epoch": 1.953330312641595, "grad_norm": 0.6984729291397564, "learning_rate": 6.112148275308933e-05, "loss": 0.8816, "step": 4311 }, { "epoch": 1.9537834164023562, "grad_norm": 0.7215757291436992, "learning_rate": 6.111073592071128e-05, "loss": 0.8934, "step": 4312 }, { "epoch": 1.9542365201631173, "grad_norm": 0.5719662248522505, "learning_rate": 6.109998697573384e-05, "loss": 0.8893, "step": 4313 }, { "epoch": 1.9546896239238785, "grad_norm": 0.4944053388289034, "learning_rate": 6.108923591923267e-05, "loss": 0.8757, "step": 4314 }, { "epoch": 1.9551427276846398, "grad_norm": 0.47444436641027166, "learning_rate": 6.107848275228365e-05, "loss": 0.9097, "step": 4315 }, { "epoch": 1.955595831445401, "grad_norm": 0.3699242467384794, "learning_rate": 6.106772747596291e-05, "loss": 0.8792, "step": 4316 }, { "epoch": 1.9560489352061623, "grad_norm": 0.359892991530378, "learning_rate": 6.10569700913467e-05, "loss": 0.8918, "step": 4317 }, { "epoch": 1.9565020389669234, "grad_norm": 0.512255896214272, "learning_rate": 6.104621059951157e-05, "loss": 0.888, "step": 4318 }, { "epoch": 1.9569551427276846, "grad_norm": 0.6781590293705814, "learning_rate": 6.103544900153425e-05, "loss": 0.8823, "step": 4319 }, { "epoch": 1.9574082464884457, "grad_norm": 0.6197310668607808, "learning_rate": 6.102468529849166e-05, "loss": 0.904, "step": 4320 }, { "epoch": 1.957861350249207, "grad_norm": 0.43295976835627553, "learning_rate": 6.101391949146095e-05, "loss": 0.885, "step": 4321 }, { "epoch": 1.9583144540099684, "grad_norm": 0.4131974696485323, "learning_rate": 6.100315158151949e-05, "loss": 0.9075, "step": 4322 }, { "epoch": 1.9587675577707295, "grad_norm": 0.4916453290773902, "learning_rate": 6.0992381569744864e-05, "loss": 0.8783, "step": 4323 }, { "epoch": 1.9592206615314907, "grad_norm": 0.6293219110883415, "learning_rate": 6.098160945721481e-05, "loss": 0.9022, "step": 4324 }, { "epoch": 1.9596737652922518, "grad_norm": 0.7601683691496849, "learning_rate": 6.097083524500735e-05, "loss": 0.8922, "step": 4325 }, { "epoch": 1.9601268690530131, "grad_norm": 0.8568898928843659, "learning_rate": 6.096005893420067e-05, "loss": 0.8862, "step": 4326 }, { "epoch": 1.9605799728137745, "grad_norm": 0.9322993257819187, "learning_rate": 6.094928052587318e-05, "loss": 0.8883, "step": 4327 }, { "epoch": 1.9610330765745356, "grad_norm": 0.9162015994295652, "learning_rate": 6.0938500021103514e-05, "loss": 0.8678, "step": 4328 }, { "epoch": 1.9614861803352968, "grad_norm": 0.8898712279640248, "learning_rate": 6.0927717420970485e-05, "loss": 0.8894, "step": 4329 }, { "epoch": 1.961939284096058, "grad_norm": 0.8240391254919859, "learning_rate": 6.0916932726553136e-05, "loss": 0.8856, "step": 4330 }, { "epoch": 1.9623923878568192, "grad_norm": 0.9317200383181901, "learning_rate": 6.090614593893072e-05, "loss": 0.9006, "step": 4331 }, { "epoch": 1.9628454916175804, "grad_norm": 0.956368623722022, "learning_rate": 6.0895357059182716e-05, "loss": 0.8834, "step": 4332 }, { "epoch": 1.9632985953783417, "grad_norm": 1.017117895102472, "learning_rate": 6.088456608838874e-05, "loss": 0.9047, "step": 4333 }, { "epoch": 1.9637516991391029, "grad_norm": 0.8289496628874008, "learning_rate": 6.0873773027628716e-05, "loss": 0.8827, "step": 4334 }, { "epoch": 1.964204802899864, "grad_norm": 0.6463756039639205, "learning_rate": 6.0862977877982705e-05, "loss": 0.8803, "step": 4335 }, { "epoch": 1.9646579066606251, "grad_norm": 0.5382204370363857, "learning_rate": 6.0852180640531016e-05, "loss": 0.8936, "step": 4336 }, { "epoch": 1.9651110104213865, "grad_norm": 0.61108465081033, "learning_rate": 6.084138131635414e-05, "loss": 0.8758, "step": 4337 }, { "epoch": 1.9655641141821478, "grad_norm": 0.6084414731390393, "learning_rate": 6.08305799065328e-05, "loss": 0.8963, "step": 4338 }, { "epoch": 1.966017217942909, "grad_norm": 0.548030951617928, "learning_rate": 6.081977641214792e-05, "loss": 0.861, "step": 4339 }, { "epoch": 1.96647032170367, "grad_norm": 0.3841664861634688, "learning_rate": 6.080897083428063e-05, "loss": 0.8892, "step": 4340 }, { "epoch": 1.9669234254644312, "grad_norm": 0.3739572992086809, "learning_rate": 6.079816317401225e-05, "loss": 0.8955, "step": 4341 }, { "epoch": 1.9673765292251926, "grad_norm": 0.4602041913811693, "learning_rate": 6.078735343242435e-05, "loss": 0.9027, "step": 4342 }, { "epoch": 1.967829632985954, "grad_norm": 0.468081822206089, "learning_rate": 6.077654161059868e-05, "loss": 0.899, "step": 4343 }, { "epoch": 1.968282736746715, "grad_norm": 0.42161287943642173, "learning_rate": 6.0765727709617194e-05, "loss": 0.8779, "step": 4344 }, { "epoch": 1.9687358405074762, "grad_norm": 0.4546570747850144, "learning_rate": 6.0754911730562075e-05, "loss": 0.8904, "step": 4345 }, { "epoch": 1.9691889442682373, "grad_norm": 0.6117418426661446, "learning_rate": 6.0744093674515695e-05, "loss": 0.8943, "step": 4346 }, { "epoch": 1.9696420480289987, "grad_norm": 0.7905009777192888, "learning_rate": 6.073327354256064e-05, "loss": 0.8875, "step": 4347 }, { "epoch": 1.9700951517897598, "grad_norm": 0.9098174331736414, "learning_rate": 6.072245133577971e-05, "loss": 0.8836, "step": 4348 }, { "epoch": 1.9705482555505212, "grad_norm": 0.9169379744797286, "learning_rate": 6.07116270552559e-05, "loss": 0.8799, "step": 4349 }, { "epoch": 1.9710013593112823, "grad_norm": 0.9108472195569998, "learning_rate": 6.070080070207245e-05, "loss": 0.8793, "step": 4350 }, { "epoch": 1.9714544630720434, "grad_norm": 0.8942894585464718, "learning_rate": 6.068997227731274e-05, "loss": 0.8591, "step": 4351 }, { "epoch": 1.9719075668328045, "grad_norm": 0.803596544060604, "learning_rate": 6.067914178206041e-05, "loss": 0.8937, "step": 4352 }, { "epoch": 1.972360670593566, "grad_norm": 0.5577554706930498, "learning_rate": 6.0668309217399294e-05, "loss": 0.8742, "step": 4353 }, { "epoch": 1.9728137743543273, "grad_norm": 0.45616585484252525, "learning_rate": 6.0657474584413433e-05, "loss": 0.88, "step": 4354 }, { "epoch": 1.9732668781150884, "grad_norm": 0.5177253645363769, "learning_rate": 6.064663788418706e-05, "loss": 0.8842, "step": 4355 }, { "epoch": 1.9737199818758495, "grad_norm": 0.5992233225500178, "learning_rate": 6.0635799117804655e-05, "loss": 0.8731, "step": 4356 }, { "epoch": 1.9741730856366106, "grad_norm": 0.7719276180207774, "learning_rate": 6.062495828635086e-05, "loss": 0.8847, "step": 4357 }, { "epoch": 1.974626189397372, "grad_norm": 0.907215193397564, "learning_rate": 6.0614115390910544e-05, "loss": 0.9185, "step": 4358 }, { "epoch": 1.9750792931581334, "grad_norm": 1.0041980882852002, "learning_rate": 6.060327043256877e-05, "loss": 0.9023, "step": 4359 }, { "epoch": 1.9755323969188945, "grad_norm": 1.0398400143418987, "learning_rate": 6.059242341241083e-05, "loss": 0.876, "step": 4360 }, { "epoch": 1.9759855006796556, "grad_norm": 0.9073548511124849, "learning_rate": 6.0581574331522216e-05, "loss": 0.891, "step": 4361 }, { "epoch": 1.9764386044404167, "grad_norm": 0.6422745875407574, "learning_rate": 6.0570723190988605e-05, "loss": 0.8938, "step": 4362 }, { "epoch": 1.976891708201178, "grad_norm": 0.46479464156574923, "learning_rate": 6.05598699918959e-05, "loss": 0.8931, "step": 4363 }, { "epoch": 1.9773448119619392, "grad_norm": 0.5752574929829909, "learning_rate": 6.0549014735330204e-05, "loss": 0.8842, "step": 4364 }, { "epoch": 1.9777979157227006, "grad_norm": 0.5987256289489314, "learning_rate": 6.053815742237785e-05, "loss": 0.8954, "step": 4365 }, { "epoch": 1.9782510194834617, "grad_norm": 0.6478018421585217, "learning_rate": 6.052729805412532e-05, "loss": 0.8948, "step": 4366 }, { "epoch": 1.9787041232442228, "grad_norm": 0.6328897638100491, "learning_rate": 6.051643663165936e-05, "loss": 0.8903, "step": 4367 }, { "epoch": 1.979157227004984, "grad_norm": 0.5820297378353545, "learning_rate": 6.0505573156066876e-05, "loss": 0.8834, "step": 4368 }, { "epoch": 1.9796103307657453, "grad_norm": 0.6066513267219013, "learning_rate": 6.0494707628435e-05, "loss": 0.8822, "step": 4369 }, { "epoch": 1.9800634345265067, "grad_norm": 0.6961140981132963, "learning_rate": 6.04838400498511e-05, "loss": 0.8758, "step": 4370 }, { "epoch": 1.9805165382872678, "grad_norm": 0.7573010528252632, "learning_rate": 6.0472970421402706e-05, "loss": 0.8911, "step": 4371 }, { "epoch": 1.980969642048029, "grad_norm": 0.7679502747524395, "learning_rate": 6.046209874417756e-05, "loss": 0.8892, "step": 4372 }, { "epoch": 1.98142274580879, "grad_norm": 0.9910289965733274, "learning_rate": 6.045122501926361e-05, "loss": 0.8958, "step": 4373 }, { "epoch": 1.9818758495695514, "grad_norm": 0.7248029856214432, "learning_rate": 6.044034924774903e-05, "loss": 0.8899, "step": 4374 }, { "epoch": 1.9823289533303128, "grad_norm": 0.6508859844131248, "learning_rate": 6.042947143072218e-05, "loss": 0.8861, "step": 4375 }, { "epoch": 1.982782057091074, "grad_norm": 0.5134890443361984, "learning_rate": 6.041859156927161e-05, "loss": 0.8847, "step": 4376 }, { "epoch": 1.983235160851835, "grad_norm": 0.4999428929790091, "learning_rate": 6.0407709664486104e-05, "loss": 0.8859, "step": 4377 }, { "epoch": 1.9836882646125962, "grad_norm": 0.516407027298636, "learning_rate": 6.039682571745466e-05, "loss": 0.8695, "step": 4378 }, { "epoch": 1.9841413683733575, "grad_norm": 0.4489755157821197, "learning_rate": 6.038593972926641e-05, "loss": 0.8927, "step": 4379 }, { "epoch": 1.9845944721341187, "grad_norm": 0.3967950008425985, "learning_rate": 6.03750517010108e-05, "loss": 0.8851, "step": 4380 }, { "epoch": 1.98504757589488, "grad_norm": 0.40541816402476333, "learning_rate": 6.0364161633777374e-05, "loss": 0.9002, "step": 4381 }, { "epoch": 1.9855006796556411, "grad_norm": 0.5191652924376073, "learning_rate": 6.035326952865594e-05, "loss": 0.903, "step": 4382 }, { "epoch": 1.9859537834164023, "grad_norm": 0.5500939009816922, "learning_rate": 6.0342375386736506e-05, "loss": 0.8809, "step": 4383 }, { "epoch": 1.9864068871771634, "grad_norm": 0.6268292857017913, "learning_rate": 6.0331479209109266e-05, "loss": 0.8742, "step": 4384 }, { "epoch": 1.9868599909379248, "grad_norm": 0.7572331875684808, "learning_rate": 6.032058099686462e-05, "loss": 0.879, "step": 4385 }, { "epoch": 1.9873130946986861, "grad_norm": 0.8032241191087895, "learning_rate": 6.030968075109317e-05, "loss": 0.8885, "step": 4386 }, { "epoch": 1.9877661984594472, "grad_norm": 0.8290540399486511, "learning_rate": 6.029877847288575e-05, "loss": 0.8788, "step": 4387 }, { "epoch": 1.9882193022202084, "grad_norm": 0.8108456297078371, "learning_rate": 6.0287874163333355e-05, "loss": 0.8821, "step": 4388 }, { "epoch": 1.9886724059809695, "grad_norm": 0.8070733152552412, "learning_rate": 6.027696782352722e-05, "loss": 0.9237, "step": 4389 }, { "epoch": 1.9891255097417309, "grad_norm": 0.7209774687106124, "learning_rate": 6.026605945455876e-05, "loss": 0.911, "step": 4390 }, { "epoch": 1.9895786135024922, "grad_norm": 0.6436159075166851, "learning_rate": 6.02551490575196e-05, "loss": 0.8851, "step": 4391 }, { "epoch": 1.9900317172632533, "grad_norm": 0.5904887987874273, "learning_rate": 6.024423663350158e-05, "loss": 0.8879, "step": 4392 }, { "epoch": 1.9904848210240145, "grad_norm": 0.5865808606580051, "learning_rate": 6.023332218359672e-05, "loss": 0.881, "step": 4393 }, { "epoch": 1.9909379247847756, "grad_norm": 0.39173782389715966, "learning_rate": 6.022240570889725e-05, "loss": 0.8967, "step": 4394 }, { "epoch": 1.991391028545537, "grad_norm": 0.3865577738118716, "learning_rate": 6.021148721049561e-05, "loss": 0.8834, "step": 4395 }, { "epoch": 1.991844132306298, "grad_norm": 0.4772457010143813, "learning_rate": 6.020056668948444e-05, "loss": 0.9053, "step": 4396 }, { "epoch": 1.9922972360670594, "grad_norm": 0.39594545964984923, "learning_rate": 6.018964414695658e-05, "loss": 0.8812, "step": 4397 }, { "epoch": 1.9927503398278206, "grad_norm": 0.4212628343743852, "learning_rate": 6.0178719584005086e-05, "loss": 0.8932, "step": 4398 }, { "epoch": 1.9932034435885817, "grad_norm": 0.4246119846405452, "learning_rate": 6.016779300172318e-05, "loss": 0.8931, "step": 4399 }, { "epoch": 1.9936565473493428, "grad_norm": 0.4274640639356403, "learning_rate": 6.015686440120433e-05, "loss": 0.8867, "step": 4400 }, { "epoch": 1.9941096511101042, "grad_norm": 0.4275946746703128, "learning_rate": 6.014593378354219e-05, "loss": 0.894, "step": 4401 }, { "epoch": 1.9945627548708655, "grad_norm": 0.43794262421429164, "learning_rate": 6.013500114983059e-05, "loss": 0.8653, "step": 4402 }, { "epoch": 1.9950158586316267, "grad_norm": 0.43560741286690735, "learning_rate": 6.01240665011636e-05, "loss": 0.8615, "step": 4403 }, { "epoch": 1.9954689623923878, "grad_norm": 0.397405267888491, "learning_rate": 6.011312983863546e-05, "loss": 0.8723, "step": 4404 }, { "epoch": 1.995922066153149, "grad_norm": 0.3996877462978369, "learning_rate": 6.010219116334064e-05, "loss": 0.872, "step": 4405 }, { "epoch": 1.9963751699139103, "grad_norm": 0.4645347601918068, "learning_rate": 6.009125047637379e-05, "loss": 0.906, "step": 4406 }, { "epoch": 1.9968282736746716, "grad_norm": 0.5193713858756344, "learning_rate": 6.008030777882978e-05, "loss": 0.8963, "step": 4407 }, { "epoch": 1.9972813774354328, "grad_norm": 0.6128363261185797, "learning_rate": 6.0069363071803664e-05, "loss": 0.8721, "step": 4408 }, { "epoch": 1.997734481196194, "grad_norm": 0.8006886332791512, "learning_rate": 6.005841635639069e-05, "loss": 0.9258, "step": 4409 }, { "epoch": 1.998187584956955, "grad_norm": 0.9939949305996999, "learning_rate": 6.0047467633686355e-05, "loss": 0.8967, "step": 4410 }, { "epoch": 1.9986406887177164, "grad_norm": 1.1370524574872372, "learning_rate": 6.003651690478631e-05, "loss": 0.863, "step": 4411 }, { "epoch": 1.9990937924784775, "grad_norm": 0.7287214552885077, "learning_rate": 6.002556417078639e-05, "loss": 0.8932, "step": 4412 }, { "epoch": 1.9995468962392389, "grad_norm": 0.4049872960808134, "learning_rate": 6.001460943278269e-05, "loss": 0.904, "step": 4413 }, { "epoch": 2.0, "grad_norm": 0.4606515955329746, "learning_rate": 6.0003652691871465e-05, "loss": 0.8961, "step": 4414 }, { "epoch": 2.000453103760761, "grad_norm": 0.6178036960388313, "learning_rate": 5.999269394914919e-05, "loss": 0.8812, "step": 4415 }, { "epoch": 2.0009062075215223, "grad_norm": 0.7905515753941853, "learning_rate": 5.998173320571252e-05, "loss": 0.8676, "step": 4416 }, { "epoch": 2.001359311282284, "grad_norm": 1.904040849941001, "learning_rate": 5.997077046265832e-05, "loss": 0.8819, "step": 4417 }, { "epoch": 2.001812415043045, "grad_norm": 1.1317295172907151, "learning_rate": 5.9959805721083676e-05, "loss": 0.8757, "step": 4418 }, { "epoch": 2.002265518803806, "grad_norm": 130.00835278139425, "learning_rate": 5.994883898208584e-05, "loss": 1.7663, "step": 4419 }, { "epoch": 2.0027186225645672, "grad_norm": 3.1598401184358553, "learning_rate": 5.993787024676228e-05, "loss": 0.9325, "step": 4420 }, { "epoch": 2.0031717263253284, "grad_norm": 1.6634021523501796, "learning_rate": 5.9926899516210665e-05, "loss": 0.8865, "step": 4421 }, { "epoch": 2.00362483008609, "grad_norm": 2.6984057102230894, "learning_rate": 5.991592679152884e-05, "loss": 0.9139, "step": 4422 }, { "epoch": 2.004077933846851, "grad_norm": 227.80193419859478, "learning_rate": 5.9904952073814915e-05, "loss": 1.1839, "step": 4423 }, { "epoch": 2.004531037607612, "grad_norm": 3.2682992251616776, "learning_rate": 5.989397536416711e-05, "loss": 0.932, "step": 4424 }, { "epoch": 2.0049841413683733, "grad_norm": 1.0999293543912736, "learning_rate": 5.9882996663683916e-05, "loss": 0.8946, "step": 4425 }, { "epoch": 2.0054372451291345, "grad_norm": 1.794241433230143, "learning_rate": 5.987201597346399e-05, "loss": 0.8846, "step": 4426 }, { "epoch": 2.0058903488898956, "grad_norm": 1.5860966417736992, "learning_rate": 5.98610332946062e-05, "loss": 0.9099, "step": 4427 }, { "epoch": 2.006343452650657, "grad_norm": 1.2696044173575851, "learning_rate": 5.985004862820959e-05, "loss": 0.8864, "step": 4428 }, { "epoch": 2.0067965564114183, "grad_norm": 2.5806006580333163, "learning_rate": 5.983906197537344e-05, "loss": 0.9064, "step": 4429 }, { "epoch": 2.0072496601721794, "grad_norm": 1.8399780603239466, "learning_rate": 5.98280733371972e-05, "loss": 0.9288, "step": 4430 }, { "epoch": 2.0077027639329406, "grad_norm": 1.169239157417277, "learning_rate": 5.981708271478052e-05, "loss": 0.8949, "step": 4431 }, { "epoch": 2.0081558676937017, "grad_norm": 1.2210862129956972, "learning_rate": 5.9806090109223275e-05, "loss": 0.897, "step": 4432 }, { "epoch": 2.0086089714544633, "grad_norm": 1.144236980488488, "learning_rate": 5.979509552162551e-05, "loss": 0.9083, "step": 4433 }, { "epoch": 2.0090620752152244, "grad_norm": 1.0984238381219205, "learning_rate": 5.978409895308748e-05, "loss": 0.8932, "step": 4434 }, { "epoch": 2.0095151789759855, "grad_norm": 0.8943032646667604, "learning_rate": 5.977310040470963e-05, "loss": 0.8924, "step": 4435 }, { "epoch": 2.0099682827367467, "grad_norm": 1.0027186736143128, "learning_rate": 5.9762099877592614e-05, "loss": 0.8773, "step": 4436 }, { "epoch": 2.010421386497508, "grad_norm": 1.1074830073676458, "learning_rate": 5.9751097372837296e-05, "loss": 0.9151, "step": 4437 }, { "epoch": 2.0108744902582694, "grad_norm": 0.8363793427629449, "learning_rate": 5.9740092891544704e-05, "loss": 0.8953, "step": 4438 }, { "epoch": 2.0113275940190305, "grad_norm": 0.7300526962432614, "learning_rate": 5.9729086434816075e-05, "loss": 0.8755, "step": 4439 }, { "epoch": 2.0117806977797916, "grad_norm": 0.6401096062896032, "learning_rate": 5.971807800375286e-05, "loss": 0.8789, "step": 4440 }, { "epoch": 2.0122338015405528, "grad_norm": 0.7460151586666511, "learning_rate": 5.97070675994567e-05, "loss": 0.8736, "step": 4441 }, { "epoch": 2.012686905301314, "grad_norm": 0.701374873819546, "learning_rate": 5.969605522302943e-05, "loss": 0.8862, "step": 4442 }, { "epoch": 2.013140009062075, "grad_norm": 0.6258801587283072, "learning_rate": 5.968504087557309e-05, "loss": 0.8971, "step": 4443 }, { "epoch": 2.0135931128228366, "grad_norm": 0.678172379734572, "learning_rate": 5.967402455818989e-05, "loss": 0.8885, "step": 4444 }, { "epoch": 2.0140462165835977, "grad_norm": 0.8388351368699954, "learning_rate": 5.9663006271982284e-05, "loss": 0.867, "step": 4445 }, { "epoch": 2.014499320344359, "grad_norm": 0.8281143854352561, "learning_rate": 5.965198601805288e-05, "loss": 0.8799, "step": 4446 }, { "epoch": 2.01495242410512, "grad_norm": 0.6157184500731122, "learning_rate": 5.96409637975045e-05, "loss": 0.8755, "step": 4447 }, { "epoch": 2.015405527865881, "grad_norm": 0.5748547880166441, "learning_rate": 5.962993961144016e-05, "loss": 0.8664, "step": 4448 }, { "epoch": 2.0158586316266427, "grad_norm": 0.5233963542423201, "learning_rate": 5.96189134609631e-05, "loss": 0.8683, "step": 4449 }, { "epoch": 2.016311735387404, "grad_norm": 0.5700899456966912, "learning_rate": 5.960788534717671e-05, "loss": 0.8934, "step": 4450 }, { "epoch": 2.016764839148165, "grad_norm": 0.5686424768947755, "learning_rate": 5.959685527118459e-05, "loss": 0.8927, "step": 4451 }, { "epoch": 2.017217942908926, "grad_norm": 0.45905759627689635, "learning_rate": 5.958582323409056e-05, "loss": 0.8612, "step": 4452 }, { "epoch": 2.017671046669687, "grad_norm": 0.492324279410492, "learning_rate": 5.957478923699864e-05, "loss": 0.8864, "step": 4453 }, { "epoch": 2.018124150430449, "grad_norm": 0.5429360192418003, "learning_rate": 5.9563753281012984e-05, "loss": 0.8782, "step": 4454 }, { "epoch": 2.01857725419121, "grad_norm": 0.4636773593000698, "learning_rate": 5.9552715367238017e-05, "loss": 0.8835, "step": 4455 }, { "epoch": 2.019030357951971, "grad_norm": 0.5270481734089594, "learning_rate": 5.954167549677831e-05, "loss": 0.8916, "step": 4456 }, { "epoch": 2.019483461712732, "grad_norm": 0.44395819848167495, "learning_rate": 5.953063367073865e-05, "loss": 0.8633, "step": 4457 }, { "epoch": 2.0199365654734933, "grad_norm": 0.4347710938521246, "learning_rate": 5.9519589890224053e-05, "loss": 0.8858, "step": 4458 }, { "epoch": 2.0203896692342544, "grad_norm": 0.45758774430281934, "learning_rate": 5.950854415633964e-05, "loss": 0.8672, "step": 4459 }, { "epoch": 2.020842772995016, "grad_norm": 0.42209931020605757, "learning_rate": 5.9497496470190815e-05, "loss": 0.8626, "step": 4460 }, { "epoch": 2.021295876755777, "grad_norm": 0.468855750959822, "learning_rate": 5.948644683288313e-05, "loss": 0.8604, "step": 4461 }, { "epoch": 2.0217489805165383, "grad_norm": 0.5279289001462253, "learning_rate": 5.9475395245522376e-05, "loss": 0.8756, "step": 4462 }, { "epoch": 2.0222020842772994, "grad_norm": 0.49516936642447607, "learning_rate": 5.9464341709214484e-05, "loss": 0.8886, "step": 4463 }, { "epoch": 2.0226551880380605, "grad_norm": 0.47290791897886, "learning_rate": 5.94532862250656e-05, "loss": 0.8698, "step": 4464 }, { "epoch": 2.023108291798822, "grad_norm": 0.4505975878131233, "learning_rate": 5.9442228794182104e-05, "loss": 0.8609, "step": 4465 }, { "epoch": 2.0235613955595833, "grad_norm": 0.5236835317370927, "learning_rate": 5.9431169417670506e-05, "loss": 0.8657, "step": 4466 }, { "epoch": 2.0240144993203444, "grad_norm": 0.5922995310996901, "learning_rate": 5.942010809663756e-05, "loss": 0.8657, "step": 4467 }, { "epoch": 2.0244676030811055, "grad_norm": 0.5321825786258756, "learning_rate": 5.9409044832190186e-05, "loss": 0.8649, "step": 4468 }, { "epoch": 2.0249207068418666, "grad_norm": 0.5345142137117115, "learning_rate": 5.9397979625435534e-05, "loss": 0.8764, "step": 4469 }, { "epoch": 2.025373810602628, "grad_norm": 0.5677832827412904, "learning_rate": 5.9386912477480894e-05, "loss": 0.885, "step": 4470 }, { "epoch": 2.0258269143633894, "grad_norm": 0.5611435941981404, "learning_rate": 5.93758433894338e-05, "loss": 0.8889, "step": 4471 }, { "epoch": 2.0262800181241505, "grad_norm": 0.505999189073966, "learning_rate": 5.936477236240196e-05, "loss": 0.8769, "step": 4472 }, { "epoch": 2.0267331218849116, "grad_norm": 0.3905573200111917, "learning_rate": 5.935369939749327e-05, "loss": 0.8663, "step": 4473 }, { "epoch": 2.0271862256456727, "grad_norm": 0.40021242382286143, "learning_rate": 5.934262449581582e-05, "loss": 0.8574, "step": 4474 }, { "epoch": 2.027639329406434, "grad_norm": 0.47858553380545055, "learning_rate": 5.933154765847792e-05, "loss": 0.8773, "step": 4475 }, { "epoch": 2.0280924331671955, "grad_norm": 0.43620986939178935, "learning_rate": 5.932046888658804e-05, "loss": 0.8779, "step": 4476 }, { "epoch": 2.0285455369279566, "grad_norm": 0.35720878519416766, "learning_rate": 5.930938818125485e-05, "loss": 0.8891, "step": 4477 }, { "epoch": 2.0289986406887177, "grad_norm": 0.39236502689976716, "learning_rate": 5.929830554358726e-05, "loss": 0.8646, "step": 4478 }, { "epoch": 2.029451744449479, "grad_norm": 0.4267768228875459, "learning_rate": 5.928722097469429e-05, "loss": 0.8828, "step": 4479 }, { "epoch": 2.02990484821024, "grad_norm": 0.47174923231837573, "learning_rate": 5.9276134475685226e-05, "loss": 0.8799, "step": 4480 }, { "epoch": 2.0303579519710016, "grad_norm": 0.48448868021091196, "learning_rate": 5.9265046047669505e-05, "loss": 0.8508, "step": 4481 }, { "epoch": 2.0308110557317627, "grad_norm": 0.43064744963725277, "learning_rate": 5.925395569175677e-05, "loss": 0.8721, "step": 4482 }, { "epoch": 2.031264159492524, "grad_norm": 0.3738625255021079, "learning_rate": 5.924286340905687e-05, "loss": 0.8819, "step": 4483 }, { "epoch": 2.031717263253285, "grad_norm": 0.3695624254107285, "learning_rate": 5.923176920067983e-05, "loss": 0.8798, "step": 4484 }, { "epoch": 2.032170367014046, "grad_norm": 0.38311849944031173, "learning_rate": 5.922067306773587e-05, "loss": 0.8678, "step": 4485 }, { "epoch": 2.0326234707748077, "grad_norm": 0.3876282993435026, "learning_rate": 5.92095750113354e-05, "loss": 0.8683, "step": 4486 }, { "epoch": 2.033076574535569, "grad_norm": 0.3488181281887409, "learning_rate": 5.919847503258904e-05, "loss": 0.8833, "step": 4487 }, { "epoch": 2.03352967829633, "grad_norm": 0.508887601575597, "learning_rate": 5.918737313260759e-05, "loss": 0.8769, "step": 4488 }, { "epoch": 2.033982782057091, "grad_norm": 0.6251158823839424, "learning_rate": 5.917626931250203e-05, "loss": 0.8758, "step": 4489 }, { "epoch": 2.034435885817852, "grad_norm": 0.6342399262170167, "learning_rate": 5.9165163573383564e-05, "loss": 0.8778, "step": 4490 }, { "epoch": 2.0348889895786133, "grad_norm": 0.6673736150150459, "learning_rate": 5.915405591636354e-05, "loss": 0.8598, "step": 4491 }, { "epoch": 2.035342093339375, "grad_norm": 0.6406868786816842, "learning_rate": 5.914294634255355e-05, "loss": 0.8933, "step": 4492 }, { "epoch": 2.035795197100136, "grad_norm": 0.6673227646098849, "learning_rate": 5.913183485306535e-05, "loss": 0.8714, "step": 4493 }, { "epoch": 2.036248300860897, "grad_norm": 0.8209993071492241, "learning_rate": 5.912072144901088e-05, "loss": 0.8651, "step": 4494 }, { "epoch": 2.0367014046216583, "grad_norm": 0.7075786305822102, "learning_rate": 5.91096061315023e-05, "loss": 0.892, "step": 4495 }, { "epoch": 2.0371545083824194, "grad_norm": 0.7576856830193177, "learning_rate": 5.909848890165194e-05, "loss": 0.8691, "step": 4496 }, { "epoch": 2.037607612143181, "grad_norm": 0.8067354790768041, "learning_rate": 5.908736976057231e-05, "loss": 0.8684, "step": 4497 }, { "epoch": 2.038060715903942, "grad_norm": 0.7720418816472774, "learning_rate": 5.907624870937616e-05, "loss": 0.8626, "step": 4498 }, { "epoch": 2.0385138196647032, "grad_norm": 0.7453992200999419, "learning_rate": 5.9065125749176376e-05, "loss": 0.8632, "step": 4499 }, { "epoch": 2.0389669234254644, "grad_norm": 0.6624617573984078, "learning_rate": 5.905400088108606e-05, "loss": 0.8606, "step": 4500 }, { "epoch": 2.0394200271862255, "grad_norm": 0.6087163868471206, "learning_rate": 5.9042874106218507e-05, "loss": 0.8791, "step": 4501 }, { "epoch": 2.0398731309469866, "grad_norm": 0.5582078962768947, "learning_rate": 5.903174542568719e-05, "loss": 0.8735, "step": 4502 }, { "epoch": 2.040326234707748, "grad_norm": 0.47252728596076915, "learning_rate": 5.9020614840605804e-05, "loss": 0.8572, "step": 4503 }, { "epoch": 2.0407793384685093, "grad_norm": 0.37466421103760345, "learning_rate": 5.9009482352088186e-05, "loss": 0.8935, "step": 4504 }, { "epoch": 2.0412324422292705, "grad_norm": 0.3372835649525867, "learning_rate": 5.8998347961248404e-05, "loss": 0.8878, "step": 4505 }, { "epoch": 2.0416855459900316, "grad_norm": 0.3684419323551228, "learning_rate": 5.8987211669200704e-05, "loss": 0.8598, "step": 4506 }, { "epoch": 2.0421386497507927, "grad_norm": 0.42373552005720544, "learning_rate": 5.8976073477059505e-05, "loss": 0.8829, "step": 4507 }, { "epoch": 2.0425917535115543, "grad_norm": 0.5548438847632204, "learning_rate": 5.896493338593945e-05, "loss": 0.8876, "step": 4508 }, { "epoch": 2.0430448572723154, "grad_norm": 0.5253145979274657, "learning_rate": 5.895379139695533e-05, "loss": 0.8648, "step": 4509 }, { "epoch": 2.0434979610330766, "grad_norm": 0.40879768328731875, "learning_rate": 5.894264751122217e-05, "loss": 0.8793, "step": 4510 }, { "epoch": 2.0439510647938377, "grad_norm": 0.3355149858312419, "learning_rate": 5.893150172985517e-05, "loss": 0.856, "step": 4511 }, { "epoch": 2.044404168554599, "grad_norm": 0.4206003549298862, "learning_rate": 5.892035405396968e-05, "loss": 0.8723, "step": 4512 }, { "epoch": 2.0448572723153604, "grad_norm": 0.4815695568144232, "learning_rate": 5.8909204484681305e-05, "loss": 0.8597, "step": 4513 }, { "epoch": 2.0453103760761215, "grad_norm": 0.4599223127808181, "learning_rate": 5.8898053023105794e-05, "loss": 0.8883, "step": 4514 }, { "epoch": 2.0457634798368827, "grad_norm": 0.4129742623662007, "learning_rate": 5.888689967035911e-05, "loss": 0.8714, "step": 4515 }, { "epoch": 2.046216583597644, "grad_norm": 0.42872579315329595, "learning_rate": 5.887574442755738e-05, "loss": 0.8658, "step": 4516 }, { "epoch": 2.046669687358405, "grad_norm": 0.3896597760610904, "learning_rate": 5.886458729581694e-05, "loss": 0.8663, "step": 4517 }, { "epoch": 2.0471227911191665, "grad_norm": 0.3950920985086765, "learning_rate": 5.885342827625431e-05, "loss": 0.8647, "step": 4518 }, { "epoch": 2.0475758948799276, "grad_norm": 0.3839782436506871, "learning_rate": 5.884226736998621e-05, "loss": 0.881, "step": 4519 }, { "epoch": 2.0480289986406888, "grad_norm": 0.38906488818976387, "learning_rate": 5.883110457812951e-05, "loss": 0.8733, "step": 4520 }, { "epoch": 2.04848210240145, "grad_norm": 0.3805554049850627, "learning_rate": 5.8819939901801314e-05, "loss": 0.892, "step": 4521 }, { "epoch": 2.048935206162211, "grad_norm": 0.3840236238017456, "learning_rate": 5.88087733421189e-05, "loss": 0.8891, "step": 4522 }, { "epoch": 2.049388309922972, "grad_norm": 0.3533047324542459, "learning_rate": 5.879760490019973e-05, "loss": 0.8839, "step": 4523 }, { "epoch": 2.0498414136837337, "grad_norm": 0.5075744417115691, "learning_rate": 5.878643457716145e-05, "loss": 0.8683, "step": 4524 }, { "epoch": 2.050294517444495, "grad_norm": 0.6032761343778219, "learning_rate": 5.87752623741219e-05, "loss": 0.8855, "step": 4525 }, { "epoch": 2.050747621205256, "grad_norm": 0.7898756300632955, "learning_rate": 5.876408829219911e-05, "loss": 0.8821, "step": 4526 }, { "epoch": 2.051200724966017, "grad_norm": 0.2959826800796188, "learning_rate": 5.875291233251128e-05, "loss": 0.8644, "step": 4527 }, { "epoch": 2.0516538287267783, "grad_norm": 0.4506502517434748, "learning_rate": 5.874173449617684e-05, "loss": 0.8687, "step": 4528 }, { "epoch": 2.05210693248754, "grad_norm": 0.6466236165817394, "learning_rate": 5.873055478431437e-05, "loss": 0.9014, "step": 4529 }, { "epoch": 2.052560036248301, "grad_norm": 0.6982625602524081, "learning_rate": 5.8719373198042644e-05, "loss": 0.8891, "step": 4530 }, { "epoch": 2.053013140009062, "grad_norm": 0.8809497830044766, "learning_rate": 5.870818973848064e-05, "loss": 0.8855, "step": 4531 }, { "epoch": 2.0534662437698232, "grad_norm": 1.0325510071109119, "learning_rate": 5.8697004406747506e-05, "loss": 0.8809, "step": 4532 }, { "epoch": 2.0539193475305844, "grad_norm": 1.0905002888247495, "learning_rate": 5.8685817203962586e-05, "loss": 0.8891, "step": 4533 }, { "epoch": 2.054372451291346, "grad_norm": 0.925362823367596, "learning_rate": 5.86746281312454e-05, "loss": 0.9036, "step": 4534 }, { "epoch": 2.054825555052107, "grad_norm": 0.9010357910573303, "learning_rate": 5.8663437189715674e-05, "loss": 0.8773, "step": 4535 }, { "epoch": 2.055278658812868, "grad_norm": 0.7311866796973343, "learning_rate": 5.86522443804933e-05, "loss": 0.8783, "step": 4536 }, { "epoch": 2.0557317625736293, "grad_norm": 0.5035673032614738, "learning_rate": 5.864104970469838e-05, "loss": 0.8673, "step": 4537 }, { "epoch": 2.0561848663343905, "grad_norm": 0.3922218754830195, "learning_rate": 5.862985316345119e-05, "loss": 0.8635, "step": 4538 }, { "epoch": 2.0566379700951516, "grad_norm": 0.5016162688981224, "learning_rate": 5.861865475787219e-05, "loss": 0.8625, "step": 4539 }, { "epoch": 2.057091073855913, "grad_norm": 0.5135437162038488, "learning_rate": 5.8607454489082026e-05, "loss": 0.8989, "step": 4540 }, { "epoch": 2.0575441776166743, "grad_norm": 0.47822601054703245, "learning_rate": 5.859625235820154e-05, "loss": 0.8884, "step": 4541 }, { "epoch": 2.0579972813774354, "grad_norm": 0.5895744644709228, "learning_rate": 5.858504836635176e-05, "loss": 0.8816, "step": 4542 }, { "epoch": 2.0584503851381966, "grad_norm": 0.7474416842515175, "learning_rate": 5.857384251465387e-05, "loss": 0.9035, "step": 4543 }, { "epoch": 2.0589034888989577, "grad_norm": 0.7827948928992966, "learning_rate": 5.8562634804229296e-05, "loss": 0.8673, "step": 4544 }, { "epoch": 2.0593565926597193, "grad_norm": 0.7718752344612912, "learning_rate": 5.85514252361996e-05, "loss": 0.8613, "step": 4545 }, { "epoch": 2.0598096964204804, "grad_norm": 0.7166919036954197, "learning_rate": 5.8540213811686555e-05, "loss": 0.8766, "step": 4546 }, { "epoch": 2.0602628001812415, "grad_norm": 0.7508845262519247, "learning_rate": 5.852900053181211e-05, "loss": 0.8986, "step": 4547 }, { "epoch": 2.0607159039420027, "grad_norm": 0.7538352704705638, "learning_rate": 5.851778539769841e-05, "loss": 0.8896, "step": 4548 }, { "epoch": 2.061169007702764, "grad_norm": 0.6007566156726805, "learning_rate": 5.850656841046778e-05, "loss": 0.8426, "step": 4549 }, { "epoch": 2.0616221114635254, "grad_norm": 0.39785761983606943, "learning_rate": 5.849534957124272e-05, "loss": 0.8748, "step": 4550 }, { "epoch": 2.0620752152242865, "grad_norm": 0.3746872503897068, "learning_rate": 5.848412888114593e-05, "loss": 0.8882, "step": 4551 }, { "epoch": 2.0625283189850476, "grad_norm": 0.35330776608062786, "learning_rate": 5.84729063413003e-05, "loss": 0.8752, "step": 4552 }, { "epoch": 2.0629814227458088, "grad_norm": 0.35961284230725943, "learning_rate": 5.846168195282888e-05, "loss": 0.8604, "step": 4553 }, { "epoch": 2.06343452650657, "grad_norm": 0.37699681892635856, "learning_rate": 5.845045571685492e-05, "loss": 0.8753, "step": 4554 }, { "epoch": 2.063887630267331, "grad_norm": 0.40814882116223555, "learning_rate": 5.8439227634501866e-05, "loss": 0.8707, "step": 4555 }, { "epoch": 2.0643407340280926, "grad_norm": 0.4561943512645532, "learning_rate": 5.842799770689333e-05, "loss": 0.8642, "step": 4556 }, { "epoch": 2.0647938377888537, "grad_norm": 0.3980102533938311, "learning_rate": 5.841676593515311e-05, "loss": 0.8732, "step": 4557 }, { "epoch": 2.065246941549615, "grad_norm": 0.5123383248403706, "learning_rate": 5.8405532320405204e-05, "loss": 0.897, "step": 4558 }, { "epoch": 2.065700045310376, "grad_norm": 0.3226752306273374, "learning_rate": 5.83942968637738e-05, "loss": 0.88, "step": 4559 }, { "epoch": 2.066153149071137, "grad_norm": 0.32974790759399325, "learning_rate": 5.838305956638323e-05, "loss": 0.8617, "step": 4560 }, { "epoch": 2.0666062528318987, "grad_norm": 0.41639520200110125, "learning_rate": 5.837182042935805e-05, "loss": 0.8753, "step": 4561 }, { "epoch": 2.06705935659266, "grad_norm": 0.43819402892964543, "learning_rate": 5.8360579453822974e-05, "loss": 0.8679, "step": 4562 }, { "epoch": 2.067512460353421, "grad_norm": 0.5087463017671663, "learning_rate": 5.834933664090293e-05, "loss": 0.8934, "step": 4563 }, { "epoch": 2.067965564114182, "grad_norm": 0.6064799428938359, "learning_rate": 5.8338091991723e-05, "loss": 0.8803, "step": 4564 }, { "epoch": 2.068418667874943, "grad_norm": 0.6242861566084189, "learning_rate": 5.832684550740846e-05, "loss": 0.8762, "step": 4565 }, { "epoch": 2.0688717716357043, "grad_norm": 0.6420490878426205, "learning_rate": 5.8315597189084784e-05, "loss": 0.8778, "step": 4566 }, { "epoch": 2.069324875396466, "grad_norm": 0.6532297449472928, "learning_rate": 5.8304347037877604e-05, "loss": 0.8645, "step": 4567 }, { "epoch": 2.069777979157227, "grad_norm": 0.6067980762397831, "learning_rate": 5.829309505491275e-05, "loss": 0.9075, "step": 4568 }, { "epoch": 2.070231082917988, "grad_norm": 0.41854075407681884, "learning_rate": 5.828184124131624e-05, "loss": 0.8581, "step": 4569 }, { "epoch": 2.0706841866787493, "grad_norm": 0.34917476725174923, "learning_rate": 5.8270585598214264e-05, "loss": 0.8628, "step": 4570 }, { "epoch": 2.0711372904395104, "grad_norm": 0.4708179587901186, "learning_rate": 5.8259328126733195e-05, "loss": 0.8842, "step": 4571 }, { "epoch": 2.071590394200272, "grad_norm": 0.4389117717370451, "learning_rate": 5.82480688279996e-05, "loss": 0.8585, "step": 4572 }, { "epoch": 2.072043497961033, "grad_norm": 0.42491011062157885, "learning_rate": 5.8236807703140234e-05, "loss": 0.8787, "step": 4573 }, { "epoch": 2.0724966017217943, "grad_norm": 0.5421259193426007, "learning_rate": 5.8225544753282e-05, "loss": 0.8581, "step": 4574 }, { "epoch": 2.0729497054825554, "grad_norm": 0.6293772912922824, "learning_rate": 5.8214279979552016e-05, "loss": 0.8721, "step": 4575 }, { "epoch": 2.0734028092433165, "grad_norm": 0.645478975412969, "learning_rate": 5.820301338307758e-05, "loss": 0.8762, "step": 4576 }, { "epoch": 2.073855913004078, "grad_norm": 0.6579150554759016, "learning_rate": 5.819174496498617e-05, "loss": 0.8775, "step": 4577 }, { "epoch": 2.0743090167648393, "grad_norm": 0.5946637276073414, "learning_rate": 5.8180474726405425e-05, "loss": 0.8709, "step": 4578 }, { "epoch": 2.0747621205256004, "grad_norm": 0.5617903059752853, "learning_rate": 5.8169202668463193e-05, "loss": 0.8618, "step": 4579 }, { "epoch": 2.0752152242863615, "grad_norm": 0.5651103618969068, "learning_rate": 5.815792879228749e-05, "loss": 0.8527, "step": 4580 }, { "epoch": 2.0756683280471226, "grad_norm": 0.3882215058347721, "learning_rate": 5.814665309900653e-05, "loss": 0.8762, "step": 4581 }, { "epoch": 2.076121431807884, "grad_norm": 0.36013816254445974, "learning_rate": 5.8135375589748686e-05, "loss": 0.8755, "step": 4582 }, { "epoch": 2.0765745355686454, "grad_norm": 0.41897871412155524, "learning_rate": 5.812409626564252e-05, "loss": 0.854, "step": 4583 }, { "epoch": 2.0770276393294065, "grad_norm": 0.4044266569601732, "learning_rate": 5.81128151278168e-05, "loss": 0.8775, "step": 4584 }, { "epoch": 2.0774807430901676, "grad_norm": 0.4128139370621715, "learning_rate": 5.8101532177400446e-05, "loss": 0.878, "step": 4585 }, { "epoch": 2.0779338468509287, "grad_norm": 0.39228615796105326, "learning_rate": 5.809024741552255e-05, "loss": 0.8639, "step": 4586 }, { "epoch": 2.07838695061169, "grad_norm": 0.4503987985356691, "learning_rate": 5.8078960843312434e-05, "loss": 0.8726, "step": 4587 }, { "epoch": 2.0788400543724515, "grad_norm": 0.57739033722956, "learning_rate": 5.806767246189954e-05, "loss": 0.8814, "step": 4588 }, { "epoch": 2.0792931581332126, "grad_norm": 0.7560076602755122, "learning_rate": 5.8056382272413554e-05, "loss": 0.8899, "step": 4589 }, { "epoch": 2.0797462618939737, "grad_norm": 0.9410826840553808, "learning_rate": 5.804509027598428e-05, "loss": 0.8723, "step": 4590 }, { "epoch": 2.080199365654735, "grad_norm": 0.9847867922907205, "learning_rate": 5.8033796473741755e-05, "loss": 0.8834, "step": 4591 }, { "epoch": 2.080652469415496, "grad_norm": 0.969822393189375, "learning_rate": 5.802250086681617e-05, "loss": 0.8734, "step": 4592 }, { "epoch": 2.0811055731762576, "grad_norm": 0.9119426713808654, "learning_rate": 5.801120345633791e-05, "loss": 0.8623, "step": 4593 }, { "epoch": 2.0815586769370187, "grad_norm": 0.7593837033441263, "learning_rate": 5.799990424343751e-05, "loss": 0.8582, "step": 4594 }, { "epoch": 2.08201178069778, "grad_norm": 0.5316805837629967, "learning_rate": 5.7988603229245724e-05, "loss": 0.8699, "step": 4595 }, { "epoch": 2.082464884458541, "grad_norm": 0.35449712897042496, "learning_rate": 5.797730041489347e-05, "loss": 0.8883, "step": 4596 }, { "epoch": 2.082917988219302, "grad_norm": 0.31386640064991295, "learning_rate": 5.796599580151184e-05, "loss": 0.8661, "step": 4597 }, { "epoch": 2.0833710919800636, "grad_norm": 0.35906423025657536, "learning_rate": 5.7954689390232123e-05, "loss": 0.8628, "step": 4598 }, { "epoch": 2.083824195740825, "grad_norm": 0.4234149200278355, "learning_rate": 5.794338118218577e-05, "loss": 0.8505, "step": 4599 }, { "epoch": 2.084277299501586, "grad_norm": 0.4790207310930615, "learning_rate": 5.793207117850442e-05, "loss": 0.886, "step": 4600 }, { "epoch": 2.084730403262347, "grad_norm": 0.48179911699576655, "learning_rate": 5.792075938031989e-05, "loss": 0.8786, "step": 4601 }, { "epoch": 2.085183507023108, "grad_norm": 0.49965510761972215, "learning_rate": 5.790944578876418e-05, "loss": 0.891, "step": 4602 }, { "epoch": 2.0856366107838693, "grad_norm": 0.6526756070694657, "learning_rate": 5.789813040496947e-05, "loss": 0.8882, "step": 4603 }, { "epoch": 2.086089714544631, "grad_norm": 0.6912142648861804, "learning_rate": 5.7886813230068106e-05, "loss": 0.8831, "step": 4604 }, { "epoch": 2.086542818305392, "grad_norm": 0.7180085687868353, "learning_rate": 5.787549426519262e-05, "loss": 0.8756, "step": 4605 }, { "epoch": 2.086995922066153, "grad_norm": 0.6632289861052785, "learning_rate": 5.786417351147575e-05, "loss": 0.8784, "step": 4606 }, { "epoch": 2.0874490258269143, "grad_norm": 0.6253644332715973, "learning_rate": 5.785285097005037e-05, "loss": 0.8614, "step": 4607 }, { "epoch": 2.0879021295876754, "grad_norm": 0.5074009404475119, "learning_rate": 5.7841526642049554e-05, "loss": 0.8653, "step": 4608 }, { "epoch": 2.088355233348437, "grad_norm": 0.3447687822319798, "learning_rate": 5.783020052860656e-05, "loss": 0.874, "step": 4609 }, { "epoch": 2.088808337109198, "grad_norm": 0.3952100652153182, "learning_rate": 5.7818872630854804e-05, "loss": 0.8592, "step": 4610 }, { "epoch": 2.0892614408699592, "grad_norm": 0.5377419064524526, "learning_rate": 5.7807542949927914e-05, "loss": 0.8729, "step": 4611 }, { "epoch": 2.0897145446307204, "grad_norm": 0.6445363351073123, "learning_rate": 5.779621148695968e-05, "loss": 0.8595, "step": 4612 }, { "epoch": 2.0901676483914815, "grad_norm": 0.7376683777882179, "learning_rate": 5.7784878243084046e-05, "loss": 0.898, "step": 4613 }, { "epoch": 2.090620752152243, "grad_norm": 0.8362922609267202, "learning_rate": 5.777354321943516e-05, "loss": 0.8652, "step": 4614 }, { "epoch": 2.091073855913004, "grad_norm": 0.8265937783819768, "learning_rate": 5.776220641714735e-05, "loss": 0.8764, "step": 4615 }, { "epoch": 2.0915269596737653, "grad_norm": 0.7214403607923383, "learning_rate": 5.7750867837355114e-05, "loss": 0.8717, "step": 4616 }, { "epoch": 2.0919800634345265, "grad_norm": 0.5134724284300936, "learning_rate": 5.7739527481193135e-05, "loss": 0.8828, "step": 4617 }, { "epoch": 2.0924331671952876, "grad_norm": 0.38092063595044223, "learning_rate": 5.7728185349796255e-05, "loss": 0.8626, "step": 4618 }, { "epoch": 2.0928862709560487, "grad_norm": 0.42213477520014603, "learning_rate": 5.771684144429952e-05, "loss": 0.8653, "step": 4619 }, { "epoch": 2.0933393747168103, "grad_norm": 0.6105873953440661, "learning_rate": 5.770549576583814e-05, "loss": 0.889, "step": 4620 }, { "epoch": 2.0937924784775714, "grad_norm": 0.7763045833258722, "learning_rate": 5.769414831554749e-05, "loss": 0.8706, "step": 4621 }, { "epoch": 2.0942455822383326, "grad_norm": 0.8777675311433213, "learning_rate": 5.768279909456315e-05, "loss": 0.8694, "step": 4622 }, { "epoch": 2.0946986859990937, "grad_norm": 0.9063549143654276, "learning_rate": 5.767144810402085e-05, "loss": 0.8872, "step": 4623 }, { "epoch": 2.095151789759855, "grad_norm": 0.7957951561267625, "learning_rate": 5.766009534505652e-05, "loss": 0.8702, "step": 4624 }, { "epoch": 2.0956048935206164, "grad_norm": 0.678755484156684, "learning_rate": 5.764874081880625e-05, "loss": 0.9006, "step": 4625 }, { "epoch": 2.0960579972813775, "grad_norm": 0.5155959365523988, "learning_rate": 5.763738452640631e-05, "loss": 0.8667, "step": 4626 }, { "epoch": 2.0965111010421387, "grad_norm": 0.4514179983349916, "learning_rate": 5.762602646899316e-05, "loss": 0.8801, "step": 4627 }, { "epoch": 2.0969642048029, "grad_norm": 0.3398826099912264, "learning_rate": 5.761466664770343e-05, "loss": 0.8776, "step": 4628 }, { "epoch": 2.097417308563661, "grad_norm": 0.3856385973405732, "learning_rate": 5.760330506367391e-05, "loss": 0.881, "step": 4629 }, { "epoch": 2.0978704123244225, "grad_norm": 0.5909373093732458, "learning_rate": 5.759194171804158e-05, "loss": 0.8925, "step": 4630 }, { "epoch": 2.0983235160851836, "grad_norm": 0.6355652724880775, "learning_rate": 5.7580576611943605e-05, "loss": 0.8768, "step": 4631 }, { "epoch": 2.0987766198459448, "grad_norm": 0.5778436477725404, "learning_rate": 5.7569209746517305e-05, "loss": 0.8833, "step": 4632 }, { "epoch": 2.099229723606706, "grad_norm": 0.4558853879333152, "learning_rate": 5.7557841122900197e-05, "loss": 0.8575, "step": 4633 }, { "epoch": 2.099682827367467, "grad_norm": 0.39300927010919673, "learning_rate": 5.754647074222997e-05, "loss": 0.8617, "step": 4634 }, { "epoch": 2.100135931128228, "grad_norm": 0.37117481505785194, "learning_rate": 5.753509860564446e-05, "loss": 0.8834, "step": 4635 }, { "epoch": 2.1005890348889897, "grad_norm": 0.35537566534011417, "learning_rate": 5.7523724714281734e-05, "loss": 0.8617, "step": 4636 }, { "epoch": 2.101042138649751, "grad_norm": 0.37979338953880615, "learning_rate": 5.751234906927998e-05, "loss": 0.8721, "step": 4637 }, { "epoch": 2.101495242410512, "grad_norm": 0.3971917207918673, "learning_rate": 5.7500971671777613e-05, "loss": 0.8569, "step": 4638 }, { "epoch": 2.101948346171273, "grad_norm": 0.3579673231704319, "learning_rate": 5.7489592522913165e-05, "loss": 0.8829, "step": 4639 }, { "epoch": 2.1024014499320343, "grad_norm": 0.4184285766652804, "learning_rate": 5.7478211623825384e-05, "loss": 0.8872, "step": 4640 }, { "epoch": 2.102854553692796, "grad_norm": 0.4463208503441737, "learning_rate": 5.746682897565317e-05, "loss": 0.9007, "step": 4641 }, { "epoch": 2.103307657453557, "grad_norm": 0.3901472029734418, "learning_rate": 5.745544457953564e-05, "loss": 0.866, "step": 4642 }, { "epoch": 2.103760761214318, "grad_norm": 0.3681742336370412, "learning_rate": 5.744405843661203e-05, "loss": 0.8706, "step": 4643 }, { "epoch": 2.1042138649750792, "grad_norm": 0.4014602195178215, "learning_rate": 5.7432670548021793e-05, "loss": 0.8713, "step": 4644 }, { "epoch": 2.1046669687358404, "grad_norm": 0.3367785362463053, "learning_rate": 5.742128091490453e-05, "loss": 0.8784, "step": 4645 }, { "epoch": 2.105120072496602, "grad_norm": 0.350128865055401, "learning_rate": 5.7409889538400044e-05, "loss": 0.8831, "step": 4646 }, { "epoch": 2.105573176257363, "grad_norm": 1.1134969144425153, "learning_rate": 5.739849641964828e-05, "loss": 0.878, "step": 4647 }, { "epoch": 2.106026280018124, "grad_norm": 0.23945218854786096, "learning_rate": 5.7387101559789364e-05, "loss": 0.8697, "step": 4648 }, { "epoch": 2.1064793837788853, "grad_norm": 0.3990384521011702, "learning_rate": 5.7375704959963626e-05, "loss": 0.8868, "step": 4649 }, { "epoch": 2.1069324875396465, "grad_norm": 0.5299346382445033, "learning_rate": 5.736430662131154e-05, "loss": 0.8773, "step": 4650 }, { "epoch": 2.1073855913004076, "grad_norm": 0.5693093482581049, "learning_rate": 5.7352906544973776e-05, "loss": 0.8799, "step": 4651 }, { "epoch": 2.107838695061169, "grad_norm": 0.6318345523180093, "learning_rate": 5.734150473209115e-05, "loss": 0.8718, "step": 4652 }, { "epoch": 2.1082917988219303, "grad_norm": 0.7325034306429262, "learning_rate": 5.733010118380467e-05, "loss": 0.8511, "step": 4653 }, { "epoch": 2.1087449025826914, "grad_norm": 0.8980293067524645, "learning_rate": 5.731869590125554e-05, "loss": 0.887, "step": 4654 }, { "epoch": 2.1091980063434526, "grad_norm": 1.0618322390515875, "learning_rate": 5.730728888558506e-05, "loss": 0.8953, "step": 4655 }, { "epoch": 2.1096511101042137, "grad_norm": 0.8248130177627827, "learning_rate": 5.729588013793481e-05, "loss": 0.8913, "step": 4656 }, { "epoch": 2.1101042138649753, "grad_norm": 1.2151249372147284, "learning_rate": 5.728446965944646e-05, "loss": 0.8911, "step": 4657 }, { "epoch": 2.1105573176257364, "grad_norm": 0.37494465863823917, "learning_rate": 5.727305745126189e-05, "loss": 0.8668, "step": 4658 }, { "epoch": 2.1110104213864975, "grad_norm": 0.6291596998141017, "learning_rate": 5.7261643514523135e-05, "loss": 0.8681, "step": 4659 }, { "epoch": 2.1114635251472587, "grad_norm": 0.9100509601651041, "learning_rate": 5.7250227850372436e-05, "loss": 0.8696, "step": 4660 }, { "epoch": 2.11191662890802, "grad_norm": 1.018165707927247, "learning_rate": 5.723881045995216e-05, "loss": 0.8798, "step": 4661 }, { "epoch": 2.1123697326687814, "grad_norm": 0.9093757742548951, "learning_rate": 5.722739134440489e-05, "loss": 0.8737, "step": 4662 }, { "epoch": 2.1128228364295425, "grad_norm": 0.7832887176064344, "learning_rate": 5.721597050487335e-05, "loss": 0.8741, "step": 4663 }, { "epoch": 2.1132759401903036, "grad_norm": 0.5453258724220422, "learning_rate": 5.7204547942500465e-05, "loss": 0.871, "step": 4664 }, { "epoch": 2.1137290439510648, "grad_norm": 0.4757370262810262, "learning_rate": 5.71931236584293e-05, "loss": 0.877, "step": 4665 }, { "epoch": 2.114182147711826, "grad_norm": 0.5415724098518981, "learning_rate": 5.718169765380312e-05, "loss": 0.8721, "step": 4666 }, { "epoch": 2.114635251472587, "grad_norm": 0.6504406248263057, "learning_rate": 5.717026992976535e-05, "loss": 0.8614, "step": 4667 }, { "epoch": 2.1150883552333486, "grad_norm": 0.7450117171873946, "learning_rate": 5.715884048745958e-05, "loss": 0.8787, "step": 4668 }, { "epoch": 2.1155414589941097, "grad_norm": 0.8045014244974074, "learning_rate": 5.714740932802958e-05, "loss": 0.8734, "step": 4669 }, { "epoch": 2.115994562754871, "grad_norm": 0.7763980914926428, "learning_rate": 5.713597645261932e-05, "loss": 0.8928, "step": 4670 }, { "epoch": 2.116447666515632, "grad_norm": 0.6158581898320518, "learning_rate": 5.712454186237287e-05, "loss": 0.8653, "step": 4671 }, { "epoch": 2.116900770276393, "grad_norm": 0.5048136634706868, "learning_rate": 5.7113105558434545e-05, "loss": 0.8815, "step": 4672 }, { "epoch": 2.1173538740371547, "grad_norm": 0.5057411130970902, "learning_rate": 5.71016675419488e-05, "loss": 0.8757, "step": 4673 }, { "epoch": 2.117806977797916, "grad_norm": 0.5242471839880435, "learning_rate": 5.709022781406026e-05, "loss": 0.8782, "step": 4674 }, { "epoch": 2.118260081558677, "grad_norm": 0.5622168664097199, "learning_rate": 5.707878637591371e-05, "loss": 0.8633, "step": 4675 }, { "epoch": 2.118713185319438, "grad_norm": 0.6007164981505231, "learning_rate": 5.706734322865414e-05, "loss": 0.8796, "step": 4676 }, { "epoch": 2.119166289080199, "grad_norm": 0.6264313973664619, "learning_rate": 5.7055898373426684e-05, "loss": 0.8887, "step": 4677 }, { "epoch": 2.119619392840961, "grad_norm": 0.6067967313250379, "learning_rate": 5.704445181137665e-05, "loss": 0.8585, "step": 4678 }, { "epoch": 2.120072496601722, "grad_norm": 0.5250098203846706, "learning_rate": 5.703300354364953e-05, "loss": 0.8815, "step": 4679 }, { "epoch": 2.120525600362483, "grad_norm": 0.4165819378743255, "learning_rate": 5.7021553571390985e-05, "loss": 0.8999, "step": 4680 }, { "epoch": 2.120978704123244, "grad_norm": 0.3690372186014333, "learning_rate": 5.701010189574683e-05, "loss": 0.8598, "step": 4681 }, { "epoch": 2.1214318078840053, "grad_norm": 0.37437947512852765, "learning_rate": 5.699864851786306e-05, "loss": 0.8789, "step": 4682 }, { "epoch": 2.1218849116447664, "grad_norm": 0.35317810645666226, "learning_rate": 5.698719343888583e-05, "loss": 0.8928, "step": 4683 }, { "epoch": 2.122338015405528, "grad_norm": 0.3523129917708894, "learning_rate": 5.6975736659961495e-05, "loss": 0.8812, "step": 4684 }, { "epoch": 2.122791119166289, "grad_norm": 0.4533676332704265, "learning_rate": 5.696427818223654e-05, "loss": 0.8734, "step": 4685 }, { "epoch": 2.1232442229270503, "grad_norm": 0.448396168561796, "learning_rate": 5.6952818006857684e-05, "loss": 0.8747, "step": 4686 }, { "epoch": 2.1236973266878114, "grad_norm": 0.4468298778926549, "learning_rate": 5.694135613497172e-05, "loss": 0.8738, "step": 4687 }, { "epoch": 2.1241504304485725, "grad_norm": 0.5124838069413314, "learning_rate": 5.6929892567725686e-05, "loss": 0.8715, "step": 4688 }, { "epoch": 2.124603534209334, "grad_norm": 0.562536573251405, "learning_rate": 5.691842730626677e-05, "loss": 0.9045, "step": 4689 }, { "epoch": 2.1250566379700953, "grad_norm": 0.618277452367663, "learning_rate": 5.690696035174233e-05, "loss": 0.8419, "step": 4690 }, { "epoch": 2.1255097417308564, "grad_norm": 0.6397761997961782, "learning_rate": 5.689549170529989e-05, "loss": 0.8658, "step": 4691 }, { "epoch": 2.1259628454916175, "grad_norm": 0.6316064839801924, "learning_rate": 5.688402136808712e-05, "loss": 0.8804, "step": 4692 }, { "epoch": 2.1264159492523786, "grad_norm": 0.9506967417840064, "learning_rate": 5.6872549341251924e-05, "loss": 0.9214, "step": 4693 }, { "epoch": 2.1268690530131398, "grad_norm": 0.5881334465972584, "learning_rate": 5.686107562594229e-05, "loss": 0.8641, "step": 4694 }, { "epoch": 2.1273221567739014, "grad_norm": 0.5628111700438275, "learning_rate": 5.6849600223306445e-05, "loss": 0.8751, "step": 4695 }, { "epoch": 2.1277752605346625, "grad_norm": 0.5617317576155165, "learning_rate": 5.6838123134492755e-05, "loss": 0.8838, "step": 4696 }, { "epoch": 2.1282283642954236, "grad_norm": 0.5524366613819399, "learning_rate": 5.682664436064976e-05, "loss": 0.8922, "step": 4697 }, { "epoch": 2.1286814680561847, "grad_norm": 0.6013153867461547, "learning_rate": 5.681516390292616e-05, "loss": 0.9014, "step": 4698 }, { "epoch": 2.129134571816946, "grad_norm": 0.5386954961035734, "learning_rate": 5.6803681762470844e-05, "loss": 0.8762, "step": 4699 }, { "epoch": 2.1295876755777075, "grad_norm": 0.6072925019352998, "learning_rate": 5.679219794043283e-05, "loss": 0.8729, "step": 4700 }, { "epoch": 2.1300407793384686, "grad_norm": 0.686055558996378, "learning_rate": 5.6780712437961356e-05, "loss": 0.8716, "step": 4701 }, { "epoch": 2.1304938830992297, "grad_norm": 0.7881345231453362, "learning_rate": 5.676922525620579e-05, "loss": 0.877, "step": 4702 }, { "epoch": 2.130946986859991, "grad_norm": 0.7756731502103842, "learning_rate": 5.675773639631568e-05, "loss": 0.8802, "step": 4703 }, { "epoch": 2.131400090620752, "grad_norm": 0.7900503764758113, "learning_rate": 5.674624585944076e-05, "loss": 0.8627, "step": 4704 }, { "epoch": 2.1318531943815135, "grad_norm": 0.7734279737107076, "learning_rate": 5.673475364673089e-05, "loss": 0.8696, "step": 4705 }, { "epoch": 2.1323062981422747, "grad_norm": 0.6685226176878676, "learning_rate": 5.6723259759336124e-05, "loss": 0.9034, "step": 4706 }, { "epoch": 2.132759401903036, "grad_norm": 0.5619144576620108, "learning_rate": 5.67117641984067e-05, "loss": 0.8653, "step": 4707 }, { "epoch": 2.133212505663797, "grad_norm": 0.5106208831854233, "learning_rate": 5.6700266965092986e-05, "loss": 0.868, "step": 4708 }, { "epoch": 2.133665609424558, "grad_norm": 0.4735174989638756, "learning_rate": 5.6688768060545553e-05, "loss": 0.8572, "step": 4709 }, { "epoch": 2.1341187131853196, "grad_norm": 0.4237273145868136, "learning_rate": 5.6677267485915104e-05, "loss": 0.8782, "step": 4710 }, { "epoch": 2.134571816946081, "grad_norm": 0.3575757080935591, "learning_rate": 5.666576524235254e-05, "loss": 0.8666, "step": 4711 }, { "epoch": 2.135024920706842, "grad_norm": 0.45713713141406, "learning_rate": 5.665426133100892e-05, "loss": 0.8856, "step": 4712 }, { "epoch": 2.135478024467603, "grad_norm": 0.5580340100872371, "learning_rate": 5.6642755753035454e-05, "loss": 0.8672, "step": 4713 }, { "epoch": 2.135931128228364, "grad_norm": 0.581054024701313, "learning_rate": 5.6631248509583544e-05, "loss": 0.8862, "step": 4714 }, { "epoch": 2.1363842319891253, "grad_norm": 0.6110966759707582, "learning_rate": 5.661973960180474e-05, "loss": 0.8725, "step": 4715 }, { "epoch": 2.136837335749887, "grad_norm": 0.5939830016624077, "learning_rate": 5.660822903085077e-05, "loss": 0.8596, "step": 4716 }, { "epoch": 2.137290439510648, "grad_norm": 0.52707932768734, "learning_rate": 5.659671679787352e-05, "loss": 0.8737, "step": 4717 }, { "epoch": 2.137743543271409, "grad_norm": 0.45130082428365487, "learning_rate": 5.6585202904025024e-05, "loss": 0.868, "step": 4718 }, { "epoch": 2.1381966470321703, "grad_norm": 0.38123088876919303, "learning_rate": 5.657368735045754e-05, "loss": 0.8815, "step": 4719 }, { "epoch": 2.1386497507929314, "grad_norm": 0.39653981187393594, "learning_rate": 5.6562170138323424e-05, "loss": 0.895, "step": 4720 }, { "epoch": 2.139102854553693, "grad_norm": 0.37809617505752585, "learning_rate": 5.6550651268775257e-05, "loss": 0.8821, "step": 4721 }, { "epoch": 2.139555958314454, "grad_norm": 0.4575113660342718, "learning_rate": 5.653913074296574e-05, "loss": 0.8771, "step": 4722 }, { "epoch": 2.1400090620752152, "grad_norm": 0.3739149445119517, "learning_rate": 5.6527608562047756e-05, "loss": 0.8888, "step": 4723 }, { "epoch": 2.1404621658359764, "grad_norm": 0.2904273109052826, "learning_rate": 5.651608472717438e-05, "loss": 0.9001, "step": 4724 }, { "epoch": 2.1409152695967375, "grad_norm": 0.3465144144941351, "learning_rate": 5.650455923949881e-05, "loss": 0.872, "step": 4725 }, { "epoch": 2.1413683733574986, "grad_norm": 0.42220519097521314, "learning_rate": 5.6493032100174416e-05, "loss": 0.9103, "step": 4726 }, { "epoch": 2.14182147711826, "grad_norm": 0.3970281788289326, "learning_rate": 5.648150331035478e-05, "loss": 0.8778, "step": 4727 }, { "epoch": 2.1422745808790213, "grad_norm": 0.45405760224376274, "learning_rate": 5.646997287119357e-05, "loss": 0.8641, "step": 4728 }, { "epoch": 2.1427276846397825, "grad_norm": 0.48285945370096145, "learning_rate": 5.6458440783844693e-05, "loss": 0.8969, "step": 4729 }, { "epoch": 2.1431807884005436, "grad_norm": 0.5051854871290692, "learning_rate": 5.64469070494622e-05, "loss": 0.8703, "step": 4730 }, { "epoch": 2.1436338921613047, "grad_norm": 0.5207260955054751, "learning_rate": 5.643537166920027e-05, "loss": 0.8633, "step": 4731 }, { "epoch": 2.1440869959220663, "grad_norm": 0.5674887205789033, "learning_rate": 5.642383464421328e-05, "loss": 0.8619, "step": 4732 }, { "epoch": 2.1445400996828274, "grad_norm": 0.6557025653782947, "learning_rate": 5.6412295975655796e-05, "loss": 0.957, "step": 4733 }, { "epoch": 2.1449932034435886, "grad_norm": 0.48914090069198396, "learning_rate": 5.640075566468248e-05, "loss": 0.8907, "step": 4734 }, { "epoch": 2.1454463072043497, "grad_norm": 0.412944760104268, "learning_rate": 5.6389213712448215e-05, "loss": 0.8612, "step": 4735 }, { "epoch": 2.145899410965111, "grad_norm": 0.31744539146108497, "learning_rate": 5.637767012010804e-05, "loss": 0.8892, "step": 4736 }, { "epoch": 2.1463525147258724, "grad_norm": 0.3651780821563718, "learning_rate": 5.636612488881712e-05, "loss": 0.8603, "step": 4737 }, { "epoch": 2.1468056184866335, "grad_norm": 0.32757534281586154, "learning_rate": 5.6354578019730844e-05, "loss": 0.8634, "step": 4738 }, { "epoch": 2.1472587222473947, "grad_norm": 0.32667831689073107, "learning_rate": 5.634302951400472e-05, "loss": 0.8955, "step": 4739 }, { "epoch": 2.147711826008156, "grad_norm": 0.46925553889736266, "learning_rate": 5.633147937279443e-05, "loss": 0.8723, "step": 4740 }, { "epoch": 2.148164929768917, "grad_norm": 0.40161607863457055, "learning_rate": 5.6319927597255816e-05, "loss": 0.876, "step": 4741 }, { "epoch": 2.1486180335296785, "grad_norm": 0.44276114561774926, "learning_rate": 5.6308374188544926e-05, "loss": 0.873, "step": 4742 }, { "epoch": 2.1490711372904396, "grad_norm": 0.6335821757846868, "learning_rate": 5.62968191478179e-05, "loss": 0.8422, "step": 4743 }, { "epoch": 2.1495242410512008, "grad_norm": 0.7798230826741983, "learning_rate": 5.628526247623108e-05, "loss": 0.8842, "step": 4744 }, { "epoch": 2.149977344811962, "grad_norm": 0.8144435412222979, "learning_rate": 5.6273704174940975e-05, "loss": 0.8688, "step": 4745 }, { "epoch": 2.150430448572723, "grad_norm": 0.8381561922038119, "learning_rate": 5.626214424510426e-05, "loss": 0.8865, "step": 4746 }, { "epoch": 2.150883552333484, "grad_norm": 0.8904018905740969, "learning_rate": 5.625058268787775e-05, "loss": 0.8862, "step": 4747 }, { "epoch": 2.1513366560942457, "grad_norm": 0.9087605483210487, "learning_rate": 5.623901950441844e-05, "loss": 0.86, "step": 4748 }, { "epoch": 2.151789759855007, "grad_norm": 0.8807015106998731, "learning_rate": 5.622745469588349e-05, "loss": 0.8636, "step": 4749 }, { "epoch": 2.152242863615768, "grad_norm": 0.7677087416363191, "learning_rate": 5.62158882634302e-05, "loss": 0.8723, "step": 4750 }, { "epoch": 2.152695967376529, "grad_norm": 0.6516175981326688, "learning_rate": 5.620432020821609e-05, "loss": 0.8664, "step": 4751 }, { "epoch": 2.1531490711372903, "grad_norm": 0.5118000605233654, "learning_rate": 5.619275053139875e-05, "loss": 0.8777, "step": 4752 }, { "epoch": 2.153602174898052, "grad_norm": 0.4846154484433814, "learning_rate": 5.6181179234136006e-05, "loss": 0.8768, "step": 4753 }, { "epoch": 2.154055278658813, "grad_norm": 0.38797157530314724, "learning_rate": 5.616960631758584e-05, "loss": 0.877, "step": 4754 }, { "epoch": 2.154508382419574, "grad_norm": 0.36269610611879693, "learning_rate": 5.615803178290636e-05, "loss": 0.8624, "step": 4755 }, { "epoch": 2.1549614861803352, "grad_norm": 0.5807355578801142, "learning_rate": 5.614645563125585e-05, "loss": 0.8937, "step": 4756 }, { "epoch": 2.1554145899410964, "grad_norm": 0.6939589432854851, "learning_rate": 5.6134877863792775e-05, "loss": 0.8815, "step": 4757 }, { "epoch": 2.1558676937018575, "grad_norm": 0.6881620035707302, "learning_rate": 5.6123298481675754e-05, "loss": 0.8641, "step": 4758 }, { "epoch": 2.156320797462619, "grad_norm": 0.5858187813252232, "learning_rate": 5.6111717486063555e-05, "loss": 0.8765, "step": 4759 }, { "epoch": 2.15677390122338, "grad_norm": 0.4208587827922099, "learning_rate": 5.610013487811512e-05, "loss": 0.862, "step": 4760 }, { "epoch": 2.1572270049841413, "grad_norm": 0.31456663719625577, "learning_rate": 5.608855065898953e-05, "loss": 0.8785, "step": 4761 }, { "epoch": 2.1576801087449025, "grad_norm": 0.37910635612992083, "learning_rate": 5.6076964829846066e-05, "loss": 0.8904, "step": 4762 }, { "epoch": 2.1581332125056636, "grad_norm": 0.516000832331608, "learning_rate": 5.6065377391844134e-05, "loss": 0.8786, "step": 4763 }, { "epoch": 2.158586316266425, "grad_norm": 0.551528206445953, "learning_rate": 5.605378834614331e-05, "loss": 0.8652, "step": 4764 }, { "epoch": 2.1590394200271863, "grad_norm": 0.5266403611242373, "learning_rate": 5.6042197693903345e-05, "loss": 0.8819, "step": 4765 }, { "epoch": 2.1594925237879474, "grad_norm": 0.5791008026419482, "learning_rate": 5.6030605436284146e-05, "loss": 0.8601, "step": 4766 }, { "epoch": 2.1599456275487086, "grad_norm": 0.6786763228098623, "learning_rate": 5.6019011574445775e-05, "loss": 0.8768, "step": 4767 }, { "epoch": 2.1603987313094697, "grad_norm": 0.6767146963637948, "learning_rate": 5.6007416109548464e-05, "loss": 0.8661, "step": 4768 }, { "epoch": 2.1608518350702313, "grad_norm": 0.5594303226634009, "learning_rate": 5.599581904275258e-05, "loss": 0.9017, "step": 4769 }, { "epoch": 2.1613049388309924, "grad_norm": 0.41597709667873217, "learning_rate": 5.5984220375218675e-05, "loss": 0.8642, "step": 4770 }, { "epoch": 2.1617580425917535, "grad_norm": 0.3708878496998883, "learning_rate": 5.5972620108107453e-05, "loss": 0.8675, "step": 4771 }, { "epoch": 2.1622111463525147, "grad_norm": 0.3784513190175112, "learning_rate": 5.5961018242579785e-05, "loss": 0.876, "step": 4772 }, { "epoch": 2.162664250113276, "grad_norm": 0.4387915497495398, "learning_rate": 5.594941477979668e-05, "loss": 0.8924, "step": 4773 }, { "epoch": 2.1631173538740374, "grad_norm": 0.44655920031457985, "learning_rate": 5.593780972091934e-05, "loss": 0.8783, "step": 4774 }, { "epoch": 2.1635704576347985, "grad_norm": 0.43321277614615245, "learning_rate": 5.5926203067109114e-05, "loss": 0.8897, "step": 4775 }, { "epoch": 2.1640235613955596, "grad_norm": 0.3903716275197271, "learning_rate": 5.591459481952749e-05, "loss": 0.8645, "step": 4776 }, { "epoch": 2.1644766651563208, "grad_norm": 0.3908485057319502, "learning_rate": 5.5902984979336125e-05, "loss": 0.8873, "step": 4777 }, { "epoch": 2.164929768917082, "grad_norm": 0.4378221741998317, "learning_rate": 5.589137354769688e-05, "loss": 0.8893, "step": 4778 }, { "epoch": 2.165382872677843, "grad_norm": 0.41132176314199775, "learning_rate": 5.58797605257717e-05, "loss": 0.8775, "step": 4779 }, { "epoch": 2.1658359764386046, "grad_norm": 0.4074046129826227, "learning_rate": 5.5868145914722736e-05, "loss": 0.8799, "step": 4780 }, { "epoch": 2.1662890801993657, "grad_norm": 0.45268704430546747, "learning_rate": 5.5856529715712295e-05, "loss": 0.8847, "step": 4781 }, { "epoch": 2.166742183960127, "grad_norm": 0.33802021190428816, "learning_rate": 5.584491192990284e-05, "loss": 0.8749, "step": 4782 }, { "epoch": 2.167195287720888, "grad_norm": 0.3499355489658203, "learning_rate": 5.583329255845697e-05, "loss": 0.8672, "step": 4783 }, { "epoch": 2.167648391481649, "grad_norm": 0.4231337122160326, "learning_rate": 5.582167160253748e-05, "loss": 0.862, "step": 4784 }, { "epoch": 2.1681014952424107, "grad_norm": 0.34938345617577216, "learning_rate": 5.58100490633073e-05, "loss": 0.8756, "step": 4785 }, { "epoch": 2.168554599003172, "grad_norm": 0.2684337151294326, "learning_rate": 5.579842494192953e-05, "loss": 0.892, "step": 4786 }, { "epoch": 2.169007702763933, "grad_norm": 0.3018415275898452, "learning_rate": 5.578679923956741e-05, "loss": 0.8741, "step": 4787 }, { "epoch": 2.169460806524694, "grad_norm": 0.3435065191733229, "learning_rate": 5.577517195738436e-05, "loss": 0.8827, "step": 4788 }, { "epoch": 2.169913910285455, "grad_norm": 0.3604695792895655, "learning_rate": 5.576354309654394e-05, "loss": 0.8748, "step": 4789 }, { "epoch": 2.170367014046217, "grad_norm": 0.29107020914229553, "learning_rate": 5.575191265820989e-05, "loss": 0.8715, "step": 4790 }, { "epoch": 2.170820117806978, "grad_norm": 0.2540482059135918, "learning_rate": 5.574028064354608e-05, "loss": 0.8931, "step": 4791 }, { "epoch": 2.171273221567739, "grad_norm": 0.3001630149744979, "learning_rate": 5.5728647053716555e-05, "loss": 0.9021, "step": 4792 }, { "epoch": 2.1717263253285, "grad_norm": 0.302742345575988, "learning_rate": 5.571701188988552e-05, "loss": 0.8846, "step": 4793 }, { "epoch": 2.1721794290892613, "grad_norm": 0.3793121853946709, "learning_rate": 5.570537515321735e-05, "loss": 0.871, "step": 4794 }, { "epoch": 2.172632532850023, "grad_norm": 0.5107871177087893, "learning_rate": 5.569373684487653e-05, "loss": 0.8679, "step": 4795 }, { "epoch": 2.173085636610784, "grad_norm": 0.5028385289383155, "learning_rate": 5.5682096966027736e-05, "loss": 0.8834, "step": 4796 }, { "epoch": 2.173538740371545, "grad_norm": 0.42990976651590573, "learning_rate": 5.567045551783579e-05, "loss": 0.8753, "step": 4797 }, { "epoch": 2.1739918441323063, "grad_norm": 0.40469991559621177, "learning_rate": 5.565881250146571e-05, "loss": 0.8715, "step": 4798 }, { "epoch": 2.1744449478930674, "grad_norm": 0.388315613094456, "learning_rate": 5.564716791808263e-05, "loss": 0.8996, "step": 4799 }, { "epoch": 2.1748980516538285, "grad_norm": 0.44558675991520746, "learning_rate": 5.563552176885183e-05, "loss": 0.8791, "step": 4800 }, { "epoch": 2.17535115541459, "grad_norm": 0.39411401451034656, "learning_rate": 5.5623874054938784e-05, "loss": 0.8614, "step": 4801 }, { "epoch": 2.1758042591753513, "grad_norm": 0.24886540820530248, "learning_rate": 5.561222477750909e-05, "loss": 0.8657, "step": 4802 }, { "epoch": 2.1762573629361124, "grad_norm": 0.3129038230204768, "learning_rate": 5.560057393772852e-05, "loss": 0.8777, "step": 4803 }, { "epoch": 2.1767104666968735, "grad_norm": 0.3901621936412802, "learning_rate": 5.558892153676303e-05, "loss": 0.8707, "step": 4804 }, { "epoch": 2.1771635704576346, "grad_norm": 0.39373888023166487, "learning_rate": 5.5577267575778666e-05, "loss": 0.868, "step": 4805 }, { "epoch": 2.177616674218396, "grad_norm": 0.39392850172385835, "learning_rate": 5.5565612055941676e-05, "loss": 0.8808, "step": 4806 }, { "epoch": 2.1780697779791574, "grad_norm": 0.33249913288532096, "learning_rate": 5.5553954978418455e-05, "loss": 0.876, "step": 4807 }, { "epoch": 2.1785228817399185, "grad_norm": 0.4319723775146683, "learning_rate": 5.554229634437556e-05, "loss": 0.8685, "step": 4808 }, { "epoch": 2.1789759855006796, "grad_norm": 0.5120003114113659, "learning_rate": 5.553063615497969e-05, "loss": 0.8634, "step": 4809 }, { "epoch": 2.1794290892614407, "grad_norm": 0.5003571058315498, "learning_rate": 5.551897441139772e-05, "loss": 0.8665, "step": 4810 }, { "epoch": 2.179882193022202, "grad_norm": 0.5092879477120739, "learning_rate": 5.550731111479664e-05, "loss": 0.878, "step": 4811 }, { "epoch": 2.1803352967829635, "grad_norm": 0.6524590352816179, "learning_rate": 5.5495646266343666e-05, "loss": 0.881, "step": 4812 }, { "epoch": 2.1807884005437246, "grad_norm": 0.8358202178536265, "learning_rate": 5.548397986720608e-05, "loss": 0.8668, "step": 4813 }, { "epoch": 2.1812415043044857, "grad_norm": 0.8671789631411924, "learning_rate": 5.5472311918551386e-05, "loss": 0.8989, "step": 4814 }, { "epoch": 2.181694608065247, "grad_norm": 0.8096527380719245, "learning_rate": 5.546064242154723e-05, "loss": 0.9015, "step": 4815 }, { "epoch": 2.182147711826008, "grad_norm": 0.7831737013061192, "learning_rate": 5.544897137736138e-05, "loss": 0.8685, "step": 4816 }, { "epoch": 2.1826008155867695, "grad_norm": 0.7007083326715926, "learning_rate": 5.543729878716181e-05, "loss": 0.8672, "step": 4817 }, { "epoch": 2.1830539193475307, "grad_norm": 0.5821528691806229, "learning_rate": 5.542562465211662e-05, "loss": 0.876, "step": 4818 }, { "epoch": 2.183507023108292, "grad_norm": 0.6294275155407829, "learning_rate": 5.541394897339405e-05, "loss": 0.8854, "step": 4819 }, { "epoch": 2.183960126869053, "grad_norm": 0.6095735397112761, "learning_rate": 5.540227175216252e-05, "loss": 0.8739, "step": 4820 }, { "epoch": 2.184413230629814, "grad_norm": 0.45687989198186174, "learning_rate": 5.539059298959061e-05, "loss": 0.8626, "step": 4821 }, { "epoch": 2.1848663343905756, "grad_norm": 0.359621256669325, "learning_rate": 5.5378912686847015e-05, "loss": 0.8875, "step": 4822 }, { "epoch": 2.185319438151337, "grad_norm": 0.46095865188766355, "learning_rate": 5.536723084510063e-05, "loss": 0.8609, "step": 4823 }, { "epoch": 2.185772541912098, "grad_norm": 0.47933020071230453, "learning_rate": 5.535554746552047e-05, "loss": 0.8782, "step": 4824 }, { "epoch": 2.186225645672859, "grad_norm": 0.4118449675678445, "learning_rate": 5.5343862549275726e-05, "loss": 0.8857, "step": 4825 }, { "epoch": 2.18667874943362, "grad_norm": 0.4132321330090307, "learning_rate": 5.533217609753573e-05, "loss": 0.8669, "step": 4826 }, { "epoch": 2.1871318531943817, "grad_norm": 0.40211921773881765, "learning_rate": 5.532048811146998e-05, "loss": 0.8825, "step": 4827 }, { "epoch": 2.187584956955143, "grad_norm": 0.4951430915224942, "learning_rate": 5.53087985922481e-05, "loss": 0.8841, "step": 4828 }, { "epoch": 2.188038060715904, "grad_norm": 0.5243052924177999, "learning_rate": 5.529710754103992e-05, "loss": 0.8836, "step": 4829 }, { "epoch": 2.188491164476665, "grad_norm": 0.49745353153915, "learning_rate": 5.528541495901536e-05, "loss": 0.8775, "step": 4830 }, { "epoch": 2.1889442682374263, "grad_norm": 0.43697619529708825, "learning_rate": 5.5273720847344534e-05, "loss": 0.8671, "step": 4831 }, { "epoch": 2.1893973719981874, "grad_norm": 0.4538761597543461, "learning_rate": 5.52620252071977e-05, "loss": 0.8633, "step": 4832 }, { "epoch": 2.189850475758949, "grad_norm": 0.5169265818506497, "learning_rate": 5.525032803974525e-05, "loss": 0.8894, "step": 4833 }, { "epoch": 2.19030357951971, "grad_norm": 0.7499712586338511, "learning_rate": 5.523862934615778e-05, "loss": 0.8867, "step": 4834 }, { "epoch": 2.1907566832804712, "grad_norm": 0.7289792269673264, "learning_rate": 5.5226929127605965e-05, "loss": 0.8763, "step": 4835 }, { "epoch": 2.1912097870412324, "grad_norm": 0.7417882328424827, "learning_rate": 5.52152273852607e-05, "loss": 0.883, "step": 4836 }, { "epoch": 2.1916628908019935, "grad_norm": 0.7586465235952196, "learning_rate": 5.520352412029301e-05, "loss": 0.8622, "step": 4837 }, { "epoch": 2.192115994562755, "grad_norm": 0.8487673367815444, "learning_rate": 5.519181933387405e-05, "loss": 0.8732, "step": 4838 }, { "epoch": 2.192569098323516, "grad_norm": 0.8986576588418204, "learning_rate": 5.518011302717515e-05, "loss": 0.8926, "step": 4839 }, { "epoch": 2.1930222020842773, "grad_norm": 0.8839403113402668, "learning_rate": 5.5168405201367784e-05, "loss": 0.8727, "step": 4840 }, { "epoch": 2.1934753058450385, "grad_norm": 0.7417302863918841, "learning_rate": 5.515669585762359e-05, "loss": 0.8632, "step": 4841 }, { "epoch": 2.1939284096057996, "grad_norm": 0.6619275221506417, "learning_rate": 5.5144984997114335e-05, "loss": 0.8729, "step": 4842 }, { "epoch": 2.1943815133665607, "grad_norm": 0.5438728474839961, "learning_rate": 5.5133272621011956e-05, "loss": 0.8902, "step": 4843 }, { "epoch": 2.1948346171273223, "grad_norm": 0.38212546616291293, "learning_rate": 5.512155873048855e-05, "loss": 0.8726, "step": 4844 }, { "epoch": 2.1952877208880834, "grad_norm": 0.3109641822923189, "learning_rate": 5.510984332671633e-05, "loss": 0.8611, "step": 4845 }, { "epoch": 2.1957408246488446, "grad_norm": 0.36214476376805915, "learning_rate": 5.509812641086771e-05, "loss": 0.8638, "step": 4846 }, { "epoch": 2.1961939284096057, "grad_norm": 0.4063237613378896, "learning_rate": 5.508640798411522e-05, "loss": 0.854, "step": 4847 }, { "epoch": 2.196647032170367, "grad_norm": 0.4193124858678961, "learning_rate": 5.5074688047631526e-05, "loss": 0.8672, "step": 4848 }, { "epoch": 2.1971001359311284, "grad_norm": 0.41126850224111877, "learning_rate": 5.506296660258949e-05, "loss": 0.8838, "step": 4849 }, { "epoch": 2.1975532396918895, "grad_norm": 0.45584458432375385, "learning_rate": 5.505124365016211e-05, "loss": 0.8637, "step": 4850 }, { "epoch": 2.1980063434526507, "grad_norm": 0.5354920444784346, "learning_rate": 5.503951919152252e-05, "loss": 0.871, "step": 4851 }, { "epoch": 2.198459447213412, "grad_norm": 0.5024464124635492, "learning_rate": 5.502779322784399e-05, "loss": 0.8828, "step": 4852 }, { "epoch": 2.198912550974173, "grad_norm": 0.48708523840112794, "learning_rate": 5.501606576030001e-05, "loss": 0.8631, "step": 4853 }, { "epoch": 2.1993656547349345, "grad_norm": 0.5009910293148352, "learning_rate": 5.500433679006414e-05, "loss": 0.8759, "step": 4854 }, { "epoch": 2.1998187584956956, "grad_norm": 0.4852811116055525, "learning_rate": 5.499260631831014e-05, "loss": 0.8654, "step": 4855 }, { "epoch": 2.2002718622564568, "grad_norm": 0.42836581598462253, "learning_rate": 5.498087434621189e-05, "loss": 0.8922, "step": 4856 }, { "epoch": 2.200724966017218, "grad_norm": 0.43844271388276107, "learning_rate": 5.496914087494347e-05, "loss": 0.8497, "step": 4857 }, { "epoch": 2.201178069777979, "grad_norm": 0.5255187981672572, "learning_rate": 5.495740590567903e-05, "loss": 0.8739, "step": 4858 }, { "epoch": 2.2016311735387406, "grad_norm": 0.44375135338024163, "learning_rate": 5.494566943959294e-05, "loss": 0.8676, "step": 4859 }, { "epoch": 2.2020842772995017, "grad_norm": 0.4105360527264789, "learning_rate": 5.493393147785968e-05, "loss": 0.8881, "step": 4860 }, { "epoch": 2.202537381060263, "grad_norm": 0.4422465548926793, "learning_rate": 5.492219202165392e-05, "loss": 0.8606, "step": 4861 }, { "epoch": 2.202990484821024, "grad_norm": 0.3981413855384484, "learning_rate": 5.491045107215043e-05, "loss": 0.871, "step": 4862 }, { "epoch": 2.203443588581785, "grad_norm": 0.40251294490288136, "learning_rate": 5.489870863052416e-05, "loss": 0.87, "step": 4863 }, { "epoch": 2.2038966923425463, "grad_norm": 0.312527459730537, "learning_rate": 5.488696469795022e-05, "loss": 0.8947, "step": 4864 }, { "epoch": 2.204349796103308, "grad_norm": 0.29641646569556074, "learning_rate": 5.487521927560384e-05, "loss": 0.8787, "step": 4865 }, { "epoch": 2.204802899864069, "grad_norm": 0.31179684422672255, "learning_rate": 5.486347236466039e-05, "loss": 0.8741, "step": 4866 }, { "epoch": 2.20525600362483, "grad_norm": 0.3594714144536846, "learning_rate": 5.4851723966295435e-05, "loss": 0.8967, "step": 4867 }, { "epoch": 2.2057091073855912, "grad_norm": 0.43012321889164895, "learning_rate": 5.4839974081684665e-05, "loss": 0.874, "step": 4868 }, { "epoch": 2.2061622111463524, "grad_norm": 0.5012910934469083, "learning_rate": 5.482822271200389e-05, "loss": 0.8691, "step": 4869 }, { "epoch": 2.206615314907114, "grad_norm": 0.4956455442061888, "learning_rate": 5.481646985842914e-05, "loss": 0.8822, "step": 4870 }, { "epoch": 2.207068418667875, "grad_norm": 0.4386385114582711, "learning_rate": 5.4804715522136515e-05, "loss": 0.8717, "step": 4871 }, { "epoch": 2.207521522428636, "grad_norm": 0.30118944106572515, "learning_rate": 5.4792959704302315e-05, "loss": 0.8743, "step": 4872 }, { "epoch": 2.2079746261893973, "grad_norm": 0.40731832449462785, "learning_rate": 5.478120240610297e-05, "loss": 0.8718, "step": 4873 }, { "epoch": 2.2084277299501585, "grad_norm": 0.28645604045486395, "learning_rate": 5.476944362871505e-05, "loss": 0.9007, "step": 4874 }, { "epoch": 2.2088808337109196, "grad_norm": 0.30714366976676194, "learning_rate": 5.475768337331529e-05, "loss": 0.8706, "step": 4875 }, { "epoch": 2.209333937471681, "grad_norm": 0.3202601761067375, "learning_rate": 5.474592164108056e-05, "loss": 0.8502, "step": 4876 }, { "epoch": 2.2097870412324423, "grad_norm": 0.3688758712122463, "learning_rate": 5.473415843318789e-05, "loss": 0.8599, "step": 4877 }, { "epoch": 2.2102401449932034, "grad_norm": 0.42721640610002676, "learning_rate": 5.472239375081445e-05, "loss": 0.869, "step": 4878 }, { "epoch": 2.2106932487539646, "grad_norm": 0.5335567303091382, "learning_rate": 5.471062759513755e-05, "loss": 0.8789, "step": 4879 }, { "epoch": 2.2111463525147257, "grad_norm": 0.678607799720587, "learning_rate": 5.469885996733467e-05, "loss": 0.867, "step": 4880 }, { "epoch": 2.2115994562754873, "grad_norm": 0.7989469241087084, "learning_rate": 5.4687090868583417e-05, "loss": 0.8615, "step": 4881 }, { "epoch": 2.2120525600362484, "grad_norm": 0.8751827794324839, "learning_rate": 5.4675320300061546e-05, "loss": 0.8648, "step": 4882 }, { "epoch": 2.2125056637970095, "grad_norm": 0.8333100385683635, "learning_rate": 5.466354826294698e-05, "loss": 0.8475, "step": 4883 }, { "epoch": 2.2129587675577707, "grad_norm": 0.6851973644189394, "learning_rate": 5.465177475841775e-05, "loss": 0.8753, "step": 4884 }, { "epoch": 2.213411871318532, "grad_norm": 0.5391978672650279, "learning_rate": 5.463999978765207e-05, "loss": 0.8876, "step": 4885 }, { "epoch": 2.2138649750792934, "grad_norm": 0.4155817268764987, "learning_rate": 5.4628223351828304e-05, "loss": 0.8608, "step": 4886 }, { "epoch": 2.2143180788400545, "grad_norm": 0.3529692983230833, "learning_rate": 5.461644545212492e-05, "loss": 0.8711, "step": 4887 }, { "epoch": 2.2147711826008156, "grad_norm": 0.4495915304816902, "learning_rate": 5.4604666089720565e-05, "loss": 0.8623, "step": 4888 }, { "epoch": 2.2152242863615768, "grad_norm": 0.5747578153682821, "learning_rate": 5.4592885265794036e-05, "loss": 0.8675, "step": 4889 }, { "epoch": 2.215677390122338, "grad_norm": 0.7638343360792084, "learning_rate": 5.4581102981524266e-05, "loss": 0.8832, "step": 4890 }, { "epoch": 2.2161304938830995, "grad_norm": 0.9019527146861916, "learning_rate": 5.4569319238090334e-05, "loss": 0.8916, "step": 4891 }, { "epoch": 2.2165835976438606, "grad_norm": 0.9020066670481287, "learning_rate": 5.455753403667146e-05, "loss": 0.8803, "step": 4892 }, { "epoch": 2.2170367014046217, "grad_norm": 0.8873593250337221, "learning_rate": 5.454574737844701e-05, "loss": 0.8679, "step": 4893 }, { "epoch": 2.217489805165383, "grad_norm": 0.8619361773296413, "learning_rate": 5.453395926459653e-05, "loss": 0.8859, "step": 4894 }, { "epoch": 2.217942908926144, "grad_norm": 0.8003470806088236, "learning_rate": 5.452216969629965e-05, "loss": 0.8784, "step": 4895 }, { "epoch": 2.218396012686905, "grad_norm": 0.7277904181310584, "learning_rate": 5.45103786747362e-05, "loss": 0.8746, "step": 4896 }, { "epoch": 2.2188491164476667, "grad_norm": 0.5343168160653397, "learning_rate": 5.4498586201086125e-05, "loss": 0.8654, "step": 4897 }, { "epoch": 2.219302220208428, "grad_norm": 0.3533041437059685, "learning_rate": 5.448679227652953e-05, "loss": 0.8806, "step": 4898 }, { "epoch": 2.219755323969189, "grad_norm": 0.3985525359417247, "learning_rate": 5.447499690224665e-05, "loss": 0.8683, "step": 4899 }, { "epoch": 2.22020842772995, "grad_norm": 0.4907676618337724, "learning_rate": 5.44632000794179e-05, "loss": 0.8794, "step": 4900 }, { "epoch": 2.220661531490711, "grad_norm": 0.47208606736433767, "learning_rate": 5.445140180922378e-05, "loss": 0.8798, "step": 4901 }, { "epoch": 2.221114635251473, "grad_norm": 0.34527579095445765, "learning_rate": 5.4439602092845e-05, "loss": 0.8819, "step": 4902 }, { "epoch": 2.221567739012234, "grad_norm": 0.3182357567547643, "learning_rate": 5.442780093146237e-05, "loss": 0.8677, "step": 4903 }, { "epoch": 2.222020842772995, "grad_norm": 0.38275375904000325, "learning_rate": 5.4415998326256854e-05, "loss": 0.8983, "step": 4904 }, { "epoch": 2.222473946533756, "grad_norm": 0.4489575707616605, "learning_rate": 5.440419427840957e-05, "loss": 0.8715, "step": 4905 }, { "epoch": 2.2229270502945173, "grad_norm": 0.5269755471627443, "learning_rate": 5.439238878910179e-05, "loss": 0.8687, "step": 4906 }, { "epoch": 2.2233801540552784, "grad_norm": 0.4999762478504887, "learning_rate": 5.438058185951489e-05, "loss": 0.864, "step": 4907 }, { "epoch": 2.22383325781604, "grad_norm": 0.3754171365295588, "learning_rate": 5.4368773490830446e-05, "loss": 0.8594, "step": 4908 }, { "epoch": 2.224286361576801, "grad_norm": 0.3530592573887493, "learning_rate": 5.435696368423013e-05, "loss": 0.8594, "step": 4909 }, { "epoch": 2.2247394653375623, "grad_norm": 0.3592318837274856, "learning_rate": 5.434515244089578e-05, "loss": 0.8865, "step": 4910 }, { "epoch": 2.2251925690983234, "grad_norm": 0.35514935971403117, "learning_rate": 5.433333976200937e-05, "loss": 0.8966, "step": 4911 }, { "epoch": 2.2256456728590845, "grad_norm": 0.40700628223968693, "learning_rate": 5.432152564875302e-05, "loss": 0.8928, "step": 4912 }, { "epoch": 2.226098776619846, "grad_norm": 0.4843696020597729, "learning_rate": 5.4309710102309003e-05, "loss": 0.8774, "step": 4913 }, { "epoch": 2.2265518803806073, "grad_norm": 0.5609635024355284, "learning_rate": 5.429789312385973e-05, "loss": 0.874, "step": 4914 }, { "epoch": 2.2270049841413684, "grad_norm": 0.7137233530757822, "learning_rate": 5.428607471458774e-05, "loss": 0.8706, "step": 4915 }, { "epoch": 2.2274580879021295, "grad_norm": 0.8585908789376594, "learning_rate": 5.427425487567575e-05, "loss": 0.8629, "step": 4916 }, { "epoch": 2.2279111916628906, "grad_norm": 0.8931615914913592, "learning_rate": 5.426243360830657e-05, "loss": 0.8555, "step": 4917 }, { "epoch": 2.228364295423652, "grad_norm": 0.8793310912389737, "learning_rate": 5.425061091366321e-05, "loss": 0.8749, "step": 4918 }, { "epoch": 2.2288173991844134, "grad_norm": 0.8165905509419864, "learning_rate": 5.423878679292877e-05, "loss": 0.8697, "step": 4919 }, { "epoch": 2.2292705029451745, "grad_norm": 0.7718434262835749, "learning_rate": 5.422696124728652e-05, "loss": 0.8641, "step": 4920 }, { "epoch": 2.2297236067059356, "grad_norm": 0.8046539715341583, "learning_rate": 5.421513427791988e-05, "loss": 0.8928, "step": 4921 }, { "epoch": 2.2301767104666967, "grad_norm": 0.9214392473560583, "learning_rate": 5.420330588601239e-05, "loss": 0.8837, "step": 4922 }, { "epoch": 2.2306298142274583, "grad_norm": 0.9377388830733265, "learning_rate": 5.419147607274776e-05, "loss": 0.8793, "step": 4923 }, { "epoch": 2.2310829179882194, "grad_norm": 0.876810119673017, "learning_rate": 5.4179644839309815e-05, "loss": 0.8524, "step": 4924 }, { "epoch": 2.2315360217489806, "grad_norm": 0.8127760412468443, "learning_rate": 5.416781218688254e-05, "loss": 0.8917, "step": 4925 }, { "epoch": 2.2319891255097417, "grad_norm": 0.8694947738685787, "learning_rate": 5.415597811665005e-05, "loss": 0.8876, "step": 4926 }, { "epoch": 2.232442229270503, "grad_norm": 0.8956733802576463, "learning_rate": 5.4144142629796607e-05, "loss": 0.8739, "step": 4927 }, { "epoch": 2.232895333031264, "grad_norm": 0.7570395180469035, "learning_rate": 5.413230572750662e-05, "loss": 0.8796, "step": 4928 }, { "epoch": 2.2333484367920255, "grad_norm": 0.40243486161082453, "learning_rate": 5.412046741096462e-05, "loss": 0.8763, "step": 4929 }, { "epoch": 2.2338015405527867, "grad_norm": 0.40630696307834885, "learning_rate": 5.4108627681355315e-05, "loss": 0.8782, "step": 4930 }, { "epoch": 2.234254644313548, "grad_norm": 0.5156498443716413, "learning_rate": 5.409678653986352e-05, "loss": 0.8801, "step": 4931 }, { "epoch": 2.234707748074309, "grad_norm": 0.524833096992224, "learning_rate": 5.4084943987674215e-05, "loss": 0.8634, "step": 4932 }, { "epoch": 2.23516085183507, "grad_norm": 0.596029118807832, "learning_rate": 5.40731000259725e-05, "loss": 0.8869, "step": 4933 }, { "epoch": 2.2356139555958316, "grad_norm": 0.6799420333880188, "learning_rate": 5.406125465594365e-05, "loss": 0.8959, "step": 4934 }, { "epoch": 2.236067059356593, "grad_norm": 0.7130118864043761, "learning_rate": 5.404940787877302e-05, "loss": 0.8803, "step": 4935 }, { "epoch": 2.236520163117354, "grad_norm": 0.6287917517233925, "learning_rate": 5.403755969564617e-05, "loss": 0.8789, "step": 4936 }, { "epoch": 2.236973266878115, "grad_norm": 0.5045728915217548, "learning_rate": 5.4025710107748765e-05, "loss": 0.8612, "step": 4937 }, { "epoch": 2.237426370638876, "grad_norm": 0.31633663814298546, "learning_rate": 5.401385911626663e-05, "loss": 0.8902, "step": 4938 }, { "epoch": 2.2378794743996373, "grad_norm": 0.3010706810814262, "learning_rate": 5.400200672238572e-05, "loss": 0.8718, "step": 4939 }, { "epoch": 2.238332578160399, "grad_norm": 0.44111460486679166, "learning_rate": 5.399015292729212e-05, "loss": 0.8864, "step": 4940 }, { "epoch": 2.23878568192116, "grad_norm": 0.44339231673568846, "learning_rate": 5.397829773217208e-05, "loss": 0.8566, "step": 4941 }, { "epoch": 2.239238785681921, "grad_norm": 0.4371318915957809, "learning_rate": 5.396644113821197e-05, "loss": 0.8825, "step": 4942 }, { "epoch": 2.2396918894426823, "grad_norm": 0.45352886710915286, "learning_rate": 5.3954583146598296e-05, "loss": 0.8684, "step": 4943 }, { "epoch": 2.2401449932034434, "grad_norm": 0.5476087383111062, "learning_rate": 5.394272375851773e-05, "loss": 0.8818, "step": 4944 }, { "epoch": 2.240598096964205, "grad_norm": 0.6214657979947548, "learning_rate": 5.393086297515706e-05, "loss": 0.8893, "step": 4945 }, { "epoch": 2.241051200724966, "grad_norm": 0.6628566890725509, "learning_rate": 5.391900079770322e-05, "loss": 0.869, "step": 4946 }, { "epoch": 2.2415043044857272, "grad_norm": 0.6409898634497884, "learning_rate": 5.3907137227343296e-05, "loss": 0.8885, "step": 4947 }, { "epoch": 2.2419574082464884, "grad_norm": 0.5438398391572614, "learning_rate": 5.389527226526449e-05, "loss": 0.8565, "step": 4948 }, { "epoch": 2.2424105120072495, "grad_norm": 0.4629483985229684, "learning_rate": 5.3883405912654155e-05, "loss": 0.8747, "step": 4949 }, { "epoch": 2.242863615768011, "grad_norm": 0.43771022823707867, "learning_rate": 5.38715381706998e-05, "loss": 0.8753, "step": 4950 }, { "epoch": 2.243316719528772, "grad_norm": 0.5495128381425656, "learning_rate": 5.385966904058904e-05, "loss": 0.8774, "step": 4951 }, { "epoch": 2.2437698232895333, "grad_norm": 0.5853534968767987, "learning_rate": 5.384779852350965e-05, "loss": 0.8637, "step": 4952 }, { "epoch": 2.2442229270502945, "grad_norm": 0.47954112829805073, "learning_rate": 5.383592662064956e-05, "loss": 0.872, "step": 4953 }, { "epoch": 2.2446760308110556, "grad_norm": 0.314343012956651, "learning_rate": 5.382405333319678e-05, "loss": 0.8674, "step": 4954 }, { "epoch": 2.245129134571817, "grad_norm": 0.3132152271332136, "learning_rate": 5.381217866233952e-05, "loss": 0.8791, "step": 4955 }, { "epoch": 2.2455822383325783, "grad_norm": 0.46096108178641093, "learning_rate": 5.38003026092661e-05, "loss": 0.8665, "step": 4956 }, { "epoch": 2.2460353420933394, "grad_norm": 0.4972029337215353, "learning_rate": 5.3788425175165e-05, "loss": 0.8928, "step": 4957 }, { "epoch": 2.2464884458541006, "grad_norm": 0.4202887200296547, "learning_rate": 5.377654636122479e-05, "loss": 0.8582, "step": 4958 }, { "epoch": 2.2469415496148617, "grad_norm": 0.3608195378522506, "learning_rate": 5.376466616863425e-05, "loss": 0.8654, "step": 4959 }, { "epoch": 2.247394653375623, "grad_norm": 0.3371079343123631, "learning_rate": 5.375278459858221e-05, "loss": 0.8767, "step": 4960 }, { "epoch": 2.2478477571363844, "grad_norm": 0.4118466991484278, "learning_rate": 5.3740901652257736e-05, "loss": 0.8673, "step": 4961 }, { "epoch": 2.2483008608971455, "grad_norm": 0.4958575705178244, "learning_rate": 5.3729017330849944e-05, "loss": 0.8672, "step": 4962 }, { "epoch": 2.2487539646579067, "grad_norm": 0.4536846789245243, "learning_rate": 5.3717131635548145e-05, "loss": 0.8665, "step": 4963 }, { "epoch": 2.249207068418668, "grad_norm": 0.437390027239519, "learning_rate": 5.370524456754176e-05, "loss": 0.8705, "step": 4964 }, { "epoch": 2.249660172179429, "grad_norm": 0.5084176722324538, "learning_rate": 5.3693356128020346e-05, "loss": 0.8652, "step": 4965 }, { "epoch": 2.2501132759401905, "grad_norm": 0.5558221110203564, "learning_rate": 5.3681466318173627e-05, "loss": 0.8824, "step": 4966 }, { "epoch": 2.2505663797009516, "grad_norm": 0.5754277523587578, "learning_rate": 5.3669575139191423e-05, "loss": 0.9009, "step": 4967 }, { "epoch": 2.2510194834617128, "grad_norm": 0.5360335415008358, "learning_rate": 5.365768259226372e-05, "loss": 0.8778, "step": 4968 }, { "epoch": 2.251472587222474, "grad_norm": 0.5218228193743054, "learning_rate": 5.3645788678580647e-05, "loss": 0.8595, "step": 4969 }, { "epoch": 2.251925690983235, "grad_norm": 0.4293367348818291, "learning_rate": 5.363389339933243e-05, "loss": 0.8778, "step": 4970 }, { "epoch": 2.252378794743996, "grad_norm": 0.36659152952450325, "learning_rate": 5.3621996755709485e-05, "loss": 0.8685, "step": 4971 }, { "epoch": 2.2528318985047577, "grad_norm": 0.4264961325747676, "learning_rate": 5.36100987489023e-05, "loss": 0.8721, "step": 4972 }, { "epoch": 2.253285002265519, "grad_norm": 0.4864884930965443, "learning_rate": 5.359819938010156e-05, "loss": 0.8671, "step": 4973 }, { "epoch": 2.25373810602628, "grad_norm": 0.5047206011479695, "learning_rate": 5.3586298650498054e-05, "loss": 0.8635, "step": 4974 }, { "epoch": 2.254191209787041, "grad_norm": 0.4847858078189771, "learning_rate": 5.3574396561282726e-05, "loss": 0.8831, "step": 4975 }, { "epoch": 2.2546443135478023, "grad_norm": 0.39437378324774164, "learning_rate": 5.356249311364662e-05, "loss": 0.8575, "step": 4976 }, { "epoch": 2.255097417308564, "grad_norm": 0.44277059685795767, "learning_rate": 5.355058830878097e-05, "loss": 0.8857, "step": 4977 }, { "epoch": 2.255550521069325, "grad_norm": 0.4662221591467112, "learning_rate": 5.35386821478771e-05, "loss": 0.8798, "step": 4978 }, { "epoch": 2.256003624830086, "grad_norm": 0.4602798773026528, "learning_rate": 5.35267746321265e-05, "loss": 0.8721, "step": 4979 }, { "epoch": 2.2564567285908472, "grad_norm": 0.5299758341742542, "learning_rate": 5.3514865762720774e-05, "loss": 0.8947, "step": 4980 }, { "epoch": 2.2569098323516084, "grad_norm": 0.6515467376949613, "learning_rate": 5.350295554085165e-05, "loss": 0.882, "step": 4981 }, { "epoch": 2.2573629361123695, "grad_norm": 0.7461460728971437, "learning_rate": 5.3491043967711043e-05, "loss": 0.8916, "step": 4982 }, { "epoch": 2.257816039873131, "grad_norm": 0.8418351777198434, "learning_rate": 5.347913104449096e-05, "loss": 0.8919, "step": 4983 }, { "epoch": 2.258269143633892, "grad_norm": 0.8662385794176996, "learning_rate": 5.346721677238355e-05, "loss": 0.8694, "step": 4984 }, { "epoch": 2.2587222473946533, "grad_norm": 0.7185439717655214, "learning_rate": 5.3455301152581106e-05, "loss": 0.8867, "step": 4985 }, { "epoch": 2.2591753511554145, "grad_norm": 0.5070946989796395, "learning_rate": 5.3443384186276044e-05, "loss": 0.8839, "step": 4986 }, { "epoch": 2.259628454916176, "grad_norm": 0.30589633181002635, "learning_rate": 5.343146587466093e-05, "loss": 0.8634, "step": 4987 }, { "epoch": 2.260081558676937, "grad_norm": 0.3536467331153795, "learning_rate": 5.3419546218928464e-05, "loss": 0.8689, "step": 4988 }, { "epoch": 2.2605346624376983, "grad_norm": 0.4308828071303492, "learning_rate": 5.3407625220271455e-05, "loss": 0.8708, "step": 4989 }, { "epoch": 2.2609877661984594, "grad_norm": 0.4701030918498077, "learning_rate": 5.3395702879882863e-05, "loss": 0.878, "step": 4990 }, { "epoch": 2.2614408699592206, "grad_norm": 0.6186083496219601, "learning_rate": 5.33837791989558e-05, "loss": 0.871, "step": 4991 }, { "epoch": 2.2618939737199817, "grad_norm": 0.5947376299964939, "learning_rate": 5.337185417868349e-05, "loss": 0.8607, "step": 4992 }, { "epoch": 2.2623470774807433, "grad_norm": 0.5381552874081078, "learning_rate": 5.335992782025929e-05, "loss": 0.8654, "step": 4993 }, { "epoch": 2.2628001812415044, "grad_norm": 0.4015895119193189, "learning_rate": 5.334800012487671e-05, "loss": 0.8618, "step": 4994 }, { "epoch": 2.2632532850022655, "grad_norm": 0.3772694666671948, "learning_rate": 5.3336071093729386e-05, "loss": 0.8682, "step": 4995 }, { "epoch": 2.2637063887630267, "grad_norm": 0.38670626274074543, "learning_rate": 5.332414072801106e-05, "loss": 0.8679, "step": 4996 }, { "epoch": 2.264159492523788, "grad_norm": 0.4180062765483464, "learning_rate": 5.3312209028915644e-05, "loss": 0.8775, "step": 4997 }, { "epoch": 2.2646125962845494, "grad_norm": 0.4105640785652557, "learning_rate": 5.330027599763717e-05, "loss": 0.8664, "step": 4998 }, { "epoch": 2.2650657000453105, "grad_norm": 0.4061348467662166, "learning_rate": 5.32883416353698e-05, "loss": 0.8686, "step": 4999 }, { "epoch": 2.2655188038060716, "grad_norm": 0.5126205760045184, "learning_rate": 5.327640594330785e-05, "loss": 0.8709, "step": 5000 }, { "epoch": 2.2659719075668328, "grad_norm": 0.5886147786534793, "learning_rate": 5.3264468922645726e-05, "loss": 0.8567, "step": 5001 }, { "epoch": 2.266425011327594, "grad_norm": 0.6095739677503562, "learning_rate": 5.325253057457801e-05, "loss": 0.8895, "step": 5002 }, { "epoch": 2.266878115088355, "grad_norm": 0.6101151588191713, "learning_rate": 5.3240590900299386e-05, "loss": 0.8847, "step": 5003 }, { "epoch": 2.2673312188491166, "grad_norm": 0.6078095085189323, "learning_rate": 5.3228649901004695e-05, "loss": 0.8734, "step": 5004 }, { "epoch": 2.2677843226098777, "grad_norm": 0.567582548207401, "learning_rate": 5.321670757788891e-05, "loss": 0.8656, "step": 5005 }, { "epoch": 2.268237426370639, "grad_norm": 0.5797288851751486, "learning_rate": 5.32047639321471e-05, "loss": 0.8563, "step": 5006 }, { "epoch": 2.2686905301314, "grad_norm": 0.617424776512293, "learning_rate": 5.319281896497452e-05, "loss": 0.8888, "step": 5007 }, { "epoch": 2.269143633892161, "grad_norm": 0.5731350654367866, "learning_rate": 5.318087267756651e-05, "loss": 0.8843, "step": 5008 }, { "epoch": 2.2695967376529227, "grad_norm": 0.38332542319124535, "learning_rate": 5.316892507111856e-05, "loss": 0.8727, "step": 5009 }, { "epoch": 2.270049841413684, "grad_norm": 0.37467986766569156, "learning_rate": 5.3156976146826325e-05, "loss": 0.8835, "step": 5010 }, { "epoch": 2.270502945174445, "grad_norm": 0.46072163138211225, "learning_rate": 5.314502590588552e-05, "loss": 0.8642, "step": 5011 }, { "epoch": 2.270956048935206, "grad_norm": 0.4263272438778774, "learning_rate": 5.313307434949206e-05, "loss": 0.8751, "step": 5012 }, { "epoch": 2.271409152695967, "grad_norm": 0.3104459679836457, "learning_rate": 5.312112147884197e-05, "loss": 0.8667, "step": 5013 }, { "epoch": 2.2718622564567283, "grad_norm": 0.3310320541121191, "learning_rate": 5.310916729513137e-05, "loss": 0.8604, "step": 5014 }, { "epoch": 2.27231536021749, "grad_norm": 2.909903918316484, "learning_rate": 5.309721179955657e-05, "loss": 0.9705, "step": 5015 }, { "epoch": 2.272768463978251, "grad_norm": 0.5028510966032759, "learning_rate": 5.3085254993313964e-05, "loss": 0.8936, "step": 5016 }, { "epoch": 2.273221567739012, "grad_norm": 0.927192502560811, "learning_rate": 5.307329687760011e-05, "loss": 0.8803, "step": 5017 }, { "epoch": 2.2736746714997733, "grad_norm": 1.0587374066621396, "learning_rate": 5.306133745361168e-05, "loss": 0.8768, "step": 5018 }, { "epoch": 2.274127775260535, "grad_norm": 0.8929226102780633, "learning_rate": 5.3049376722545486e-05, "loss": 0.8922, "step": 5019 }, { "epoch": 2.274580879021296, "grad_norm": 0.8417821994761989, "learning_rate": 5.303741468559846e-05, "loss": 0.8818, "step": 5020 }, { "epoch": 2.275033982782057, "grad_norm": 0.7143471569040695, "learning_rate": 5.3025451343967674e-05, "loss": 0.8787, "step": 5021 }, { "epoch": 2.2754870865428183, "grad_norm": 0.6876130573930918, "learning_rate": 5.3013486698850325e-05, "loss": 0.8726, "step": 5022 }, { "epoch": 2.2759401903035794, "grad_norm": 0.634212262847447, "learning_rate": 5.300152075144373e-05, "loss": 0.8738, "step": 5023 }, { "epoch": 2.2763932940643405, "grad_norm": 0.6620753647506542, "learning_rate": 5.2989553502945365e-05, "loss": 0.868, "step": 5024 }, { "epoch": 2.276846397825102, "grad_norm": 0.6582988826726103, "learning_rate": 5.297758495455282e-05, "loss": 0.8748, "step": 5025 }, { "epoch": 2.2772995015858633, "grad_norm": 0.5256972091380337, "learning_rate": 5.29656151074638e-05, "loss": 0.874, "step": 5026 }, { "epoch": 2.2777526053466244, "grad_norm": 0.6028230924473869, "learning_rate": 5.2953643962876166e-05, "loss": 0.8699, "step": 5027 }, { "epoch": 2.2782057091073855, "grad_norm": 0.6540996103460612, "learning_rate": 5.2941671521987905e-05, "loss": 0.873, "step": 5028 }, { "epoch": 2.2786588128681466, "grad_norm": 0.47807355229021303, "learning_rate": 5.2929697785997116e-05, "loss": 0.874, "step": 5029 }, { "epoch": 2.279111916628908, "grad_norm": 0.5188193808219204, "learning_rate": 5.291772275610203e-05, "loss": 0.8636, "step": 5030 }, { "epoch": 2.2795650203896693, "grad_norm": 0.6021392990031753, "learning_rate": 5.2905746433501044e-05, "loss": 0.8596, "step": 5031 }, { "epoch": 2.2800181241504305, "grad_norm": 0.4471953984295619, "learning_rate": 5.2893768819392627e-05, "loss": 0.8647, "step": 5032 }, { "epoch": 2.2804712279111916, "grad_norm": 0.5818963896811609, "learning_rate": 5.288178991497541e-05, "loss": 0.8746, "step": 5033 }, { "epoch": 2.2809243316719527, "grad_norm": 0.6125478351761938, "learning_rate": 5.2869809721448154e-05, "loss": 0.8652, "step": 5034 }, { "epoch": 2.281377435432714, "grad_norm": 0.5390439656613791, "learning_rate": 5.285782824000976e-05, "loss": 0.8705, "step": 5035 }, { "epoch": 2.2818305391934754, "grad_norm": 0.636797302373421, "learning_rate": 5.284584547185921e-05, "loss": 0.899, "step": 5036 }, { "epoch": 2.2822836429542366, "grad_norm": 0.6646421388786684, "learning_rate": 5.2833861418195666e-05, "loss": 0.8799, "step": 5037 }, { "epoch": 2.2827367467149977, "grad_norm": 0.5531556233115682, "learning_rate": 5.282187608021841e-05, "loss": 0.8922, "step": 5038 }, { "epoch": 2.283189850475759, "grad_norm": 0.4764314012150944, "learning_rate": 5.280988945912682e-05, "loss": 0.8754, "step": 5039 }, { "epoch": 2.28364295423652, "grad_norm": 0.4381825916806677, "learning_rate": 5.2797901556120445e-05, "loss": 0.8503, "step": 5040 }, { "epoch": 2.2840960579972815, "grad_norm": 0.33883110126988836, "learning_rate": 5.278591237239893e-05, "loss": 0.8699, "step": 5041 }, { "epoch": 2.2845491617580427, "grad_norm": 0.316324514766024, "learning_rate": 5.277392190916205e-05, "loss": 0.8688, "step": 5042 }, { "epoch": 2.285002265518804, "grad_norm": 0.33693167058013723, "learning_rate": 5.276193016760973e-05, "loss": 0.8537, "step": 5043 }, { "epoch": 2.285455369279565, "grad_norm": 0.35908937846659306, "learning_rate": 5.274993714894201e-05, "loss": 0.8612, "step": 5044 }, { "epoch": 2.285908473040326, "grad_norm": 0.49763003872421757, "learning_rate": 5.273794285435907e-05, "loss": 0.8684, "step": 5045 }, { "epoch": 2.286361576801087, "grad_norm": 0.6208445499013954, "learning_rate": 5.2725947285061184e-05, "loss": 0.8812, "step": 5046 }, { "epoch": 2.286814680561849, "grad_norm": 0.5742346570933186, "learning_rate": 5.271395044224879e-05, "loss": 0.8739, "step": 5047 }, { "epoch": 2.28726778432261, "grad_norm": 0.4693436054167508, "learning_rate": 5.270195232712245e-05, "loss": 0.8766, "step": 5048 }, { "epoch": 2.287720888083371, "grad_norm": 0.41053361120846127, "learning_rate": 5.2689952940882816e-05, "loss": 0.856, "step": 5049 }, { "epoch": 2.288173991844132, "grad_norm": 0.46855183465262223, "learning_rate": 5.2677952284730696e-05, "loss": 0.8633, "step": 5050 }, { "epoch": 2.2886270956048937, "grad_norm": 0.5214811592181191, "learning_rate": 5.2665950359867054e-05, "loss": 0.8983, "step": 5051 }, { "epoch": 2.289080199365655, "grad_norm": 0.4578530411345687, "learning_rate": 5.265394716749292e-05, "loss": 0.8683, "step": 5052 }, { "epoch": 2.289533303126416, "grad_norm": 0.37551613333162964, "learning_rate": 5.264194270880949e-05, "loss": 0.8692, "step": 5053 }, { "epoch": 2.289986406887177, "grad_norm": 0.31475061376794816, "learning_rate": 5.262993698501808e-05, "loss": 0.8772, "step": 5054 }, { "epoch": 2.2904395106479383, "grad_norm": 0.3338477730324424, "learning_rate": 5.261792999732012e-05, "loss": 0.8684, "step": 5055 }, { "epoch": 2.2908926144086994, "grad_norm": 0.3252135462219049, "learning_rate": 5.26059217469172e-05, "loss": 0.8848, "step": 5056 }, { "epoch": 2.291345718169461, "grad_norm": 0.30788549696273826, "learning_rate": 5.2593912235010996e-05, "loss": 0.8778, "step": 5057 }, { "epoch": 2.291798821930222, "grad_norm": 0.3656175689845402, "learning_rate": 5.2581901462803334e-05, "loss": 0.8743, "step": 5058 }, { "epoch": 2.2922519256909832, "grad_norm": 0.4153414596256738, "learning_rate": 5.256988943149614e-05, "loss": 0.8899, "step": 5059 }, { "epoch": 2.2927050294517444, "grad_norm": 0.37943803827946915, "learning_rate": 5.2557876142291494e-05, "loss": 0.8561, "step": 5060 }, { "epoch": 2.2931581332125055, "grad_norm": 0.40339905363633605, "learning_rate": 5.254586159639161e-05, "loss": 0.8701, "step": 5061 }, { "epoch": 2.293611236973267, "grad_norm": 0.43399142018624176, "learning_rate": 5.2533845794998803e-05, "loss": 0.8781, "step": 5062 }, { "epoch": 2.294064340734028, "grad_norm": 0.345749198936694, "learning_rate": 5.252182873931551e-05, "loss": 0.9001, "step": 5063 }, { "epoch": 2.2945174444947893, "grad_norm": 0.3504459136648701, "learning_rate": 5.250981043054431e-05, "loss": 0.8566, "step": 5064 }, { "epoch": 2.2949705482555505, "grad_norm": 0.3551436973362582, "learning_rate": 5.249779086988791e-05, "loss": 0.8818, "step": 5065 }, { "epoch": 2.2954236520163116, "grad_norm": 0.3608312014558719, "learning_rate": 5.248577005854914e-05, "loss": 0.8806, "step": 5066 }, { "epoch": 2.2958767557770727, "grad_norm": 0.4217624129850842, "learning_rate": 5.247374799773093e-05, "loss": 0.87, "step": 5067 }, { "epoch": 2.2963298595378343, "grad_norm": 0.4191296111154129, "learning_rate": 5.2461724688636375e-05, "loss": 0.8903, "step": 5068 }, { "epoch": 2.2967829632985954, "grad_norm": 0.36869276819179403, "learning_rate": 5.2449700132468655e-05, "loss": 0.8777, "step": 5069 }, { "epoch": 2.2972360670593566, "grad_norm": 0.41422183912635463, "learning_rate": 5.2437674330431114e-05, "loss": 0.8609, "step": 5070 }, { "epoch": 2.2976891708201177, "grad_norm": 0.4810416615366433, "learning_rate": 5.242564728372719e-05, "loss": 0.884, "step": 5071 }, { "epoch": 2.298142274580879, "grad_norm": 0.4804035360497293, "learning_rate": 5.241361899356046e-05, "loss": 0.8826, "step": 5072 }, { "epoch": 2.2985953783416404, "grad_norm": 0.5670388187075721, "learning_rate": 5.240158946113461e-05, "loss": 0.897, "step": 5073 }, { "epoch": 2.2990484821024015, "grad_norm": 0.6900535750405051, "learning_rate": 5.238955868765349e-05, "loss": 0.8766, "step": 5074 }, { "epoch": 2.2995015858631627, "grad_norm": 0.7013097173142572, "learning_rate": 5.2377526674321034e-05, "loss": 0.8773, "step": 5075 }, { "epoch": 2.299954689623924, "grad_norm": 0.6324228047778333, "learning_rate": 5.236549342234129e-05, "loss": 0.8916, "step": 5076 }, { "epoch": 2.300407793384685, "grad_norm": 0.5862005906089041, "learning_rate": 5.235345893291849e-05, "loss": 0.8558, "step": 5077 }, { "epoch": 2.300860897145446, "grad_norm": 0.49874044978265075, "learning_rate": 5.234142320725693e-05, "loss": 0.8833, "step": 5078 }, { "epoch": 2.3013140009062076, "grad_norm": 0.350330161339743, "learning_rate": 5.232938624656106e-05, "loss": 0.8729, "step": 5079 }, { "epoch": 2.3017671046669688, "grad_norm": 0.3168861476047358, "learning_rate": 5.231734805203545e-05, "loss": 0.8735, "step": 5080 }, { "epoch": 2.30222020842773, "grad_norm": 0.41468889894501415, "learning_rate": 5.230530862488477e-05, "loss": 0.882, "step": 5081 }, { "epoch": 2.302673312188491, "grad_norm": 0.4650421228396974, "learning_rate": 5.2293267966313874e-05, "loss": 0.8794, "step": 5082 }, { "epoch": 2.3031264159492526, "grad_norm": 0.5055808406864533, "learning_rate": 5.2281226077527646e-05, "loss": 0.8827, "step": 5083 }, { "epoch": 2.3035795197100137, "grad_norm": 0.5892392608456417, "learning_rate": 5.226918295973118e-05, "loss": 0.8797, "step": 5084 }, { "epoch": 2.304032623470775, "grad_norm": 0.6964053845060285, "learning_rate": 5.225713861412965e-05, "loss": 0.8621, "step": 5085 }, { "epoch": 2.304485727231536, "grad_norm": 0.733703450255958, "learning_rate": 5.224509304192836e-05, "loss": 0.8691, "step": 5086 }, { "epoch": 2.304938830992297, "grad_norm": 0.7262519289854132, "learning_rate": 5.223304624433273e-05, "loss": 0.8692, "step": 5087 }, { "epoch": 2.3053919347530583, "grad_norm": 0.7809623329674079, "learning_rate": 5.222099822254833e-05, "loss": 0.8792, "step": 5088 }, { "epoch": 2.30584503851382, "grad_norm": 0.8086883737029797, "learning_rate": 5.220894897778081e-05, "loss": 0.8877, "step": 5089 }, { "epoch": 2.306298142274581, "grad_norm": 0.8278363107007248, "learning_rate": 5.2196898511235974e-05, "loss": 0.8699, "step": 5090 }, { "epoch": 2.306751246035342, "grad_norm": 0.8120385696176831, "learning_rate": 5.2184846824119745e-05, "loss": 0.8938, "step": 5091 }, { "epoch": 2.3072043497961032, "grad_norm": 0.7035626604694873, "learning_rate": 5.2172793917638175e-05, "loss": 0.8837, "step": 5092 }, { "epoch": 2.3076574535568644, "grad_norm": 0.3929025700885302, "learning_rate": 5.21607397929974e-05, "loss": 0.872, "step": 5093 }, { "epoch": 2.308110557317626, "grad_norm": 0.45015074562125706, "learning_rate": 5.2148684451403706e-05, "loss": 0.8823, "step": 5094 }, { "epoch": 2.308563661078387, "grad_norm": 0.4047908923229847, "learning_rate": 5.2136627894063515e-05, "loss": 0.8766, "step": 5095 }, { "epoch": 2.309016764839148, "grad_norm": 0.43425375476619477, "learning_rate": 5.212457012218334e-05, "loss": 0.8846, "step": 5096 }, { "epoch": 2.3094698685999093, "grad_norm": 0.47880976194028146, "learning_rate": 5.211251113696984e-05, "loss": 0.8699, "step": 5097 }, { "epoch": 2.3099229723606705, "grad_norm": 0.4965892625745924, "learning_rate": 5.210045093962977e-05, "loss": 0.8639, "step": 5098 }, { "epoch": 2.3103760761214316, "grad_norm": 0.5857380748825316, "learning_rate": 5.2088389531370044e-05, "loss": 0.8577, "step": 5099 }, { "epoch": 2.310829179882193, "grad_norm": 0.636591151771424, "learning_rate": 5.2076326913397666e-05, "loss": 0.8746, "step": 5100 }, { "epoch": 2.3112822836429543, "grad_norm": 0.5242305588301875, "learning_rate": 5.206426308691976e-05, "loss": 0.8858, "step": 5101 }, { "epoch": 2.3117353874037154, "grad_norm": 0.4581766129374588, "learning_rate": 5.205219805314359e-05, "loss": 0.8745, "step": 5102 }, { "epoch": 2.3121884911644766, "grad_norm": 0.5187953877620788, "learning_rate": 5.204013181327652e-05, "loss": 0.8672, "step": 5103 }, { "epoch": 2.3126415949252377, "grad_norm": 0.5581493502961125, "learning_rate": 5.202806436852606e-05, "loss": 0.8848, "step": 5104 }, { "epoch": 2.3130946986859993, "grad_norm": 0.5561480752687848, "learning_rate": 5.201599572009982e-05, "loss": 0.8717, "step": 5105 }, { "epoch": 2.3135478024467604, "grad_norm": 0.48150611433640117, "learning_rate": 5.200392586920553e-05, "loss": 0.8683, "step": 5106 }, { "epoch": 2.3140009062075215, "grad_norm": 0.3412361336884979, "learning_rate": 5.1991854817051064e-05, "loss": 0.8798, "step": 5107 }, { "epoch": 2.3144540099682827, "grad_norm": 0.3284017449475696, "learning_rate": 5.197978256484439e-05, "loss": 0.8718, "step": 5108 }, { "epoch": 2.314907113729044, "grad_norm": 0.35922292941004325, "learning_rate": 5.1967709113793617e-05, "loss": 0.8817, "step": 5109 }, { "epoch": 2.315360217489805, "grad_norm": 0.3706381664719294, "learning_rate": 5.195563446510694e-05, "loss": 0.9002, "step": 5110 }, { "epoch": 2.3158133212505665, "grad_norm": 0.39700232663745555, "learning_rate": 5.194355861999271e-05, "loss": 0.8804, "step": 5111 }, { "epoch": 2.3162664250113276, "grad_norm": 0.44759232733113696, "learning_rate": 5.1931481579659385e-05, "loss": 0.8809, "step": 5112 }, { "epoch": 2.3167195287720888, "grad_norm": 0.44443831175430804, "learning_rate": 5.191940334531554e-05, "loss": 0.901, "step": 5113 }, { "epoch": 2.31717263253285, "grad_norm": 0.4131023829396727, "learning_rate": 5.190732391816988e-05, "loss": 0.9031, "step": 5114 }, { "epoch": 2.3176257362936115, "grad_norm": 0.38722291087791827, "learning_rate": 5.189524329943121e-05, "loss": 0.9032, "step": 5115 }, { "epoch": 2.3180788400543726, "grad_norm": 0.4282172206327747, "learning_rate": 5.1883161490308455e-05, "loss": 0.8833, "step": 5116 }, { "epoch": 2.3185319438151337, "grad_norm": 0.4397221379411558, "learning_rate": 5.1871078492010694e-05, "loss": 0.8695, "step": 5117 }, { "epoch": 2.318985047575895, "grad_norm": 0.4699028513803115, "learning_rate": 5.185899430574708e-05, "loss": 0.864, "step": 5118 }, { "epoch": 2.319438151336656, "grad_norm": 0.43968626474829703, "learning_rate": 5.184690893272693e-05, "loss": 0.8567, "step": 5119 }, { "epoch": 2.319891255097417, "grad_norm": 0.5457114618188823, "learning_rate": 5.1834822374159614e-05, "loss": 0.8981, "step": 5120 }, { "epoch": 2.3203443588581787, "grad_norm": 0.688384948661717, "learning_rate": 5.182273463125469e-05, "loss": 0.8883, "step": 5121 }, { "epoch": 2.32079746261894, "grad_norm": 0.6669839018662252, "learning_rate": 5.181064570522181e-05, "loss": 0.8819, "step": 5122 }, { "epoch": 2.321250566379701, "grad_norm": 0.6018980936274555, "learning_rate": 5.179855559727072e-05, "loss": 0.8692, "step": 5123 }, { "epoch": 2.321703670140462, "grad_norm": 0.5554278110032214, "learning_rate": 5.178646430861133e-05, "loss": 0.8768, "step": 5124 }, { "epoch": 2.322156773901223, "grad_norm": 0.5937194775703984, "learning_rate": 5.177437184045362e-05, "loss": 0.8646, "step": 5125 }, { "epoch": 2.322609877661985, "grad_norm": 0.6692864774632815, "learning_rate": 5.1762278194007723e-05, "loss": 0.8552, "step": 5126 }, { "epoch": 2.323062981422746, "grad_norm": 0.6538111841881292, "learning_rate": 5.1750183370483874e-05, "loss": 0.8865, "step": 5127 }, { "epoch": 2.323516085183507, "grad_norm": 0.565619780550321, "learning_rate": 5.173808737109243e-05, "loss": 0.8559, "step": 5128 }, { "epoch": 2.323969188944268, "grad_norm": 0.43335543470153076, "learning_rate": 5.1725990197043863e-05, "loss": 0.8755, "step": 5129 }, { "epoch": 2.3244222927050293, "grad_norm": 0.37427606446186396, "learning_rate": 5.1713891849548766e-05, "loss": 0.8719, "step": 5130 }, { "epoch": 2.3248753964657904, "grad_norm": 0.32842228757236946, "learning_rate": 5.170179232981786e-05, "loss": 0.8669, "step": 5131 }, { "epoch": 2.325328500226552, "grad_norm": 0.31470583205750435, "learning_rate": 5.168969163906195e-05, "loss": 0.8633, "step": 5132 }, { "epoch": 2.325781603987313, "grad_norm": 0.3486252240337019, "learning_rate": 5.167758977849199e-05, "loss": 0.8744, "step": 5133 }, { "epoch": 2.3262347077480743, "grad_norm": 0.363841231684572, "learning_rate": 5.1665486749319034e-05, "loss": 0.8604, "step": 5134 }, { "epoch": 2.3266878115088354, "grad_norm": 0.3507154855835516, "learning_rate": 5.165338255275428e-05, "loss": 0.8903, "step": 5135 }, { "epoch": 2.3271409152695965, "grad_norm": 0.36354830288645645, "learning_rate": 5.1641277190009006e-05, "loss": 0.8722, "step": 5136 }, { "epoch": 2.327594019030358, "grad_norm": 0.395001835418837, "learning_rate": 5.1629170662294616e-05, "loss": 0.878, "step": 5137 }, { "epoch": 2.3280471227911192, "grad_norm": 0.36990824837288533, "learning_rate": 5.161706297082266e-05, "loss": 0.854, "step": 5138 }, { "epoch": 2.3285002265518804, "grad_norm": 0.4019503713832862, "learning_rate": 5.160495411680476e-05, "loss": 0.8934, "step": 5139 }, { "epoch": 2.3289533303126415, "grad_norm": 0.33443816444815344, "learning_rate": 5.15928441014527e-05, "loss": 0.8614, "step": 5140 }, { "epoch": 2.3294064340734026, "grad_norm": 0.3802786406184352, "learning_rate": 5.1580732925978335e-05, "loss": 0.8572, "step": 5141 }, { "epoch": 2.3298595378341638, "grad_norm": 0.3852677850346242, "learning_rate": 5.156862059159366e-05, "loss": 0.8509, "step": 5142 }, { "epoch": 2.3303126415949253, "grad_norm": 0.35719076683939815, "learning_rate": 5.155650709951081e-05, "loss": 0.8801, "step": 5143 }, { "epoch": 2.3307657453556865, "grad_norm": 0.3736290222339697, "learning_rate": 5.1544392450941976e-05, "loss": 0.8672, "step": 5144 }, { "epoch": 2.3312188491164476, "grad_norm": 0.3319490407351445, "learning_rate": 5.153227664709952e-05, "loss": 0.8568, "step": 5145 }, { "epoch": 2.3316719528772087, "grad_norm": 0.3729642372355619, "learning_rate": 5.152015968919589e-05, "loss": 0.8897, "step": 5146 }, { "epoch": 2.3321250566379703, "grad_norm": 0.3922852072926206, "learning_rate": 5.1508041578443646e-05, "loss": 0.8604, "step": 5147 }, { "epoch": 2.3325781603987314, "grad_norm": 0.5089856679581778, "learning_rate": 5.1495922316055494e-05, "loss": 0.8809, "step": 5148 }, { "epoch": 2.3330312641594926, "grad_norm": 0.513434016499116, "learning_rate": 5.1483801903244235e-05, "loss": 0.8978, "step": 5149 }, { "epoch": 2.3334843679202537, "grad_norm": 0.4679343240920314, "learning_rate": 5.147168034122277e-05, "loss": 0.8678, "step": 5150 }, { "epoch": 2.333937471681015, "grad_norm": 0.49431350966056226, "learning_rate": 5.145955763120415e-05, "loss": 0.8784, "step": 5151 }, { "epoch": 2.334390575441776, "grad_norm": 0.49183580991792974, "learning_rate": 5.144743377440151e-05, "loss": 0.895, "step": 5152 }, { "epoch": 2.3348436792025375, "grad_norm": 0.4160992879545229, "learning_rate": 5.1435308772028126e-05, "loss": 0.8633, "step": 5153 }, { "epoch": 2.3352967829632987, "grad_norm": 0.4397870093959188, "learning_rate": 5.142318262529736e-05, "loss": 0.875, "step": 5154 }, { "epoch": 2.33574988672406, "grad_norm": 0.511227082276664, "learning_rate": 5.14110553354227e-05, "loss": 0.8819, "step": 5155 }, { "epoch": 2.336202990484821, "grad_norm": 0.4631584497775334, "learning_rate": 5.139892690361777e-05, "loss": 0.8813, "step": 5156 }, { "epoch": 2.336656094245582, "grad_norm": 0.41875995547221045, "learning_rate": 5.138679733109627e-05, "loss": 0.8714, "step": 5157 }, { "epoch": 2.3371091980063436, "grad_norm": 0.5021191655979288, "learning_rate": 5.137466661907204e-05, "loss": 0.8801, "step": 5158 }, { "epoch": 2.3375623017671048, "grad_norm": 0.6100351870713804, "learning_rate": 5.1362534768759034e-05, "loss": 0.8779, "step": 5159 }, { "epoch": 2.338015405527866, "grad_norm": 0.6163695273960207, "learning_rate": 5.135040178137131e-05, "loss": 0.8739, "step": 5160 }, { "epoch": 2.338468509288627, "grad_norm": 0.6802371462082188, "learning_rate": 5.133826765812306e-05, "loss": 0.8879, "step": 5161 }, { "epoch": 2.338921613049388, "grad_norm": 0.6551277988626176, "learning_rate": 5.132613240022855e-05, "loss": 0.8929, "step": 5162 }, { "epoch": 2.3393747168101493, "grad_norm": 0.553233539400175, "learning_rate": 5.131399600890218e-05, "loss": 0.8968, "step": 5163 }, { "epoch": 2.339827820570911, "grad_norm": 0.4400766772831694, "learning_rate": 5.1301858485358474e-05, "loss": 0.878, "step": 5164 }, { "epoch": 2.340280924331672, "grad_norm": 0.40345772663384627, "learning_rate": 5.128971983081208e-05, "loss": 0.8749, "step": 5165 }, { "epoch": 2.340734028092433, "grad_norm": 0.3795833976045401, "learning_rate": 5.127758004647772e-05, "loss": 0.8647, "step": 5166 }, { "epoch": 2.3411871318531943, "grad_norm": 0.389473283841231, "learning_rate": 5.126543913357025e-05, "loss": 0.8824, "step": 5167 }, { "epoch": 2.341640235613956, "grad_norm": 0.38997542966948034, "learning_rate": 5.125329709330464e-05, "loss": 0.8798, "step": 5168 }, { "epoch": 2.342093339374717, "grad_norm": 0.4483875221337762, "learning_rate": 5.124115392689598e-05, "loss": 0.8643, "step": 5169 }, { "epoch": 2.342546443135478, "grad_norm": 0.4551331060835129, "learning_rate": 5.122900963555948e-05, "loss": 0.8682, "step": 5170 }, { "epoch": 2.3429995468962392, "grad_norm": 0.4403773349105869, "learning_rate": 5.121686422051042e-05, "loss": 0.8629, "step": 5171 }, { "epoch": 2.3434526506570004, "grad_norm": 0.38694835752189355, "learning_rate": 5.120471768296421e-05, "loss": 0.8814, "step": 5172 }, { "epoch": 2.3439057544177615, "grad_norm": 0.4125183030145876, "learning_rate": 5.119257002413641e-05, "loss": 0.8943, "step": 5173 }, { "epoch": 2.344358858178523, "grad_norm": 0.455468175100084, "learning_rate": 5.118042124524264e-05, "loss": 0.8732, "step": 5174 }, { "epoch": 2.344811961939284, "grad_norm": 0.5115885362225197, "learning_rate": 5.116827134749867e-05, "loss": 0.8639, "step": 5175 }, { "epoch": 2.3452650657000453, "grad_norm": 0.5869768176092565, "learning_rate": 5.115612033212037e-05, "loss": 0.8821, "step": 5176 }, { "epoch": 2.3457181694608065, "grad_norm": 0.5790036859514094, "learning_rate": 5.1143968200323715e-05, "loss": 0.8582, "step": 5177 }, { "epoch": 2.3461712732215676, "grad_norm": 0.7228063055208369, "learning_rate": 5.1131814953324797e-05, "loss": 0.8835, "step": 5178 }, { "epoch": 2.346624376982329, "grad_norm": 0.4635358249545105, "learning_rate": 5.1119660592339825e-05, "loss": 0.8734, "step": 5179 }, { "epoch": 2.3470774807430903, "grad_norm": 0.3284420186906669, "learning_rate": 5.110750511858509e-05, "loss": 0.8901, "step": 5180 }, { "epoch": 2.3475305845038514, "grad_norm": 0.35353413263602973, "learning_rate": 5.109534853327705e-05, "loss": 0.8674, "step": 5181 }, { "epoch": 2.3479836882646126, "grad_norm": 0.44921705212868257, "learning_rate": 5.108319083763222e-05, "loss": 0.8704, "step": 5182 }, { "epoch": 2.3484367920253737, "grad_norm": 0.6470245911800296, "learning_rate": 5.107103203286726e-05, "loss": 0.8769, "step": 5183 }, { "epoch": 2.348889895786135, "grad_norm": 0.8452328891497386, "learning_rate": 5.105887212019892e-05, "loss": 0.8819, "step": 5184 }, { "epoch": 2.3493429995468964, "grad_norm": 0.9419976757945898, "learning_rate": 5.1046711100844074e-05, "loss": 0.8678, "step": 5185 }, { "epoch": 2.3497961033076575, "grad_norm": 0.9128837730425111, "learning_rate": 5.103454897601971e-05, "loss": 0.8868, "step": 5186 }, { "epoch": 2.3502492070684187, "grad_norm": 0.8961744451096468, "learning_rate": 5.102238574694291e-05, "loss": 0.8581, "step": 5187 }, { "epoch": 2.35070231082918, "grad_norm": 0.9374331946310118, "learning_rate": 5.101022141483089e-05, "loss": 0.8844, "step": 5188 }, { "epoch": 2.351155414589941, "grad_norm": 0.9737699318166317, "learning_rate": 5.0998055980900954e-05, "loss": 0.8753, "step": 5189 }, { "epoch": 2.3516085183507025, "grad_norm": 0.8700536301336104, "learning_rate": 5.098588944637051e-05, "loss": 0.8795, "step": 5190 }, { "epoch": 2.3520616221114636, "grad_norm": 0.6204874296773873, "learning_rate": 5.09737218124571e-05, "loss": 0.8744, "step": 5191 }, { "epoch": 2.3525147258722248, "grad_norm": 0.4370834560717441, "learning_rate": 5.096155308037838e-05, "loss": 0.8678, "step": 5192 }, { "epoch": 2.352967829632986, "grad_norm": 0.3769077844764583, "learning_rate": 5.094938325135209e-05, "loss": 0.8693, "step": 5193 }, { "epoch": 2.353420933393747, "grad_norm": 0.4860639642737307, "learning_rate": 5.093721232659609e-05, "loss": 0.8852, "step": 5194 }, { "epoch": 2.353874037154508, "grad_norm": 0.561071237604667, "learning_rate": 5.092504030732836e-05, "loss": 0.8687, "step": 5195 }, { "epoch": 2.3543271409152697, "grad_norm": 0.5869538653290851, "learning_rate": 5.0912867194766995e-05, "loss": 0.8844, "step": 5196 }, { "epoch": 2.354780244676031, "grad_norm": 0.6062350526557133, "learning_rate": 5.0900692990130155e-05, "loss": 0.8878, "step": 5197 }, { "epoch": 2.355233348436792, "grad_norm": 0.6804734516517165, "learning_rate": 5.088851769463616e-05, "loss": 0.8895, "step": 5198 }, { "epoch": 2.355686452197553, "grad_norm": 0.5686063330348078, "learning_rate": 5.087634130950341e-05, "loss": 0.8724, "step": 5199 }, { "epoch": 2.3561395559583147, "grad_norm": 0.4481820444698741, "learning_rate": 5.086416383595044e-05, "loss": 0.8833, "step": 5200 }, { "epoch": 2.356592659719076, "grad_norm": 0.37625354799717703, "learning_rate": 5.085198527519586e-05, "loss": 0.8601, "step": 5201 }, { "epoch": 2.357045763479837, "grad_norm": 0.3064688064281122, "learning_rate": 5.083980562845841e-05, "loss": 0.8595, "step": 5202 }, { "epoch": 2.357498867240598, "grad_norm": 0.3034696272957792, "learning_rate": 5.082762489695694e-05, "loss": 0.8803, "step": 5203 }, { "epoch": 2.3579519710013592, "grad_norm": 0.3053240409211632, "learning_rate": 5.0815443081910396e-05, "loss": 0.8664, "step": 5204 }, { "epoch": 2.3584050747621204, "grad_norm": 0.3734450764168325, "learning_rate": 5.080326018453785e-05, "loss": 0.8597, "step": 5205 }, { "epoch": 2.358858178522882, "grad_norm": 0.5743859658178921, "learning_rate": 5.0791076206058474e-05, "loss": 0.8623, "step": 5206 }, { "epoch": 2.359311282283643, "grad_norm": 0.4181639749671156, "learning_rate": 5.0778891147691536e-05, "loss": 0.9118, "step": 5207 }, { "epoch": 2.359764386044404, "grad_norm": 15.777786376722842, "learning_rate": 5.0766705010656416e-05, "loss": 0.9405, "step": 5208 }, { "epoch": 2.3602174898051653, "grad_norm": 0.9268248929205486, "learning_rate": 5.075451779617263e-05, "loss": 0.8635, "step": 5209 }, { "epoch": 2.3606705935659265, "grad_norm": 1.3628553073885221, "learning_rate": 5.074232950545976e-05, "loss": 0.8698, "step": 5210 }, { "epoch": 2.361123697326688, "grad_norm": 0.6222715561830081, "learning_rate": 5.073014013973753e-05, "loss": 0.8677, "step": 5211 }, { "epoch": 2.361576801087449, "grad_norm": 1.1588323489070362, "learning_rate": 5.071794970022575e-05, "loss": 0.8749, "step": 5212 }, { "epoch": 2.3620299048482103, "grad_norm": 1.037129936116999, "learning_rate": 5.070575818814435e-05, "loss": 0.8766, "step": 5213 }, { "epoch": 2.3624830086089714, "grad_norm": 1.1049799224647432, "learning_rate": 5.069356560471337e-05, "loss": 0.8633, "step": 5214 }, { "epoch": 2.3629361123697326, "grad_norm": 0.8173509345775306, "learning_rate": 5.0681371951152934e-05, "loss": 0.8891, "step": 5215 }, { "epoch": 2.3633892161304937, "grad_norm": 0.7425286974255587, "learning_rate": 5.066917722868329e-05, "loss": 0.8846, "step": 5216 }, { "epoch": 2.3638423198912553, "grad_norm": 0.8328795082597624, "learning_rate": 5.06569814385248e-05, "loss": 0.8957, "step": 5217 }, { "epoch": 2.3642954236520164, "grad_norm": 0.9506221350958476, "learning_rate": 5.0644784581897916e-05, "loss": 0.8759, "step": 5218 }, { "epoch": 2.3647485274127775, "grad_norm": 0.9503912533657678, "learning_rate": 5.0632586660023206e-05, "loss": 0.8916, "step": 5219 }, { "epoch": 2.3652016311735387, "grad_norm": 0.7360237445646192, "learning_rate": 5.0620387674121354e-05, "loss": 0.8636, "step": 5220 }, { "epoch": 2.3656547349343, "grad_norm": 0.4497813432297956, "learning_rate": 5.060818762541313e-05, "loss": 0.872, "step": 5221 }, { "epoch": 2.3661078386950614, "grad_norm": 0.7685879721727157, "learning_rate": 5.059598651511944e-05, "loss": 0.8848, "step": 5222 }, { "epoch": 2.3665609424558225, "grad_norm": 0.9694978080401918, "learning_rate": 5.058378434446124e-05, "loss": 0.8657, "step": 5223 }, { "epoch": 2.3670140462165836, "grad_norm": 0.7194033357430631, "learning_rate": 5.057158111465966e-05, "loss": 0.8761, "step": 5224 }, { "epoch": 2.3674671499773448, "grad_norm": 0.5535752897806872, "learning_rate": 5.0559376826935895e-05, "loss": 0.8724, "step": 5225 }, { "epoch": 2.367920253738106, "grad_norm": 0.7418864083214548, "learning_rate": 5.054717148251126e-05, "loss": 0.8853, "step": 5226 }, { "epoch": 2.368373357498867, "grad_norm": 0.7507402517552154, "learning_rate": 5.0534965082607146e-05, "loss": 0.8832, "step": 5227 }, { "epoch": 2.3688264612596286, "grad_norm": 0.6394510665019082, "learning_rate": 5.052275762844513e-05, "loss": 0.8693, "step": 5228 }, { "epoch": 2.3692795650203897, "grad_norm": 0.5577479440010433, "learning_rate": 5.051054912124679e-05, "loss": 0.8762, "step": 5229 }, { "epoch": 2.369732668781151, "grad_norm": 0.5151790596009119, "learning_rate": 5.049833956223388e-05, "loss": 0.8691, "step": 5230 }, { "epoch": 2.370185772541912, "grad_norm": 0.5714012715618089, "learning_rate": 5.048612895262822e-05, "loss": 0.8903, "step": 5231 }, { "epoch": 2.3706388763026736, "grad_norm": 0.5903517076677318, "learning_rate": 5.0473917293651786e-05, "loss": 0.8901, "step": 5232 }, { "epoch": 2.3710919800634347, "grad_norm": 0.443244603582095, "learning_rate": 5.0461704586526605e-05, "loss": 0.8789, "step": 5233 }, { "epoch": 2.371545083824196, "grad_norm": 0.36724388565641713, "learning_rate": 5.044949083247482e-05, "loss": 0.8835, "step": 5234 }, { "epoch": 2.371998187584957, "grad_norm": 0.5386002314858753, "learning_rate": 5.0437276032718714e-05, "loss": 0.8649, "step": 5235 }, { "epoch": 2.372451291345718, "grad_norm": 0.4966220998624612, "learning_rate": 5.0425060188480636e-05, "loss": 0.8667, "step": 5236 }, { "epoch": 2.372904395106479, "grad_norm": 0.38579218850418145, "learning_rate": 5.041284330098306e-05, "loss": 0.8777, "step": 5237 }, { "epoch": 2.373357498867241, "grad_norm": 0.45689526424202487, "learning_rate": 5.040062537144854e-05, "loss": 0.8924, "step": 5238 }, { "epoch": 2.373810602628002, "grad_norm": 0.38057587954283295, "learning_rate": 5.038840640109978e-05, "loss": 0.8824, "step": 5239 }, { "epoch": 2.374263706388763, "grad_norm": 0.39965920261567917, "learning_rate": 5.0376186391159546e-05, "loss": 0.8945, "step": 5240 }, { "epoch": 2.374716810149524, "grad_norm": 0.4280489426751433, "learning_rate": 5.036396534285072e-05, "loss": 0.881, "step": 5241 }, { "epoch": 2.3751699139102853, "grad_norm": 0.5835532867153678, "learning_rate": 5.035174325739628e-05, "loss": 0.8722, "step": 5242 }, { "epoch": 2.375623017671047, "grad_norm": 0.5746225823126639, "learning_rate": 5.033952013601934e-05, "loss": 0.8796, "step": 5243 }, { "epoch": 2.376076121431808, "grad_norm": 0.5337478763187489, "learning_rate": 5.0327295979943084e-05, "loss": 0.8612, "step": 5244 }, { "epoch": 2.376529225192569, "grad_norm": 0.5301927721509354, "learning_rate": 5.031507079039081e-05, "loss": 0.8428, "step": 5245 }, { "epoch": 2.3769823289533303, "grad_norm": 0.5298579656175734, "learning_rate": 5.0302844568585925e-05, "loss": 0.8701, "step": 5246 }, { "epoch": 2.3774354327140914, "grad_norm": 0.4374925976880921, "learning_rate": 5.029061731575193e-05, "loss": 0.863, "step": 5247 }, { "epoch": 2.3778885364748525, "grad_norm": 0.27853835386307263, "learning_rate": 5.027838903311244e-05, "loss": 0.87, "step": 5248 }, { "epoch": 2.378341640235614, "grad_norm": 0.41110337888797005, "learning_rate": 5.026615972189118e-05, "loss": 0.8454, "step": 5249 }, { "epoch": 2.3787947439963752, "grad_norm": 0.409183976674253, "learning_rate": 5.025392938331193e-05, "loss": 0.8573, "step": 5250 }, { "epoch": 2.3792478477571364, "grad_norm": 0.41207170892405914, "learning_rate": 5.024169801859863e-05, "loss": 0.8905, "step": 5251 }, { "epoch": 2.3797009515178975, "grad_norm": 0.4545853869723989, "learning_rate": 5.02294656289753e-05, "loss": 0.8865, "step": 5252 }, { "epoch": 2.3801540552786586, "grad_norm": 0.4388878856812527, "learning_rate": 5.021723221566606e-05, "loss": 0.8934, "step": 5253 }, { "epoch": 2.38060715903942, "grad_norm": 0.40396617049785305, "learning_rate": 5.020499777989514e-05, "loss": 0.8666, "step": 5254 }, { "epoch": 2.3810602628001813, "grad_norm": 0.4239176292506707, "learning_rate": 5.019276232288686e-05, "loss": 0.8716, "step": 5255 }, { "epoch": 2.3815133665609425, "grad_norm": 0.38668495854358564, "learning_rate": 5.0180525845865655e-05, "loss": 0.8903, "step": 5256 }, { "epoch": 2.3819664703217036, "grad_norm": 0.2947946524627571, "learning_rate": 5.0168288350056057e-05, "loss": 0.8773, "step": 5257 }, { "epoch": 2.3824195740824647, "grad_norm": 0.32377286567131186, "learning_rate": 5.01560498366827e-05, "loss": 0.8629, "step": 5258 }, { "epoch": 2.382872677843226, "grad_norm": 0.4067258206385915, "learning_rate": 5.0143810306970324e-05, "loss": 0.8595, "step": 5259 }, { "epoch": 2.3833257816039874, "grad_norm": 0.3783231700893364, "learning_rate": 5.013156976214375e-05, "loss": 0.8655, "step": 5260 }, { "epoch": 2.3837788853647486, "grad_norm": 0.34021766402023623, "learning_rate": 5.011932820342793e-05, "loss": 0.862, "step": 5261 }, { "epoch": 2.3842319891255097, "grad_norm": 0.3818142886757717, "learning_rate": 5.010708563204791e-05, "loss": 0.8545, "step": 5262 }, { "epoch": 2.384685092886271, "grad_norm": 0.4112520208878074, "learning_rate": 5.0094842049228825e-05, "loss": 0.8557, "step": 5263 }, { "epoch": 2.3851381966470324, "grad_norm": 0.3223594237013375, "learning_rate": 5.008259745619591e-05, "loss": 0.8723, "step": 5264 }, { "epoch": 2.3855913004077935, "grad_norm": 0.339168151341854, "learning_rate": 5.0070351854174526e-05, "loss": 0.8767, "step": 5265 }, { "epoch": 2.3860444041685547, "grad_norm": 0.35155026617814905, "learning_rate": 5.00581052443901e-05, "loss": 0.8814, "step": 5266 }, { "epoch": 2.386497507929316, "grad_norm": 0.32541345224199325, "learning_rate": 5.004585762806821e-05, "loss": 0.8764, "step": 5267 }, { "epoch": 2.386950611690077, "grad_norm": 0.3626320679535202, "learning_rate": 5.003360900643447e-05, "loss": 0.8985, "step": 5268 }, { "epoch": 2.387403715450838, "grad_norm": 0.3697206880936694, "learning_rate": 5.0021359380714635e-05, "loss": 0.8873, "step": 5269 }, { "epoch": 2.3878568192115996, "grad_norm": 0.36749851290433855, "learning_rate": 5.000910875213455e-05, "loss": 0.8833, "step": 5270 }, { "epoch": 2.3883099229723608, "grad_norm": 0.30068359258971333, "learning_rate": 4.999685712192018e-05, "loss": 0.8751, "step": 5271 }, { "epoch": 2.388763026733122, "grad_norm": 0.4460531236714641, "learning_rate": 4.998460449129757e-05, "loss": 0.8859, "step": 5272 }, { "epoch": 2.389216130493883, "grad_norm": 0.3067641175709765, "learning_rate": 4.997235086149284e-05, "loss": 0.8547, "step": 5273 }, { "epoch": 2.389669234254644, "grad_norm": 0.26475401287643435, "learning_rate": 4.996009623373228e-05, "loss": 0.9078, "step": 5274 }, { "epoch": 2.3901223380154057, "grad_norm": 0.26584694718328755, "learning_rate": 4.9947840609242216e-05, "loss": 0.8708, "step": 5275 }, { "epoch": 2.390575441776167, "grad_norm": 0.3105294325462565, "learning_rate": 4.99355839892491e-05, "loss": 0.8641, "step": 5276 }, { "epoch": 2.391028545536928, "grad_norm": 0.36614779267840336, "learning_rate": 4.9923326374979475e-05, "loss": 0.8769, "step": 5277 }, { "epoch": 2.391481649297689, "grad_norm": 0.4038395619723789, "learning_rate": 4.991106776765999e-05, "loss": 0.8591, "step": 5278 }, { "epoch": 2.3919347530584503, "grad_norm": 0.32359387172924675, "learning_rate": 4.98988081685174e-05, "loss": 0.8661, "step": 5279 }, { "epoch": 2.3923878568192114, "grad_norm": 0.3609250233097924, "learning_rate": 4.9886547578778534e-05, "loss": 0.8626, "step": 5280 }, { "epoch": 2.392840960579973, "grad_norm": 0.308916791541856, "learning_rate": 4.987428599967036e-05, "loss": 0.8609, "step": 5281 }, { "epoch": 2.393294064340734, "grad_norm": 0.27469365961149894, "learning_rate": 4.9862023432419916e-05, "loss": 0.8802, "step": 5282 }, { "epoch": 2.3937471681014952, "grad_norm": 0.3329695202324411, "learning_rate": 4.984975987825433e-05, "loss": 0.8821, "step": 5283 }, { "epoch": 2.3942002718622564, "grad_norm": 0.36239751933370307, "learning_rate": 4.983749533840086e-05, "loss": 0.8529, "step": 5284 }, { "epoch": 2.3946533756230175, "grad_norm": 0.29160689466126505, "learning_rate": 4.982522981408684e-05, "loss": 0.8628, "step": 5285 }, { "epoch": 2.395106479383779, "grad_norm": 0.2903415253810334, "learning_rate": 4.98129633065397e-05, "loss": 0.8814, "step": 5286 }, { "epoch": 2.39555958314454, "grad_norm": 0.3441600241241724, "learning_rate": 4.9800695816986986e-05, "loss": 0.8777, "step": 5287 }, { "epoch": 2.3960126869053013, "grad_norm": 0.39938225183812615, "learning_rate": 4.978842734665634e-05, "loss": 0.8855, "step": 5288 }, { "epoch": 2.3964657906660625, "grad_norm": 0.3713200257951387, "learning_rate": 4.977615789677549e-05, "loss": 0.8749, "step": 5289 }, { "epoch": 2.3969188944268236, "grad_norm": 0.3217005293677561, "learning_rate": 4.976388746857226e-05, "loss": 0.8683, "step": 5290 }, { "epoch": 2.3973719981875847, "grad_norm": 0.3280048096171072, "learning_rate": 4.975161606327459e-05, "loss": 0.866, "step": 5291 }, { "epoch": 2.3978251019483463, "grad_norm": 0.34346166542937084, "learning_rate": 4.97393436821105e-05, "loss": 0.8806, "step": 5292 }, { "epoch": 2.3982782057091074, "grad_norm": 0.4128363057611224, "learning_rate": 4.972707032630814e-05, "loss": 0.8831, "step": 5293 }, { "epoch": 2.3987313094698686, "grad_norm": 0.46981627605669335, "learning_rate": 4.971479599709569e-05, "loss": 0.8576, "step": 5294 }, { "epoch": 2.3991844132306297, "grad_norm": 0.5016892890375245, "learning_rate": 4.9702520695701494e-05, "loss": 0.8569, "step": 5295 }, { "epoch": 2.3996375169913913, "grad_norm": 0.4796065358213695, "learning_rate": 4.969024442335397e-05, "loss": 0.9022, "step": 5296 }, { "epoch": 2.4000906207521524, "grad_norm": 0.40216934951279315, "learning_rate": 4.9677967181281636e-05, "loss": 0.8821, "step": 5297 }, { "epoch": 2.4005437245129135, "grad_norm": 0.3447385246254171, "learning_rate": 4.9665688970713095e-05, "loss": 0.8767, "step": 5298 }, { "epoch": 2.4009968282736747, "grad_norm": 0.3615546713218113, "learning_rate": 4.965340979287706e-05, "loss": 0.8727, "step": 5299 }, { "epoch": 2.401449932034436, "grad_norm": 0.40880251401759715, "learning_rate": 4.9641129649002323e-05, "loss": 0.8722, "step": 5300 }, { "epoch": 2.401903035795197, "grad_norm": 0.42615143914717823, "learning_rate": 4.962884854031782e-05, "loss": 0.8681, "step": 5301 }, { "epoch": 2.4023561395559585, "grad_norm": 0.44213079633071733, "learning_rate": 4.961656646805251e-05, "loss": 0.8805, "step": 5302 }, { "epoch": 2.4028092433167196, "grad_norm": 0.42554963668181345, "learning_rate": 4.960428343343551e-05, "loss": 0.8743, "step": 5303 }, { "epoch": 2.4032623470774808, "grad_norm": 0.3981188786837031, "learning_rate": 4.9591999437695994e-05, "loss": 0.8818, "step": 5304 }, { "epoch": 2.403715450838242, "grad_norm": 0.4309348649368985, "learning_rate": 4.957971448206327e-05, "loss": 0.8707, "step": 5305 }, { "epoch": 2.404168554599003, "grad_norm": 0.3428508981705459, "learning_rate": 4.956742856776671e-05, "loss": 0.8664, "step": 5306 }, { "epoch": 2.4046216583597646, "grad_norm": 0.3580955412091772, "learning_rate": 4.955514169603581e-05, "loss": 0.8681, "step": 5307 }, { "epoch": 2.4050747621205257, "grad_norm": 0.31385173174256786, "learning_rate": 4.954285386810011e-05, "loss": 0.8927, "step": 5308 }, { "epoch": 2.405527865881287, "grad_norm": 0.27727787032903156, "learning_rate": 4.953056508518932e-05, "loss": 0.8532, "step": 5309 }, { "epoch": 2.405980969642048, "grad_norm": 0.3156904075129117, "learning_rate": 4.951827534853318e-05, "loss": 0.8956, "step": 5310 }, { "epoch": 2.406434073402809, "grad_norm": 0.3193776359648529, "learning_rate": 4.950598465936157e-05, "loss": 0.8695, "step": 5311 }, { "epoch": 2.4068871771635703, "grad_norm": 0.31030060362586215, "learning_rate": 4.9493693018904425e-05, "loss": 0.8681, "step": 5312 }, { "epoch": 2.407340280924332, "grad_norm": 0.34301918643418455, "learning_rate": 4.9481400428391816e-05, "loss": 0.8926, "step": 5313 }, { "epoch": 2.407793384685093, "grad_norm": 0.3386982499662249, "learning_rate": 4.946910688905389e-05, "loss": 0.8721, "step": 5314 }, { "epoch": 2.408246488445854, "grad_norm": 0.36710957296239377, "learning_rate": 4.9456812402120886e-05, "loss": 0.879, "step": 5315 }, { "epoch": 2.4086995922066152, "grad_norm": 0.37840055603583655, "learning_rate": 4.9444516968823133e-05, "loss": 0.8878, "step": 5316 }, { "epoch": 2.4091526959673764, "grad_norm": 0.40565958159639204, "learning_rate": 4.943222059039107e-05, "loss": 0.8837, "step": 5317 }, { "epoch": 2.409605799728138, "grad_norm": 0.47209440042773976, "learning_rate": 4.941992326805522e-05, "loss": 0.8943, "step": 5318 }, { "epoch": 2.410058903488899, "grad_norm": 0.5790270216252357, "learning_rate": 4.9407625003046215e-05, "loss": 0.8767, "step": 5319 }, { "epoch": 2.41051200724966, "grad_norm": 0.618781477034772, "learning_rate": 4.9395325796594764e-05, "loss": 0.8795, "step": 5320 }, { "epoch": 2.4109651110104213, "grad_norm": 0.5981547278762459, "learning_rate": 4.938302564993167e-05, "loss": 0.862, "step": 5321 }, { "epoch": 2.4114182147711825, "grad_norm": 0.561064423446561, "learning_rate": 4.937072456428784e-05, "loss": 0.8875, "step": 5322 }, { "epoch": 2.4118713185319436, "grad_norm": 0.5547348743038152, "learning_rate": 4.935842254089427e-05, "loss": 0.8695, "step": 5323 }, { "epoch": 2.412324422292705, "grad_norm": 0.5647159665312051, "learning_rate": 4.934611958098206e-05, "loss": 0.8722, "step": 5324 }, { "epoch": 2.4127775260534663, "grad_norm": 0.5792631355727218, "learning_rate": 4.933381568578239e-05, "loss": 0.8756, "step": 5325 }, { "epoch": 2.4132306298142274, "grad_norm": 0.5325886272905261, "learning_rate": 4.932151085652653e-05, "loss": 0.8618, "step": 5326 }, { "epoch": 2.4136837335749886, "grad_norm": 0.3916974915682354, "learning_rate": 4.930920509444588e-05, "loss": 0.8799, "step": 5327 }, { "epoch": 2.41413683733575, "grad_norm": 0.2746975152812174, "learning_rate": 4.9296898400771884e-05, "loss": 0.8915, "step": 5328 }, { "epoch": 2.4145899410965113, "grad_norm": 0.30086245383832205, "learning_rate": 4.92845907767361e-05, "loss": 0.8714, "step": 5329 }, { "epoch": 2.4150430448572724, "grad_norm": 0.3329705806517024, "learning_rate": 4.927228222357018e-05, "loss": 0.8744, "step": 5330 }, { "epoch": 2.4154961486180335, "grad_norm": 0.39306961149826836, "learning_rate": 4.925997274250588e-05, "loss": 0.8724, "step": 5331 }, { "epoch": 2.4159492523787947, "grad_norm": 0.4462164559345308, "learning_rate": 4.9247662334775016e-05, "loss": 0.8652, "step": 5332 }, { "epoch": 2.416402356139556, "grad_norm": 0.4433368842067668, "learning_rate": 4.923535100160954e-05, "loss": 0.8908, "step": 5333 }, { "epoch": 2.4168554599003174, "grad_norm": 0.4736387896469187, "learning_rate": 4.922303874424147e-05, "loss": 0.8768, "step": 5334 }, { "epoch": 2.4173085636610785, "grad_norm": 0.4543243764529906, "learning_rate": 4.921072556390292e-05, "loss": 0.8626, "step": 5335 }, { "epoch": 2.4177616674218396, "grad_norm": 0.7399424624652179, "learning_rate": 4.91984114618261e-05, "loss": 0.9328, "step": 5336 }, { "epoch": 2.4182147711826008, "grad_norm": 0.5796138858051824, "learning_rate": 4.918609643924331e-05, "loss": 0.8702, "step": 5337 }, { "epoch": 2.418667874943362, "grad_norm": 0.6061520596083118, "learning_rate": 4.917378049738694e-05, "loss": 0.8611, "step": 5338 }, { "epoch": 2.4191209787041235, "grad_norm": 0.48175521781225367, "learning_rate": 4.916146363748946e-05, "loss": 0.8692, "step": 5339 }, { "epoch": 2.4195740824648846, "grad_norm": 0.32144715475738517, "learning_rate": 4.914914586078347e-05, "loss": 0.8684, "step": 5340 }, { "epoch": 2.4200271862256457, "grad_norm": 0.3173132563823196, "learning_rate": 4.9136827168501634e-05, "loss": 0.8859, "step": 5341 }, { "epoch": 2.420480289986407, "grad_norm": 0.311399583645999, "learning_rate": 4.912450756187671e-05, "loss": 0.8726, "step": 5342 }, { "epoch": 2.420933393747168, "grad_norm": 0.3107303828145189, "learning_rate": 4.911218704214154e-05, "loss": 0.883, "step": 5343 }, { "epoch": 2.421386497507929, "grad_norm": 0.31693093613048373, "learning_rate": 4.909986561052907e-05, "loss": 0.876, "step": 5344 }, { "epoch": 2.4218396012686907, "grad_norm": 0.28126886067615214, "learning_rate": 4.9087543268272334e-05, "loss": 0.8638, "step": 5345 }, { "epoch": 2.422292705029452, "grad_norm": 0.3288570533651426, "learning_rate": 4.9075220016604463e-05, "loss": 0.8753, "step": 5346 }, { "epoch": 2.422745808790213, "grad_norm": 0.3664462895557663, "learning_rate": 4.906289585675867e-05, "loss": 0.8802, "step": 5347 }, { "epoch": 2.423198912550974, "grad_norm": 0.5098897911492416, "learning_rate": 4.9050570789968255e-05, "loss": 0.8553, "step": 5348 }, { "epoch": 2.423652016311735, "grad_norm": 0.5599860824691859, "learning_rate": 4.903824481746662e-05, "loss": 0.8795, "step": 5349 }, { "epoch": 2.424105120072497, "grad_norm": 0.5637907051930915, "learning_rate": 4.902591794048725e-05, "loss": 0.8852, "step": 5350 }, { "epoch": 2.424558223833258, "grad_norm": 0.5714353779028915, "learning_rate": 4.9013590160263737e-05, "loss": 0.8731, "step": 5351 }, { "epoch": 2.425011327594019, "grad_norm": 0.5693886757660828, "learning_rate": 4.900126147802973e-05, "loss": 0.8585, "step": 5352 }, { "epoch": 2.42546443135478, "grad_norm": 0.6298055495144953, "learning_rate": 4.8988931895019e-05, "loss": 0.8431, "step": 5353 }, { "epoch": 2.4259175351155413, "grad_norm": 0.6751217680548255, "learning_rate": 4.8976601412465415e-05, "loss": 0.8528, "step": 5354 }, { "epoch": 2.4263706388763024, "grad_norm": 0.6250353147067769, "learning_rate": 4.896427003160287e-05, "loss": 0.8804, "step": 5355 }, { "epoch": 2.426823742637064, "grad_norm": 0.5891490637596136, "learning_rate": 4.895193775366543e-05, "loss": 0.848, "step": 5356 }, { "epoch": 2.427276846397825, "grad_norm": 0.5111241994236472, "learning_rate": 4.89396045798872e-05, "loss": 0.9004, "step": 5357 }, { "epoch": 2.4277299501585863, "grad_norm": 0.4768559921387701, "learning_rate": 4.8927270511502386e-05, "loss": 0.8909, "step": 5358 }, { "epoch": 2.4281830539193474, "grad_norm": 0.4854995972529617, "learning_rate": 4.8914935549745305e-05, "loss": 0.8831, "step": 5359 }, { "epoch": 2.428636157680109, "grad_norm": 0.5035746072953265, "learning_rate": 4.890259969585032e-05, "loss": 0.8598, "step": 5360 }, { "epoch": 2.42908926144087, "grad_norm": 0.3923763341570211, "learning_rate": 4.8890262951051925e-05, "loss": 0.8726, "step": 5361 }, { "epoch": 2.4295423652016312, "grad_norm": 0.39514369528424553, "learning_rate": 4.8877925316584686e-05, "loss": 0.8814, "step": 5362 }, { "epoch": 2.4299954689623924, "grad_norm": 0.4491254891606789, "learning_rate": 4.8865586793683245e-05, "loss": 0.8684, "step": 5363 }, { "epoch": 2.4304485727231535, "grad_norm": 0.4473655345900605, "learning_rate": 4.8853247383582364e-05, "loss": 0.8833, "step": 5364 }, { "epoch": 2.4309016764839146, "grad_norm": 0.4112319768098744, "learning_rate": 4.8840907087516855e-05, "loss": 0.8739, "step": 5365 }, { "epoch": 2.431354780244676, "grad_norm": 0.4280771976472455, "learning_rate": 4.882856590672165e-05, "loss": 0.8781, "step": 5366 }, { "epoch": 2.4318078840054373, "grad_norm": 0.41615150517154864, "learning_rate": 4.881622384243177e-05, "loss": 0.8879, "step": 5367 }, { "epoch": 2.4322609877661985, "grad_norm": 0.5061389291375664, "learning_rate": 4.880388089588229e-05, "loss": 0.8872, "step": 5368 }, { "epoch": 2.4327140915269596, "grad_norm": 0.6106837002203067, "learning_rate": 4.879153706830842e-05, "loss": 0.8768, "step": 5369 }, { "epoch": 2.4331671952877207, "grad_norm": 0.6844168951676151, "learning_rate": 4.877919236094542e-05, "loss": 0.8686, "step": 5370 }, { "epoch": 2.4336202990484823, "grad_norm": 0.6190834677909006, "learning_rate": 4.8766846775028656e-05, "loss": 0.8559, "step": 5371 }, { "epoch": 2.4340734028092434, "grad_norm": 0.5567855628484362, "learning_rate": 4.8754500311793596e-05, "loss": 0.8623, "step": 5372 }, { "epoch": 2.4345265065700046, "grad_norm": 0.6014829177609532, "learning_rate": 4.8742152972475746e-05, "loss": 0.8737, "step": 5373 }, { "epoch": 2.4349796103307657, "grad_norm": 0.6195449956482829, "learning_rate": 4.872980475831075e-05, "loss": 0.8911, "step": 5374 }, { "epoch": 2.435432714091527, "grad_norm": 0.5374049104522158, "learning_rate": 4.8717455670534326e-05, "loss": 0.8816, "step": 5375 }, { "epoch": 2.435885817852288, "grad_norm": 0.5288675713568143, "learning_rate": 4.870510571038227e-05, "loss": 0.8985, "step": 5376 }, { "epoch": 2.4363389216130495, "grad_norm": 0.5193789978050866, "learning_rate": 4.869275487909047e-05, "loss": 0.8726, "step": 5377 }, { "epoch": 2.4367920253738107, "grad_norm": 0.4505284108942441, "learning_rate": 4.868040317789491e-05, "loss": 0.8872, "step": 5378 }, { "epoch": 2.437245129134572, "grad_norm": 0.34364235695969303, "learning_rate": 4.866805060803163e-05, "loss": 0.8909, "step": 5379 }, { "epoch": 2.437698232895333, "grad_norm": 0.3317295675425565, "learning_rate": 4.865569717073682e-05, "loss": 0.8937, "step": 5380 }, { "epoch": 2.438151336656094, "grad_norm": 0.3684416986801113, "learning_rate": 4.864334286724668e-05, "loss": 0.8581, "step": 5381 }, { "epoch": 2.4386044404168556, "grad_norm": 0.4132351678903935, "learning_rate": 4.863098769879754e-05, "loss": 0.8855, "step": 5382 }, { "epoch": 2.4390575441776168, "grad_norm": 0.5203524689095851, "learning_rate": 4.861863166662581e-05, "loss": 0.874, "step": 5383 }, { "epoch": 2.439510647938378, "grad_norm": 0.6447524659597162, "learning_rate": 4.8606274771968005e-05, "loss": 0.8826, "step": 5384 }, { "epoch": 2.439963751699139, "grad_norm": 0.7918075208413734, "learning_rate": 4.859391701606069e-05, "loss": 0.8807, "step": 5385 }, { "epoch": 2.4404168554599, "grad_norm": 0.8387962462576661, "learning_rate": 4.858155840014054e-05, "loss": 0.9, "step": 5386 }, { "epoch": 2.4408699592206613, "grad_norm": 0.7632606738970893, "learning_rate": 4.85691989254443e-05, "loss": 0.8914, "step": 5387 }, { "epoch": 2.441323062981423, "grad_norm": 0.5807358735032588, "learning_rate": 4.8556838593208845e-05, "loss": 0.8743, "step": 5388 }, { "epoch": 2.441776166742184, "grad_norm": 0.395075004569632, "learning_rate": 4.8544477404671054e-05, "loss": 0.8588, "step": 5389 }, { "epoch": 2.442229270502945, "grad_norm": 0.32522358532289114, "learning_rate": 4.853211536106796e-05, "loss": 0.8712, "step": 5390 }, { "epoch": 2.4426823742637063, "grad_norm": 0.4162673980258829, "learning_rate": 4.851975246363666e-05, "loss": 0.9007, "step": 5391 }, { "epoch": 2.443135478024468, "grad_norm": 0.5466877328201308, "learning_rate": 4.850738871361434e-05, "loss": 0.8825, "step": 5392 }, { "epoch": 2.443588581785229, "grad_norm": 0.4970903935612619, "learning_rate": 4.849502411223827e-05, "loss": 0.8736, "step": 5393 }, { "epoch": 2.44404168554599, "grad_norm": 0.4330189697773245, "learning_rate": 4.84826586607458e-05, "loss": 0.8718, "step": 5394 }, { "epoch": 2.4444947893067512, "grad_norm": 0.4620127368044252, "learning_rate": 4.847029236037437e-05, "loss": 0.8782, "step": 5395 }, { "epoch": 2.4449478930675124, "grad_norm": 0.5654955552372406, "learning_rate": 4.84579252123615e-05, "loss": 0.8863, "step": 5396 }, { "epoch": 2.4454009968282735, "grad_norm": 0.5823195069979051, "learning_rate": 4.844555721794482e-05, "loss": 0.8929, "step": 5397 }, { "epoch": 2.445854100589035, "grad_norm": 0.5225246680607515, "learning_rate": 4.843318837836199e-05, "loss": 0.8691, "step": 5398 }, { "epoch": 2.446307204349796, "grad_norm": 0.5516818514918737, "learning_rate": 4.842081869485081e-05, "loss": 0.8736, "step": 5399 }, { "epoch": 2.4467603081105573, "grad_norm": 0.7090771116452328, "learning_rate": 4.840844816864914e-05, "loss": 0.8901, "step": 5400 }, { "epoch": 2.4472134118713185, "grad_norm": 0.7001838865769408, "learning_rate": 4.839607680099491e-05, "loss": 0.8702, "step": 5401 }, { "epoch": 2.4476665156320796, "grad_norm": 0.6691880889569649, "learning_rate": 4.838370459312616e-05, "loss": 0.8767, "step": 5402 }, { "epoch": 2.448119619392841, "grad_norm": 0.5332697466906319, "learning_rate": 4.8371331546281025e-05, "loss": 0.8745, "step": 5403 }, { "epoch": 2.4485727231536023, "grad_norm": 0.37151861258596697, "learning_rate": 4.835895766169768e-05, "loss": 0.882, "step": 5404 }, { "epoch": 2.4490258269143634, "grad_norm": 0.3200934741263881, "learning_rate": 4.8346582940614414e-05, "loss": 0.8772, "step": 5405 }, { "epoch": 2.4494789306751246, "grad_norm": 0.45884764023114677, "learning_rate": 4.8334207384269596e-05, "loss": 0.869, "step": 5406 }, { "epoch": 2.4499320344358857, "grad_norm": 0.5715819445020256, "learning_rate": 4.832183099390168e-05, "loss": 0.8748, "step": 5407 }, { "epoch": 2.450385138196647, "grad_norm": 0.45239141286337237, "learning_rate": 4.830945377074919e-05, "loss": 0.8744, "step": 5408 }, { "epoch": 2.4508382419574084, "grad_norm": 0.3652817038180547, "learning_rate": 4.829707571605075e-05, "loss": 0.8496, "step": 5409 }, { "epoch": 2.4512913457181695, "grad_norm": 0.348755629671941, "learning_rate": 4.828469683104505e-05, "loss": 0.8939, "step": 5410 }, { "epoch": 2.4517444494789307, "grad_norm": 0.45854226660621256, "learning_rate": 4.827231711697089e-05, "loss": 0.867, "step": 5411 }, { "epoch": 2.452197553239692, "grad_norm": 0.5156567934428881, "learning_rate": 4.825993657506712e-05, "loss": 0.8518, "step": 5412 }, { "epoch": 2.452650657000453, "grad_norm": 0.4603157232499018, "learning_rate": 4.82475552065727e-05, "loss": 0.8917, "step": 5413 }, { "epoch": 2.4531037607612145, "grad_norm": 0.36340245972129753, "learning_rate": 4.823517301272666e-05, "loss": 0.8817, "step": 5414 }, { "epoch": 2.4535568645219756, "grad_norm": 0.3670750956054307, "learning_rate": 4.822278999476811e-05, "loss": 0.8776, "step": 5415 }, { "epoch": 2.4540099682827368, "grad_norm": 0.3725095957570125, "learning_rate": 4.821040615393624e-05, "loss": 0.8915, "step": 5416 }, { "epoch": 2.454463072043498, "grad_norm": 0.3355653673992236, "learning_rate": 4.8198021491470355e-05, "loss": 0.8787, "step": 5417 }, { "epoch": 2.454916175804259, "grad_norm": 0.2833565902875523, "learning_rate": 4.8185636008609785e-05, "loss": 0.8722, "step": 5418 }, { "epoch": 2.45536927956502, "grad_norm": 0.35762023808714544, "learning_rate": 4.8173249706594e-05, "loss": 0.8589, "step": 5419 }, { "epoch": 2.4558223833257817, "grad_norm": 0.4279036585141253, "learning_rate": 4.81608625866625e-05, "loss": 0.8866, "step": 5420 }, { "epoch": 2.456275487086543, "grad_norm": 0.37061500080170257, "learning_rate": 4.814847465005491e-05, "loss": 0.8712, "step": 5421 }, { "epoch": 2.456728590847304, "grad_norm": 0.4051958923277958, "learning_rate": 4.813608589801092e-05, "loss": 0.8698, "step": 5422 }, { "epoch": 2.457181694608065, "grad_norm": 0.4461012416561926, "learning_rate": 4.81236963317703e-05, "loss": 0.8723, "step": 5423 }, { "epoch": 2.4576347983688267, "grad_norm": 0.5264998164594696, "learning_rate": 4.811130595257289e-05, "loss": 0.8709, "step": 5424 }, { "epoch": 2.458087902129588, "grad_norm": 0.5949290407040879, "learning_rate": 4.809891476165863e-05, "loss": 0.8601, "step": 5425 }, { "epoch": 2.458541005890349, "grad_norm": 0.601725679819511, "learning_rate": 4.8086522760267535e-05, "loss": 0.8818, "step": 5426 }, { "epoch": 2.45899410965111, "grad_norm": 0.53333291385781, "learning_rate": 4.807412994963969e-05, "loss": 0.8701, "step": 5427 }, { "epoch": 2.4594472134118712, "grad_norm": 0.48200744413411684, "learning_rate": 4.806173633101531e-05, "loss": 0.8865, "step": 5428 }, { "epoch": 2.4599003171726324, "grad_norm": 0.42056901663137125, "learning_rate": 4.804934190563461e-05, "loss": 0.8581, "step": 5429 }, { "epoch": 2.460353420933394, "grad_norm": 0.3310369748498965, "learning_rate": 4.803694667473794e-05, "loss": 0.8822, "step": 5430 }, { "epoch": 2.460806524694155, "grad_norm": 0.2911042053753294, "learning_rate": 4.802455063956572e-05, "loss": 0.8694, "step": 5431 }, { "epoch": 2.461259628454916, "grad_norm": 0.32796318918263606, "learning_rate": 4.801215380135847e-05, "loss": 0.8765, "step": 5432 }, { "epoch": 2.4617127322156773, "grad_norm": 0.4053061349867875, "learning_rate": 4.799975616135674e-05, "loss": 0.891, "step": 5433 }, { "epoch": 2.4621658359764385, "grad_norm": 0.5536562124983555, "learning_rate": 4.798735772080121e-05, "loss": 0.8761, "step": 5434 }, { "epoch": 2.4626189397372, "grad_norm": 0.7172609384640879, "learning_rate": 4.7974958480932594e-05, "loss": 0.8894, "step": 5435 }, { "epoch": 2.463072043497961, "grad_norm": 0.7126465356187238, "learning_rate": 4.7962558442991745e-05, "loss": 0.8792, "step": 5436 }, { "epoch": 2.4635251472587223, "grad_norm": 1.0168247546295897, "learning_rate": 4.7950157608219546e-05, "loss": 0.8612, "step": 5437 }, { "epoch": 2.4639782510194834, "grad_norm": 0.34633646398360685, "learning_rate": 4.793775597785698e-05, "loss": 0.8587, "step": 5438 }, { "epoch": 2.4644313547802446, "grad_norm": 0.4008391174097015, "learning_rate": 4.792535355314511e-05, "loss": 0.8807, "step": 5439 }, { "epoch": 2.4648844585410057, "grad_norm": 0.4381546898576736, "learning_rate": 4.791295033532506e-05, "loss": 0.8842, "step": 5440 }, { "epoch": 2.4653375623017673, "grad_norm": 0.5904201347925546, "learning_rate": 4.7900546325638075e-05, "loss": 0.8619, "step": 5441 }, { "epoch": 2.4657906660625284, "grad_norm": 0.7724695414768737, "learning_rate": 4.788814152532542e-05, "loss": 0.8852, "step": 5442 }, { "epoch": 2.4662437698232895, "grad_norm": 0.8060504210098489, "learning_rate": 4.787573593562849e-05, "loss": 0.8756, "step": 5443 }, { "epoch": 2.4666968735840507, "grad_norm": 0.8600227936314203, "learning_rate": 4.786332955778875e-05, "loss": 0.8765, "step": 5444 }, { "epoch": 2.467149977344812, "grad_norm": 3.0662507556492273, "learning_rate": 4.785092239304771e-05, "loss": 0.9309, "step": 5445 }, { "epoch": 2.4676030811055734, "grad_norm": 1.528339660755391, "learning_rate": 4.7838514442647e-05, "loss": 0.8843, "step": 5446 }, { "epoch": 2.4680561848663345, "grad_norm": 1.486544761258212, "learning_rate": 4.782610570782831e-05, "loss": 0.8891, "step": 5447 }, { "epoch": 2.4685092886270956, "grad_norm": 0.7483167839449644, "learning_rate": 4.78136961898334e-05, "loss": 0.8996, "step": 5448 }, { "epoch": 2.4689623923878568, "grad_norm": 1.8739203585859867, "learning_rate": 4.780128588990414e-05, "loss": 0.8866, "step": 5449 }, { "epoch": 2.469415496148618, "grad_norm": 0.9405431815773823, "learning_rate": 4.7788874809282445e-05, "loss": 0.9017, "step": 5450 }, { "epoch": 2.469868599909379, "grad_norm": 2.3367028889103456, "learning_rate": 4.777646294921031e-05, "loss": 0.8852, "step": 5451 }, { "epoch": 2.4703217036701406, "grad_norm": 2.129248954955108, "learning_rate": 4.776405031092983e-05, "loss": 0.9084, "step": 5452 }, { "epoch": 2.4707748074309017, "grad_norm": 1.5818871409953001, "learning_rate": 4.775163689568317e-05, "loss": 0.8902, "step": 5453 }, { "epoch": 2.471227911191663, "grad_norm": 1.648004220686059, "learning_rate": 4.773922270471255e-05, "loss": 0.8966, "step": 5454 }, { "epoch": 2.471681014952424, "grad_norm": 1.5373697879978925, "learning_rate": 4.772680773926031e-05, "loss": 0.8836, "step": 5455 }, { "epoch": 2.4721341187131856, "grad_norm": 1.3573932422451676, "learning_rate": 4.771439200056882e-05, "loss": 0.8774, "step": 5456 }, { "epoch": 2.4725872224739467, "grad_norm": 1.4201347979406167, "learning_rate": 4.770197548988057e-05, "loss": 0.8914, "step": 5457 }, { "epoch": 2.473040326234708, "grad_norm": 1.3167866780308652, "learning_rate": 4.76895582084381e-05, "loss": 0.8575, "step": 5458 }, { "epoch": 2.473493429995469, "grad_norm": 1.0365890603678296, "learning_rate": 4.767714015748405e-05, "loss": 0.8831, "step": 5459 }, { "epoch": 2.47394653375623, "grad_norm": 1.8919345714197524, "learning_rate": 4.76647213382611e-05, "loss": 0.89, "step": 5460 }, { "epoch": 2.474399637516991, "grad_norm": 1.3713525518690555, "learning_rate": 4.765230175201203e-05, "loss": 0.878, "step": 5461 }, { "epoch": 2.474852741277753, "grad_norm": 1.7675774381397071, "learning_rate": 4.7639881399979724e-05, "loss": 0.8743, "step": 5462 }, { "epoch": 2.475305845038514, "grad_norm": 1.5001036210255934, "learning_rate": 4.762746028340708e-05, "loss": 0.8923, "step": 5463 }, { "epoch": 2.475758948799275, "grad_norm": 1.5580225545665773, "learning_rate": 4.761503840353713e-05, "loss": 0.8865, "step": 5464 }, { "epoch": 2.476212052560036, "grad_norm": 1.1708359657059257, "learning_rate": 4.760261576161295e-05, "loss": 0.8548, "step": 5465 }, { "epoch": 2.4766651563207973, "grad_norm": 1.7405280193094932, "learning_rate": 4.759019235887771e-05, "loss": 0.8854, "step": 5466 }, { "epoch": 2.477118260081559, "grad_norm": 1.0849740915216246, "learning_rate": 4.7577768196574646e-05, "loss": 0.8824, "step": 5467 }, { "epoch": 2.47757136384232, "grad_norm": 1.9578038210044768, "learning_rate": 4.756534327594708e-05, "loss": 0.8888, "step": 5468 }, { "epoch": 2.478024467603081, "grad_norm": 1.5080722644086766, "learning_rate": 4.7552917598238374e-05, "loss": 0.8785, "step": 5469 }, { "epoch": 2.4784775713638423, "grad_norm": 1.7604692687904264, "learning_rate": 4.754049116469202e-05, "loss": 0.8939, "step": 5470 }, { "epoch": 2.4789306751246034, "grad_norm": 1.4810115579797947, "learning_rate": 4.752806397655155e-05, "loss": 0.8804, "step": 5471 }, { "epoch": 2.4793837788853645, "grad_norm": 1.6825394037530237, "learning_rate": 4.7515636035060586e-05, "loss": 0.8907, "step": 5472 }, { "epoch": 2.479836882646126, "grad_norm": 1.3263839579687537, "learning_rate": 4.7503207341462815e-05, "loss": 0.867, "step": 5473 }, { "epoch": 2.4802899864068872, "grad_norm": 1.8947391461988503, "learning_rate": 4.7490777897002e-05, "loss": 0.8919, "step": 5474 }, { "epoch": 2.4807430901676484, "grad_norm": 1.529942538319513, "learning_rate": 4.7478347702922005e-05, "loss": 0.8575, "step": 5475 }, { "epoch": 2.4811961939284095, "grad_norm": 1.5859936876662668, "learning_rate": 4.746591676046673e-05, "loss": 0.8795, "step": 5476 }, { "epoch": 2.4816492976891706, "grad_norm": 1.4519221294268705, "learning_rate": 4.745348507088017e-05, "loss": 0.8797, "step": 5477 }, { "epoch": 2.482102401449932, "grad_norm": 1.6862998661557418, "learning_rate": 4.744105263540639e-05, "loss": 0.8905, "step": 5478 }, { "epoch": 2.4825555052106933, "grad_norm": 1.4223216243877694, "learning_rate": 4.742861945528954e-05, "loss": 0.894, "step": 5479 }, { "epoch": 2.4830086089714545, "grad_norm": 1.5818460293449852, "learning_rate": 4.7416185531773824e-05, "loss": 0.8772, "step": 5480 }, { "epoch": 2.4834617127322156, "grad_norm": 1.3902688314100196, "learning_rate": 4.7403750866103553e-05, "loss": 0.8668, "step": 5481 }, { "epoch": 2.4839148164929767, "grad_norm": 1.7315805687557722, "learning_rate": 4.739131545952306e-05, "loss": 0.8655, "step": 5482 }, { "epoch": 2.484367920253738, "grad_norm": 1.4678361393715218, "learning_rate": 4.737887931327683e-05, "loss": 0.8868, "step": 5483 }, { "epoch": 2.4848210240144994, "grad_norm": 1.6532717359241547, "learning_rate": 4.7366442428609347e-05, "loss": 0.8772, "step": 5484 }, { "epoch": 2.4852741277752606, "grad_norm": 1.450959994372625, "learning_rate": 4.7354004806765196e-05, "loss": 0.8635, "step": 5485 }, { "epoch": 2.4857272315360217, "grad_norm": 1.654149270527945, "learning_rate": 4.7341566448989046e-05, "loss": 0.8825, "step": 5486 }, { "epoch": 2.486180335296783, "grad_norm": 1.4716568161105752, "learning_rate": 4.732912735652564e-05, "loss": 0.888, "step": 5487 }, { "epoch": 2.4866334390575444, "grad_norm": 1.637322722251652, "learning_rate": 4.731668753061977e-05, "loss": 0.8748, "step": 5488 }, { "epoch": 2.4870865428183055, "grad_norm": 1.4808763063860613, "learning_rate": 4.7304246972516324e-05, "loss": 0.8573, "step": 5489 }, { "epoch": 2.4875396465790667, "grad_norm": 1.5334054619357165, "learning_rate": 4.7291805683460254e-05, "loss": 0.8817, "step": 5490 }, { "epoch": 2.487992750339828, "grad_norm": 1.3084484241777607, "learning_rate": 4.727936366469659e-05, "loss": 0.8637, "step": 5491 }, { "epoch": 2.488445854100589, "grad_norm": 1.696414689061972, "learning_rate": 4.726692091747044e-05, "loss": 0.8847, "step": 5492 }, { "epoch": 2.48889895786135, "grad_norm": 1.4074541582277538, "learning_rate": 4.725447744302697e-05, "loss": 0.8809, "step": 5493 }, { "epoch": 2.4893520616221116, "grad_norm": 1.6841669226152802, "learning_rate": 4.724203324261145e-05, "loss": 0.8817, "step": 5494 }, { "epoch": 2.4898051653828728, "grad_norm": 1.4743589016726997, "learning_rate": 4.7229588317469155e-05, "loss": 0.8896, "step": 5495 }, { "epoch": 2.490258269143634, "grad_norm": 1.6199497460531593, "learning_rate": 4.7217142668845507e-05, "loss": 0.8724, "step": 5496 }, { "epoch": 2.490711372904395, "grad_norm": 1.3869869292061945, "learning_rate": 4.7204696297985975e-05, "loss": 0.8754, "step": 5497 }, { "epoch": 2.491164476665156, "grad_norm": 1.599582957081405, "learning_rate": 4.7192249206136065e-05, "loss": 0.8965, "step": 5498 }, { "epoch": 2.4916175804259177, "grad_norm": 1.3946034469283568, "learning_rate": 4.717980139454143e-05, "loss": 0.8596, "step": 5499 }, { "epoch": 2.492070684186679, "grad_norm": 1.640403096339969, "learning_rate": 4.71673528644477e-05, "loss": 0.8903, "step": 5500 }, { "epoch": 2.49252378794744, "grad_norm": 1.4194550277731781, "learning_rate": 4.7154903617100676e-05, "loss": 0.883, "step": 5501 }, { "epoch": 2.492976891708201, "grad_norm": 1.5330530783129477, "learning_rate": 4.7142453653746164e-05, "loss": 0.8567, "step": 5502 }, { "epoch": 2.4934299954689623, "grad_norm": 1.4351740234686716, "learning_rate": 4.713000297563005e-05, "loss": 0.8989, "step": 5503 }, { "epoch": 2.4938830992297234, "grad_norm": 1.5495282478649908, "learning_rate": 4.711755158399831e-05, "loss": 0.8816, "step": 5504 }, { "epoch": 2.494336202990485, "grad_norm": 1.3428971812691375, "learning_rate": 4.710509948009698e-05, "loss": 0.876, "step": 5505 }, { "epoch": 2.494789306751246, "grad_norm": 1.519788405806767, "learning_rate": 4.709264666517217e-05, "loss": 0.8707, "step": 5506 }, { "epoch": 2.4952424105120072, "grad_norm": 1.3587652813723112, "learning_rate": 4.7080193140470065e-05, "loss": 0.8985, "step": 5507 }, { "epoch": 2.4956955142727684, "grad_norm": 1.5739628396740368, "learning_rate": 4.7067738907236926e-05, "loss": 0.8792, "step": 5508 }, { "epoch": 2.4961486180335295, "grad_norm": 1.3630123244292207, "learning_rate": 4.705528396671906e-05, "loss": 0.8652, "step": 5509 }, { "epoch": 2.496601721794291, "grad_norm": 1.540129802729648, "learning_rate": 4.7042828320162875e-05, "loss": 0.8754, "step": 5510 }, { "epoch": 2.497054825555052, "grad_norm": 1.3890068687045967, "learning_rate": 4.703037196881482e-05, "loss": 0.8803, "step": 5511 }, { "epoch": 2.4975079293158133, "grad_norm": 1.5197746339550395, "learning_rate": 4.7017914913921446e-05, "loss": 0.8869, "step": 5512 }, { "epoch": 2.4979610330765745, "grad_norm": 1.3065927530869372, "learning_rate": 4.7005457156729344e-05, "loss": 0.8414, "step": 5513 }, { "epoch": 2.4984141368373356, "grad_norm": 1.5108907434501435, "learning_rate": 4.699299869848521e-05, "loss": 0.8821, "step": 5514 }, { "epoch": 2.4988672405980967, "grad_norm": 1.2929608558972345, "learning_rate": 4.698053954043579e-05, "loss": 0.8931, "step": 5515 }, { "epoch": 2.4993203443588583, "grad_norm": 1.526273473276304, "learning_rate": 4.696807968382786e-05, "loss": 0.8732, "step": 5516 }, { "epoch": 2.4997734481196194, "grad_norm": 1.3596949688512132, "learning_rate": 4.695561912990835e-05, "loss": 0.8747, "step": 5517 }, { "epoch": 2.5002265518803806, "grad_norm": 1.4798282209682854, "learning_rate": 4.69431578799242e-05, "loss": 0.8728, "step": 5518 }, { "epoch": 2.5006796556411417, "grad_norm": 1.3389049257282715, "learning_rate": 4.6930695935122437e-05, "loss": 0.8767, "step": 5519 }, { "epoch": 2.5011327594019033, "grad_norm": 1.3874296278641387, "learning_rate": 4.691823329675017e-05, "loss": 0.8689, "step": 5520 }, { "epoch": 2.5015858631626644, "grad_norm": 1.2413214502937249, "learning_rate": 4.690576996605454e-05, "loss": 0.8634, "step": 5521 }, { "epoch": 2.5020389669234255, "grad_norm": 1.6079620573676334, "learning_rate": 4.6893305944282786e-05, "loss": 0.8976, "step": 5522 }, { "epoch": 2.5024920706841867, "grad_norm": 1.3791454573419517, "learning_rate": 4.688084123268222e-05, "loss": 0.8512, "step": 5523 }, { "epoch": 2.502945174444948, "grad_norm": 1.3753091585955235, "learning_rate": 4.68683758325002e-05, "loss": 0.8655, "step": 5524 }, { "epoch": 2.503398278205709, "grad_norm": 1.1704467856326686, "learning_rate": 4.6855909744984184e-05, "loss": 0.8739, "step": 5525 }, { "epoch": 2.50385138196647, "grad_norm": 1.6103945149916725, "learning_rate": 4.684344297138168e-05, "loss": 0.8697, "step": 5526 }, { "epoch": 2.5043044857272316, "grad_norm": 1.4086075659616548, "learning_rate": 4.6830975512940253e-05, "loss": 0.8852, "step": 5527 }, { "epoch": 2.5047575894879928, "grad_norm": 1.3471089629283686, "learning_rate": 4.681850737090757e-05, "loss": 0.8774, "step": 5528 }, { "epoch": 2.505210693248754, "grad_norm": 1.295225676213785, "learning_rate": 4.680603854653132e-05, "loss": 0.8791, "step": 5529 }, { "epoch": 2.505663797009515, "grad_norm": 1.3275773210668655, "learning_rate": 4.679356904105931e-05, "loss": 0.8698, "step": 5530 }, { "epoch": 2.5061169007702766, "grad_norm": 1.0162197531300123, "learning_rate": 4.678109885573938e-05, "loss": 0.8894, "step": 5531 }, { "epoch": 2.5065700045310377, "grad_norm": 1.6874338041683503, "learning_rate": 4.676862799181946e-05, "loss": 0.8507, "step": 5532 }, { "epoch": 2.507023108291799, "grad_norm": 1.4988032137851133, "learning_rate": 4.6756156450547515e-05, "loss": 0.8522, "step": 5533 }, { "epoch": 2.50747621205256, "grad_norm": 1.335646959826851, "learning_rate": 4.674368423317162e-05, "loss": 0.8654, "step": 5534 }, { "epoch": 2.507929315813321, "grad_norm": 1.2821957517292795, "learning_rate": 4.6731211340939897e-05, "loss": 0.884, "step": 5535 }, { "epoch": 2.5083824195740823, "grad_norm": 1.3111160328715885, "learning_rate": 4.6718737775100536e-05, "loss": 0.877, "step": 5536 }, { "epoch": 2.508835523334844, "grad_norm": 1.1022806124933127, "learning_rate": 4.6706263536901804e-05, "loss": 0.887, "step": 5537 }, { "epoch": 2.509288627095605, "grad_norm": 1.6048870448039136, "learning_rate": 4.6693788627592004e-05, "loss": 0.8892, "step": 5538 }, { "epoch": 2.509741730856366, "grad_norm": 1.4092333129139973, "learning_rate": 4.668131304841954e-05, "loss": 0.8812, "step": 5539 }, { "epoch": 2.5101948346171272, "grad_norm": 1.2802456368464703, "learning_rate": 4.666883680063286e-05, "loss": 0.868, "step": 5540 }, { "epoch": 2.510647938377889, "grad_norm": 1.2261410700565636, "learning_rate": 4.6656359885480515e-05, "loss": 0.8963, "step": 5541 }, { "epoch": 2.51110104213865, "grad_norm": 1.4097064991496704, "learning_rate": 4.664388230421109e-05, "loss": 0.8741, "step": 5542 }, { "epoch": 2.511554145899411, "grad_norm": 1.1472675125758354, "learning_rate": 4.6631404058073236e-05, "loss": 0.8802, "step": 5543 }, { "epoch": 2.512007249660172, "grad_norm": 1.5384090324194395, "learning_rate": 4.6618925148315676e-05, "loss": 0.8758, "step": 5544 }, { "epoch": 2.5124603534209333, "grad_norm": 1.3416219728850356, "learning_rate": 4.660644557618722e-05, "loss": 0.8808, "step": 5545 }, { "epoch": 2.5129134571816945, "grad_norm": 1.3129782530075997, "learning_rate": 4.659396534293671e-05, "loss": 0.8724, "step": 5546 }, { "epoch": 2.5133665609424556, "grad_norm": 1.2015895844834847, "learning_rate": 4.65814844498131e-05, "loss": 0.8445, "step": 5547 }, { "epoch": 2.513819664703217, "grad_norm": 1.2669226652637622, "learning_rate": 4.656900289806534e-05, "loss": 0.8603, "step": 5548 }, { "epoch": 2.5142727684639783, "grad_norm": 1.0662441305630905, "learning_rate": 4.65565206889425e-05, "loss": 0.8696, "step": 5549 }, { "epoch": 2.5147258722247394, "grad_norm": 1.4476978770997853, "learning_rate": 4.6544037823693735e-05, "loss": 0.8653, "step": 5550 }, { "epoch": 2.5151789759855006, "grad_norm": 1.2668058396774957, "learning_rate": 4.6531554303568186e-05, "loss": 0.8857, "step": 5551 }, { "epoch": 2.515632079746262, "grad_norm": 1.2117190951376167, "learning_rate": 4.651907012981514e-05, "loss": 0.8788, "step": 5552 }, { "epoch": 2.5160851835070233, "grad_norm": 1.1352371087446222, "learning_rate": 4.6506585303683916e-05, "loss": 0.8862, "step": 5553 }, { "epoch": 2.5165382872677844, "grad_norm": 1.256317860011909, "learning_rate": 4.649409982642389e-05, "loss": 0.8691, "step": 5554 }, { "epoch": 2.5169913910285455, "grad_norm": 1.0845964639125454, "learning_rate": 4.648161369928451e-05, "loss": 0.8749, "step": 5555 }, { "epoch": 2.5174444947893067, "grad_norm": 1.2392979697738922, "learning_rate": 4.646912692351529e-05, "loss": 0.863, "step": 5556 }, { "epoch": 2.517897598550068, "grad_norm": 0.989565300825066, "learning_rate": 4.645663950036581e-05, "loss": 0.8792, "step": 5557 }, { "epoch": 2.518350702310829, "grad_norm": 1.3961832700591552, "learning_rate": 4.644415143108573e-05, "loss": 0.8825, "step": 5558 }, { "epoch": 2.5188038060715905, "grad_norm": 1.33327120726678, "learning_rate": 4.6431662716924743e-05, "loss": 0.8744, "step": 5559 }, { "epoch": 2.5192569098323516, "grad_norm": 1.0310272947833767, "learning_rate": 4.641917335913262e-05, "loss": 0.8754, "step": 5560 }, { "epoch": 2.5197100135931128, "grad_norm": 0.889225331811043, "learning_rate": 4.640668335895922e-05, "loss": 0.8812, "step": 5561 }, { "epoch": 2.520163117353874, "grad_norm": 1.033073713356036, "learning_rate": 4.6394192717654426e-05, "loss": 0.8897, "step": 5562 }, { "epoch": 2.5206162211146355, "grad_norm": 0.7077943617403512, "learning_rate": 4.6381701436468224e-05, "loss": 0.9075, "step": 5563 }, { "epoch": 2.5210693248753966, "grad_norm": 1.053964888849025, "learning_rate": 4.636920951665063e-05, "loss": 0.8779, "step": 5564 }, { "epoch": 2.5215224286361577, "grad_norm": 0.4492257516068177, "learning_rate": 4.6356716959451736e-05, "loss": 0.8571, "step": 5565 }, { "epoch": 2.521975532396919, "grad_norm": 0.8854827630326603, "learning_rate": 4.634422376612171e-05, "loss": 0.8883, "step": 5566 }, { "epoch": 2.52242863615768, "grad_norm": 0.49963349568898785, "learning_rate": 4.633172993791077e-05, "loss": 0.9015, "step": 5567 }, { "epoch": 2.522881739918441, "grad_norm": 0.6136878173923287, "learning_rate": 4.63192354760692e-05, "loss": 0.8658, "step": 5568 }, { "epoch": 2.5233348436792027, "grad_norm": 0.4708953825095812, "learning_rate": 4.6306740381847356e-05, "loss": 0.8694, "step": 5569 }, { "epoch": 2.523787947439964, "grad_norm": 0.5461220558835465, "learning_rate": 4.629424465649564e-05, "loss": 0.8956, "step": 5570 }, { "epoch": 2.524241051200725, "grad_norm": 0.4463017306995093, "learning_rate": 4.628174830126455e-05, "loss": 0.8633, "step": 5571 }, { "epoch": 2.524694154961486, "grad_norm": 0.4574725432907616, "learning_rate": 4.6269251317404604e-05, "loss": 0.8603, "step": 5572 }, { "epoch": 2.5251472587222477, "grad_norm": 0.4178322210688436, "learning_rate": 4.6256753706166413e-05, "loss": 0.8703, "step": 5573 }, { "epoch": 2.525600362483009, "grad_norm": 0.4371286217756273, "learning_rate": 4.624425546880063e-05, "loss": 0.8675, "step": 5574 }, { "epoch": 2.52605346624377, "grad_norm": 0.3709641463305943, "learning_rate": 4.623175660655799e-05, "loss": 0.8667, "step": 5575 }, { "epoch": 2.526506570004531, "grad_norm": 0.3680712723415576, "learning_rate": 4.6219257120689275e-05, "loss": 0.8762, "step": 5576 }, { "epoch": 2.526959673765292, "grad_norm": 0.34510734751147093, "learning_rate": 4.620675701244535e-05, "loss": 0.8646, "step": 5577 }, { "epoch": 2.5274127775260533, "grad_norm": 0.3835291426960551, "learning_rate": 4.6194256283077116e-05, "loss": 0.8922, "step": 5578 }, { "epoch": 2.5278658812868144, "grad_norm": 0.3729633021167964, "learning_rate": 4.618175493383556e-05, "loss": 0.8844, "step": 5579 }, { "epoch": 2.528318985047576, "grad_norm": 0.3861174091158015, "learning_rate": 4.616925296597172e-05, "loss": 0.886, "step": 5580 }, { "epoch": 2.528772088808337, "grad_norm": 0.3147955964022457, "learning_rate": 4.6156750380736703e-05, "loss": 0.8745, "step": 5581 }, { "epoch": 2.5292251925690983, "grad_norm": 0.36096049960209897, "learning_rate": 4.6144247179381656e-05, "loss": 0.8531, "step": 5582 }, { "epoch": 2.5296782963298594, "grad_norm": 0.31132624085447375, "learning_rate": 4.613174336315781e-05, "loss": 0.8654, "step": 5583 }, { "epoch": 2.530131400090621, "grad_norm": 0.4348448690510856, "learning_rate": 4.6119238933316434e-05, "loss": 0.869, "step": 5584 }, { "epoch": 2.530584503851382, "grad_norm": 0.32353163433695653, "learning_rate": 4.61067338911089e-05, "loss": 0.8914, "step": 5585 }, { "epoch": 2.5310376076121432, "grad_norm": 0.33369890318833145, "learning_rate": 4.609422823778661e-05, "loss": 0.8586, "step": 5586 }, { "epoch": 2.5314907113729044, "grad_norm": 0.3389085507912515, "learning_rate": 4.6081721974601016e-05, "loss": 0.8564, "step": 5587 }, { "epoch": 2.5319438151336655, "grad_norm": 0.3662868460551837, "learning_rate": 4.6069215102803674e-05, "loss": 0.8831, "step": 5588 }, { "epoch": 2.5323969188944266, "grad_norm": 0.3043323450263237, "learning_rate": 4.605670762364616e-05, "loss": 0.88, "step": 5589 }, { "epoch": 2.5328500226551878, "grad_norm": 0.30909211017609245, "learning_rate": 4.604419953838014e-05, "loss": 0.8769, "step": 5590 }, { "epoch": 2.5333031264159493, "grad_norm": 0.3246329864569024, "learning_rate": 4.6031690848257284e-05, "loss": 0.8738, "step": 5591 }, { "epoch": 2.5337562301767105, "grad_norm": 0.28853346134098007, "learning_rate": 4.6019181554529414e-05, "loss": 0.8798, "step": 5592 }, { "epoch": 2.5342093339374716, "grad_norm": 0.32718430288695566, "learning_rate": 4.600667165844835e-05, "loss": 0.8638, "step": 5593 }, { "epoch": 2.5346624376982327, "grad_norm": 0.2860924519462773, "learning_rate": 4.599416116126597e-05, "loss": 0.8665, "step": 5594 }, { "epoch": 2.5351155414589943, "grad_norm": 0.28305174158955526, "learning_rate": 4.598165006423424e-05, "loss": 0.8737, "step": 5595 }, { "epoch": 2.5355686452197554, "grad_norm": 0.26031239835875375, "learning_rate": 4.596913836860517e-05, "loss": 0.8593, "step": 5596 }, { "epoch": 2.5360217489805166, "grad_norm": 0.2766738371062197, "learning_rate": 4.5956626075630844e-05, "loss": 0.8964, "step": 5597 }, { "epoch": 2.5364748527412777, "grad_norm": 0.2470569179050254, "learning_rate": 4.59441131865634e-05, "loss": 0.8824, "step": 5598 }, { "epoch": 2.536927956502039, "grad_norm": 0.2882471531758648, "learning_rate": 4.593159970265501e-05, "loss": 0.8763, "step": 5599 }, { "epoch": 2.5373810602628, "grad_norm": 0.23655024429662372, "learning_rate": 4.5919085625157936e-05, "loss": 0.8884, "step": 5600 }, { "epoch": 2.5378341640235615, "grad_norm": 0.25875995724388373, "learning_rate": 4.5906570955324486e-05, "loss": 0.8767, "step": 5601 }, { "epoch": 2.5382872677843227, "grad_norm": 0.2783547492842727, "learning_rate": 4.589405569440704e-05, "loss": 0.8701, "step": 5602 }, { "epoch": 2.538740371545084, "grad_norm": 0.26779249971404434, "learning_rate": 4.588153984365802e-05, "loss": 0.8726, "step": 5603 }, { "epoch": 2.539193475305845, "grad_norm": 0.27033212342935364, "learning_rate": 4.586902340432992e-05, "loss": 0.875, "step": 5604 }, { "epoch": 2.5396465790666065, "grad_norm": 0.2450887275577372, "learning_rate": 4.5856506377675295e-05, "loss": 0.8748, "step": 5605 }, { "epoch": 2.5400996828273676, "grad_norm": 0.26223455844673255, "learning_rate": 4.584398876494674e-05, "loss": 0.8824, "step": 5606 }, { "epoch": 2.5405527865881288, "grad_norm": 0.26726989886137914, "learning_rate": 4.583147056739695e-05, "loss": 0.8754, "step": 5607 }, { "epoch": 2.54100589034889, "grad_norm": 0.232032690191317, "learning_rate": 4.5818951786278597e-05, "loss": 0.8639, "step": 5608 }, { "epoch": 2.541458994109651, "grad_norm": 0.28190393230847227, "learning_rate": 4.580643242284451e-05, "loss": 0.8612, "step": 5609 }, { "epoch": 2.541912097870412, "grad_norm": 0.27447806625597215, "learning_rate": 4.5793912478347506e-05, "loss": 0.8844, "step": 5610 }, { "epoch": 2.5423652016311733, "grad_norm": 0.27628195447589243, "learning_rate": 4.5781391954040494e-05, "loss": 0.8639, "step": 5611 }, { "epoch": 2.542818305391935, "grad_norm": 0.22838656338140134, "learning_rate": 4.5768870851176435e-05, "loss": 0.8785, "step": 5612 }, { "epoch": 2.543271409152696, "grad_norm": 0.25407794588722554, "learning_rate": 4.575634917100833e-05, "loss": 0.8665, "step": 5613 }, { "epoch": 2.543724512913457, "grad_norm": 0.25352446527977435, "learning_rate": 4.5743826914789266e-05, "loss": 0.8646, "step": 5614 }, { "epoch": 2.5441776166742183, "grad_norm": 0.2512618459133219, "learning_rate": 4.573130408377237e-05, "loss": 0.883, "step": 5615 }, { "epoch": 2.54463072043498, "grad_norm": 0.2158595180313097, "learning_rate": 4.571878067921083e-05, "loss": 0.8844, "step": 5616 }, { "epoch": 2.545083824195741, "grad_norm": 0.25401777241345525, "learning_rate": 4.5706256702357896e-05, "loss": 0.8681, "step": 5617 }, { "epoch": 2.545536927956502, "grad_norm": 0.28916433288400256, "learning_rate": 4.569373215446685e-05, "loss": 0.8597, "step": 5618 }, { "epoch": 2.5459900317172632, "grad_norm": 0.23336046529222598, "learning_rate": 4.568120703679108e-05, "loss": 0.8824, "step": 5619 }, { "epoch": 2.5464431354780244, "grad_norm": 0.2706495706353766, "learning_rate": 4.566868135058399e-05, "loss": 0.8948, "step": 5620 }, { "epoch": 2.5468962392387855, "grad_norm": 0.37398308568387323, "learning_rate": 4.565615509709905e-05, "loss": 0.8571, "step": 5621 }, { "epoch": 2.5473493429995466, "grad_norm": 0.344764254776235, "learning_rate": 4.564362827758979e-05, "loss": 0.8658, "step": 5622 }, { "epoch": 2.547802446760308, "grad_norm": 0.2745618210993354, "learning_rate": 4.563110089330981e-05, "loss": 0.8608, "step": 5623 }, { "epoch": 2.5482555505210693, "grad_norm": 0.34033127250041223, "learning_rate": 4.561857294551275e-05, "loss": 0.8672, "step": 5624 }, { "epoch": 2.5487086542818305, "grad_norm": 0.2998805506378872, "learning_rate": 4.560604443545231e-05, "loss": 0.8835, "step": 5625 }, { "epoch": 2.5491617580425916, "grad_norm": 0.2621677751747806, "learning_rate": 4.559351536438223e-05, "loss": 0.8826, "step": 5626 }, { "epoch": 2.549614861803353, "grad_norm": 0.29041516962057184, "learning_rate": 4.558098573355634e-05, "loss": 0.8821, "step": 5627 }, { "epoch": 2.5500679655641143, "grad_norm": 0.2419210688209254, "learning_rate": 4.5568455544228514e-05, "loss": 0.8719, "step": 5628 }, { "epoch": 2.5505210693248754, "grad_norm": 0.3176580628436435, "learning_rate": 4.555592479765265e-05, "loss": 0.8792, "step": 5629 }, { "epoch": 2.5509741730856366, "grad_norm": 0.2650679701807559, "learning_rate": 4.554339349508276e-05, "loss": 0.8543, "step": 5630 }, { "epoch": 2.5514272768463977, "grad_norm": 0.24933545301302154, "learning_rate": 4.553086163777285e-05, "loss": 0.884, "step": 5631 }, { "epoch": 2.551880380607159, "grad_norm": 0.379014700177186, "learning_rate": 4.5518329226977034e-05, "loss": 0.8875, "step": 5632 }, { "epoch": 2.5523334843679204, "grad_norm": 0.3140135658539811, "learning_rate": 4.550579626394946e-05, "loss": 0.8681, "step": 5633 }, { "epoch": 2.5527865881286815, "grad_norm": 0.2654424368498857, "learning_rate": 4.549326274994432e-05, "loss": 0.8695, "step": 5634 }, { "epoch": 2.5532396918894427, "grad_norm": 0.6545252326602434, "learning_rate": 4.548072868621587e-05, "loss": 0.8897, "step": 5635 }, { "epoch": 2.553692795650204, "grad_norm": 0.23630165679049053, "learning_rate": 4.546819407401842e-05, "loss": 0.8862, "step": 5636 }, { "epoch": 2.5541458994109654, "grad_norm": 0.3131110049278691, "learning_rate": 4.545565891460635e-05, "loss": 0.8891, "step": 5637 }, { "epoch": 2.5545990031717265, "grad_norm": 0.29912462709010024, "learning_rate": 4.544312320923406e-05, "loss": 0.8503, "step": 5638 }, { "epoch": 2.5550521069324876, "grad_norm": 0.23987586829515967, "learning_rate": 4.543058695915605e-05, "loss": 0.8785, "step": 5639 }, { "epoch": 2.5555052106932488, "grad_norm": 0.3549659510233424, "learning_rate": 4.541805016562684e-05, "loss": 0.8703, "step": 5640 }, { "epoch": 2.55595831445401, "grad_norm": 0.431547729353181, "learning_rate": 4.5405512829901025e-05, "loss": 0.8757, "step": 5641 }, { "epoch": 2.556411418214771, "grad_norm": 0.31690821922577866, "learning_rate": 4.5392974953233246e-05, "loss": 0.8877, "step": 5642 }, { "epoch": 2.556864521975532, "grad_norm": 0.25714288612736885, "learning_rate": 4.5380436536878175e-05, "loss": 0.8846, "step": 5643 }, { "epoch": 2.5573176257362937, "grad_norm": 0.3442283355372257, "learning_rate": 4.5367897582090564e-05, "loss": 0.8682, "step": 5644 }, { "epoch": 2.557770729497055, "grad_norm": 0.3014269905182232, "learning_rate": 4.5355358090125235e-05, "loss": 0.8693, "step": 5645 }, { "epoch": 2.558223833257816, "grad_norm": 0.30176868756801883, "learning_rate": 4.534281806223703e-05, "loss": 0.8721, "step": 5646 }, { "epoch": 2.558676937018577, "grad_norm": 0.2988315985134873, "learning_rate": 4.533027749968086e-05, "loss": 0.876, "step": 5647 }, { "epoch": 2.5591300407793387, "grad_norm": 0.2586910602592691, "learning_rate": 4.531773640371169e-05, "loss": 0.8668, "step": 5648 }, { "epoch": 2.5595831445401, "grad_norm": 0.24549094326961665, "learning_rate": 4.530519477558454e-05, "loss": 0.8854, "step": 5649 }, { "epoch": 2.560036248300861, "grad_norm": 0.29844164909664245, "learning_rate": 4.5292652616554465e-05, "loss": 0.8871, "step": 5650 }, { "epoch": 2.560489352061622, "grad_norm": 0.34973215527542556, "learning_rate": 4.5280109927876596e-05, "loss": 0.8585, "step": 5651 }, { "epoch": 2.5609424558223832, "grad_norm": 0.23740498601645854, "learning_rate": 4.526756671080611e-05, "loss": 0.8487, "step": 5652 }, { "epoch": 2.5613955595831444, "grad_norm": 0.3336948277303818, "learning_rate": 4.5255022966598234e-05, "loss": 0.8534, "step": 5653 }, { "epoch": 2.5618486633439055, "grad_norm": 0.36088829830156033, "learning_rate": 4.5242478696508246e-05, "loss": 0.8682, "step": 5654 }, { "epoch": 2.562301767104667, "grad_norm": 0.24192889660496705, "learning_rate": 4.522993390179149e-05, "loss": 0.8728, "step": 5655 }, { "epoch": 2.562754870865428, "grad_norm": 0.23729203340626437, "learning_rate": 4.521738858370333e-05, "loss": 0.8489, "step": 5656 }, { "epoch": 2.5632079746261893, "grad_norm": 0.26946121577574383, "learning_rate": 4.5204842743499235e-05, "loss": 0.8794, "step": 5657 }, { "epoch": 2.5636610783869505, "grad_norm": 0.22916922775366447, "learning_rate": 4.5192296382434664e-05, "loss": 0.8628, "step": 5658 }, { "epoch": 2.564114182147712, "grad_norm": 0.24517821102610288, "learning_rate": 4.517974950176519e-05, "loss": 0.8681, "step": 5659 }, { "epoch": 2.564567285908473, "grad_norm": 0.24686832383012333, "learning_rate": 4.516720210274639e-05, "loss": 0.8648, "step": 5660 }, { "epoch": 2.5650203896692343, "grad_norm": 0.425790444084783, "learning_rate": 4.5154654186633916e-05, "loss": 0.9016, "step": 5661 }, { "epoch": 2.5654734934299954, "grad_norm": 0.28048250266076413, "learning_rate": 4.5142105754683455e-05, "loss": 0.8693, "step": 5662 }, { "epoch": 2.5659265971907566, "grad_norm": 0.301062029104736, "learning_rate": 4.512955680815076e-05, "loss": 0.8796, "step": 5663 }, { "epoch": 2.5663797009515177, "grad_norm": 0.3238647944569381, "learning_rate": 4.511700734829166e-05, "loss": 0.8765, "step": 5664 }, { "epoch": 2.5668328047122793, "grad_norm": 0.3235325618388274, "learning_rate": 4.510445737636198e-05, "loss": 0.8697, "step": 5665 }, { "epoch": 2.5672859084730404, "grad_norm": 0.2777847240969096, "learning_rate": 4.509190689361762e-05, "loss": 0.8783, "step": 5666 }, { "epoch": 2.5677390122338015, "grad_norm": 0.24050415388595806, "learning_rate": 4.507935590131456e-05, "loss": 0.8788, "step": 5667 }, { "epoch": 2.5681921159945627, "grad_norm": 0.2915433590738713, "learning_rate": 4.50668044007088e-05, "loss": 0.8692, "step": 5668 }, { "epoch": 2.5686452197553242, "grad_norm": 0.33417939812717523, "learning_rate": 4.505425239305638e-05, "loss": 0.8853, "step": 5669 }, { "epoch": 2.5690983235160854, "grad_norm": 0.27281204886876936, "learning_rate": 4.5041699879613416e-05, "loss": 0.87, "step": 5670 }, { "epoch": 2.5695514272768465, "grad_norm": 0.25308714359711953, "learning_rate": 4.502914686163608e-05, "loss": 0.8591, "step": 5671 }, { "epoch": 2.5700045310376076, "grad_norm": 0.348574117068304, "learning_rate": 4.5016593340380555e-05, "loss": 0.8719, "step": 5672 }, { "epoch": 2.5704576347983688, "grad_norm": 0.26353971052841796, "learning_rate": 4.5004039317103126e-05, "loss": 0.8857, "step": 5673 }, { "epoch": 2.57091073855913, "grad_norm": 0.27119785769040705, "learning_rate": 4.49914847930601e-05, "loss": 0.8925, "step": 5674 }, { "epoch": 2.571363842319891, "grad_norm": 0.3006016933025354, "learning_rate": 4.497892976950781e-05, "loss": 0.8542, "step": 5675 }, { "epoch": 2.5718169460806526, "grad_norm": 0.29339762892671084, "learning_rate": 4.4966374247702696e-05, "loss": 0.8769, "step": 5676 }, { "epoch": 2.5722700498414137, "grad_norm": 0.2487163340893313, "learning_rate": 4.495381822890122e-05, "loss": 0.8714, "step": 5677 }, { "epoch": 2.572723153602175, "grad_norm": 0.23158016442719598, "learning_rate": 4.4941261714359855e-05, "loss": 0.8784, "step": 5678 }, { "epoch": 2.573176257362936, "grad_norm": 0.24975577860244438, "learning_rate": 4.49287047053352e-05, "loss": 0.8651, "step": 5679 }, { "epoch": 2.5736293611236976, "grad_norm": 0.28172302716031233, "learning_rate": 4.491614720308383e-05, "loss": 0.8701, "step": 5680 }, { "epoch": 2.5740824648844587, "grad_norm": 0.2701936563835579, "learning_rate": 4.4903589208862426e-05, "loss": 0.8602, "step": 5681 }, { "epoch": 2.57453556864522, "grad_norm": 0.29001194937931213, "learning_rate": 4.4891030723927696e-05, "loss": 0.8605, "step": 5682 }, { "epoch": 2.574988672405981, "grad_norm": 0.284654963671033, "learning_rate": 4.487847174953638e-05, "loss": 0.8727, "step": 5683 }, { "epoch": 2.575441776166742, "grad_norm": 0.25904209731677574, "learning_rate": 4.48659122869453e-05, "loss": 0.8817, "step": 5684 }, { "epoch": 2.575894879927503, "grad_norm": 0.324018504673439, "learning_rate": 4.485335233741131e-05, "loss": 0.8805, "step": 5685 }, { "epoch": 2.5763479836882643, "grad_norm": 0.28672938350222454, "learning_rate": 4.484079190219128e-05, "loss": 0.8742, "step": 5686 }, { "epoch": 2.576801087449026, "grad_norm": 0.24344637439932293, "learning_rate": 4.482823098254221e-05, "loss": 0.8596, "step": 5687 }, { "epoch": 2.577254191209787, "grad_norm": 0.20883800125989843, "learning_rate": 4.4815669579721066e-05, "loss": 0.8839, "step": 5688 }, { "epoch": 2.577707294970548, "grad_norm": 0.2812199540253256, "learning_rate": 4.4803107694984906e-05, "loss": 0.8764, "step": 5689 }, { "epoch": 2.5781603987313093, "grad_norm": 0.23551063204978698, "learning_rate": 4.479054532959082e-05, "loss": 0.8675, "step": 5690 }, { "epoch": 2.578613502492071, "grad_norm": 0.25881565160055003, "learning_rate": 4.477798248479596e-05, "loss": 0.8827, "step": 5691 }, { "epoch": 2.579066606252832, "grad_norm": 0.34914171645969794, "learning_rate": 4.476541916185753e-05, "loss": 0.8753, "step": 5692 }, { "epoch": 2.579519710013593, "grad_norm": 0.28276895053217, "learning_rate": 4.475285536203274e-05, "loss": 0.8534, "step": 5693 }, { "epoch": 2.5799728137743543, "grad_norm": 0.2655577484907486, "learning_rate": 4.47402910865789e-05, "loss": 0.8328, "step": 5694 }, { "epoch": 2.5804259175351154, "grad_norm": 0.2955477717196231, "learning_rate": 4.472772633675336e-05, "loss": 0.864, "step": 5695 }, { "epoch": 2.5808790212958765, "grad_norm": 0.24199176788699292, "learning_rate": 4.471516111381347e-05, "loss": 0.8852, "step": 5696 }, { "epoch": 2.581332125056638, "grad_norm": 0.24596399801982322, "learning_rate": 4.470259541901666e-05, "loss": 0.8737, "step": 5697 }, { "epoch": 2.5817852288173992, "grad_norm": 0.25130421812138953, "learning_rate": 4.469002925362045e-05, "loss": 0.8549, "step": 5698 }, { "epoch": 2.5822383325781604, "grad_norm": 0.2590790814502454, "learning_rate": 4.467746261888232e-05, "loss": 0.8668, "step": 5699 }, { "epoch": 2.5826914363389215, "grad_norm": 0.23354733327332264, "learning_rate": 4.4664895516059864e-05, "loss": 0.8629, "step": 5700 }, { "epoch": 2.583144540099683, "grad_norm": 0.21080685426032877, "learning_rate": 4.4652327946410695e-05, "loss": 0.8992, "step": 5701 }, { "epoch": 2.583597643860444, "grad_norm": 0.21170688515201433, "learning_rate": 4.463975991119248e-05, "loss": 0.8812, "step": 5702 }, { "epoch": 2.5840507476212053, "grad_norm": 0.20245869681028192, "learning_rate": 4.462719141166293e-05, "loss": 0.8748, "step": 5703 }, { "epoch": 2.5845038513819665, "grad_norm": 0.24339135581409904, "learning_rate": 4.4614622449079804e-05, "loss": 0.8854, "step": 5704 }, { "epoch": 2.5849569551427276, "grad_norm": 0.2217486810983419, "learning_rate": 4.4602053024700906e-05, "loss": 0.8628, "step": 5705 }, { "epoch": 2.5854100589034887, "grad_norm": 0.24932704719129842, "learning_rate": 4.458948313978408e-05, "loss": 0.8716, "step": 5706 }, { "epoch": 2.58586316266425, "grad_norm": 0.24046961399512387, "learning_rate": 4.457691279558724e-05, "loss": 0.8549, "step": 5707 }, { "epoch": 2.5863162664250114, "grad_norm": 0.28372904349389405, "learning_rate": 4.4564341993368306e-05, "loss": 0.8717, "step": 5708 }, { "epoch": 2.5867693701857726, "grad_norm": 0.3027748682126054, "learning_rate": 4.4551770734385284e-05, "loss": 0.8795, "step": 5709 }, { "epoch": 2.5872224739465337, "grad_norm": 0.2658703118127412, "learning_rate": 4.4539199019896214e-05, "loss": 0.8511, "step": 5710 }, { "epoch": 2.587675577707295, "grad_norm": 0.2678707591042744, "learning_rate": 4.452662685115916e-05, "loss": 0.8705, "step": 5711 }, { "epoch": 2.5881286814680564, "grad_norm": 0.21550964711880136, "learning_rate": 4.451405422943226e-05, "loss": 0.87, "step": 5712 }, { "epoch": 2.5885817852288175, "grad_norm": 0.23649160303504208, "learning_rate": 4.450148115597367e-05, "loss": 0.8905, "step": 5713 }, { "epoch": 2.5890348889895787, "grad_norm": 0.3042062957581173, "learning_rate": 4.448890763204162e-05, "loss": 0.8832, "step": 5714 }, { "epoch": 2.58948799275034, "grad_norm": 0.29647878600581756, "learning_rate": 4.447633365889436e-05, "loss": 0.872, "step": 5715 }, { "epoch": 2.589941096511101, "grad_norm": 0.294224766722582, "learning_rate": 4.44637592377902e-05, "loss": 0.8721, "step": 5716 }, { "epoch": 2.590394200271862, "grad_norm": 0.2875731354387807, "learning_rate": 4.44511843699875e-05, "loss": 0.8698, "step": 5717 }, { "epoch": 2.590847304032623, "grad_norm": 0.2959677441822498, "learning_rate": 4.4438609056744644e-05, "loss": 0.8952, "step": 5718 }, { "epoch": 2.5913004077933848, "grad_norm": 0.28983666907095035, "learning_rate": 4.442603329932007e-05, "loss": 0.8661, "step": 5719 }, { "epoch": 2.591753511554146, "grad_norm": 0.25054222640886475, "learning_rate": 4.441345709897228e-05, "loss": 0.8876, "step": 5720 }, { "epoch": 2.592206615314907, "grad_norm": 0.27034151164193587, "learning_rate": 4.44008804569598e-05, "loss": 0.903, "step": 5721 }, { "epoch": 2.592659719075668, "grad_norm": 0.2639737998960957, "learning_rate": 4.438830337454119e-05, "loss": 0.8735, "step": 5722 }, { "epoch": 2.5931128228364297, "grad_norm": 0.22291730364930984, "learning_rate": 4.4375725852975064e-05, "loss": 0.8622, "step": 5723 }, { "epoch": 2.593565926597191, "grad_norm": 0.23824264632132428, "learning_rate": 4.436314789352009e-05, "loss": 0.8694, "step": 5724 }, { "epoch": 2.594019030357952, "grad_norm": 0.27886599354154307, "learning_rate": 4.435056949743499e-05, "loss": 0.8844, "step": 5725 }, { "epoch": 2.594472134118713, "grad_norm": 0.2592333667214438, "learning_rate": 4.433799066597848e-05, "loss": 0.8839, "step": 5726 }, { "epoch": 2.5949252378794743, "grad_norm": 0.29564399939210967, "learning_rate": 4.432541140040938e-05, "loss": 0.8693, "step": 5727 }, { "epoch": 2.5953783416402354, "grad_norm": 0.2964177905532497, "learning_rate": 4.4312831701986516e-05, "loss": 0.8477, "step": 5728 }, { "epoch": 2.595831445400997, "grad_norm": 0.3379773855921209, "learning_rate": 4.4300251571968775e-05, "loss": 0.867, "step": 5729 }, { "epoch": 2.596284549161758, "grad_norm": 0.3436850480713352, "learning_rate": 4.428767101161506e-05, "loss": 0.8699, "step": 5730 }, { "epoch": 2.5967376529225192, "grad_norm": 0.3008349187756875, "learning_rate": 4.4275090022184354e-05, "loss": 0.8636, "step": 5731 }, { "epoch": 2.5971907566832804, "grad_norm": 0.22306729866474423, "learning_rate": 4.426250860493565e-05, "loss": 0.8751, "step": 5732 }, { "epoch": 2.597643860444042, "grad_norm": 0.2696489740245366, "learning_rate": 4.4249926761128013e-05, "loss": 0.859, "step": 5733 }, { "epoch": 2.598096964204803, "grad_norm": 0.2947036222459248, "learning_rate": 4.4237344492020525e-05, "loss": 0.8587, "step": 5734 }, { "epoch": 2.598550067965564, "grad_norm": 0.236479347140167, "learning_rate": 4.422476179887234e-05, "loss": 0.8676, "step": 5735 }, { "epoch": 2.5990031717263253, "grad_norm": 0.2458316032139386, "learning_rate": 4.421217868294261e-05, "loss": 0.8605, "step": 5736 }, { "epoch": 2.5994562754870865, "grad_norm": 0.2195449697295513, "learning_rate": 4.419959514549059e-05, "loss": 0.8883, "step": 5737 }, { "epoch": 2.5999093792478476, "grad_norm": 0.25925914376354486, "learning_rate": 4.418701118777552e-05, "loss": 0.8958, "step": 5738 }, { "epoch": 2.6003624830086087, "grad_norm": 0.22498684094019386, "learning_rate": 4.4174426811056706e-05, "loss": 0.8662, "step": 5739 }, { "epoch": 2.6008155867693703, "grad_norm": 0.2697181112979227, "learning_rate": 4.41618420165935e-05, "loss": 0.9025, "step": 5740 }, { "epoch": 2.6012686905301314, "grad_norm": 0.2688289724971182, "learning_rate": 4.414925680564529e-05, "loss": 0.8709, "step": 5741 }, { "epoch": 2.6017217942908926, "grad_norm": 0.25364172669883417, "learning_rate": 4.413667117947151e-05, "loss": 0.8566, "step": 5742 }, { "epoch": 2.6021748980516537, "grad_norm": 0.2918519483528357, "learning_rate": 4.412408513933163e-05, "loss": 0.8808, "step": 5743 }, { "epoch": 2.6026280018124153, "grad_norm": 0.31458923048303006, "learning_rate": 4.4111498686485164e-05, "loss": 0.8647, "step": 5744 }, { "epoch": 2.6030811055731764, "grad_norm": 0.22756201214572028, "learning_rate": 4.4098911822191655e-05, "loss": 0.8797, "step": 5745 }, { "epoch": 2.6035342093339375, "grad_norm": 0.2667084422529707, "learning_rate": 4.408632454771072e-05, "loss": 0.8551, "step": 5746 }, { "epoch": 2.6039873130946987, "grad_norm": 0.23773445892692738, "learning_rate": 4.407373686430199e-05, "loss": 0.8681, "step": 5747 }, { "epoch": 2.60444041685546, "grad_norm": 0.24139248445010003, "learning_rate": 4.4061148773225134e-05, "loss": 0.8768, "step": 5748 }, { "epoch": 2.604893520616221, "grad_norm": 0.2315900741107735, "learning_rate": 4.404856027573987e-05, "loss": 0.8722, "step": 5749 }, { "epoch": 2.605346624376982, "grad_norm": 0.23000528629632094, "learning_rate": 4.403597137310597e-05, "loss": 0.8834, "step": 5750 }, { "epoch": 2.6057997281377436, "grad_norm": 0.2081107213249972, "learning_rate": 4.402338206658322e-05, "loss": 0.8617, "step": 5751 }, { "epoch": 2.6062528318985048, "grad_norm": 0.20904311964833555, "learning_rate": 4.4010792357431485e-05, "loss": 0.8677, "step": 5752 }, { "epoch": 2.606705935659266, "grad_norm": 0.23543010245220433, "learning_rate": 4.399820224691062e-05, "loss": 0.8735, "step": 5753 }, { "epoch": 2.607159039420027, "grad_norm": 0.2655837992541495, "learning_rate": 4.3985611736280555e-05, "loss": 0.8556, "step": 5754 }, { "epoch": 2.6076121431807886, "grad_norm": 0.30498781303810557, "learning_rate": 4.397302082680125e-05, "loss": 0.8612, "step": 5755 }, { "epoch": 2.6080652469415497, "grad_norm": 0.30908439694199125, "learning_rate": 4.3960429519732714e-05, "loss": 0.8764, "step": 5756 }, { "epoch": 2.608518350702311, "grad_norm": 0.28310779397730584, "learning_rate": 4.394783781633498e-05, "loss": 0.8776, "step": 5757 }, { "epoch": 2.608971454463072, "grad_norm": 0.21891884884387194, "learning_rate": 4.393524571786813e-05, "loss": 0.8732, "step": 5758 }, { "epoch": 2.609424558223833, "grad_norm": 0.23927341969861188, "learning_rate": 4.392265322559228e-05, "loss": 0.8875, "step": 5759 }, { "epoch": 2.6098776619845943, "grad_norm": 0.2516464891062393, "learning_rate": 4.39100603407676e-05, "loss": 0.8634, "step": 5760 }, { "epoch": 2.610330765745356, "grad_norm": 0.28078820289647916, "learning_rate": 4.389746706465427e-05, "loss": 0.8627, "step": 5761 }, { "epoch": 2.610783869506117, "grad_norm": 0.22504456142224308, "learning_rate": 4.3884873398512546e-05, "loss": 0.8875, "step": 5762 }, { "epoch": 2.611236973266878, "grad_norm": 0.2539466769049345, "learning_rate": 4.3872279343602695e-05, "loss": 0.8801, "step": 5763 }, { "epoch": 2.6116900770276392, "grad_norm": 0.33252947198468674, "learning_rate": 4.385968490118505e-05, "loss": 0.8717, "step": 5764 }, { "epoch": 2.612143180788401, "grad_norm": 0.30201733928061747, "learning_rate": 4.384709007251994e-05, "loss": 0.8924, "step": 5765 }, { "epoch": 2.612596284549162, "grad_norm": 0.27245376720293746, "learning_rate": 4.383449485886777e-05, "loss": 0.8504, "step": 5766 }, { "epoch": 2.613049388309923, "grad_norm": 0.26815838782622214, "learning_rate": 4.3821899261488976e-05, "loss": 0.8629, "step": 5767 }, { "epoch": 2.613502492070684, "grad_norm": 0.2643866450171154, "learning_rate": 4.380930328164402e-05, "loss": 0.8796, "step": 5768 }, { "epoch": 2.6139555958314453, "grad_norm": 0.3383116682172315, "learning_rate": 4.3796706920593405e-05, "loss": 0.8705, "step": 5769 }, { "epoch": 2.6144086995922065, "grad_norm": 0.2797085076850769, "learning_rate": 4.3784110179597704e-05, "loss": 0.8669, "step": 5770 }, { "epoch": 2.6148618033529676, "grad_norm": 0.24613458002436545, "learning_rate": 4.3771513059917486e-05, "loss": 0.8739, "step": 5771 }, { "epoch": 2.615314907113729, "grad_norm": 0.3022181177385973, "learning_rate": 4.375891556281335e-05, "loss": 0.8829, "step": 5772 }, { "epoch": 2.6157680108744903, "grad_norm": 0.22879706199156463, "learning_rate": 4.3746317689546e-05, "loss": 0.8686, "step": 5773 }, { "epoch": 2.6162211146352514, "grad_norm": 0.39484336912347867, "learning_rate": 4.3733719441376105e-05, "loss": 0.8609, "step": 5774 }, { "epoch": 2.6166742183960126, "grad_norm": 0.34757376446754845, "learning_rate": 4.3721120819564406e-05, "loss": 0.8778, "step": 5775 }, { "epoch": 2.617127322156774, "grad_norm": 0.3209930015007411, "learning_rate": 4.370852182537167e-05, "loss": 0.875, "step": 5776 }, { "epoch": 2.6175804259175353, "grad_norm": 0.25871752928370007, "learning_rate": 4.369592246005872e-05, "loss": 0.887, "step": 5777 }, { "epoch": 2.6180335296782964, "grad_norm": 0.28978238251416316, "learning_rate": 4.368332272488639e-05, "loss": 0.8616, "step": 5778 }, { "epoch": 2.6184866334390575, "grad_norm": 0.3155357252815631, "learning_rate": 4.367072262111558e-05, "loss": 0.8694, "step": 5779 }, { "epoch": 2.6189397371998187, "grad_norm": 0.21028105013503592, "learning_rate": 4.365812215000719e-05, "loss": 0.8819, "step": 5780 }, { "epoch": 2.61939284096058, "grad_norm": 0.3317971508921712, "learning_rate": 4.36455213128222e-05, "loss": 0.8971, "step": 5781 }, { "epoch": 2.619845944721341, "grad_norm": 0.3347032897117345, "learning_rate": 4.36329201108216e-05, "loss": 0.8526, "step": 5782 }, { "epoch": 2.6202990484821025, "grad_norm": 0.32502315970054496, "learning_rate": 4.362031854526641e-05, "loss": 0.8777, "step": 5783 }, { "epoch": 2.6207521522428636, "grad_norm": 0.3353395556728812, "learning_rate": 4.36077166174177e-05, "loss": 0.8689, "step": 5784 }, { "epoch": 2.6212052560036248, "grad_norm": 0.31527099056225605, "learning_rate": 4.359511432853658e-05, "loss": 0.8745, "step": 5785 }, { "epoch": 2.621658359764386, "grad_norm": 0.24098721908914134, "learning_rate": 4.358251167988418e-05, "loss": 0.8745, "step": 5786 }, { "epoch": 2.6221114635251475, "grad_norm": 0.34437982031627545, "learning_rate": 4.356990867272168e-05, "loss": 0.8574, "step": 5787 }, { "epoch": 2.6225645672859086, "grad_norm": 0.2912668388334511, "learning_rate": 4.3557305308310293e-05, "loss": 0.8771, "step": 5788 }, { "epoch": 2.6230176710466697, "grad_norm": 0.41677716086163447, "learning_rate": 4.354470158791127e-05, "loss": 0.8959, "step": 5789 }, { "epoch": 2.623470774807431, "grad_norm": 0.36927167443750997, "learning_rate": 4.353209751278591e-05, "loss": 0.8866, "step": 5790 }, { "epoch": 2.623923878568192, "grad_norm": 0.2730651461354044, "learning_rate": 4.351949308419549e-05, "loss": 0.8746, "step": 5791 }, { "epoch": 2.624376982328953, "grad_norm": 0.38951800675422504, "learning_rate": 4.350688830340139e-05, "loss": 0.8556, "step": 5792 }, { "epoch": 2.6248300860897147, "grad_norm": 0.33866537255679885, "learning_rate": 4.349428317166499e-05, "loss": 0.8799, "step": 5793 }, { "epoch": 2.625283189850476, "grad_norm": 0.2879880918478578, "learning_rate": 4.348167769024773e-05, "loss": 0.8648, "step": 5794 }, { "epoch": 2.625736293611237, "grad_norm": 0.3742849612052289, "learning_rate": 4.3469071860411054e-05, "loss": 0.8846, "step": 5795 }, { "epoch": 2.626189397371998, "grad_norm": 0.33007315377654056, "learning_rate": 4.345646568341646e-05, "loss": 0.8649, "step": 5796 }, { "epoch": 2.6266425011327597, "grad_norm": 0.294687037789201, "learning_rate": 4.3443859160525477e-05, "loss": 0.8504, "step": 5797 }, { "epoch": 2.627095604893521, "grad_norm": 0.3498927445343495, "learning_rate": 4.343125229299968e-05, "loss": 0.8649, "step": 5798 }, { "epoch": 2.627548708654282, "grad_norm": 0.31353506827358313, "learning_rate": 4.341864508210065e-05, "loss": 0.8584, "step": 5799 }, { "epoch": 2.628001812415043, "grad_norm": 0.2947554848459301, "learning_rate": 4.340603752909003e-05, "loss": 0.8604, "step": 5800 }, { "epoch": 2.628454916175804, "grad_norm": 0.3494772144551321, "learning_rate": 4.3393429635229475e-05, "loss": 0.8846, "step": 5801 }, { "epoch": 2.6289080199365653, "grad_norm": 0.3967687356149581, "learning_rate": 4.338082140178069e-05, "loss": 0.8625, "step": 5802 }, { "epoch": 2.6293611236973264, "grad_norm": 0.28161916764336076, "learning_rate": 4.336821283000541e-05, "loss": 0.8619, "step": 5803 }, { "epoch": 2.629814227458088, "grad_norm": 0.2927572454782279, "learning_rate": 4.33556039211654e-05, "loss": 0.8849, "step": 5804 }, { "epoch": 2.630267331218849, "grad_norm": 0.4772721853547909, "learning_rate": 4.334299467652248e-05, "loss": 0.8983, "step": 5805 }, { "epoch": 2.6307204349796103, "grad_norm": 0.43545303598164975, "learning_rate": 4.333038509733848e-05, "loss": 0.8727, "step": 5806 }, { "epoch": 2.6311735387403714, "grad_norm": 0.24578320658903385, "learning_rate": 4.331777518487524e-05, "loss": 0.8726, "step": 5807 }, { "epoch": 2.631626642501133, "grad_norm": 0.3650801577994707, "learning_rate": 4.3305164940394716e-05, "loss": 0.9011, "step": 5808 }, { "epoch": 2.632079746261894, "grad_norm": 0.3488019048679638, "learning_rate": 4.329255436515879e-05, "loss": 0.8955, "step": 5809 }, { "epoch": 2.6325328500226552, "grad_norm": 0.2691748316852804, "learning_rate": 4.327994346042946e-05, "loss": 0.857, "step": 5810 }, { "epoch": 2.6329859537834164, "grad_norm": 0.2613253803545095, "learning_rate": 4.326733222746872e-05, "loss": 0.8573, "step": 5811 }, { "epoch": 2.6334390575441775, "grad_norm": 0.3055455265964215, "learning_rate": 4.325472066753861e-05, "loss": 0.8662, "step": 5812 }, { "epoch": 2.6338921613049386, "grad_norm": 0.24029252138264123, "learning_rate": 4.324210878190119e-05, "loss": 0.8693, "step": 5813 }, { "epoch": 2.6343452650656998, "grad_norm": 0.305493058069059, "learning_rate": 4.322949657181857e-05, "loss": 0.884, "step": 5814 }, { "epoch": 2.6347983688264613, "grad_norm": 0.32689438409570737, "learning_rate": 4.321688403855287e-05, "loss": 0.8637, "step": 5815 }, { "epoch": 2.6352514725872225, "grad_norm": 0.2391447211897717, "learning_rate": 4.3204271183366285e-05, "loss": 0.8866, "step": 5816 }, { "epoch": 2.6357045763479836, "grad_norm": 0.31722679177357904, "learning_rate": 4.319165800752097e-05, "loss": 0.89, "step": 5817 }, { "epoch": 2.6361576801087447, "grad_norm": 0.33450577413165683, "learning_rate": 4.317904451227917e-05, "loss": 0.8743, "step": 5818 }, { "epoch": 2.6366107838695063, "grad_norm": 0.24435802450835462, "learning_rate": 4.3166430698903164e-05, "loss": 0.8615, "step": 5819 }, { "epoch": 2.6370638876302674, "grad_norm": 0.24507327835075365, "learning_rate": 4.315381656865522e-05, "loss": 0.8767, "step": 5820 }, { "epoch": 2.6375169913910286, "grad_norm": 0.3329618487163967, "learning_rate": 4.314120212279767e-05, "loss": 0.8698, "step": 5821 }, { "epoch": 2.6379700951517897, "grad_norm": 0.29615691092967916, "learning_rate": 4.312858736259288e-05, "loss": 0.891, "step": 5822 }, { "epoch": 2.638423198912551, "grad_norm": 0.20305248962935332, "learning_rate": 4.311597228930323e-05, "loss": 0.881, "step": 5823 }, { "epoch": 2.638876302673312, "grad_norm": 0.26524175426240176, "learning_rate": 4.3103356904191135e-05, "loss": 0.8688, "step": 5824 }, { "epoch": 2.6393294064340735, "grad_norm": 0.2678894204156909, "learning_rate": 4.309074120851907e-05, "loss": 0.8694, "step": 5825 }, { "epoch": 2.6397825101948347, "grad_norm": 0.3137017067472724, "learning_rate": 4.307812520354948e-05, "loss": 0.8641, "step": 5826 }, { "epoch": 2.640235613955596, "grad_norm": 0.3156565817347327, "learning_rate": 4.3065508890544906e-05, "loss": 0.8566, "step": 5827 }, { "epoch": 2.640688717716357, "grad_norm": 0.27575918546127864, "learning_rate": 4.305289227076787e-05, "loss": 0.8695, "step": 5828 }, { "epoch": 2.6411418214771185, "grad_norm": 0.23220884721798285, "learning_rate": 4.304027534548096e-05, "loss": 0.8794, "step": 5829 }, { "epoch": 2.6415949252378796, "grad_norm": 0.21830447798938726, "learning_rate": 4.302765811594678e-05, "loss": 0.8646, "step": 5830 }, { "epoch": 2.6420480289986408, "grad_norm": 0.25880283432709167, "learning_rate": 4.3015040583427956e-05, "loss": 0.886, "step": 5831 }, { "epoch": 2.642501132759402, "grad_norm": 0.22504081889451719, "learning_rate": 4.3002422749187144e-05, "loss": 0.8818, "step": 5832 }, { "epoch": 2.642954236520163, "grad_norm": 0.28418693667372846, "learning_rate": 4.2989804614487076e-05, "loss": 0.8684, "step": 5833 }, { "epoch": 2.643407340280924, "grad_norm": 0.25587051065363753, "learning_rate": 4.297718618059044e-05, "loss": 0.8682, "step": 5834 }, { "epoch": 2.6438604440416853, "grad_norm": 0.2628359934709993, "learning_rate": 4.296456744876002e-05, "loss": 0.8499, "step": 5835 }, { "epoch": 2.644313547802447, "grad_norm": 0.2335540116869879, "learning_rate": 4.295194842025858e-05, "loss": 0.869, "step": 5836 }, { "epoch": 2.644766651563208, "grad_norm": 0.23935604235281902, "learning_rate": 4.293932909634894e-05, "loss": 0.8819, "step": 5837 }, { "epoch": 2.645219755323969, "grad_norm": 0.21980821912633794, "learning_rate": 4.2926709478293954e-05, "loss": 0.8593, "step": 5838 }, { "epoch": 2.6456728590847303, "grad_norm": 0.257006525163332, "learning_rate": 4.291408956735648e-05, "loss": 0.8661, "step": 5839 }, { "epoch": 2.646125962845492, "grad_norm": 0.2905342778260467, "learning_rate": 4.290146936479943e-05, "loss": 0.8703, "step": 5840 }, { "epoch": 2.646579066606253, "grad_norm": 0.35619428000490894, "learning_rate": 4.288884887188574e-05, "loss": 0.8844, "step": 5841 }, { "epoch": 2.647032170367014, "grad_norm": 0.28924469797567703, "learning_rate": 4.2876228089878366e-05, "loss": 0.8768, "step": 5842 }, { "epoch": 2.6474852741277752, "grad_norm": 0.2824283449398914, "learning_rate": 4.286360702004031e-05, "loss": 0.8752, "step": 5843 }, { "epoch": 2.6479383778885364, "grad_norm": 0.3733391838514582, "learning_rate": 4.285098566363457e-05, "loss": 0.8659, "step": 5844 }, { "epoch": 2.6483914816492975, "grad_norm": 0.284327201455786, "learning_rate": 4.2838364021924216e-05, "loss": 0.8786, "step": 5845 }, { "epoch": 2.6488445854100586, "grad_norm": 0.2821923591551608, "learning_rate": 4.282574209617231e-05, "loss": 0.8866, "step": 5846 }, { "epoch": 2.64929768917082, "grad_norm": 0.24824922679479489, "learning_rate": 4.281311988764196e-05, "loss": 0.8792, "step": 5847 }, { "epoch": 2.6497507929315813, "grad_norm": 0.27646558322397635, "learning_rate": 4.280049739759629e-05, "loss": 0.8771, "step": 5848 }, { "epoch": 2.6502038966923425, "grad_norm": 0.27798415919618014, "learning_rate": 4.278787462729848e-05, "loss": 0.8776, "step": 5849 }, { "epoch": 2.6506570004531036, "grad_norm": 0.305986717449478, "learning_rate": 4.277525157801171e-05, "loss": 0.8736, "step": 5850 }, { "epoch": 2.651110104213865, "grad_norm": 0.33550748521782875, "learning_rate": 4.2762628250999206e-05, "loss": 0.8887, "step": 5851 }, { "epoch": 2.6515632079746263, "grad_norm": 0.2976445109372052, "learning_rate": 4.2750004647524196e-05, "loss": 0.8741, "step": 5852 }, { "epoch": 2.6520163117353874, "grad_norm": 0.2999910446477184, "learning_rate": 4.273738076884996e-05, "loss": 0.8573, "step": 5853 }, { "epoch": 2.6524694154961486, "grad_norm": 0.24808562181406446, "learning_rate": 4.2724756616239794e-05, "loss": 0.8833, "step": 5854 }, { "epoch": 2.6529225192569097, "grad_norm": 0.29972247644846295, "learning_rate": 4.271213219095704e-05, "loss": 0.8817, "step": 5855 }, { "epoch": 2.653375623017671, "grad_norm": 0.3481936011750187, "learning_rate": 4.269950749426505e-05, "loss": 0.8686, "step": 5856 }, { "epoch": 2.6538287267784324, "grad_norm": 0.2977897157697596, "learning_rate": 4.268688252742719e-05, "loss": 0.8785, "step": 5857 }, { "epoch": 2.6542818305391935, "grad_norm": 0.2713652395252644, "learning_rate": 4.267425729170688e-05, "loss": 0.8516, "step": 5858 }, { "epoch": 2.6547349342999547, "grad_norm": 0.2592761110759463, "learning_rate": 4.2661631788367564e-05, "loss": 0.8592, "step": 5859 }, { "epoch": 2.655188038060716, "grad_norm": 0.25879137167687705, "learning_rate": 4.264900601867269e-05, "loss": 0.9033, "step": 5860 }, { "epoch": 2.6556411418214774, "grad_norm": 0.26685972746211567, "learning_rate": 4.2636379983885756e-05, "loss": 0.8539, "step": 5861 }, { "epoch": 2.6560942455822385, "grad_norm": 0.30836852834676226, "learning_rate": 4.262375368527028e-05, "loss": 0.8737, "step": 5862 }, { "epoch": 2.6565473493429996, "grad_norm": 0.29776956557051926, "learning_rate": 4.26111271240898e-05, "loss": 0.8903, "step": 5863 }, { "epoch": 2.6570004531037608, "grad_norm": 0.3189200072539116, "learning_rate": 4.259850030160788e-05, "loss": 0.8658, "step": 5864 }, { "epoch": 2.657453556864522, "grad_norm": 0.347746026133354, "learning_rate": 4.2585873219088115e-05, "loss": 0.8836, "step": 5865 }, { "epoch": 2.657906660625283, "grad_norm": 0.37506291231815503, "learning_rate": 4.257324587779414e-05, "loss": 0.8792, "step": 5866 }, { "epoch": 2.658359764386044, "grad_norm": 0.3129372499853787, "learning_rate": 4.2560618278989596e-05, "loss": 0.8616, "step": 5867 }, { "epoch": 2.6588128681468057, "grad_norm": 0.26459798192322725, "learning_rate": 4.2547990423938146e-05, "loss": 0.869, "step": 5868 }, { "epoch": 2.659265971907567, "grad_norm": 0.3064889122373345, "learning_rate": 4.2535362313903494e-05, "loss": 0.8908, "step": 5869 }, { "epoch": 2.659719075668328, "grad_norm": 0.28575453731054934, "learning_rate": 4.252273395014937e-05, "loss": 0.8635, "step": 5870 }, { "epoch": 2.660172179429089, "grad_norm": 0.23779403796944248, "learning_rate": 4.251010533393951e-05, "loss": 0.8785, "step": 5871 }, { "epoch": 2.6606252831898507, "grad_norm": 0.27549681013452787, "learning_rate": 4.24974764665377e-05, "loss": 0.8743, "step": 5872 }, { "epoch": 2.661078386950612, "grad_norm": 0.2265452488002809, "learning_rate": 4.2484847349207736e-05, "loss": 0.8875, "step": 5873 }, { "epoch": 2.661531490711373, "grad_norm": 0.3214127606263601, "learning_rate": 4.247221798321344e-05, "loss": 0.8737, "step": 5874 }, { "epoch": 2.661984594472134, "grad_norm": 0.3208302045707578, "learning_rate": 4.245958836981866e-05, "loss": 0.8734, "step": 5875 }, { "epoch": 2.6624376982328952, "grad_norm": 0.21109848948259066, "learning_rate": 4.244695851028728e-05, "loss": 0.8633, "step": 5876 }, { "epoch": 2.6628908019936564, "grad_norm": 0.2498512345447968, "learning_rate": 4.24343284058832e-05, "loss": 0.8964, "step": 5877 }, { "epoch": 2.6633439057544175, "grad_norm": 0.23178452810616296, "learning_rate": 4.242169805787033e-05, "loss": 0.857, "step": 5878 }, { "epoch": 2.663797009515179, "grad_norm": 0.21211492827008221, "learning_rate": 4.240906746751261e-05, "loss": 0.8753, "step": 5879 }, { "epoch": 2.66425011327594, "grad_norm": 0.21991696279501285, "learning_rate": 4.239643663607403e-05, "loss": 0.872, "step": 5880 }, { "epoch": 2.6647032170367013, "grad_norm": 0.23264387612213644, "learning_rate": 4.2383805564818584e-05, "loss": 0.8871, "step": 5881 }, { "epoch": 2.6651563207974625, "grad_norm": 0.23398320946530576, "learning_rate": 4.237117425501029e-05, "loss": 0.872, "step": 5882 }, { "epoch": 2.665609424558224, "grad_norm": 0.38106657054323023, "learning_rate": 4.235854270791319e-05, "loss": 0.89, "step": 5883 }, { "epoch": 2.666062528318985, "grad_norm": 0.2354090268066528, "learning_rate": 4.2345910924791356e-05, "loss": 0.862, "step": 5884 }, { "epoch": 2.6665156320797463, "grad_norm": 0.22613041294198205, "learning_rate": 4.2333278906908876e-05, "loss": 0.8656, "step": 5885 }, { "epoch": 2.6669687358405074, "grad_norm": 0.2574504808425522, "learning_rate": 4.2320646655529865e-05, "loss": 0.8844, "step": 5886 }, { "epoch": 2.6674218396012686, "grad_norm": 0.2150421911026645, "learning_rate": 4.230801417191847e-05, "loss": 0.8683, "step": 5887 }, { "epoch": 2.6678749433620297, "grad_norm": 0.22410671030790774, "learning_rate": 4.229538145733883e-05, "loss": 0.8603, "step": 5888 }, { "epoch": 2.6683280471227913, "grad_norm": 0.23868132538452697, "learning_rate": 4.2282748513055146e-05, "loss": 0.8788, "step": 5889 }, { "epoch": 2.6687811508835524, "grad_norm": 0.24298732967011435, "learning_rate": 4.227011534033162e-05, "loss": 0.8824, "step": 5890 }, { "epoch": 2.6692342546443135, "grad_norm": 0.20730245477202497, "learning_rate": 4.2257481940432487e-05, "loss": 0.8584, "step": 5891 }, { "epoch": 2.6696873584050747, "grad_norm": 0.24123092781328256, "learning_rate": 4.2244848314621996e-05, "loss": 0.8955, "step": 5892 }, { "epoch": 2.6701404621658362, "grad_norm": 0.2524666860379319, "learning_rate": 4.223221446416442e-05, "loss": 0.869, "step": 5893 }, { "epoch": 2.6705935659265974, "grad_norm": 0.24804118312382173, "learning_rate": 4.221958039032407e-05, "loss": 0.8511, "step": 5894 }, { "epoch": 2.6710466696873585, "grad_norm": 0.2387132787573057, "learning_rate": 4.2206946094365254e-05, "loss": 0.8885, "step": 5895 }, { "epoch": 2.6714997734481196, "grad_norm": 0.21617368599573608, "learning_rate": 4.219431157755232e-05, "loss": 0.8879, "step": 5896 }, { "epoch": 2.6719528772088808, "grad_norm": 0.22450768452178224, "learning_rate": 4.2181676841149643e-05, "loss": 0.8659, "step": 5897 }, { "epoch": 2.672405980969642, "grad_norm": 0.27875446385698344, "learning_rate": 4.2169041886421586e-05, "loss": 0.8721, "step": 5898 }, { "epoch": 2.672859084730403, "grad_norm": 0.37301498985961834, "learning_rate": 4.2156406714632575e-05, "loss": 0.8655, "step": 5899 }, { "epoch": 2.6733121884911646, "grad_norm": 0.24999276657594124, "learning_rate": 4.2143771327047026e-05, "loss": 0.8911, "step": 5900 }, { "epoch": 2.6737652922519257, "grad_norm": 0.2650013843669078, "learning_rate": 4.2131135724929404e-05, "loss": 0.8757, "step": 5901 }, { "epoch": 2.674218396012687, "grad_norm": 0.2324882953246424, "learning_rate": 4.2118499909544186e-05, "loss": 0.8987, "step": 5902 }, { "epoch": 2.674671499773448, "grad_norm": 0.2220448603592014, "learning_rate": 4.210586388215586e-05, "loss": 0.8722, "step": 5903 }, { "epoch": 2.6751246035342096, "grad_norm": 0.2865737749001807, "learning_rate": 4.2093227644028944e-05, "loss": 0.8564, "step": 5904 }, { "epoch": 2.6755777072949707, "grad_norm": 0.2389952067228228, "learning_rate": 4.208059119642797e-05, "loss": 0.8826, "step": 5905 }, { "epoch": 2.676030811055732, "grad_norm": 0.24589659852969126, "learning_rate": 4.206795454061749e-05, "loss": 0.874, "step": 5906 }, { "epoch": 2.676483914816493, "grad_norm": 0.29161510725816847, "learning_rate": 4.205531767786211e-05, "loss": 0.8926, "step": 5907 }, { "epoch": 2.676937018577254, "grad_norm": 0.26649367144473723, "learning_rate": 4.20426806094264e-05, "loss": 0.8585, "step": 5908 }, { "epoch": 2.677390122338015, "grad_norm": 0.2689170830580432, "learning_rate": 4.2030043336574986e-05, "loss": 0.8788, "step": 5909 }, { "epoch": 2.677843226098777, "grad_norm": 0.31008185488333295, "learning_rate": 4.201740586057252e-05, "loss": 0.8704, "step": 5910 }, { "epoch": 2.678296329859538, "grad_norm": 0.30054256036268867, "learning_rate": 4.200476818268366e-05, "loss": 0.8433, "step": 5911 }, { "epoch": 2.678749433620299, "grad_norm": 0.3206908507586916, "learning_rate": 4.199213030417309e-05, "loss": 0.8511, "step": 5912 }, { "epoch": 2.67920253738106, "grad_norm": 0.3538962777669488, "learning_rate": 4.1979492226305504e-05, "loss": 0.8749, "step": 5913 }, { "epoch": 2.6796556411418218, "grad_norm": 0.30706449607993325, "learning_rate": 4.1966853950345635e-05, "loss": 0.8802, "step": 5914 }, { "epoch": 2.680108744902583, "grad_norm": 0.27526209964196635, "learning_rate": 4.195421547755819e-05, "loss": 0.8795, "step": 5915 }, { "epoch": 2.680561848663344, "grad_norm": 0.2764625221035011, "learning_rate": 4.194157680920797e-05, "loss": 0.8719, "step": 5916 }, { "epoch": 2.681014952424105, "grad_norm": 0.2913229141330508, "learning_rate": 4.192893794655973e-05, "loss": 0.8555, "step": 5917 }, { "epoch": 2.6814680561848663, "grad_norm": 0.358243311234618, "learning_rate": 4.1916298890878285e-05, "loss": 0.8515, "step": 5918 }, { "epoch": 2.6819211599456274, "grad_norm": 0.38291331443217963, "learning_rate": 4.190365964342844e-05, "loss": 0.8882, "step": 5919 }, { "epoch": 2.6823742637063885, "grad_norm": 0.37199463520493803, "learning_rate": 4.189102020547505e-05, "loss": 0.868, "step": 5920 }, { "epoch": 2.68282736746715, "grad_norm": 0.28504364617001815, "learning_rate": 4.187838057828297e-05, "loss": 0.8819, "step": 5921 }, { "epoch": 2.6832804712279112, "grad_norm": 0.22858047668769074, "learning_rate": 4.186574076311706e-05, "loss": 0.8584, "step": 5922 }, { "epoch": 2.6837335749886724, "grad_norm": 0.262548590389573, "learning_rate": 4.185310076124223e-05, "loss": 0.8645, "step": 5923 }, { "epoch": 2.6841866787494335, "grad_norm": 0.3911337659197468, "learning_rate": 4.1840460573923387e-05, "loss": 0.8926, "step": 5924 }, { "epoch": 2.684639782510195, "grad_norm": 0.41618639993081313, "learning_rate": 4.182782020242546e-05, "loss": 0.8555, "step": 5925 }, { "epoch": 2.685092886270956, "grad_norm": 0.33043084642555764, "learning_rate": 4.181517964801341e-05, "loss": 0.8822, "step": 5926 }, { "epoch": 2.6855459900317173, "grad_norm": 0.22599479045132936, "learning_rate": 4.18025389119522e-05, "loss": 0.8621, "step": 5927 }, { "epoch": 2.6859990937924785, "grad_norm": 0.2740248774938987, "learning_rate": 4.1789897995506814e-05, "loss": 0.8797, "step": 5928 }, { "epoch": 2.6864521975532396, "grad_norm": 0.2863513000623292, "learning_rate": 4.1777256899942274e-05, "loss": 0.8521, "step": 5929 }, { "epoch": 2.6869053013140007, "grad_norm": 0.26886945079644126, "learning_rate": 4.1764615626523584e-05, "loss": 0.872, "step": 5930 }, { "epoch": 2.687358405074762, "grad_norm": 0.2652398690716769, "learning_rate": 4.175197417651579e-05, "loss": 0.8613, "step": 5931 }, { "epoch": 2.6878115088355234, "grad_norm": 0.2553422732533474, "learning_rate": 4.173933255118395e-05, "loss": 0.8666, "step": 5932 }, { "epoch": 2.6882646125962846, "grad_norm": 0.22486690406341436, "learning_rate": 4.1726690751793146e-05, "loss": 0.8735, "step": 5933 }, { "epoch": 2.6887177163570457, "grad_norm": 0.2854684589764244, "learning_rate": 4.1714048779608466e-05, "loss": 0.8512, "step": 5934 }, { "epoch": 2.689170820117807, "grad_norm": 0.30699115072672417, "learning_rate": 4.1701406635895027e-05, "loss": 0.8785, "step": 5935 }, { "epoch": 2.6896239238785684, "grad_norm": 0.2544550456883686, "learning_rate": 4.168876432191795e-05, "loss": 0.8786, "step": 5936 }, { "epoch": 2.6900770276393295, "grad_norm": 0.2407938555387644, "learning_rate": 4.167612183894238e-05, "loss": 0.8725, "step": 5937 }, { "epoch": 2.6905301314000907, "grad_norm": 0.2626659934427123, "learning_rate": 4.166347918823349e-05, "loss": 0.8762, "step": 5938 }, { "epoch": 2.690983235160852, "grad_norm": 0.29538961046441353, "learning_rate": 4.165083637105645e-05, "loss": 0.8822, "step": 5939 }, { "epoch": 2.691436338921613, "grad_norm": 0.2331550335918487, "learning_rate": 4.1638193388676455e-05, "loss": 0.8663, "step": 5940 }, { "epoch": 2.691889442682374, "grad_norm": 0.37891366803108295, "learning_rate": 4.162555024235872e-05, "loss": 0.8852, "step": 5941 }, { "epoch": 2.6923425464431356, "grad_norm": 0.40131592056320375, "learning_rate": 4.161290693336848e-05, "loss": 0.8933, "step": 5942 }, { "epoch": 2.6927956502038968, "grad_norm": 0.2878126650966456, "learning_rate": 4.1600263462970964e-05, "loss": 0.8527, "step": 5943 }, { "epoch": 2.693248753964658, "grad_norm": 0.25528942367855023, "learning_rate": 4.1587619832431434e-05, "loss": 0.8835, "step": 5944 }, { "epoch": 2.693701857725419, "grad_norm": 0.2842500064336783, "learning_rate": 4.157497604301518e-05, "loss": 0.8634, "step": 5945 }, { "epoch": 2.6941549614861806, "grad_norm": 0.27806847997490486, "learning_rate": 4.156233209598749e-05, "loss": 0.869, "step": 5946 }, { "epoch": 2.6946080652469417, "grad_norm": 0.2690591815352601, "learning_rate": 4.154968799261367e-05, "loss": 0.8647, "step": 5947 }, { "epoch": 2.695061169007703, "grad_norm": 0.24851178229157733, "learning_rate": 4.153704373415905e-05, "loss": 0.8617, "step": 5948 }, { "epoch": 2.695514272768464, "grad_norm": 0.25869227195003786, "learning_rate": 4.1524399321888955e-05, "loss": 0.8644, "step": 5949 }, { "epoch": 2.695967376529225, "grad_norm": 0.24712816687596315, "learning_rate": 4.151175475706876e-05, "loss": 0.8664, "step": 5950 }, { "epoch": 2.6964204802899863, "grad_norm": 0.27871176601125164, "learning_rate": 4.149911004096382e-05, "loss": 0.8572, "step": 5951 }, { "epoch": 2.6968735840507474, "grad_norm": 0.2506867109308759, "learning_rate": 4.1486465174839514e-05, "loss": 0.8786, "step": 5952 }, { "epoch": 2.697326687811509, "grad_norm": 0.254474945659635, "learning_rate": 4.147382015996127e-05, "loss": 0.8783, "step": 5953 }, { "epoch": 2.69777979157227, "grad_norm": 0.22696760810844333, "learning_rate": 4.146117499759449e-05, "loss": 0.8646, "step": 5954 }, { "epoch": 2.6982328953330312, "grad_norm": 0.2783042161559857, "learning_rate": 4.144852968900459e-05, "loss": 0.8533, "step": 5955 }, { "epoch": 2.6986859990937924, "grad_norm": 0.21959327767352865, "learning_rate": 4.143588423545704e-05, "loss": 0.872, "step": 5956 }, { "epoch": 2.699139102854554, "grad_norm": 0.2924363408073572, "learning_rate": 4.142323863821727e-05, "loss": 0.8671, "step": 5957 }, { "epoch": 2.699592206615315, "grad_norm": 0.28194202310053074, "learning_rate": 4.141059289855078e-05, "loss": 0.8687, "step": 5958 }, { "epoch": 2.700045310376076, "grad_norm": 0.2300536409462394, "learning_rate": 4.1397947017723054e-05, "loss": 0.8829, "step": 5959 }, { "epoch": 2.7004984141368373, "grad_norm": 0.2898422035374467, "learning_rate": 4.138530099699958e-05, "loss": 0.8612, "step": 5960 }, { "epoch": 2.7009515178975985, "grad_norm": 0.2533972753575715, "learning_rate": 4.137265483764588e-05, "loss": 0.8573, "step": 5961 }, { "epoch": 2.7014046216583596, "grad_norm": 0.30157628289196736, "learning_rate": 4.136000854092749e-05, "loss": 0.854, "step": 5962 }, { "epoch": 2.7018577254191207, "grad_norm": 0.3419747126967074, "learning_rate": 4.134736210810995e-05, "loss": 0.8669, "step": 5963 }, { "epoch": 2.7023108291798823, "grad_norm": 0.3716709757803535, "learning_rate": 4.1334715540458815e-05, "loss": 0.8784, "step": 5964 }, { "epoch": 2.7027639329406434, "grad_norm": 0.2896088796093044, "learning_rate": 4.132206883923966e-05, "loss": 0.8727, "step": 5965 }, { "epoch": 2.7032170367014046, "grad_norm": 0.24400793589363823, "learning_rate": 4.130942200571808e-05, "loss": 0.8706, "step": 5966 }, { "epoch": 2.7036701404621657, "grad_norm": 0.2864924747979619, "learning_rate": 4.1296775041159634e-05, "loss": 0.884, "step": 5967 }, { "epoch": 2.7041232442229273, "grad_norm": 0.30365849393376354, "learning_rate": 4.128412794682997e-05, "loss": 0.8666, "step": 5968 }, { "epoch": 2.7045763479836884, "grad_norm": 0.39467269678804817, "learning_rate": 4.12714807239947e-05, "loss": 0.8622, "step": 5969 }, { "epoch": 2.7050294517444495, "grad_norm": 0.38293983627929645, "learning_rate": 4.1258833373919465e-05, "loss": 0.8614, "step": 5970 }, { "epoch": 2.7054825555052107, "grad_norm": 0.2796589559973184, "learning_rate": 4.1246185897869904e-05, "loss": 0.8715, "step": 5971 }, { "epoch": 2.705935659265972, "grad_norm": 0.2767343294552254, "learning_rate": 4.123353829711168e-05, "loss": 0.889, "step": 5972 }, { "epoch": 2.706388763026733, "grad_norm": 0.3985629227034665, "learning_rate": 4.1220890572910476e-05, "loss": 0.8638, "step": 5973 }, { "epoch": 2.7068418667874945, "grad_norm": 0.2959593900549946, "learning_rate": 4.120824272653197e-05, "loss": 0.8597, "step": 5974 }, { "epoch": 2.7072949705482556, "grad_norm": 0.32230951927248847, "learning_rate": 4.119559475924187e-05, "loss": 0.8846, "step": 5975 }, { "epoch": 2.7077480743090168, "grad_norm": 0.2958646826783052, "learning_rate": 4.1182946672305874e-05, "loss": 0.8944, "step": 5976 }, { "epoch": 2.708201178069778, "grad_norm": 0.28962855455837977, "learning_rate": 4.117029846698971e-05, "loss": 0.8758, "step": 5977 }, { "epoch": 2.7086542818305395, "grad_norm": 0.36509758468907777, "learning_rate": 4.115765014455912e-05, "loss": 0.8785, "step": 5978 }, { "epoch": 2.7091073855913006, "grad_norm": 0.3171385829205896, "learning_rate": 4.1145001706279845e-05, "loss": 0.8744, "step": 5979 }, { "epoch": 2.7095604893520617, "grad_norm": 0.2831513657473552, "learning_rate": 4.113235315341764e-05, "loss": 0.8622, "step": 5980 }, { "epoch": 2.710013593112823, "grad_norm": 0.38075034244289135, "learning_rate": 4.111970448723828e-05, "loss": 0.8686, "step": 5981 }, { "epoch": 2.710466696873584, "grad_norm": 0.38814698908773965, "learning_rate": 4.110705570900754e-05, "loss": 0.8899, "step": 5982 }, { "epoch": 2.710919800634345, "grad_norm": 0.27945751538051905, "learning_rate": 4.1094406819991224e-05, "loss": 0.8555, "step": 5983 }, { "epoch": 2.7113729043951063, "grad_norm": 0.3621845294961135, "learning_rate": 4.1081757821455127e-05, "loss": 0.8862, "step": 5984 }, { "epoch": 2.711826008155868, "grad_norm": 0.26960797579331824, "learning_rate": 4.106910871466505e-05, "loss": 0.8549, "step": 5985 }, { "epoch": 2.712279111916629, "grad_norm": 0.295662126810483, "learning_rate": 4.105645950088684e-05, "loss": 0.8695, "step": 5986 }, { "epoch": 2.71273221567739, "grad_norm": 0.32003677492775573, "learning_rate": 4.1043810181386314e-05, "loss": 0.8643, "step": 5987 }, { "epoch": 2.713185319438151, "grad_norm": 0.3224307460780828, "learning_rate": 4.103116075742933e-05, "loss": 0.8703, "step": 5988 }, { "epoch": 2.713638423198913, "grad_norm": 0.3308595711521685, "learning_rate": 4.101851123028175e-05, "loss": 0.8562, "step": 5989 }, { "epoch": 2.714091526959674, "grad_norm": 0.252182255918661, "learning_rate": 4.1005861601209424e-05, "loss": 0.8682, "step": 5990 }, { "epoch": 2.714544630720435, "grad_norm": 0.315457417774626, "learning_rate": 4.0993211871478246e-05, "loss": 0.8752, "step": 5991 }, { "epoch": 2.714997734481196, "grad_norm": 0.4266149988049669, "learning_rate": 4.098056204235409e-05, "loss": 0.8563, "step": 5992 }, { "epoch": 2.7154508382419573, "grad_norm": 0.4218842000071335, "learning_rate": 4.096791211510286e-05, "loss": 0.9021, "step": 5993 }, { "epoch": 2.7159039420027185, "grad_norm": 0.435176944616699, "learning_rate": 4.095526209099046e-05, "loss": 0.8761, "step": 5994 }, { "epoch": 2.7163570457634796, "grad_norm": 0.3794940965249908, "learning_rate": 4.094261197128281e-05, "loss": 0.8915, "step": 5995 }, { "epoch": 2.716810149524241, "grad_norm": 0.3277314848703036, "learning_rate": 4.092996175724583e-05, "loss": 0.8811, "step": 5996 }, { "epoch": 2.7172632532850023, "grad_norm": 0.26317738937064394, "learning_rate": 4.091731145014547e-05, "loss": 0.874, "step": 5997 }, { "epoch": 2.7177163570457634, "grad_norm": 0.36702781057826406, "learning_rate": 4.0904661051247656e-05, "loss": 0.8705, "step": 5998 }, { "epoch": 2.7181694608065246, "grad_norm": 0.4155095672198218, "learning_rate": 4.0892010561818355e-05, "loss": 0.8467, "step": 5999 }, { "epoch": 2.718622564567286, "grad_norm": 0.4087155905515761, "learning_rate": 4.087935998312353e-05, "loss": 0.8567, "step": 6000 }, { "epoch": 2.7190756683280473, "grad_norm": 0.41395956596734984, "learning_rate": 4.086670931642916e-05, "loss": 0.8708, "step": 6001 }, { "epoch": 2.7195287720888084, "grad_norm": 0.32029610139700027, "learning_rate": 4.0854058563001195e-05, "loss": 0.8811, "step": 6002 }, { "epoch": 2.7199818758495695, "grad_norm": 0.23711238683365649, "learning_rate": 4.0841407724105655e-05, "loss": 0.8631, "step": 6003 }, { "epoch": 2.7204349796103307, "grad_norm": 0.26194481799014824, "learning_rate": 4.082875680100852e-05, "loss": 0.8767, "step": 6004 }, { "epoch": 2.720888083371092, "grad_norm": 0.3058333392326134, "learning_rate": 4.081610579497581e-05, "loss": 0.8694, "step": 6005 }, { "epoch": 2.7213411871318534, "grad_norm": 0.31952927680459353, "learning_rate": 4.0803454707273545e-05, "loss": 0.871, "step": 6006 }, { "epoch": 2.7217942908926145, "grad_norm": 0.2509486504258794, "learning_rate": 4.079080353916773e-05, "loss": 0.8574, "step": 6007 }, { "epoch": 2.7222473946533756, "grad_norm": 0.23338231131384873, "learning_rate": 4.077815229192441e-05, "loss": 0.8677, "step": 6008 }, { "epoch": 2.7227004984141367, "grad_norm": 0.2732281553173304, "learning_rate": 4.0765500966809614e-05, "loss": 0.8853, "step": 6009 }, { "epoch": 2.7231536021748983, "grad_norm": 0.38691646428764814, "learning_rate": 4.07528495650894e-05, "loss": 0.8886, "step": 6010 }, { "epoch": 2.7236067059356595, "grad_norm": 0.4864206812122278, "learning_rate": 4.0740198088029804e-05, "loss": 0.8595, "step": 6011 }, { "epoch": 2.7240598096964206, "grad_norm": 0.4796710557265463, "learning_rate": 4.072754653689691e-05, "loss": 0.8847, "step": 6012 }, { "epoch": 2.7245129134571817, "grad_norm": 0.3725279336708697, "learning_rate": 4.071489491295678e-05, "loss": 0.8895, "step": 6013 }, { "epoch": 2.724966017217943, "grad_norm": 0.3732552903107998, "learning_rate": 4.0702243217475474e-05, "loss": 0.8791, "step": 6014 }, { "epoch": 2.725419120978704, "grad_norm": 0.2719099927177403, "learning_rate": 4.0689591451719106e-05, "loss": 0.8778, "step": 6015 }, { "epoch": 2.725872224739465, "grad_norm": 0.28537833606216145, "learning_rate": 4.0676939616953744e-05, "loss": 0.8565, "step": 6016 }, { "epoch": 2.7263253285002267, "grad_norm": 0.3048854709386283, "learning_rate": 4.0664287714445505e-05, "loss": 0.8736, "step": 6017 }, { "epoch": 2.726778432260988, "grad_norm": 0.2827352153132275, "learning_rate": 4.0651635745460474e-05, "loss": 0.876, "step": 6018 }, { "epoch": 2.727231536021749, "grad_norm": 0.3204617482662181, "learning_rate": 4.0638983711264776e-05, "loss": 0.8693, "step": 6019 }, { "epoch": 2.72768463978251, "grad_norm": 0.3217331204063285, "learning_rate": 4.062633161312451e-05, "loss": 0.8583, "step": 6020 }, { "epoch": 2.7281377435432717, "grad_norm": 0.34969346736507473, "learning_rate": 4.0613679452305826e-05, "loss": 0.8699, "step": 6021 }, { "epoch": 2.728590847304033, "grad_norm": 0.2827316488014365, "learning_rate": 4.0601027230074843e-05, "loss": 0.881, "step": 6022 }, { "epoch": 2.729043951064794, "grad_norm": 0.3152637474508504, "learning_rate": 4.058837494769769e-05, "loss": 0.8819, "step": 6023 }, { "epoch": 2.729497054825555, "grad_norm": 0.38026921541190806, "learning_rate": 4.057572260644052e-05, "loss": 0.8605, "step": 6024 }, { "epoch": 2.729950158586316, "grad_norm": 0.3175882283390028, "learning_rate": 4.056307020756948e-05, "loss": 0.8628, "step": 6025 }, { "epoch": 2.7304032623470773, "grad_norm": 0.2809624461302234, "learning_rate": 4.055041775235072e-05, "loss": 0.8689, "step": 6026 }, { "epoch": 2.7308563661078384, "grad_norm": 0.36814140097727127, "learning_rate": 4.053776524205041e-05, "loss": 0.8706, "step": 6027 }, { "epoch": 2.7313094698686, "grad_norm": 0.40728157185623076, "learning_rate": 4.052511267793471e-05, "loss": 0.8598, "step": 6028 }, { "epoch": 2.731762573629361, "grad_norm": 0.3240425588394687, "learning_rate": 4.0512460061269775e-05, "loss": 0.8728, "step": 6029 }, { "epoch": 2.7322156773901223, "grad_norm": 0.3809617409902837, "learning_rate": 4.04998073933218e-05, "loss": 0.8782, "step": 6030 }, { "epoch": 2.7326687811508834, "grad_norm": 0.30588302697385444, "learning_rate": 4.048715467535696e-05, "loss": 0.8661, "step": 6031 }, { "epoch": 2.733121884911645, "grad_norm": 0.2695201676345449, "learning_rate": 4.047450190864145e-05, "loss": 0.8679, "step": 6032 }, { "epoch": 2.733574988672406, "grad_norm": 0.37803067075339253, "learning_rate": 4.046184909444144e-05, "loss": 0.8584, "step": 6033 }, { "epoch": 2.7340280924331672, "grad_norm": 0.2902454578816658, "learning_rate": 4.044919623402316e-05, "loss": 0.8842, "step": 6034 }, { "epoch": 2.7344811961939284, "grad_norm": 0.3007006071297407, "learning_rate": 4.043654332865278e-05, "loss": 0.8597, "step": 6035 }, { "epoch": 2.7349342999546895, "grad_norm": 0.35435548563462904, "learning_rate": 4.042389037959652e-05, "loss": 0.8787, "step": 6036 }, { "epoch": 2.7353874037154506, "grad_norm": 0.29428613040887996, "learning_rate": 4.041123738812058e-05, "loss": 0.8602, "step": 6037 }, { "epoch": 2.735840507476212, "grad_norm": 0.2749050591254965, "learning_rate": 4.0398584355491176e-05, "loss": 0.8771, "step": 6038 }, { "epoch": 2.7362936112369733, "grad_norm": 0.3457492702141727, "learning_rate": 4.0385931282974525e-05, "loss": 0.8923, "step": 6039 }, { "epoch": 2.7367467149977345, "grad_norm": 0.34376467108219444, "learning_rate": 4.037327817183686e-05, "loss": 0.8611, "step": 6040 }, { "epoch": 2.7371998187584956, "grad_norm": 0.33689797222221096, "learning_rate": 4.0360625023344394e-05, "loss": 0.8761, "step": 6041 }, { "epoch": 2.737652922519257, "grad_norm": 0.2876201979018234, "learning_rate": 4.0347971838763366e-05, "loss": 0.8852, "step": 6042 }, { "epoch": 2.7381060262800183, "grad_norm": 0.29467158044468766, "learning_rate": 4.033531861936e-05, "loss": 0.8623, "step": 6043 }, { "epoch": 2.7385591300407794, "grad_norm": 0.35570582148396124, "learning_rate": 4.0322665366400544e-05, "loss": 0.8708, "step": 6044 }, { "epoch": 2.7390122338015406, "grad_norm": 0.3353423153880636, "learning_rate": 4.031001208115123e-05, "loss": 0.8924, "step": 6045 }, { "epoch": 2.7394653375623017, "grad_norm": 0.33475072335211387, "learning_rate": 4.029735876487829e-05, "loss": 0.8976, "step": 6046 }, { "epoch": 2.739918441323063, "grad_norm": 0.33140380090307675, "learning_rate": 4.028470541884799e-05, "loss": 0.8427, "step": 6047 }, { "epoch": 2.740371545083824, "grad_norm": 0.3511415609275893, "learning_rate": 4.027205204432656e-05, "loss": 0.8901, "step": 6048 }, { "epoch": 2.7408246488445855, "grad_norm": 0.3414156961684191, "learning_rate": 4.025939864258027e-05, "loss": 0.8603, "step": 6049 }, { "epoch": 2.7412777526053467, "grad_norm": 0.39962071658988074, "learning_rate": 4.024674521487537e-05, "loss": 0.8734, "step": 6050 }, { "epoch": 2.741730856366108, "grad_norm": 0.2843707706730997, "learning_rate": 4.023409176247811e-05, "loss": 0.8811, "step": 6051 }, { "epoch": 2.742183960126869, "grad_norm": 0.2879674256812999, "learning_rate": 4.0221438286654765e-05, "loss": 0.8695, "step": 6052 }, { "epoch": 2.7426370638876305, "grad_norm": 0.26138448470513564, "learning_rate": 4.020878478867159e-05, "loss": 0.8672, "step": 6053 }, { "epoch": 2.7430901676483916, "grad_norm": 0.24132227384839286, "learning_rate": 4.019613126979483e-05, "loss": 0.8981, "step": 6054 }, { "epoch": 2.7435432714091528, "grad_norm": 0.28273309097664, "learning_rate": 4.018347773129077e-05, "loss": 0.8413, "step": 6055 }, { "epoch": 2.743996375169914, "grad_norm": 0.29905187706675396, "learning_rate": 4.017082417442569e-05, "loss": 0.8833, "step": 6056 }, { "epoch": 2.744449478930675, "grad_norm": 0.24722896172894274, "learning_rate": 4.015817060046583e-05, "loss": 0.8695, "step": 6057 }, { "epoch": 2.744902582691436, "grad_norm": 0.30161727098610924, "learning_rate": 4.0145517010677484e-05, "loss": 0.8703, "step": 6058 }, { "epoch": 2.7453556864521973, "grad_norm": 0.30081099258969635, "learning_rate": 4.013286340632692e-05, "loss": 0.8697, "step": 6059 }, { "epoch": 2.745808790212959, "grad_norm": 0.2663858114980627, "learning_rate": 4.012020978868041e-05, "loss": 0.8731, "step": 6060 }, { "epoch": 2.74626189397372, "grad_norm": 0.3391971906409004, "learning_rate": 4.0107556159004234e-05, "loss": 0.8685, "step": 6061 }, { "epoch": 2.746714997734481, "grad_norm": 0.36120672708236456, "learning_rate": 4.009490251856467e-05, "loss": 0.8803, "step": 6062 }, { "epoch": 2.7471681014952423, "grad_norm": 0.3108781021472091, "learning_rate": 4.008224886862799e-05, "loss": 0.8758, "step": 6063 }, { "epoch": 2.747621205256004, "grad_norm": 0.3237054430491846, "learning_rate": 4.006959521046049e-05, "loss": 0.854, "step": 6064 }, { "epoch": 2.748074309016765, "grad_norm": 0.3237639477335632, "learning_rate": 4.005694154532842e-05, "loss": 0.8717, "step": 6065 }, { "epoch": 2.748527412777526, "grad_norm": 0.2790374051807214, "learning_rate": 4.004428787449809e-05, "loss": 0.8808, "step": 6066 }, { "epoch": 2.7489805165382872, "grad_norm": 0.3174660223942424, "learning_rate": 4.003163419923577e-05, "loss": 0.8556, "step": 6067 }, { "epoch": 2.7494336202990484, "grad_norm": 0.3833808113371036, "learning_rate": 4.0018980520807745e-05, "loss": 0.8745, "step": 6068 }, { "epoch": 2.7498867240598095, "grad_norm": 0.2813135594290376, "learning_rate": 4.0006326840480295e-05, "loss": 0.8754, "step": 6069 }, { "epoch": 2.750339827820571, "grad_norm": 0.28639896405821724, "learning_rate": 3.999367315951971e-05, "loss": 0.881, "step": 6070 }, { "epoch": 2.750792931581332, "grad_norm": 0.40090000752032423, "learning_rate": 3.998101947919227e-05, "loss": 0.8809, "step": 6071 }, { "epoch": 2.7512460353420933, "grad_norm": 0.32832762560091944, "learning_rate": 3.996836580076424e-05, "loss": 0.8917, "step": 6072 }, { "epoch": 2.7516991391028545, "grad_norm": 0.2822682686990215, "learning_rate": 3.995571212550192e-05, "loss": 0.877, "step": 6073 }, { "epoch": 2.752152242863616, "grad_norm": 0.3477107516662863, "learning_rate": 3.994305845467159e-05, "loss": 0.8622, "step": 6074 }, { "epoch": 2.752605346624377, "grad_norm": 0.35783940372590606, "learning_rate": 3.993040478953953e-05, "loss": 0.8603, "step": 6075 }, { "epoch": 2.7530584503851383, "grad_norm": 0.3892134319720549, "learning_rate": 3.991775113137202e-05, "loss": 0.898, "step": 6076 }, { "epoch": 2.7535115541458994, "grad_norm": 0.4648126655945333, "learning_rate": 3.990509748143534e-05, "loss": 0.8893, "step": 6077 }, { "epoch": 2.7539646579066606, "grad_norm": 0.42396890278563953, "learning_rate": 3.989244384099578e-05, "loss": 0.8831, "step": 6078 }, { "epoch": 2.7544177616674217, "grad_norm": 0.33609021245887116, "learning_rate": 3.9879790211319606e-05, "loss": 0.8726, "step": 6079 }, { "epoch": 2.754870865428183, "grad_norm": 0.29148982499796827, "learning_rate": 3.9867136593673096e-05, "loss": 0.857, "step": 6080 }, { "epoch": 2.7553239691889444, "grad_norm": 0.3094449080864272, "learning_rate": 3.985448298932253e-05, "loss": 0.8444, "step": 6081 }, { "epoch": 2.7557770729497055, "grad_norm": 0.4067065164065901, "learning_rate": 3.984182939953418e-05, "loss": 0.8896, "step": 6082 }, { "epoch": 2.7562301767104667, "grad_norm": 0.49532600402905974, "learning_rate": 3.982917582557432e-05, "loss": 0.8581, "step": 6083 }, { "epoch": 2.756683280471228, "grad_norm": 0.43518774872880844, "learning_rate": 3.981652226870923e-05, "loss": 0.8661, "step": 6084 }, { "epoch": 2.7571363842319894, "grad_norm": 0.2878997466406975, "learning_rate": 3.9803868730205175e-05, "loss": 0.8732, "step": 6085 }, { "epoch": 2.7575894879927505, "grad_norm": 0.29528668110614603, "learning_rate": 3.9791215211328426e-05, "loss": 0.8848, "step": 6086 }, { "epoch": 2.7580425917535116, "grad_norm": 0.3958373638041033, "learning_rate": 3.977856171334524e-05, "loss": 0.8592, "step": 6087 }, { "epoch": 2.7584956955142728, "grad_norm": 0.4280347622678202, "learning_rate": 3.976590823752188e-05, "loss": 0.8567, "step": 6088 }, { "epoch": 2.758948799275034, "grad_norm": 0.3678831956921213, "learning_rate": 3.975325478512463e-05, "loss": 0.8806, "step": 6089 }, { "epoch": 2.759401903035795, "grad_norm": 0.33111058831219803, "learning_rate": 3.974060135741972e-05, "loss": 0.8638, "step": 6090 }, { "epoch": 2.759855006796556, "grad_norm": 0.3386437047506555, "learning_rate": 3.972794795567343e-05, "loss": 0.8768, "step": 6091 }, { "epoch": 2.7603081105573177, "grad_norm": 0.3022471021533651, "learning_rate": 3.9715294581152015e-05, "loss": 0.8492, "step": 6092 }, { "epoch": 2.760761214318079, "grad_norm": 0.3753656600619124, "learning_rate": 3.970264123512171e-05, "loss": 0.8704, "step": 6093 }, { "epoch": 2.76121431807884, "grad_norm": 0.5069186479624651, "learning_rate": 3.968998791884878e-05, "loss": 0.8643, "step": 6094 }, { "epoch": 2.761667421839601, "grad_norm": 0.4870017953705029, "learning_rate": 3.967733463359946e-05, "loss": 0.8806, "step": 6095 }, { "epoch": 2.7621205256003627, "grad_norm": 0.34256505937545234, "learning_rate": 3.966468138064e-05, "loss": 0.8502, "step": 6096 }, { "epoch": 2.762573629361124, "grad_norm": 0.26202982453803597, "learning_rate": 3.965202816123664e-05, "loss": 0.8691, "step": 6097 }, { "epoch": 2.763026733121885, "grad_norm": 0.3300982143208575, "learning_rate": 3.9639374976655606e-05, "loss": 0.8791, "step": 6098 }, { "epoch": 2.763479836882646, "grad_norm": 0.343925285058004, "learning_rate": 3.9626721828163156e-05, "loss": 0.8556, "step": 6099 }, { "epoch": 2.763932940643407, "grad_norm": 0.35148079184246794, "learning_rate": 3.961406871702549e-05, "loss": 0.8705, "step": 6100 }, { "epoch": 2.7643860444041684, "grad_norm": 0.35788616493836745, "learning_rate": 3.9601415644508844e-05, "loss": 0.8499, "step": 6101 }, { "epoch": 2.76483914816493, "grad_norm": 0.24538699955429236, "learning_rate": 3.958876261187944e-05, "loss": 0.8625, "step": 6102 }, { "epoch": 2.765292251925691, "grad_norm": 0.26383850974168016, "learning_rate": 3.95761096204035e-05, "loss": 0.8705, "step": 6103 }, { "epoch": 2.765745355686452, "grad_norm": 0.3255376778578457, "learning_rate": 3.9563456671347236e-05, "loss": 0.8684, "step": 6104 }, { "epoch": 2.7661984594472133, "grad_norm": 0.32148438175249183, "learning_rate": 3.955080376597686e-05, "loss": 0.8705, "step": 6105 }, { "epoch": 2.766651563207975, "grad_norm": 0.33849334014804705, "learning_rate": 3.953815090555858e-05, "loss": 0.8677, "step": 6106 }, { "epoch": 2.767104666968736, "grad_norm": 0.2837626880707358, "learning_rate": 3.952549809135857e-05, "loss": 0.8788, "step": 6107 }, { "epoch": 2.767557770729497, "grad_norm": 0.29759025208206613, "learning_rate": 3.951284532464306e-05, "loss": 0.8865, "step": 6108 }, { "epoch": 2.7680108744902583, "grad_norm": 0.25764389678644944, "learning_rate": 3.950019260667822e-05, "loss": 0.8726, "step": 6109 }, { "epoch": 2.7684639782510194, "grad_norm": 0.30925396570524866, "learning_rate": 3.9487539938730245e-05, "loss": 0.8898, "step": 6110 }, { "epoch": 2.7689170820117806, "grad_norm": 0.3810956017076035, "learning_rate": 3.947488732206531e-05, "loss": 0.8598, "step": 6111 }, { "epoch": 2.7693701857725417, "grad_norm": 0.3895187813292658, "learning_rate": 3.946223475794961e-05, "loss": 0.8599, "step": 6112 }, { "epoch": 2.7698232895333033, "grad_norm": 0.27226584522570374, "learning_rate": 3.9449582247649286e-05, "loss": 0.8674, "step": 6113 }, { "epoch": 2.7702763932940644, "grad_norm": 0.26784514059184283, "learning_rate": 3.9436929792430535e-05, "loss": 0.8601, "step": 6114 }, { "epoch": 2.7707294970548255, "grad_norm": 0.2877620191178111, "learning_rate": 3.942427739355949e-05, "loss": 0.8792, "step": 6115 }, { "epoch": 2.7711826008155866, "grad_norm": 0.35573036565703525, "learning_rate": 3.941162505230232e-05, "loss": 0.869, "step": 6116 }, { "epoch": 2.7716357045763482, "grad_norm": 0.3630479492936979, "learning_rate": 3.939897276992517e-05, "loss": 0.863, "step": 6117 }, { "epoch": 2.7720888083371094, "grad_norm": 0.3669917496559888, "learning_rate": 3.938632054769418e-05, "loss": 0.8846, "step": 6118 }, { "epoch": 2.7725419120978705, "grad_norm": 0.43828155671924013, "learning_rate": 3.937366838687549e-05, "loss": 0.8642, "step": 6119 }, { "epoch": 2.7729950158586316, "grad_norm": 0.3787731611497032, "learning_rate": 3.936101628873524e-05, "loss": 0.8814, "step": 6120 }, { "epoch": 2.7734481196193927, "grad_norm": 0.2472679047712021, "learning_rate": 3.934836425453953e-05, "loss": 0.8641, "step": 6121 }, { "epoch": 2.773901223380154, "grad_norm": 0.4274190204696975, "learning_rate": 3.93357122855545e-05, "loss": 0.8765, "step": 6122 }, { "epoch": 2.774354327140915, "grad_norm": 0.5218702113882142, "learning_rate": 3.932306038304626e-05, "loss": 0.8604, "step": 6123 }, { "epoch": 2.7748074309016766, "grad_norm": 0.5639495199145546, "learning_rate": 3.93104085482809e-05, "loss": 0.8653, "step": 6124 }, { "epoch": 2.7752605346624377, "grad_norm": 0.5173520037004926, "learning_rate": 3.929775678252453e-05, "loss": 0.8657, "step": 6125 }, { "epoch": 2.775713638423199, "grad_norm": 0.4564630591986277, "learning_rate": 3.9285105087043235e-05, "loss": 0.8698, "step": 6126 }, { "epoch": 2.77616674218396, "grad_norm": 0.36534926467599504, "learning_rate": 3.92724534631031e-05, "loss": 0.8747, "step": 6127 }, { "epoch": 2.7766198459447216, "grad_norm": 0.3695891193697033, "learning_rate": 3.92598019119702e-05, "loss": 0.8821, "step": 6128 }, { "epoch": 2.7770729497054827, "grad_norm": 0.5245594807135451, "learning_rate": 3.9247150434910614e-05, "loss": 0.8784, "step": 6129 }, { "epoch": 2.777526053466244, "grad_norm": 0.4783906676305149, "learning_rate": 3.92344990331904e-05, "loss": 0.8728, "step": 6130 }, { "epoch": 2.777979157227005, "grad_norm": 0.3435508661800548, "learning_rate": 3.9221847708075606e-05, "loss": 0.8748, "step": 6131 }, { "epoch": 2.778432260987766, "grad_norm": 0.3863242287502164, "learning_rate": 3.920919646083228e-05, "loss": 0.8807, "step": 6132 }, { "epoch": 2.778885364748527, "grad_norm": 0.4685789814826585, "learning_rate": 3.919654529272647e-05, "loss": 0.8687, "step": 6133 }, { "epoch": 2.779338468509289, "grad_norm": 0.4086725943235172, "learning_rate": 3.9183894205024196e-05, "loss": 0.8485, "step": 6134 }, { "epoch": 2.77979157227005, "grad_norm": 0.41953509253720556, "learning_rate": 3.9171243198991485e-05, "loss": 0.8771, "step": 6135 }, { "epoch": 2.780244676030811, "grad_norm": 0.4423116325912041, "learning_rate": 3.915859227589436e-05, "loss": 0.8812, "step": 6136 }, { "epoch": 2.780697779791572, "grad_norm": 0.3200230809377739, "learning_rate": 3.914594143699881e-05, "loss": 0.8588, "step": 6137 }, { "epoch": 2.7811508835523338, "grad_norm": 0.2558301465234832, "learning_rate": 3.9133290683570855e-05, "loss": 0.8764, "step": 6138 }, { "epoch": 2.781603987313095, "grad_norm": 0.31495807785444613, "learning_rate": 3.912064001687648e-05, "loss": 0.8569, "step": 6139 }, { "epoch": 2.782057091073856, "grad_norm": 0.32759956227256953, "learning_rate": 3.910798943818165e-05, "loss": 0.8566, "step": 6140 }, { "epoch": 2.782510194834617, "grad_norm": 0.334320270470963, "learning_rate": 3.909533894875235e-05, "loss": 0.8681, "step": 6141 }, { "epoch": 2.7829632985953783, "grad_norm": 0.3777244370539555, "learning_rate": 3.908268854985453e-05, "loss": 0.8574, "step": 6142 }, { "epoch": 2.7834164023561394, "grad_norm": 0.3362374593726536, "learning_rate": 3.907003824275417e-05, "loss": 0.8638, "step": 6143 }, { "epoch": 2.7838695061169005, "grad_norm": 0.3391271578740109, "learning_rate": 3.905738802871719e-05, "loss": 0.875, "step": 6144 }, { "epoch": 2.784322609877662, "grad_norm": 0.3907414394935393, "learning_rate": 3.9044737909009545e-05, "loss": 0.8827, "step": 6145 }, { "epoch": 2.7847757136384232, "grad_norm": 0.47380784428746847, "learning_rate": 3.903208788489714e-05, "loss": 0.8845, "step": 6146 }, { "epoch": 2.7852288173991844, "grad_norm": 0.39397713134496415, "learning_rate": 3.901943795764592e-05, "loss": 0.878, "step": 6147 }, { "epoch": 2.7856819211599455, "grad_norm": 0.32491129723769574, "learning_rate": 3.900678812852176e-05, "loss": 0.8724, "step": 6148 }, { "epoch": 2.786135024920707, "grad_norm": 0.37394907096742813, "learning_rate": 3.899413839879058e-05, "loss": 0.867, "step": 6149 }, { "epoch": 2.786588128681468, "grad_norm": 0.3790220553571639, "learning_rate": 3.898148876971826e-05, "loss": 0.8648, "step": 6150 }, { "epoch": 2.7870412324422293, "grad_norm": 0.5192542196964562, "learning_rate": 3.896883924257067e-05, "loss": 0.8688, "step": 6151 }, { "epoch": 2.7874943362029905, "grad_norm": 0.4581359100359331, "learning_rate": 3.8956189818613685e-05, "loss": 0.8816, "step": 6152 }, { "epoch": 2.7879474399637516, "grad_norm": 0.2962875047074123, "learning_rate": 3.894354049911317e-05, "loss": 0.8664, "step": 6153 }, { "epoch": 2.7884005437245127, "grad_norm": 0.31984053403976054, "learning_rate": 3.893089128533495e-05, "loss": 0.8726, "step": 6154 }, { "epoch": 2.788853647485274, "grad_norm": 0.35998468814390877, "learning_rate": 3.8918242178544894e-05, "loss": 0.8847, "step": 6155 }, { "epoch": 2.7893067512460354, "grad_norm": 0.43529380783013305, "learning_rate": 3.8905593180008796e-05, "loss": 0.8579, "step": 6156 }, { "epoch": 2.7897598550067966, "grad_norm": 0.4766885443226048, "learning_rate": 3.889294429099247e-05, "loss": 0.8725, "step": 6157 }, { "epoch": 2.7902129587675577, "grad_norm": 0.4061538267850164, "learning_rate": 3.8880295512761736e-05, "loss": 0.8743, "step": 6158 }, { "epoch": 2.790666062528319, "grad_norm": 0.33296461417095763, "learning_rate": 3.886764684658238e-05, "loss": 0.8749, "step": 6159 }, { "epoch": 2.7911191662890804, "grad_norm": 0.33042371142403043, "learning_rate": 3.8854998293720175e-05, "loss": 0.8484, "step": 6160 }, { "epoch": 2.7915722700498415, "grad_norm": 0.3212435290294456, "learning_rate": 3.88423498554409e-05, "loss": 0.8762, "step": 6161 }, { "epoch": 2.7920253738106027, "grad_norm": 0.39908591208291466, "learning_rate": 3.882970153301031e-05, "loss": 0.8799, "step": 6162 }, { "epoch": 2.792478477571364, "grad_norm": 0.476682626668373, "learning_rate": 3.8817053327694146e-05, "loss": 0.8895, "step": 6163 }, { "epoch": 2.792931581332125, "grad_norm": 0.35803675182796557, "learning_rate": 3.880440524075815e-05, "loss": 0.8657, "step": 6164 }, { "epoch": 2.793384685092886, "grad_norm": 0.299616224319746, "learning_rate": 3.879175727346804e-05, "loss": 0.872, "step": 6165 }, { "epoch": 2.7938377888536476, "grad_norm": 0.3172631288990671, "learning_rate": 3.877910942708954e-05, "loss": 0.872, "step": 6166 }, { "epoch": 2.7942908926144088, "grad_norm": 0.26672477682436735, "learning_rate": 3.876646170288834e-05, "loss": 0.8807, "step": 6167 }, { "epoch": 2.79474399637517, "grad_norm": 0.2866115847173372, "learning_rate": 3.8753814102130117e-05, "loss": 0.8894, "step": 6168 }, { "epoch": 2.795197100135931, "grad_norm": 0.28988125843534807, "learning_rate": 3.8741166626080555e-05, "loss": 0.8784, "step": 6169 }, { "epoch": 2.7956502038966926, "grad_norm": 0.30807655802779976, "learning_rate": 3.872851927600531e-05, "loss": 0.8769, "step": 6170 }, { "epoch": 2.7961033076574537, "grad_norm": 0.27172387743845866, "learning_rate": 3.8715872053170043e-05, "loss": 0.8933, "step": 6171 }, { "epoch": 2.796556411418215, "grad_norm": 0.3415843903159542, "learning_rate": 3.870322495884037e-05, "loss": 0.8658, "step": 6172 }, { "epoch": 2.797009515178976, "grad_norm": 0.3914259668483527, "learning_rate": 3.8690577994281936e-05, "loss": 0.8812, "step": 6173 }, { "epoch": 2.797462618939737, "grad_norm": 0.36795356755997255, "learning_rate": 3.8677931160760344e-05, "loss": 0.8654, "step": 6174 }, { "epoch": 2.7979157227004983, "grad_norm": 0.24691677369211362, "learning_rate": 3.86652844595412e-05, "loss": 0.8831, "step": 6175 }, { "epoch": 2.7983688264612594, "grad_norm": 0.3203525927812474, "learning_rate": 3.865263789189006e-05, "loss": 0.8648, "step": 6176 }, { "epoch": 2.798821930222021, "grad_norm": 0.35806588147866264, "learning_rate": 3.863999145907252e-05, "loss": 0.8806, "step": 6177 }, { "epoch": 2.799275033982782, "grad_norm": 0.30818127330487344, "learning_rate": 3.862734516235413e-05, "loss": 0.8862, "step": 6178 }, { "epoch": 2.7997281377435432, "grad_norm": 0.28043551847906484, "learning_rate": 3.861469900300043e-05, "loss": 0.8746, "step": 6179 }, { "epoch": 2.8001812415043044, "grad_norm": 0.28665625692043994, "learning_rate": 3.860205298227696e-05, "loss": 0.873, "step": 6180 }, { "epoch": 2.800634345265066, "grad_norm": 0.31231952284456366, "learning_rate": 3.8589407101449224e-05, "loss": 0.8664, "step": 6181 }, { "epoch": 2.801087449025827, "grad_norm": 0.30557656144931433, "learning_rate": 3.8576761361782733e-05, "loss": 0.8781, "step": 6182 }, { "epoch": 2.801540552786588, "grad_norm": 0.31417418008945674, "learning_rate": 3.856411576454297e-05, "loss": 0.865, "step": 6183 }, { "epoch": 2.8019936565473493, "grad_norm": 0.3298853739627325, "learning_rate": 3.855147031099543e-05, "loss": 0.8692, "step": 6184 }, { "epoch": 2.8024467603081105, "grad_norm": 0.30625019239533346, "learning_rate": 3.8538825002405526e-05, "loss": 0.8785, "step": 6185 }, { "epoch": 2.8028998640688716, "grad_norm": 0.27022866169191156, "learning_rate": 3.8526179840038745e-05, "loss": 0.8538, "step": 6186 }, { "epoch": 2.8033529678296327, "grad_norm": 0.2775198113478391, "learning_rate": 3.851353482516049e-05, "loss": 0.8668, "step": 6187 }, { "epoch": 2.8038060715903943, "grad_norm": 0.3208189726487793, "learning_rate": 3.85008899590362e-05, "loss": 0.8503, "step": 6188 }, { "epoch": 2.8042591753511554, "grad_norm": 0.29410903129782234, "learning_rate": 3.848824524293126e-05, "loss": 0.8789, "step": 6189 }, { "epoch": 2.8047122791119166, "grad_norm": 0.24661496104912223, "learning_rate": 3.847560067811106e-05, "loss": 0.8882, "step": 6190 }, { "epoch": 2.8051653828726777, "grad_norm": 0.3377735168410793, "learning_rate": 3.8462956265840966e-05, "loss": 0.8791, "step": 6191 }, { "epoch": 2.8056184866334393, "grad_norm": 0.3749501930159103, "learning_rate": 3.8450312007386344e-05, "loss": 0.8839, "step": 6192 }, { "epoch": 2.8060715903942004, "grad_norm": 0.306186071093824, "learning_rate": 3.8437667904012524e-05, "loss": 0.8656, "step": 6193 }, { "epoch": 2.8065246941549615, "grad_norm": 0.22900070289721192, "learning_rate": 3.842502395698483e-05, "loss": 0.8705, "step": 6194 }, { "epoch": 2.8069777979157227, "grad_norm": 0.2980436921599602, "learning_rate": 3.841238016756858e-05, "loss": 0.878, "step": 6195 }, { "epoch": 2.807430901676484, "grad_norm": 0.3075015865677053, "learning_rate": 3.839973653702905e-05, "loss": 0.8641, "step": 6196 }, { "epoch": 2.807884005437245, "grad_norm": 0.2585808254687258, "learning_rate": 3.838709306663153e-05, "loss": 0.8884, "step": 6197 }, { "epoch": 2.8083371091980065, "grad_norm": 0.24291438203639285, "learning_rate": 3.837444975764128e-05, "loss": 0.8658, "step": 6198 }, { "epoch": 2.8087902129587676, "grad_norm": 0.22417281757752594, "learning_rate": 3.8361806611323545e-05, "loss": 0.8542, "step": 6199 }, { "epoch": 2.8092433167195288, "grad_norm": 0.26549089703403267, "learning_rate": 3.834916362894356e-05, "loss": 0.8797, "step": 6200 }, { "epoch": 2.80969642048029, "grad_norm": 0.2937313856425692, "learning_rate": 3.8336520811766516e-05, "loss": 0.8917, "step": 6201 }, { "epoch": 2.8101495242410515, "grad_norm": 0.2161400664190621, "learning_rate": 3.8323878161057623e-05, "loss": 0.8643, "step": 6202 }, { "epoch": 2.8106026280018126, "grad_norm": 0.2995670072473001, "learning_rate": 3.831123567808206e-05, "loss": 0.8541, "step": 6203 }, { "epoch": 2.8110557317625737, "grad_norm": 0.25453323460668703, "learning_rate": 3.829859336410498e-05, "loss": 0.8846, "step": 6204 }, { "epoch": 2.811508835523335, "grad_norm": 0.5308585881129131, "learning_rate": 3.828595122039154e-05, "loss": 0.8854, "step": 6205 }, { "epoch": 2.811961939284096, "grad_norm": 0.33119946981518156, "learning_rate": 3.8273309248206854e-05, "loss": 0.8765, "step": 6206 }, { "epoch": 2.812415043044857, "grad_norm": 0.38799835284349826, "learning_rate": 3.8260667448816054e-05, "loss": 0.8859, "step": 6207 }, { "epoch": 2.8128681468056183, "grad_norm": 0.32233139187270027, "learning_rate": 3.8248025823484224e-05, "loss": 0.8671, "step": 6208 }, { "epoch": 2.81332125056638, "grad_norm": 0.24589810841146126, "learning_rate": 3.823538437347643e-05, "loss": 0.8648, "step": 6209 }, { "epoch": 2.813774354327141, "grad_norm": 0.2544014404057459, "learning_rate": 3.822274310005773e-05, "loss": 0.8634, "step": 6210 }, { "epoch": 2.814227458087902, "grad_norm": 0.21139519036822121, "learning_rate": 3.8210102004493185e-05, "loss": 0.8853, "step": 6211 }, { "epoch": 2.814680561848663, "grad_norm": 0.30718712019510824, "learning_rate": 3.81974610880478e-05, "loss": 0.8927, "step": 6212 }, { "epoch": 2.815133665609425, "grad_norm": 0.36487525078659877, "learning_rate": 3.818482035198661e-05, "loss": 0.8621, "step": 6213 }, { "epoch": 2.815586769370186, "grad_norm": 0.38023154991426905, "learning_rate": 3.817217979757456e-05, "loss": 0.8677, "step": 6214 }, { "epoch": 2.816039873130947, "grad_norm": 0.32720348189100096, "learning_rate": 3.815953942607664e-05, "loss": 0.8754, "step": 6215 }, { "epoch": 2.816492976891708, "grad_norm": 0.30143314838591734, "learning_rate": 3.8146899238757794e-05, "loss": 0.8739, "step": 6216 }, { "epoch": 2.8169460806524693, "grad_norm": 0.24780789887296942, "learning_rate": 3.8134259236882955e-05, "loss": 0.8754, "step": 6217 }, { "epoch": 2.8173991844132305, "grad_norm": 0.21284071138961594, "learning_rate": 3.812161942171705e-05, "loss": 0.8811, "step": 6218 }, { "epoch": 2.8178522881739916, "grad_norm": 0.24375727244821965, "learning_rate": 3.810897979452496e-05, "loss": 0.8551, "step": 6219 }, { "epoch": 2.818305391934753, "grad_norm": 0.24885414353391366, "learning_rate": 3.8096340356571576e-05, "loss": 0.8749, "step": 6220 }, { "epoch": 2.8187584956955143, "grad_norm": 0.2601256717670804, "learning_rate": 3.8083701109121735e-05, "loss": 0.8552, "step": 6221 }, { "epoch": 2.8192115994562754, "grad_norm": 0.3069725041190893, "learning_rate": 3.807106205344029e-05, "loss": 0.8578, "step": 6222 }, { "epoch": 2.8196647032170365, "grad_norm": 0.2699645202115264, "learning_rate": 3.8058423190792055e-05, "loss": 0.8597, "step": 6223 }, { "epoch": 2.820117806977798, "grad_norm": 0.224253900521068, "learning_rate": 3.804578452244183e-05, "loss": 0.8773, "step": 6224 }, { "epoch": 2.8205709107385593, "grad_norm": 0.2623418368185755, "learning_rate": 3.803314604965439e-05, "loss": 0.8589, "step": 6225 }, { "epoch": 2.8210240144993204, "grad_norm": 0.3498890652107185, "learning_rate": 3.802050777369451e-05, "loss": 0.8577, "step": 6226 }, { "epoch": 2.8214771182600815, "grad_norm": 0.284055462867579, "learning_rate": 3.8007869695826915e-05, "loss": 0.8624, "step": 6227 }, { "epoch": 2.8219302220208426, "grad_norm": 0.27537415832555606, "learning_rate": 3.799523181731635e-05, "loss": 0.8759, "step": 6228 }, { "epoch": 2.822383325781604, "grad_norm": 0.3014684221371488, "learning_rate": 3.798259413942749e-05, "loss": 0.8764, "step": 6229 }, { "epoch": 2.8228364295423654, "grad_norm": 0.2181724419011742, "learning_rate": 3.796995666342503e-05, "loss": 0.8619, "step": 6230 }, { "epoch": 2.8232895333031265, "grad_norm": 0.24099811424883402, "learning_rate": 3.7957319390573616e-05, "loss": 0.8634, "step": 6231 }, { "epoch": 2.8237426370638876, "grad_norm": 0.2605364878343017, "learning_rate": 3.7944682322137906e-05, "loss": 0.875, "step": 6232 }, { "epoch": 2.8241957408246487, "grad_norm": 0.2515564259031017, "learning_rate": 3.793204545938252e-05, "loss": 0.8596, "step": 6233 }, { "epoch": 2.8246488445854103, "grad_norm": 0.2610323297133143, "learning_rate": 3.791940880357204e-05, "loss": 0.8668, "step": 6234 }, { "epoch": 2.8251019483461715, "grad_norm": 0.19312252791142323, "learning_rate": 3.790677235597107e-05, "loss": 0.8711, "step": 6235 }, { "epoch": 2.8255550521069326, "grad_norm": 0.2987459600969955, "learning_rate": 3.7894136117844155e-05, "loss": 0.8815, "step": 6236 }, { "epoch": 2.8260081558676937, "grad_norm": 0.566368357948998, "learning_rate": 3.788150009045582e-05, "loss": 0.9035, "step": 6237 }, { "epoch": 2.826461259628455, "grad_norm": 0.26462648631389, "learning_rate": 3.78688642750706e-05, "loss": 0.8891, "step": 6238 }, { "epoch": 2.826914363389216, "grad_norm": 0.2977259126912304, "learning_rate": 3.785622867295299e-05, "loss": 0.8669, "step": 6239 }, { "epoch": 2.827367467149977, "grad_norm": 0.3599885325539847, "learning_rate": 3.784359328536744e-05, "loss": 0.857, "step": 6240 }, { "epoch": 2.8278205709107387, "grad_norm": 0.3208517117402885, "learning_rate": 3.783095811357843e-05, "loss": 0.875, "step": 6241 }, { "epoch": 2.8282736746715, "grad_norm": 0.30501558765349424, "learning_rate": 3.781832315885037e-05, "loss": 0.8744, "step": 6242 }, { "epoch": 2.828726778432261, "grad_norm": 0.27428987255300824, "learning_rate": 3.7805688422447684e-05, "loss": 0.8793, "step": 6243 }, { "epoch": 2.829179882193022, "grad_norm": 0.31305438073547714, "learning_rate": 3.779305390563475e-05, "loss": 0.8727, "step": 6244 }, { "epoch": 2.8296329859537837, "grad_norm": 0.34450322179630516, "learning_rate": 3.7780419609675945e-05, "loss": 0.8754, "step": 6245 }, { "epoch": 2.830086089714545, "grad_norm": 0.3197034906068491, "learning_rate": 3.776778553583559e-05, "loss": 0.8782, "step": 6246 }, { "epoch": 2.830539193475306, "grad_norm": 0.3692621945246486, "learning_rate": 3.775515168537802e-05, "loss": 0.8731, "step": 6247 }, { "epoch": 2.830992297236067, "grad_norm": 0.40470526438055743, "learning_rate": 3.774251805956753e-05, "loss": 0.8644, "step": 6248 }, { "epoch": 2.831445400996828, "grad_norm": 0.36048622577533485, "learning_rate": 3.7729884659668394e-05, "loss": 0.8534, "step": 6249 }, { "epoch": 2.8318985047575893, "grad_norm": 0.28985293918987254, "learning_rate": 3.771725148694487e-05, "loss": 0.8645, "step": 6250 }, { "epoch": 2.8323516085183504, "grad_norm": 0.255659214046325, "learning_rate": 3.770461854266118e-05, "loss": 0.8627, "step": 6251 }, { "epoch": 2.832804712279112, "grad_norm": 0.24316366387687013, "learning_rate": 3.7691985828081544e-05, "loss": 0.8657, "step": 6252 }, { "epoch": 2.833257816039873, "grad_norm": 0.2807714434191611, "learning_rate": 3.767935334447014e-05, "loss": 0.8577, "step": 6253 }, { "epoch": 2.8337109198006343, "grad_norm": 0.30602584722776294, "learning_rate": 3.766672109309113e-05, "loss": 0.8753, "step": 6254 }, { "epoch": 2.8341640235613954, "grad_norm": 0.2795575093305523, "learning_rate": 3.765408907520865e-05, "loss": 0.8787, "step": 6255 }, { "epoch": 2.834617127322157, "grad_norm": 0.33921405253589365, "learning_rate": 3.764145729208681e-05, "loss": 0.8871, "step": 6256 }, { "epoch": 2.835070231082918, "grad_norm": 0.2758959851421866, "learning_rate": 3.762882574498971e-05, "loss": 0.8539, "step": 6257 }, { "epoch": 2.8355233348436792, "grad_norm": 0.2458598532903117, "learning_rate": 3.7616194435181416e-05, "loss": 0.891, "step": 6258 }, { "epoch": 2.8359764386044404, "grad_norm": 0.3029063193120051, "learning_rate": 3.760356336392597e-05, "loss": 0.8554, "step": 6259 }, { "epoch": 2.8364295423652015, "grad_norm": 0.35281846607692463, "learning_rate": 3.759093253248739e-05, "loss": 0.8838, "step": 6260 }, { "epoch": 2.8368826461259626, "grad_norm": 0.4654494363913354, "learning_rate": 3.757830194212969e-05, "loss": 0.8933, "step": 6261 }, { "epoch": 2.837335749886724, "grad_norm": 0.37194561841215457, "learning_rate": 3.7565671594116816e-05, "loss": 0.8761, "step": 6262 }, { "epoch": 2.8377888536474853, "grad_norm": 0.31561662576272415, "learning_rate": 3.7553041489712725e-05, "loss": 0.8567, "step": 6263 }, { "epoch": 2.8382419574082465, "grad_norm": 0.34586829770003363, "learning_rate": 3.754041163018134e-05, "loss": 0.8668, "step": 6264 }, { "epoch": 2.8386950611690076, "grad_norm": 0.39470743426473825, "learning_rate": 3.752778201678657e-05, "loss": 0.8445, "step": 6265 }, { "epoch": 2.839148164929769, "grad_norm": 0.45332071048843775, "learning_rate": 3.751515265079227e-05, "loss": 0.8712, "step": 6266 }, { "epoch": 2.8396012686905303, "grad_norm": 0.4755885850532261, "learning_rate": 3.75025235334623e-05, "loss": 0.8783, "step": 6267 }, { "epoch": 2.8400543724512914, "grad_norm": 0.403528820644161, "learning_rate": 3.7489894666060494e-05, "loss": 0.8591, "step": 6268 }, { "epoch": 2.8405074762120526, "grad_norm": 0.31985369875412106, "learning_rate": 3.747726604985064e-05, "loss": 0.8663, "step": 6269 }, { "epoch": 2.8409605799728137, "grad_norm": 0.23402685552147234, "learning_rate": 3.746463768609652e-05, "loss": 0.8538, "step": 6270 }, { "epoch": 2.841413683733575, "grad_norm": 0.23213413410847344, "learning_rate": 3.7452009576061874e-05, "loss": 0.8534, "step": 6271 }, { "epoch": 2.841866787494336, "grad_norm": 0.27958165483020414, "learning_rate": 3.7439381721010424e-05, "loss": 0.8752, "step": 6272 }, { "epoch": 2.8423198912550975, "grad_norm": 0.29191383963761797, "learning_rate": 3.742675412220588e-05, "loss": 0.8578, "step": 6273 }, { "epoch": 2.8427729950158587, "grad_norm": 0.38682021597865757, "learning_rate": 3.7414126780911905e-05, "loss": 0.8559, "step": 6274 }, { "epoch": 2.84322609877662, "grad_norm": 0.4125423658136584, "learning_rate": 3.7401499698392145e-05, "loss": 0.8697, "step": 6275 }, { "epoch": 2.843679202537381, "grad_norm": 0.36224828839365175, "learning_rate": 3.7388872875910225e-05, "loss": 0.8902, "step": 6276 }, { "epoch": 2.8441323062981425, "grad_norm": 0.3330307221313172, "learning_rate": 3.737624631472974e-05, "loss": 0.8664, "step": 6277 }, { "epoch": 2.8445854100589036, "grad_norm": 0.2502734826958987, "learning_rate": 3.736362001611426e-05, "loss": 0.8809, "step": 6278 }, { "epoch": 2.8450385138196648, "grad_norm": 0.2909899017099211, "learning_rate": 3.735099398132733e-05, "loss": 0.8764, "step": 6279 }, { "epoch": 2.845491617580426, "grad_norm": 0.3751440993555395, "learning_rate": 3.733836821163245e-05, "loss": 0.8638, "step": 6280 }, { "epoch": 2.845944721341187, "grad_norm": 0.38359538757666384, "learning_rate": 3.7325742708293134e-05, "loss": 0.8913, "step": 6281 }, { "epoch": 2.846397825101948, "grad_norm": 0.40692786548352383, "learning_rate": 3.731311747257283e-05, "loss": 0.8864, "step": 6282 }, { "epoch": 2.8468509288627093, "grad_norm": 0.374341047954178, "learning_rate": 3.730049250573497e-05, "loss": 0.863, "step": 6283 }, { "epoch": 2.847304032623471, "grad_norm": 0.3031515678112397, "learning_rate": 3.728786780904297e-05, "loss": 0.8867, "step": 6284 }, { "epoch": 2.847757136384232, "grad_norm": 0.22866170234827693, "learning_rate": 3.727524338376021e-05, "loss": 0.8588, "step": 6285 }, { "epoch": 2.848210240144993, "grad_norm": 0.2920789643625583, "learning_rate": 3.7262619231150057e-05, "loss": 0.8749, "step": 6286 }, { "epoch": 2.8486633439057543, "grad_norm": 0.2607815163252198, "learning_rate": 3.724999535247582e-05, "loss": 0.8818, "step": 6287 }, { "epoch": 2.849116447666516, "grad_norm": 0.2447149026257135, "learning_rate": 3.723737174900081e-05, "loss": 0.8905, "step": 6288 }, { "epoch": 2.849569551427277, "grad_norm": 0.2554386887319941, "learning_rate": 3.72247484219883e-05, "loss": 0.8648, "step": 6289 }, { "epoch": 2.850022655188038, "grad_norm": 0.37632317689906675, "learning_rate": 3.721212537270153e-05, "loss": 0.8869, "step": 6290 }, { "epoch": 2.8504757589487992, "grad_norm": 0.28533680931766625, "learning_rate": 3.719950260240372e-05, "loss": 0.8759, "step": 6291 }, { "epoch": 2.8509288627095604, "grad_norm": 0.28048174683608246, "learning_rate": 3.718688011235806e-05, "loss": 0.8736, "step": 6292 }, { "epoch": 2.8513819664703215, "grad_norm": 0.27108058131733503, "learning_rate": 3.717425790382771e-05, "loss": 0.8638, "step": 6293 }, { "epoch": 2.851835070231083, "grad_norm": 0.3525230025332842, "learning_rate": 3.71616359780758e-05, "loss": 0.8703, "step": 6294 }, { "epoch": 2.852288173991844, "grad_norm": 0.441260716631327, "learning_rate": 3.7149014336365434e-05, "loss": 0.8904, "step": 6295 }, { "epoch": 2.8527412777526053, "grad_norm": 0.4320545109371847, "learning_rate": 3.7136392979959695e-05, "loss": 0.8505, "step": 6296 }, { "epoch": 2.8531943815133665, "grad_norm": 0.4488050678102766, "learning_rate": 3.712377191012164e-05, "loss": 0.8767, "step": 6297 }, { "epoch": 2.853647485274128, "grad_norm": 0.43913193523784855, "learning_rate": 3.7111151128114266e-05, "loss": 0.8714, "step": 6298 }, { "epoch": 2.854100589034889, "grad_norm": 0.4769016152375151, "learning_rate": 3.709853063520058e-05, "loss": 0.8549, "step": 6299 }, { "epoch": 2.8545536927956503, "grad_norm": 0.45802527708624863, "learning_rate": 3.708591043264353e-05, "loss": 0.8722, "step": 6300 }, { "epoch": 2.8550067965564114, "grad_norm": 0.39040261579468755, "learning_rate": 3.707329052170606e-05, "loss": 0.8773, "step": 6301 }, { "epoch": 2.8554599003171726, "grad_norm": 0.42951946232479354, "learning_rate": 3.706067090365107e-05, "loss": 0.879, "step": 6302 }, { "epoch": 2.8559130040779337, "grad_norm": 0.4740279577254549, "learning_rate": 3.704805157974143e-05, "loss": 0.8654, "step": 6303 }, { "epoch": 2.856366107838695, "grad_norm": 0.39965195766778466, "learning_rate": 3.703543255123999e-05, "loss": 0.8722, "step": 6304 }, { "epoch": 2.8568192115994564, "grad_norm": 0.35072520833335685, "learning_rate": 3.702281381940957e-05, "loss": 0.86, "step": 6305 }, { "epoch": 2.8572723153602175, "grad_norm": 0.35914147677588987, "learning_rate": 3.701019538551294e-05, "loss": 0.8637, "step": 6306 }, { "epoch": 2.8577254191209787, "grad_norm": 0.3301632972063245, "learning_rate": 3.699757725081286e-05, "loss": 0.8665, "step": 6307 }, { "epoch": 2.85817852288174, "grad_norm": 0.3671999110198262, "learning_rate": 3.6984959416572064e-05, "loss": 0.8739, "step": 6308 }, { "epoch": 2.8586316266425014, "grad_norm": 0.3525700873466139, "learning_rate": 3.697234188405324e-05, "loss": 0.8664, "step": 6309 }, { "epoch": 2.8590847304032625, "grad_norm": 0.30686421771262956, "learning_rate": 3.6959724654519045e-05, "loss": 0.8857, "step": 6310 }, { "epoch": 2.8595378341640236, "grad_norm": 0.28592842922703654, "learning_rate": 3.6947107729232135e-05, "loss": 0.8581, "step": 6311 }, { "epoch": 2.8599909379247848, "grad_norm": 0.36620833802993097, "learning_rate": 3.69344911094551e-05, "loss": 0.8855, "step": 6312 }, { "epoch": 2.860444041685546, "grad_norm": 0.38785170784102946, "learning_rate": 3.6921874796450523e-05, "loss": 0.8638, "step": 6313 }, { "epoch": 2.860897145446307, "grad_norm": 0.3657391389324479, "learning_rate": 3.690925879148094e-05, "loss": 0.8909, "step": 6314 }, { "epoch": 2.861350249207068, "grad_norm": 0.3674627061719567, "learning_rate": 3.6896643095808865e-05, "loss": 0.8807, "step": 6315 }, { "epoch": 2.8618033529678297, "grad_norm": 0.40723165277300233, "learning_rate": 3.6884027710696776e-05, "loss": 0.8818, "step": 6316 }, { "epoch": 2.862256456728591, "grad_norm": 0.39574624814362275, "learning_rate": 3.687141263740712e-05, "loss": 0.8725, "step": 6317 }, { "epoch": 2.862709560489352, "grad_norm": 0.4238019439052389, "learning_rate": 3.6858797877202335e-05, "loss": 0.8813, "step": 6318 }, { "epoch": 2.863162664250113, "grad_norm": 0.39586647195879415, "learning_rate": 3.6846183431344786e-05, "loss": 0.8681, "step": 6319 }, { "epoch": 2.8636157680108747, "grad_norm": 0.2869377702011253, "learning_rate": 3.683356930109684e-05, "loss": 0.8595, "step": 6320 }, { "epoch": 2.864068871771636, "grad_norm": 0.2969613431090147, "learning_rate": 3.682095548772083e-05, "loss": 0.8954, "step": 6321 }, { "epoch": 2.864521975532397, "grad_norm": 0.46667704240120417, "learning_rate": 3.6808341992479045e-05, "loss": 0.9071, "step": 6322 }, { "epoch": 2.864975079293158, "grad_norm": 0.35787031588550217, "learning_rate": 3.679572881663373e-05, "loss": 0.8627, "step": 6323 }, { "epoch": 2.865428183053919, "grad_norm": 0.38387183854493173, "learning_rate": 3.678311596144713e-05, "loss": 0.877, "step": 6324 }, { "epoch": 2.8658812868146804, "grad_norm": 0.39913990802136995, "learning_rate": 3.677050342818143e-05, "loss": 0.8518, "step": 6325 }, { "epoch": 2.866334390575442, "grad_norm": 0.3830800456795963, "learning_rate": 3.6757891218098805e-05, "loss": 0.8561, "step": 6326 }, { "epoch": 2.866787494336203, "grad_norm": 0.28838303111861824, "learning_rate": 3.674527933246141e-05, "loss": 0.8709, "step": 6327 }, { "epoch": 2.867240598096964, "grad_norm": 0.36446589178394645, "learning_rate": 3.6732667772531295e-05, "loss": 0.8776, "step": 6328 }, { "epoch": 2.8676937018577253, "grad_norm": 0.29681400883044595, "learning_rate": 3.672005653957056e-05, "loss": 0.8476, "step": 6329 }, { "epoch": 2.868146805618487, "grad_norm": 0.4020141144432962, "learning_rate": 3.670744563484123e-05, "loss": 0.8919, "step": 6330 }, { "epoch": 2.868599909379248, "grad_norm": 0.33671150390745486, "learning_rate": 3.6694835059605304e-05, "loss": 0.8687, "step": 6331 }, { "epoch": 2.869053013140009, "grad_norm": 0.34010246078705125, "learning_rate": 3.668222481512477e-05, "loss": 0.8705, "step": 6332 }, { "epoch": 2.8695061169007703, "grad_norm": 0.30175625113033394, "learning_rate": 3.666961490266155e-05, "loss": 0.8839, "step": 6333 }, { "epoch": 2.8699592206615314, "grad_norm": 0.2912983693511446, "learning_rate": 3.665700532347754e-05, "loss": 0.8831, "step": 6334 }, { "epoch": 2.8704123244222925, "grad_norm": 0.32896758135481724, "learning_rate": 3.664439607883462e-05, "loss": 0.8741, "step": 6335 }, { "epoch": 2.8708654281830537, "grad_norm": 0.32404726953401924, "learning_rate": 3.6631787169994615e-05, "loss": 0.8739, "step": 6336 }, { "epoch": 2.8713185319438153, "grad_norm": 0.32036219544878314, "learning_rate": 3.661917859821933e-05, "loss": 0.8755, "step": 6337 }, { "epoch": 2.8717716357045764, "grad_norm": 0.4036896787484114, "learning_rate": 3.660657036477055e-05, "loss": 0.8739, "step": 6338 }, { "epoch": 2.8722247394653375, "grad_norm": 0.3644070878175807, "learning_rate": 3.659396247090999e-05, "loss": 0.893, "step": 6339 }, { "epoch": 2.8726778432260986, "grad_norm": 0.39566346550062165, "learning_rate": 3.6581354917899365e-05, "loss": 0.8721, "step": 6340 }, { "epoch": 2.8731309469868602, "grad_norm": 0.30321138282508076, "learning_rate": 3.656874770700033e-05, "loss": 0.8645, "step": 6341 }, { "epoch": 2.8735840507476214, "grad_norm": 0.32234375076244026, "learning_rate": 3.655614083947453e-05, "loss": 0.8761, "step": 6342 }, { "epoch": 2.8740371545083825, "grad_norm": 0.4450484913241956, "learning_rate": 3.654353431658355e-05, "loss": 0.8676, "step": 6343 }, { "epoch": 2.8744902582691436, "grad_norm": 0.4356144328530542, "learning_rate": 3.653092813958896e-05, "loss": 0.873, "step": 6344 }, { "epoch": 2.8749433620299047, "grad_norm": 0.37478705134778045, "learning_rate": 3.6518322309752284e-05, "loss": 0.8765, "step": 6345 }, { "epoch": 2.875396465790666, "grad_norm": 0.32131766209687856, "learning_rate": 3.6505716828335016e-05, "loss": 0.8848, "step": 6346 }, { "epoch": 2.875849569551427, "grad_norm": 0.2443163084518915, "learning_rate": 3.649311169659862e-05, "loss": 0.8619, "step": 6347 }, { "epoch": 2.8763026733121886, "grad_norm": 0.22591565867301408, "learning_rate": 3.648050691580452e-05, "loss": 0.8889, "step": 6348 }, { "epoch": 2.8767557770729497, "grad_norm": 0.3759119520802731, "learning_rate": 3.6467902487214106e-05, "loss": 0.8918, "step": 6349 }, { "epoch": 2.877208880833711, "grad_norm": 0.40230105786117565, "learning_rate": 3.6455298412088736e-05, "loss": 0.8964, "step": 6350 }, { "epoch": 2.877661984594472, "grad_norm": 0.3558002559715785, "learning_rate": 3.644269469168971e-05, "loss": 0.8816, "step": 6351 }, { "epoch": 2.8781150883552336, "grad_norm": 0.3375171571296183, "learning_rate": 3.643009132727833e-05, "loss": 0.8808, "step": 6352 }, { "epoch": 2.8785681921159947, "grad_norm": 0.2636127561184914, "learning_rate": 3.641748832011584e-05, "loss": 0.8541, "step": 6353 }, { "epoch": 2.879021295876756, "grad_norm": 0.3111708359284778, "learning_rate": 3.640488567146344e-05, "loss": 0.867, "step": 6354 }, { "epoch": 2.879474399637517, "grad_norm": 0.2859674667400318, "learning_rate": 3.639228338258232e-05, "loss": 0.9031, "step": 6355 }, { "epoch": 2.879927503398278, "grad_norm": 0.30709433745611564, "learning_rate": 3.63796814547336e-05, "loss": 0.8829, "step": 6356 }, { "epoch": 2.880380607159039, "grad_norm": 0.3400417890653372, "learning_rate": 3.636707988917841e-05, "loss": 0.8816, "step": 6357 }, { "epoch": 2.880833710919801, "grad_norm": 0.33188581508200116, "learning_rate": 3.635447868717781e-05, "loss": 0.8767, "step": 6358 }, { "epoch": 2.881286814680562, "grad_norm": 0.24948215087329015, "learning_rate": 3.6341877849992815e-05, "loss": 0.8709, "step": 6359 }, { "epoch": 2.881739918441323, "grad_norm": 0.32283986704422896, "learning_rate": 3.632927737888443e-05, "loss": 0.8637, "step": 6360 }, { "epoch": 2.882193022202084, "grad_norm": 0.3761503284321755, "learning_rate": 3.631667727511362e-05, "loss": 0.8826, "step": 6361 }, { "epoch": 2.8826461259628458, "grad_norm": 0.2625856765395313, "learning_rate": 3.6304077539941295e-05, "loss": 0.8553, "step": 6362 }, { "epoch": 2.883099229723607, "grad_norm": 0.386232045435121, "learning_rate": 3.629147817462834e-05, "loss": 0.8706, "step": 6363 }, { "epoch": 2.883552333484368, "grad_norm": 0.41324279914990114, "learning_rate": 3.627887918043561e-05, "loss": 0.8656, "step": 6364 }, { "epoch": 2.884005437245129, "grad_norm": 0.4837817943998368, "learning_rate": 3.626628055862391e-05, "loss": 0.8673, "step": 6365 }, { "epoch": 2.8844585410058903, "grad_norm": 0.47400161004031477, "learning_rate": 3.625368231045402e-05, "loss": 0.8578, "step": 6366 }, { "epoch": 2.8849116447666514, "grad_norm": 0.4483951704052567, "learning_rate": 3.6241084437186654e-05, "loss": 0.8852, "step": 6367 }, { "epoch": 2.8853647485274125, "grad_norm": 0.41551266606433584, "learning_rate": 3.622848694008253e-05, "loss": 0.8751, "step": 6368 }, { "epoch": 2.885817852288174, "grad_norm": 0.3145236255630245, "learning_rate": 3.62158898204023e-05, "loss": 0.8538, "step": 6369 }, { "epoch": 2.8862709560489352, "grad_norm": 0.26195573143467926, "learning_rate": 3.620329307940659e-05, "loss": 0.86, "step": 6370 }, { "epoch": 2.8867240598096964, "grad_norm": 0.3389893914645555, "learning_rate": 3.6190696718355986e-05, "loss": 0.8871, "step": 6371 }, { "epoch": 2.8871771635704575, "grad_norm": 0.4081853208819388, "learning_rate": 3.617810073851103e-05, "loss": 0.8699, "step": 6372 }, { "epoch": 2.887630267331219, "grad_norm": 0.43746617486238343, "learning_rate": 3.616550514113223e-05, "loss": 0.8843, "step": 6373 }, { "epoch": 2.88808337109198, "grad_norm": 0.49239121389325585, "learning_rate": 3.615290992748007e-05, "loss": 0.8613, "step": 6374 }, { "epoch": 2.8885364748527413, "grad_norm": 0.4851302076800286, "learning_rate": 3.614031509881496e-05, "loss": 0.8485, "step": 6375 }, { "epoch": 2.8889895786135025, "grad_norm": 0.42340078153265215, "learning_rate": 3.6127720656397304e-05, "loss": 0.8848, "step": 6376 }, { "epoch": 2.8894426823742636, "grad_norm": 0.3360523136536932, "learning_rate": 3.6115126601487454e-05, "loss": 0.8748, "step": 6377 }, { "epoch": 2.8898957861350247, "grad_norm": 0.30349569726519543, "learning_rate": 3.610253293534573e-05, "loss": 0.8637, "step": 6378 }, { "epoch": 2.890348889895786, "grad_norm": 0.30244721437372885, "learning_rate": 3.608993965923241e-05, "loss": 0.8607, "step": 6379 }, { "epoch": 2.8908019936565474, "grad_norm": 0.277577661983794, "learning_rate": 3.607734677440772e-05, "loss": 0.8845, "step": 6380 }, { "epoch": 2.8912550974173086, "grad_norm": 0.2939765013537106, "learning_rate": 3.606475428213188e-05, "loss": 0.862, "step": 6381 }, { "epoch": 2.8917082011780697, "grad_norm": 0.25361822920564786, "learning_rate": 3.605216218366502e-05, "loss": 0.8901, "step": 6382 }, { "epoch": 2.892161304938831, "grad_norm": 0.2511078206939107, "learning_rate": 3.603957048026729e-05, "loss": 0.8709, "step": 6383 }, { "epoch": 2.8926144086995924, "grad_norm": 0.29558925512685125, "learning_rate": 3.6026979173198764e-05, "loss": 0.8705, "step": 6384 }, { "epoch": 2.8930675124603535, "grad_norm": 0.28686347765988185, "learning_rate": 3.601438826371946e-05, "loss": 0.8607, "step": 6385 }, { "epoch": 2.8935206162211147, "grad_norm": 0.29910087209293323, "learning_rate": 3.60017977530894e-05, "loss": 0.8785, "step": 6386 }, { "epoch": 2.893973719981876, "grad_norm": 0.3408322096550382, "learning_rate": 3.5989207642568535e-05, "loss": 0.8729, "step": 6387 }, { "epoch": 2.894426823742637, "grad_norm": 0.32163212688982173, "learning_rate": 3.597661793341679e-05, "loss": 0.8611, "step": 6388 }, { "epoch": 2.894879927503398, "grad_norm": 0.39104454739238187, "learning_rate": 3.596402862689405e-05, "loss": 0.8657, "step": 6389 }, { "epoch": 2.8953330312641596, "grad_norm": 0.43184956869999164, "learning_rate": 3.5951439724260146e-05, "loss": 0.8738, "step": 6390 }, { "epoch": 2.8957861350249208, "grad_norm": 0.3159721615503925, "learning_rate": 3.593885122677489e-05, "loss": 0.8806, "step": 6391 }, { "epoch": 2.896239238785682, "grad_norm": 0.23003937166918048, "learning_rate": 3.592626313569803e-05, "loss": 0.8656, "step": 6392 }, { "epoch": 2.896692342546443, "grad_norm": 0.2853276720410437, "learning_rate": 3.5913675452289295e-05, "loss": 0.8538, "step": 6393 }, { "epoch": 2.8971454463072046, "grad_norm": 0.3212883196481529, "learning_rate": 3.5901088177808365e-05, "loss": 0.8622, "step": 6394 }, { "epoch": 2.8975985500679657, "grad_norm": 0.26894282611491777, "learning_rate": 3.588850131351486e-05, "loss": 0.8441, "step": 6395 }, { "epoch": 2.898051653828727, "grad_norm": 0.27654216404885623, "learning_rate": 3.587591486066839e-05, "loss": 0.8685, "step": 6396 }, { "epoch": 2.898504757589488, "grad_norm": 0.24655950409892943, "learning_rate": 3.586332882052851e-05, "loss": 0.864, "step": 6397 }, { "epoch": 2.898957861350249, "grad_norm": 0.22168975662007376, "learning_rate": 3.585074319435472e-05, "loss": 0.8657, "step": 6398 }, { "epoch": 2.8994109651110103, "grad_norm": 0.32972292580557927, "learning_rate": 3.583815798340651e-05, "loss": 0.8851, "step": 6399 }, { "epoch": 2.8998640688717714, "grad_norm": 0.4701134114988265, "learning_rate": 3.58255731889433e-05, "loss": 0.8823, "step": 6400 }, { "epoch": 2.900317172632533, "grad_norm": 0.4351562038565672, "learning_rate": 3.581298881222449e-05, "loss": 0.8837, "step": 6401 }, { "epoch": 2.900770276393294, "grad_norm": 0.28415049267171666, "learning_rate": 3.580040485450942e-05, "loss": 0.8585, "step": 6402 }, { "epoch": 2.9012233801540552, "grad_norm": 0.302271498243651, "learning_rate": 3.5787821317057396e-05, "loss": 0.8791, "step": 6403 }, { "epoch": 2.9016764839148164, "grad_norm": 0.33441268503120586, "learning_rate": 3.5775238201127675e-05, "loss": 0.893, "step": 6404 }, { "epoch": 2.902129587675578, "grad_norm": 0.2548111729642943, "learning_rate": 3.576265550797948e-05, "loss": 0.8581, "step": 6405 }, { "epoch": 2.902582691436339, "grad_norm": 0.2521507869722304, "learning_rate": 3.5750073238872e-05, "loss": 0.8857, "step": 6406 }, { "epoch": 2.9030357951971, "grad_norm": 0.33031644192633874, "learning_rate": 3.573749139506436e-05, "loss": 0.8598, "step": 6407 }, { "epoch": 2.9034888989578613, "grad_norm": 0.3586855108160195, "learning_rate": 3.572490997781566e-05, "loss": 0.8721, "step": 6408 }, { "epoch": 2.9039420027186225, "grad_norm": 0.26544885618865005, "learning_rate": 3.571232898838495e-05, "loss": 0.8864, "step": 6409 }, { "epoch": 2.9043951064793836, "grad_norm": 0.26291783378098177, "learning_rate": 3.569974842803124e-05, "loss": 0.8528, "step": 6410 }, { "epoch": 2.9048482102401447, "grad_norm": 0.35441086026659047, "learning_rate": 3.56871682980135e-05, "loss": 0.8651, "step": 6411 }, { "epoch": 2.9053013140009063, "grad_norm": 0.38186656932010044, "learning_rate": 3.567458859959063e-05, "loss": 0.8669, "step": 6412 }, { "epoch": 2.9057544177616674, "grad_norm": 0.3405472022531274, "learning_rate": 3.566200933402152e-05, "loss": 0.8741, "step": 6413 }, { "epoch": 2.9062075215224286, "grad_norm": 0.5484095125769961, "learning_rate": 3.5649430502565024e-05, "loss": 0.8896, "step": 6414 }, { "epoch": 2.9066606252831897, "grad_norm": 0.3268861013992468, "learning_rate": 3.5636852106479915e-05, "loss": 0.8845, "step": 6415 }, { "epoch": 2.9071137290439513, "grad_norm": 0.25742127315314467, "learning_rate": 3.562427414702494e-05, "loss": 0.8695, "step": 6416 }, { "epoch": 2.9075668328047124, "grad_norm": 0.2156255236018921, "learning_rate": 3.5611696625458824e-05, "loss": 0.8691, "step": 6417 }, { "epoch": 2.9080199365654735, "grad_norm": 0.2733940371136518, "learning_rate": 3.55991195430402e-05, "loss": 0.8938, "step": 6418 }, { "epoch": 2.9084730403262347, "grad_norm": 0.25150206519077867, "learning_rate": 3.5586542901027724e-05, "loss": 0.8746, "step": 6419 }, { "epoch": 2.908926144086996, "grad_norm": 0.21491599451167792, "learning_rate": 3.5573966700679934e-05, "loss": 0.8649, "step": 6420 }, { "epoch": 2.909379247847757, "grad_norm": 0.2167400775535284, "learning_rate": 3.556139094325536e-05, "loss": 0.8884, "step": 6421 }, { "epoch": 2.9098323516085185, "grad_norm": 0.21817742958187963, "learning_rate": 3.554881563001251e-05, "loss": 0.8691, "step": 6422 }, { "epoch": 2.9102854553692796, "grad_norm": 0.20271061779083419, "learning_rate": 3.5536240762209804e-05, "loss": 0.8819, "step": 6423 }, { "epoch": 2.9107385591300408, "grad_norm": 0.20853893206297233, "learning_rate": 3.552366634110565e-05, "loss": 0.8631, "step": 6424 }, { "epoch": 2.911191662890802, "grad_norm": 0.22740489406752287, "learning_rate": 3.551109236795839e-05, "loss": 0.8781, "step": 6425 }, { "epoch": 2.9116447666515635, "grad_norm": 0.2646429039670345, "learning_rate": 3.5498518844026334e-05, "loss": 0.8771, "step": 6426 }, { "epoch": 2.9120978704123246, "grad_norm": 0.30324414184571885, "learning_rate": 3.5485945770567754e-05, "loss": 0.876, "step": 6427 }, { "epoch": 2.9125509741730857, "grad_norm": 0.23770969840620404, "learning_rate": 3.5473373148840844e-05, "loss": 0.8579, "step": 6428 }, { "epoch": 2.913004077933847, "grad_norm": 0.2931224950749816, "learning_rate": 3.546080098010379e-05, "loss": 0.8726, "step": 6429 }, { "epoch": 2.913457181694608, "grad_norm": 0.3056253807100892, "learning_rate": 3.5448229265614716e-05, "loss": 0.8941, "step": 6430 }, { "epoch": 2.913910285455369, "grad_norm": 0.24731106617845328, "learning_rate": 3.5435658006631694e-05, "loss": 0.8814, "step": 6431 }, { "epoch": 2.9143633892161303, "grad_norm": 0.22879876435616983, "learning_rate": 3.542308720441277e-05, "loss": 0.8774, "step": 6432 }, { "epoch": 2.914816492976892, "grad_norm": 0.255645150970466, "learning_rate": 3.5410516860215924e-05, "loss": 0.8443, "step": 6433 }, { "epoch": 2.915269596737653, "grad_norm": 0.2630619628611858, "learning_rate": 3.53979469752991e-05, "loss": 0.8825, "step": 6434 }, { "epoch": 2.915722700498414, "grad_norm": 0.22322357065220122, "learning_rate": 3.538537755092021e-05, "loss": 0.8785, "step": 6435 }, { "epoch": 2.916175804259175, "grad_norm": 0.20758272207427048, "learning_rate": 3.537280858833708e-05, "loss": 0.8692, "step": 6436 }, { "epoch": 2.916628908019937, "grad_norm": 0.2346055837568327, "learning_rate": 3.5360240088807525e-05, "loss": 0.889, "step": 6437 }, { "epoch": 2.917082011780698, "grad_norm": 0.26997257712470724, "learning_rate": 3.534767205358931e-05, "loss": 0.8696, "step": 6438 }, { "epoch": 2.917535115541459, "grad_norm": 0.3012642387119678, "learning_rate": 3.533510448394014e-05, "loss": 0.877, "step": 6439 }, { "epoch": 2.91798821930222, "grad_norm": 0.31442428932927885, "learning_rate": 3.5322537381117684e-05, "loss": 0.8549, "step": 6440 }, { "epoch": 2.9184413230629813, "grad_norm": 0.25071983579189444, "learning_rate": 3.530997074637958e-05, "loss": 0.862, "step": 6441 }, { "epoch": 2.9188944268237424, "grad_norm": 0.2645066387539897, "learning_rate": 3.529740458098335e-05, "loss": 0.8829, "step": 6442 }, { "epoch": 2.9193475305845036, "grad_norm": 0.2823058859444915, "learning_rate": 3.528483888618655e-05, "loss": 0.843, "step": 6443 }, { "epoch": 2.919800634345265, "grad_norm": 0.24939543054630872, "learning_rate": 3.527227366324666e-05, "loss": 0.8627, "step": 6444 }, { "epoch": 2.9202537381060263, "grad_norm": 0.19904923589083176, "learning_rate": 3.5259708913421105e-05, "loss": 0.8657, "step": 6445 }, { "epoch": 2.9207068418667874, "grad_norm": 0.2569854422765667, "learning_rate": 3.524714463796727e-05, "loss": 0.8388, "step": 6446 }, { "epoch": 2.9211599456275485, "grad_norm": 0.3019377369221893, "learning_rate": 3.52345808381425e-05, "loss": 0.8571, "step": 6447 }, { "epoch": 2.92161304938831, "grad_norm": 0.2548520581972284, "learning_rate": 3.522201751520406e-05, "loss": 0.8753, "step": 6448 }, { "epoch": 2.9220661531490713, "grad_norm": 0.2057150656178715, "learning_rate": 3.52094546704092e-05, "loss": 0.8776, "step": 6449 }, { "epoch": 2.9225192569098324, "grad_norm": 0.2687187281844998, "learning_rate": 3.5196892305015114e-05, "loss": 0.8567, "step": 6450 }, { "epoch": 2.9229723606705935, "grad_norm": 0.2846322251182245, "learning_rate": 3.5184330420278954e-05, "loss": 0.8479, "step": 6451 }, { "epoch": 2.9234254644313546, "grad_norm": 0.20632427245907442, "learning_rate": 3.5171769017457806e-05, "loss": 0.8746, "step": 6452 }, { "epoch": 2.923878568192116, "grad_norm": 0.2713192903049213, "learning_rate": 3.5159208097808726e-05, "loss": 0.8698, "step": 6453 }, { "epoch": 2.9243316719528774, "grad_norm": 0.23262157772028558, "learning_rate": 3.5146647662588706e-05, "loss": 0.8912, "step": 6454 }, { "epoch": 2.9247847757136385, "grad_norm": 0.2433253549881604, "learning_rate": 3.5134087713054714e-05, "loss": 0.8838, "step": 6455 }, { "epoch": 2.9252378794743996, "grad_norm": 0.2859245349316857, "learning_rate": 3.5121528250463627e-05, "loss": 0.8614, "step": 6456 }, { "epoch": 2.9256909832351607, "grad_norm": 0.24127739209620963, "learning_rate": 3.510896927607232e-05, "loss": 0.843, "step": 6457 }, { "epoch": 2.9261440869959223, "grad_norm": 0.2905919550340053, "learning_rate": 3.509641079113758e-05, "loss": 0.8863, "step": 6458 }, { "epoch": 2.9265971907566835, "grad_norm": 0.2882460331238998, "learning_rate": 3.5083852796916176e-05, "loss": 0.8787, "step": 6459 }, { "epoch": 2.9270502945174446, "grad_norm": 0.3084632907772496, "learning_rate": 3.5071295294664816e-05, "loss": 0.8656, "step": 6460 }, { "epoch": 2.9275033982782057, "grad_norm": 0.3365468917041526, "learning_rate": 3.505873828564015e-05, "loss": 0.8651, "step": 6461 }, { "epoch": 2.927956502038967, "grad_norm": 0.32251953647291, "learning_rate": 3.50461817710988e-05, "loss": 0.8742, "step": 6462 }, { "epoch": 2.928409605799728, "grad_norm": 0.33599447959993695, "learning_rate": 3.503362575229732e-05, "loss": 0.8762, "step": 6463 }, { "epoch": 2.928862709560489, "grad_norm": 0.3317242454972115, "learning_rate": 3.50210702304922e-05, "loss": 0.8726, "step": 6464 }, { "epoch": 2.9293158133212507, "grad_norm": 0.26514692759923447, "learning_rate": 3.5008515206939923e-05, "loss": 0.8558, "step": 6465 }, { "epoch": 2.929768917082012, "grad_norm": 0.2488425880373957, "learning_rate": 3.499596068289688e-05, "loss": 0.8618, "step": 6466 }, { "epoch": 2.930222020842773, "grad_norm": 0.287955377169897, "learning_rate": 3.498340665961945e-05, "loss": 0.8656, "step": 6467 }, { "epoch": 2.930675124603534, "grad_norm": 0.24976464100302845, "learning_rate": 3.4970853138363935e-05, "loss": 0.8708, "step": 6468 }, { "epoch": 2.9311282283642957, "grad_norm": 0.3104737596897892, "learning_rate": 3.49583001203866e-05, "loss": 0.8914, "step": 6469 }, { "epoch": 2.931581332125057, "grad_norm": 0.28027741742155654, "learning_rate": 3.494574760694363e-05, "loss": 0.8767, "step": 6470 }, { "epoch": 2.932034435885818, "grad_norm": 0.2850772719762885, "learning_rate": 3.493319559929121e-05, "loss": 0.8756, "step": 6471 }, { "epoch": 2.932487539646579, "grad_norm": 0.24967452565497594, "learning_rate": 3.492064409868545e-05, "loss": 0.8635, "step": 6472 }, { "epoch": 2.93294064340734, "grad_norm": 0.21961052422732952, "learning_rate": 3.4908093106382384e-05, "loss": 0.887, "step": 6473 }, { "epoch": 2.9333937471681013, "grad_norm": 0.28468519288538363, "learning_rate": 3.4895542623638036e-05, "loss": 0.8836, "step": 6474 }, { "epoch": 2.9338468509288624, "grad_norm": 0.2575123079245645, "learning_rate": 3.488299265170835e-05, "loss": 0.8741, "step": 6475 }, { "epoch": 2.934299954689624, "grad_norm": 0.2543980859514728, "learning_rate": 3.4870443191849244e-05, "loss": 0.8771, "step": 6476 }, { "epoch": 2.934753058450385, "grad_norm": 0.2658735019743026, "learning_rate": 3.485789424531656e-05, "loss": 0.8613, "step": 6477 }, { "epoch": 2.9352061622111463, "grad_norm": 0.28299376112051483, "learning_rate": 3.4845345813366105e-05, "loss": 0.8887, "step": 6478 }, { "epoch": 2.9356592659719074, "grad_norm": 0.2982952461684055, "learning_rate": 3.483279789725362e-05, "loss": 0.8679, "step": 6479 }, { "epoch": 2.936112369732669, "grad_norm": 0.21361648881819645, "learning_rate": 3.482025049823483e-05, "loss": 0.862, "step": 6480 }, { "epoch": 2.93656547349343, "grad_norm": 0.23644517650228636, "learning_rate": 3.480770361756534e-05, "loss": 0.8805, "step": 6481 }, { "epoch": 2.9370185772541912, "grad_norm": 0.2591583624092629, "learning_rate": 3.479515725650078e-05, "loss": 0.8668, "step": 6482 }, { "epoch": 2.9374716810149524, "grad_norm": 0.2261117364037585, "learning_rate": 3.4782611416296675e-05, "loss": 0.8883, "step": 6483 }, { "epoch": 2.9379247847757135, "grad_norm": 0.2135295166394325, "learning_rate": 3.477006609820852e-05, "loss": 0.8829, "step": 6484 }, { "epoch": 2.9383778885364746, "grad_norm": 0.2843410692740452, "learning_rate": 3.4757521303491754e-05, "loss": 0.8882, "step": 6485 }, { "epoch": 2.938830992297236, "grad_norm": 0.26387255306293367, "learning_rate": 3.474497703340177e-05, "loss": 0.867, "step": 6486 }, { "epoch": 2.9392840960579973, "grad_norm": 0.2605980042086717, "learning_rate": 3.473243328919389e-05, "loss": 0.8796, "step": 6487 }, { "epoch": 2.9397371998187585, "grad_norm": 0.24346350031518885, "learning_rate": 3.471989007212341e-05, "loss": 0.8645, "step": 6488 }, { "epoch": 2.9401903035795196, "grad_norm": 0.20462817918434084, "learning_rate": 3.470734738344554e-05, "loss": 0.882, "step": 6489 }, { "epoch": 2.940643407340281, "grad_norm": 0.21097542284175508, "learning_rate": 3.469480522441547e-05, "loss": 0.884, "step": 6490 }, { "epoch": 2.9410965111010423, "grad_norm": 0.20987709583607897, "learning_rate": 3.4682263596288315e-05, "loss": 0.8537, "step": 6491 }, { "epoch": 2.9415496148618034, "grad_norm": 0.22834943141599606, "learning_rate": 3.466972250031914e-05, "loss": 0.8652, "step": 6492 }, { "epoch": 2.9420027186225646, "grad_norm": 0.2160650905831672, "learning_rate": 3.465718193776297e-05, "loss": 0.8714, "step": 6493 }, { "epoch": 2.9424558223833257, "grad_norm": 0.20794698745095316, "learning_rate": 3.4644641909874765e-05, "loss": 0.8556, "step": 6494 }, { "epoch": 2.942908926144087, "grad_norm": 0.24187133699434568, "learning_rate": 3.463210241790943e-05, "loss": 0.8802, "step": 6495 }, { "epoch": 2.943362029904848, "grad_norm": 0.25895989005138037, "learning_rate": 3.461956346312184e-05, "loss": 0.8743, "step": 6496 }, { "epoch": 2.9438151336656095, "grad_norm": 0.26677316477330665, "learning_rate": 3.4607025046766774e-05, "loss": 0.8777, "step": 6497 }, { "epoch": 2.9442682374263707, "grad_norm": 0.28330540546592803, "learning_rate": 3.459448717009899e-05, "loss": 0.8679, "step": 6498 }, { "epoch": 2.944721341187132, "grad_norm": 0.28613593390616326, "learning_rate": 3.458194983437317e-05, "loss": 0.862, "step": 6499 }, { "epoch": 2.945174444947893, "grad_norm": 0.2581227721478216, "learning_rate": 3.4569413040843966e-05, "loss": 0.8767, "step": 6500 }, { "epoch": 2.9456275487086545, "grad_norm": 0.323414172614761, "learning_rate": 3.455687679076596e-05, "loss": 0.8861, "step": 6501 }, { "epoch": 2.9460806524694156, "grad_norm": 0.3576227159458921, "learning_rate": 3.4544341085393673e-05, "loss": 0.8672, "step": 6502 }, { "epoch": 2.9465337562301768, "grad_norm": 0.3477871306198772, "learning_rate": 3.45318059259816e-05, "loss": 0.8992, "step": 6503 }, { "epoch": 2.946986859990938, "grad_norm": 0.28267165151867607, "learning_rate": 3.4519271313784156e-05, "loss": 0.8712, "step": 6504 }, { "epoch": 2.947439963751699, "grad_norm": 0.3525712422892493, "learning_rate": 3.45067372500557e-05, "loss": 0.8668, "step": 6505 }, { "epoch": 2.94789306751246, "grad_norm": 0.3436829227978633, "learning_rate": 3.449420373605056e-05, "loss": 0.8671, "step": 6506 }, { "epoch": 2.9483461712732213, "grad_norm": 0.26080811887738237, "learning_rate": 3.448167077302297e-05, "loss": 0.8682, "step": 6507 }, { "epoch": 2.948799275033983, "grad_norm": 0.2758245691039191, "learning_rate": 3.4469138362227164e-05, "loss": 0.8796, "step": 6508 }, { "epoch": 2.949252378794744, "grad_norm": 0.2673115378162089, "learning_rate": 3.445660650491727e-05, "loss": 0.8725, "step": 6509 }, { "epoch": 2.949705482555505, "grad_norm": 0.31466453096476477, "learning_rate": 3.4444075202347365e-05, "loss": 0.8782, "step": 6510 }, { "epoch": 2.9501585863162663, "grad_norm": 0.4064896964700517, "learning_rate": 3.4431544455771506e-05, "loss": 0.8602, "step": 6511 }, { "epoch": 2.950611690077028, "grad_norm": 0.3909767913907306, "learning_rate": 3.4419014266443675e-05, "loss": 0.8637, "step": 6512 }, { "epoch": 2.951064793837789, "grad_norm": 0.3205004353404766, "learning_rate": 3.4406484635617785e-05, "loss": 0.8742, "step": 6513 }, { "epoch": 2.95151789759855, "grad_norm": 0.32625928352074246, "learning_rate": 3.439395556454771e-05, "loss": 0.8584, "step": 6514 }, { "epoch": 2.9519710013593112, "grad_norm": 0.299242663797953, "learning_rate": 3.438142705448726e-05, "loss": 0.8812, "step": 6515 }, { "epoch": 2.9524241051200724, "grad_norm": 0.24958371427219098, "learning_rate": 3.43688991066902e-05, "loss": 0.873, "step": 6516 }, { "epoch": 2.9528772088808335, "grad_norm": 0.23470424234757226, "learning_rate": 3.435637172241022e-05, "loss": 0.8709, "step": 6517 }, { "epoch": 2.953330312641595, "grad_norm": 0.288019246370403, "learning_rate": 3.4343844902900965e-05, "loss": 0.8658, "step": 6518 }, { "epoch": 2.953783416402356, "grad_norm": 0.289263987150974, "learning_rate": 3.4331318649416024e-05, "loss": 0.8707, "step": 6519 }, { "epoch": 2.9542365201631173, "grad_norm": 0.26904323426156185, "learning_rate": 3.431879296320893e-05, "loss": 0.8577, "step": 6520 }, { "epoch": 2.9546896239238785, "grad_norm": 0.2633427440447182, "learning_rate": 3.4306267845533154e-05, "loss": 0.8713, "step": 6521 }, { "epoch": 2.95514272768464, "grad_norm": 0.23392232696526405, "learning_rate": 3.429374329764212e-05, "loss": 0.8904, "step": 6522 }, { "epoch": 2.955595831445401, "grad_norm": 0.25290403322020477, "learning_rate": 3.428121932078918e-05, "loss": 0.8716, "step": 6523 }, { "epoch": 2.9560489352061623, "grad_norm": 0.2575664459041097, "learning_rate": 3.4268695916227644e-05, "loss": 0.8441, "step": 6524 }, { "epoch": 2.9565020389669234, "grad_norm": 0.2849845851113263, "learning_rate": 3.425617308521075e-05, "loss": 0.8845, "step": 6525 }, { "epoch": 2.9569551427276846, "grad_norm": 0.23794109355098733, "learning_rate": 3.424365082899168e-05, "loss": 0.8744, "step": 6526 }, { "epoch": 2.9574082464884457, "grad_norm": 0.24845659508018228, "learning_rate": 3.4231129148823585e-05, "loss": 0.8628, "step": 6527 }, { "epoch": 2.957861350249207, "grad_norm": 0.2729097070835412, "learning_rate": 3.421860804595951e-05, "loss": 0.8598, "step": 6528 }, { "epoch": 2.9583144540099684, "grad_norm": 0.2062029472690857, "learning_rate": 3.420608752165251e-05, "loss": 0.8644, "step": 6529 }, { "epoch": 2.9587675577707295, "grad_norm": 0.3293467745470828, "learning_rate": 3.4193567577155505e-05, "loss": 0.8863, "step": 6530 }, { "epoch": 2.9592206615314907, "grad_norm": 0.2616363180531302, "learning_rate": 3.418104821372141e-05, "loss": 0.8679, "step": 6531 }, { "epoch": 2.959673765292252, "grad_norm": 0.1975721083804294, "learning_rate": 3.416852943260307e-05, "loss": 0.8761, "step": 6532 }, { "epoch": 2.9601268690530134, "grad_norm": 0.24347086226535603, "learning_rate": 3.415601123505327e-05, "loss": 0.8631, "step": 6533 }, { "epoch": 2.9605799728137745, "grad_norm": 0.21377235354383883, "learning_rate": 3.414349362232472e-05, "loss": 0.8661, "step": 6534 }, { "epoch": 2.9610330765745356, "grad_norm": 0.2469626062176701, "learning_rate": 3.413097659567009e-05, "loss": 0.883, "step": 6535 }, { "epoch": 2.9614861803352968, "grad_norm": 0.2544659074074298, "learning_rate": 3.411846015634199e-05, "loss": 0.8751, "step": 6536 }, { "epoch": 2.961939284096058, "grad_norm": 0.19050226097850104, "learning_rate": 3.4105944305592976e-05, "loss": 0.8743, "step": 6537 }, { "epoch": 2.962392387856819, "grad_norm": 0.24032537937414944, "learning_rate": 3.409342904467553e-05, "loss": 0.8639, "step": 6538 }, { "epoch": 2.96284549161758, "grad_norm": 0.26471146571257004, "learning_rate": 3.408091437484208e-05, "loss": 0.8565, "step": 6539 }, { "epoch": 2.9632985953783417, "grad_norm": 0.25145042419120556, "learning_rate": 3.4068400297345e-05, "loss": 0.8752, "step": 6540 }, { "epoch": 2.963751699139103, "grad_norm": 0.2514610250836933, "learning_rate": 3.4055886813436614e-05, "loss": 0.8917, "step": 6541 }, { "epoch": 2.964204802899864, "grad_norm": 0.2838969608008055, "learning_rate": 3.4043373924369155e-05, "loss": 0.8653, "step": 6542 }, { "epoch": 2.964657906660625, "grad_norm": 0.28435095950220973, "learning_rate": 3.4030861631394825e-05, "loss": 0.87, "step": 6543 }, { "epoch": 2.9651110104213867, "grad_norm": 0.31604419819722, "learning_rate": 3.401834993576577e-05, "loss": 0.8802, "step": 6544 }, { "epoch": 2.965564114182148, "grad_norm": 0.32903770821535716, "learning_rate": 3.400583883873403e-05, "loss": 0.8747, "step": 6545 }, { "epoch": 2.966017217942909, "grad_norm": 0.2826264729814623, "learning_rate": 3.399332834155166e-05, "loss": 0.8718, "step": 6546 }, { "epoch": 2.96647032170367, "grad_norm": 0.3567271444481031, "learning_rate": 3.3980818445470586e-05, "loss": 0.8578, "step": 6547 }, { "epoch": 2.966923425464431, "grad_norm": 0.44213114661005476, "learning_rate": 3.3968309151742716e-05, "loss": 0.8665, "step": 6548 }, { "epoch": 2.9673765292251923, "grad_norm": 0.3678284364965101, "learning_rate": 3.395580046161988e-05, "loss": 0.8771, "step": 6549 }, { "epoch": 2.967829632985954, "grad_norm": 0.2832155595810091, "learning_rate": 3.3943292376353844e-05, "loss": 0.8517, "step": 6550 }, { "epoch": 2.968282736746715, "grad_norm": 0.3098251710468907, "learning_rate": 3.393078489719633e-05, "loss": 0.8625, "step": 6551 }, { "epoch": 2.968735840507476, "grad_norm": 0.3037995263468712, "learning_rate": 3.391827802539898e-05, "loss": 0.8504, "step": 6552 }, { "epoch": 2.9691889442682373, "grad_norm": 0.3040744056502003, "learning_rate": 3.3905771762213396e-05, "loss": 0.893, "step": 6553 }, { "epoch": 2.969642048028999, "grad_norm": 0.2741763425926884, "learning_rate": 3.3893266108891096e-05, "loss": 0.8732, "step": 6554 }, { "epoch": 2.97009515178976, "grad_norm": 0.2914683979725527, "learning_rate": 3.3880761066683586e-05, "loss": 0.8781, "step": 6555 }, { "epoch": 2.970548255550521, "grad_norm": 0.34091712063060503, "learning_rate": 3.386825663684222e-05, "loss": 0.8603, "step": 6556 }, { "epoch": 2.9710013593112823, "grad_norm": 0.27582057493408657, "learning_rate": 3.3855752820618364e-05, "loss": 0.8724, "step": 6557 }, { "epoch": 2.9714544630720434, "grad_norm": 0.24621020118343825, "learning_rate": 3.384324961926331e-05, "loss": 0.8772, "step": 6558 }, { "epoch": 2.9719075668328045, "grad_norm": 0.2900795664432069, "learning_rate": 3.3830747034028285e-05, "loss": 0.842, "step": 6559 }, { "epoch": 2.9723606705935657, "grad_norm": 0.2799431536876848, "learning_rate": 3.3818245066164445e-05, "loss": 0.8815, "step": 6560 }, { "epoch": 2.9728137743543273, "grad_norm": 0.2661443724187626, "learning_rate": 3.38057437169229e-05, "loss": 0.8758, "step": 6561 }, { "epoch": 2.9732668781150884, "grad_norm": 0.3035869474347661, "learning_rate": 3.379324298755467e-05, "loss": 0.8772, "step": 6562 }, { "epoch": 2.9737199818758495, "grad_norm": 0.2729634268045068, "learning_rate": 3.3780742879310745e-05, "loss": 0.8674, "step": 6563 }, { "epoch": 2.9741730856366106, "grad_norm": 0.25548488238448375, "learning_rate": 3.376824339344204e-05, "loss": 0.8682, "step": 6564 }, { "epoch": 2.9746261893973722, "grad_norm": 0.2809733464281243, "learning_rate": 3.375574453119939e-05, "loss": 0.8737, "step": 6565 }, { "epoch": 2.9750792931581334, "grad_norm": 0.2161484858044596, "learning_rate": 3.374324629383361e-05, "loss": 0.8785, "step": 6566 }, { "epoch": 2.9755323969188945, "grad_norm": 0.28002688399204073, "learning_rate": 3.373074868259541e-05, "loss": 0.8717, "step": 6567 }, { "epoch": 2.9759855006796556, "grad_norm": 0.29188773649760746, "learning_rate": 3.371825169873546e-05, "loss": 0.8958, "step": 6568 }, { "epoch": 2.9764386044404167, "grad_norm": 0.2773444069380518, "learning_rate": 3.370575534350437e-05, "loss": 0.8749, "step": 6569 }, { "epoch": 2.976891708201178, "grad_norm": 0.21209313751242745, "learning_rate": 3.369325961815266e-05, "loss": 0.8896, "step": 6570 }, { "epoch": 2.977344811961939, "grad_norm": 0.2823693329522958, "learning_rate": 3.368076452393081e-05, "loss": 0.8622, "step": 6571 }, { "epoch": 2.9777979157227006, "grad_norm": 0.3321631430432864, "learning_rate": 3.3668270062089245e-05, "loss": 0.8717, "step": 6572 }, { "epoch": 2.9782510194834617, "grad_norm": 0.2390796071293208, "learning_rate": 3.3655776233878304e-05, "loss": 0.8615, "step": 6573 }, { "epoch": 2.978704123244223, "grad_norm": 0.23716366147528414, "learning_rate": 3.364328304054828e-05, "loss": 0.862, "step": 6574 }, { "epoch": 2.979157227004984, "grad_norm": 0.262218418447444, "learning_rate": 3.3630790483349386e-05, "loss": 0.8883, "step": 6575 }, { "epoch": 2.9796103307657456, "grad_norm": 0.21032329403867367, "learning_rate": 3.361829856353178e-05, "loss": 0.8882, "step": 6576 }, { "epoch": 2.9800634345265067, "grad_norm": 0.24933448139812625, "learning_rate": 3.360580728234559e-05, "loss": 0.8498, "step": 6577 }, { "epoch": 2.980516538287268, "grad_norm": 0.2229620965227731, "learning_rate": 3.359331664104079e-05, "loss": 0.8772, "step": 6578 }, { "epoch": 2.980969642048029, "grad_norm": 0.25832699880492255, "learning_rate": 3.358082664086739e-05, "loss": 0.8565, "step": 6579 }, { "epoch": 2.98142274580879, "grad_norm": 0.2284829120850873, "learning_rate": 3.356833728307527e-05, "loss": 0.8479, "step": 6580 }, { "epoch": 2.981875849569551, "grad_norm": 0.2839150722880179, "learning_rate": 3.355584856891428e-05, "loss": 0.8587, "step": 6581 }, { "epoch": 2.982328953330313, "grad_norm": 0.21692520691539333, "learning_rate": 3.3543360499634194e-05, "loss": 0.8741, "step": 6582 }, { "epoch": 2.982782057091074, "grad_norm": 0.271380762704496, "learning_rate": 3.3530873076484724e-05, "loss": 0.855, "step": 6583 }, { "epoch": 2.983235160851835, "grad_norm": 0.3082193537339624, "learning_rate": 3.35183863007155e-05, "loss": 0.8817, "step": 6584 }, { "epoch": 2.983688264612596, "grad_norm": 0.2708931868561585, "learning_rate": 3.3505900173576127e-05, "loss": 0.8843, "step": 6585 }, { "epoch": 2.9841413683733577, "grad_norm": 0.2939102168466019, "learning_rate": 3.349341469631609e-05, "loss": 0.8753, "step": 6586 }, { "epoch": 2.984594472134119, "grad_norm": 0.3065748047436184, "learning_rate": 3.3480929870184864e-05, "loss": 0.8679, "step": 6587 }, { "epoch": 2.98504757589488, "grad_norm": 0.29355516676784266, "learning_rate": 3.346844569643182e-05, "loss": 0.8613, "step": 6588 }, { "epoch": 2.985500679655641, "grad_norm": 0.2655057423902087, "learning_rate": 3.345596217630628e-05, "loss": 0.8712, "step": 6589 }, { "epoch": 2.9859537834164023, "grad_norm": 0.25078169879318446, "learning_rate": 3.3443479311057503e-05, "loss": 0.8731, "step": 6590 }, { "epoch": 2.9864068871771634, "grad_norm": 0.2524436852529728, "learning_rate": 3.343099710193467e-05, "loss": 0.8795, "step": 6591 }, { "epoch": 2.9868599909379245, "grad_norm": 0.2007039884370848, "learning_rate": 3.341851555018692e-05, "loss": 0.857, "step": 6592 }, { "epoch": 2.987313094698686, "grad_norm": 0.3184323762698371, "learning_rate": 3.3406034657063296e-05, "loss": 0.9057, "step": 6593 }, { "epoch": 2.9877661984594472, "grad_norm": 0.27824714428247704, "learning_rate": 3.3393554423812785e-05, "loss": 0.8706, "step": 6594 }, { "epoch": 2.9882193022202084, "grad_norm": 0.5733525335985875, "learning_rate": 3.338107485168433e-05, "loss": 0.8949, "step": 6595 }, { "epoch": 2.9886724059809695, "grad_norm": 0.2856684499684627, "learning_rate": 3.336859594192677e-05, "loss": 0.8794, "step": 6596 }, { "epoch": 2.989125509741731, "grad_norm": 0.2355204577245487, "learning_rate": 3.3356117695788917e-05, "loss": 0.8639, "step": 6597 }, { "epoch": 2.989578613502492, "grad_norm": 0.320967744677627, "learning_rate": 3.3343640114519484e-05, "loss": 0.8712, "step": 6598 }, { "epoch": 2.9900317172632533, "grad_norm": 0.34875694596160683, "learning_rate": 3.3331163199367137e-05, "loss": 0.8766, "step": 6599 }, { "epoch": 2.9904848210240145, "grad_norm": 0.24417962101547444, "learning_rate": 3.331868695158047e-05, "loss": 0.8658, "step": 6600 }, { "epoch": 2.9909379247847756, "grad_norm": 0.35893050627843304, "learning_rate": 3.3306211372408e-05, "loss": 0.8848, "step": 6601 }, { "epoch": 2.9913910285455367, "grad_norm": 0.37977594383070484, "learning_rate": 3.329373646309821e-05, "loss": 0.8614, "step": 6602 }, { "epoch": 2.991844132306298, "grad_norm": 0.2659308234911431, "learning_rate": 3.3281262224899464e-05, "loss": 0.8883, "step": 6603 }, { "epoch": 2.9922972360670594, "grad_norm": 0.38345779441326355, "learning_rate": 3.32687886590601e-05, "loss": 0.8611, "step": 6604 }, { "epoch": 2.9927503398278206, "grad_norm": 0.4203992787813878, "learning_rate": 3.3256315766828376e-05, "loss": 0.8839, "step": 6605 }, { "epoch": 2.9932034435885817, "grad_norm": 0.2953520452140685, "learning_rate": 3.324384354945249e-05, "loss": 0.8599, "step": 6606 }, { "epoch": 2.993656547349343, "grad_norm": 0.31891492192426946, "learning_rate": 3.323137200818055e-05, "loss": 0.8925, "step": 6607 }, { "epoch": 2.9941096511101044, "grad_norm": 0.33997995262854763, "learning_rate": 3.321890114426062e-05, "loss": 0.8584, "step": 6608 }, { "epoch": 2.9945627548708655, "grad_norm": 0.30874016889729905, "learning_rate": 3.320643095894069e-05, "loss": 0.8839, "step": 6609 }, { "epoch": 2.9950158586316267, "grad_norm": 0.24263225163729799, "learning_rate": 3.3193961453468684e-05, "loss": 0.8499, "step": 6610 }, { "epoch": 2.995468962392388, "grad_norm": 0.31269831474999904, "learning_rate": 3.318149262909244e-05, "loss": 0.8608, "step": 6611 }, { "epoch": 2.995922066153149, "grad_norm": 0.33757883636437847, "learning_rate": 3.316902448705976e-05, "loss": 0.8826, "step": 6612 }, { "epoch": 2.99637516991391, "grad_norm": 0.32670514194894684, "learning_rate": 3.3156557028618346e-05, "loss": 0.8515, "step": 6613 }, { "epoch": 2.9968282736746716, "grad_norm": 0.31317675661836847, "learning_rate": 3.3144090255015836e-05, "loss": 0.8707, "step": 6614 }, { "epoch": 2.9972813774354328, "grad_norm": 0.2938641904168847, "learning_rate": 3.3131624167499815e-05, "loss": 0.8714, "step": 6615 }, { "epoch": 2.997734481196194, "grad_norm": 0.2559101309584173, "learning_rate": 3.311915876731781e-05, "loss": 0.8762, "step": 6616 }, { "epoch": 2.998187584956955, "grad_norm": 0.30390613204316846, "learning_rate": 3.310669405571724e-05, "loss": 0.8588, "step": 6617 }, { "epoch": 2.9986406887177166, "grad_norm": 0.37094450023292724, "learning_rate": 3.309423003394549e-05, "loss": 0.8536, "step": 6618 }, { "epoch": 2.9990937924784777, "grad_norm": 0.34069748346313783, "learning_rate": 3.308176670324985e-05, "loss": 0.8916, "step": 6619 }, { "epoch": 2.999546896239239, "grad_norm": 0.288415733584224, "learning_rate": 3.306930406487757e-05, "loss": 0.8687, "step": 6620 }, { "epoch": 3.0, "grad_norm": 0.2708275863292067, "learning_rate": 3.305684212007581e-05, "loss": 0.8672, "step": 6621 }, { "epoch": 3.000453103760761, "grad_norm": 0.2305337874738349, "learning_rate": 3.304438087009167e-05, "loss": 0.8331, "step": 6622 }, { "epoch": 3.0009062075215223, "grad_norm": 0.2943695454499855, "learning_rate": 3.303192031617216e-05, "loss": 0.855, "step": 6623 }, { "epoch": 3.001359311282284, "grad_norm": 0.28405736033118545, "learning_rate": 3.301946045956424e-05, "loss": 0.8642, "step": 6624 }, { "epoch": 3.001812415043045, "grad_norm": 0.3025646886617903, "learning_rate": 3.30070013015148e-05, "loss": 0.8624, "step": 6625 }, { "epoch": 3.002265518803806, "grad_norm": 0.24572077575630263, "learning_rate": 3.299454284327066e-05, "loss": 0.8365, "step": 6626 }, { "epoch": 3.0027186225645672, "grad_norm": 0.3615248040978342, "learning_rate": 3.298208508607857e-05, "loss": 0.872, "step": 6627 }, { "epoch": 3.0031717263253284, "grad_norm": 0.20345109187462862, "learning_rate": 3.2969628031185185e-05, "loss": 0.8532, "step": 6628 }, { "epoch": 3.00362483008609, "grad_norm": 0.2675364266460973, "learning_rate": 3.295717167983714e-05, "loss": 0.8452, "step": 6629 }, { "epoch": 3.004077933846851, "grad_norm": 0.2925255042395626, "learning_rate": 3.2944716033280955e-05, "loss": 0.8585, "step": 6630 }, { "epoch": 3.004531037607612, "grad_norm": 0.23597060415283247, "learning_rate": 3.293226109276309e-05, "loss": 0.8671, "step": 6631 }, { "epoch": 3.0049841413683733, "grad_norm": 0.2294620275609992, "learning_rate": 3.291980685952994e-05, "loss": 0.8649, "step": 6632 }, { "epoch": 3.0054372451291345, "grad_norm": 0.29296500266497855, "learning_rate": 3.290735333482784e-05, "loss": 0.816, "step": 6633 }, { "epoch": 3.0058903488898956, "grad_norm": 0.3602511579118954, "learning_rate": 3.289490051990303e-05, "loss": 0.8676, "step": 6634 }, { "epoch": 3.006343452650657, "grad_norm": 0.3466061278537806, "learning_rate": 3.2882448416001706e-05, "loss": 0.8604, "step": 6635 }, { "epoch": 3.0067965564114183, "grad_norm": 0.3102227803977284, "learning_rate": 3.2869997024369965e-05, "loss": 0.839, "step": 6636 }, { "epoch": 3.0072496601721794, "grad_norm": 0.27099586306046464, "learning_rate": 3.285754634625384e-05, "loss": 0.8485, "step": 6637 }, { "epoch": 3.0077027639329406, "grad_norm": 0.23948176167401786, "learning_rate": 3.284509638289933e-05, "loss": 0.8454, "step": 6638 }, { "epoch": 3.0081558676937017, "grad_norm": 0.3067405100391821, "learning_rate": 3.2832647135552305e-05, "loss": 0.8724, "step": 6639 }, { "epoch": 3.0086089714544633, "grad_norm": 0.35926847980400334, "learning_rate": 3.282019860545859e-05, "loss": 0.8578, "step": 6640 }, { "epoch": 3.0090620752152244, "grad_norm": 0.3121860513330946, "learning_rate": 3.280775079386394e-05, "loss": 0.8623, "step": 6641 }, { "epoch": 3.0095151789759855, "grad_norm": 0.301764252340525, "learning_rate": 3.2795303702014045e-05, "loss": 0.8453, "step": 6642 }, { "epoch": 3.0099682827367467, "grad_norm": 0.24607056270272856, "learning_rate": 3.27828573311545e-05, "loss": 0.848, "step": 6643 }, { "epoch": 3.010421386497508, "grad_norm": 0.2514263976618719, "learning_rate": 3.277041168253086e-05, "loss": 0.8599, "step": 6644 }, { "epoch": 3.0108744902582694, "grad_norm": 0.35373371282697413, "learning_rate": 3.275796675738857e-05, "loss": 0.8583, "step": 6645 }, { "epoch": 3.0113275940190305, "grad_norm": 0.431760794068099, "learning_rate": 3.274552255697304e-05, "loss": 0.865, "step": 6646 }, { "epoch": 3.0117806977797916, "grad_norm": 0.38513139722435535, "learning_rate": 3.2733079082529574e-05, "loss": 0.8547, "step": 6647 }, { "epoch": 3.0122338015405528, "grad_norm": 0.28214585480444776, "learning_rate": 3.272063633530342e-05, "loss": 0.8453, "step": 6648 }, { "epoch": 3.012686905301314, "grad_norm": 0.19906939173650368, "learning_rate": 3.270819431653976e-05, "loss": 0.8636, "step": 6649 }, { "epoch": 3.013140009062075, "grad_norm": 0.24568430448802966, "learning_rate": 3.2695753027483696e-05, "loss": 0.8581, "step": 6650 }, { "epoch": 3.0135931128228366, "grad_norm": 0.31999438444177863, "learning_rate": 3.268331246938025e-05, "loss": 0.8474, "step": 6651 }, { "epoch": 3.0140462165835977, "grad_norm": 0.3546794878824226, "learning_rate": 3.267087264347438e-05, "loss": 0.8402, "step": 6652 }, { "epoch": 3.014499320344359, "grad_norm": 0.2953115523083946, "learning_rate": 3.265843355101096e-05, "loss": 0.8313, "step": 6653 }, { "epoch": 3.01495242410512, "grad_norm": 0.25102708501534887, "learning_rate": 3.264599519323482e-05, "loss": 0.8476, "step": 6654 }, { "epoch": 3.015405527865881, "grad_norm": 0.2312876346157363, "learning_rate": 3.263355757139067e-05, "loss": 0.8527, "step": 6655 }, { "epoch": 3.0158586316266427, "grad_norm": 0.26656744668569665, "learning_rate": 3.262112068672317e-05, "loss": 0.8542, "step": 6656 }, { "epoch": 3.016311735387404, "grad_norm": 0.27266950646534766, "learning_rate": 3.260868454047693e-05, "loss": 0.8539, "step": 6657 }, { "epoch": 3.016764839148165, "grad_norm": 0.31538759019215873, "learning_rate": 3.259624913389645e-05, "loss": 0.8451, "step": 6658 }, { "epoch": 3.017217942908926, "grad_norm": 0.3562893214597452, "learning_rate": 3.2583814468226176e-05, "loss": 0.8534, "step": 6659 }, { "epoch": 3.017671046669687, "grad_norm": 0.28332299112793924, "learning_rate": 3.2571380544710466e-05, "loss": 0.8448, "step": 6660 }, { "epoch": 3.018124150430449, "grad_norm": 0.2847418622242126, "learning_rate": 3.2558947364593615e-05, "loss": 0.8673, "step": 6661 }, { "epoch": 3.01857725419121, "grad_norm": 0.27858078418551463, "learning_rate": 3.2546514929119834e-05, "loss": 0.8521, "step": 6662 }, { "epoch": 3.019030357951971, "grad_norm": 0.29436923369118756, "learning_rate": 3.253408323953328e-05, "loss": 0.853, "step": 6663 }, { "epoch": 3.019483461712732, "grad_norm": 0.37016746698641445, "learning_rate": 3.2521652297078e-05, "loss": 0.8437, "step": 6664 }, { "epoch": 3.0199365654734933, "grad_norm": 0.4636607631912439, "learning_rate": 3.2509222102998e-05, "loss": 0.8464, "step": 6665 }, { "epoch": 3.0203896692342544, "grad_norm": 0.5396535175541376, "learning_rate": 3.249679265853719e-05, "loss": 0.852, "step": 6666 }, { "epoch": 3.020842772995016, "grad_norm": 0.42004478926194855, "learning_rate": 3.248436396493942e-05, "loss": 0.8448, "step": 6667 }, { "epoch": 3.021295876755777, "grad_norm": 0.2762097569135902, "learning_rate": 3.247193602344845e-05, "loss": 0.8327, "step": 6668 }, { "epoch": 3.0217489805165383, "grad_norm": 0.311806291232975, "learning_rate": 3.2459508835308e-05, "loss": 0.8596, "step": 6669 }, { "epoch": 3.0222020842772994, "grad_norm": 0.2897381909431408, "learning_rate": 3.2447082401761646e-05, "loss": 0.8518, "step": 6670 }, { "epoch": 3.0226551880380605, "grad_norm": 0.30239391801443877, "learning_rate": 3.243465672405294e-05, "loss": 0.8533, "step": 6671 }, { "epoch": 3.023108291798822, "grad_norm": 0.2525599468917428, "learning_rate": 3.242223180342537e-05, "loss": 0.8562, "step": 6672 }, { "epoch": 3.0235613955595833, "grad_norm": 0.24745777707087394, "learning_rate": 3.24098076411223e-05, "loss": 0.8671, "step": 6673 }, { "epoch": 3.0240144993203444, "grad_norm": 0.2641453898055411, "learning_rate": 3.239738423838707e-05, "loss": 0.8486, "step": 6674 }, { "epoch": 3.0244676030811055, "grad_norm": 0.27328657800306105, "learning_rate": 3.238496159646289e-05, "loss": 0.8545, "step": 6675 }, { "epoch": 3.0249207068418666, "grad_norm": 0.35490903244885075, "learning_rate": 3.237253971659294e-05, "loss": 0.8523, "step": 6676 }, { "epoch": 3.025373810602628, "grad_norm": 0.3898935519360656, "learning_rate": 3.23601186000203e-05, "loss": 0.8611, "step": 6677 }, { "epoch": 3.0258269143633894, "grad_norm": 0.3719884033300934, "learning_rate": 3.234769824798799e-05, "loss": 0.8529, "step": 6678 }, { "epoch": 3.0262800181241505, "grad_norm": 0.3152538546073932, "learning_rate": 3.233527866173892e-05, "loss": 0.8658, "step": 6679 }, { "epoch": 3.0267331218849116, "grad_norm": 0.20292379405461716, "learning_rate": 3.232285984251597e-05, "loss": 0.857, "step": 6680 }, { "epoch": 3.0271862256456727, "grad_norm": 0.2444494369152077, "learning_rate": 3.2310441791561905e-05, "loss": 0.8411, "step": 6681 }, { "epoch": 3.027639329406434, "grad_norm": 0.5152643015377918, "learning_rate": 3.229802451011945e-05, "loss": 0.8911, "step": 6682 }, { "epoch": 3.0280924331671955, "grad_norm": 0.35989561052545405, "learning_rate": 3.22856079994312e-05, "loss": 0.8766, "step": 6683 }, { "epoch": 3.0285455369279566, "grad_norm": 0.3490859838851154, "learning_rate": 3.227319226073971e-05, "loss": 0.8827, "step": 6684 }, { "epoch": 3.0289986406887177, "grad_norm": 0.2891893961242168, "learning_rate": 3.226077729528746e-05, "loss": 0.8612, "step": 6685 }, { "epoch": 3.029451744449479, "grad_norm": 0.2314353863448066, "learning_rate": 3.2248363104316844e-05, "loss": 0.85, "step": 6686 }, { "epoch": 3.02990484821024, "grad_norm": 0.31140086196844285, "learning_rate": 3.223594968907018e-05, "loss": 0.841, "step": 6687 }, { "epoch": 3.0303579519710016, "grad_norm": 0.5139635217924717, "learning_rate": 3.22235370507897e-05, "loss": 0.8568, "step": 6688 }, { "epoch": 3.0308110557317627, "grad_norm": 0.4001827743183911, "learning_rate": 3.221112519071756e-05, "loss": 0.8469, "step": 6689 }, { "epoch": 3.031264159492524, "grad_norm": 0.3156624946633444, "learning_rate": 3.2198714110095864e-05, "loss": 0.8675, "step": 6690 }, { "epoch": 3.031717263253285, "grad_norm": 0.22466904229083828, "learning_rate": 3.2186303810166606e-05, "loss": 0.8475, "step": 6691 }, { "epoch": 3.032170367014046, "grad_norm": 0.35112344413832086, "learning_rate": 3.217389429217171e-05, "loss": 0.8815, "step": 6692 }, { "epoch": 3.0326234707748077, "grad_norm": 0.458137131766628, "learning_rate": 3.216148555735301e-05, "loss": 0.8501, "step": 6693 }, { "epoch": 3.033076574535569, "grad_norm": 0.41327330259838985, "learning_rate": 3.21490776069523e-05, "loss": 0.8534, "step": 6694 }, { "epoch": 3.03352967829633, "grad_norm": 0.3819272404761825, "learning_rate": 3.213667044221127e-05, "loss": 0.8511, "step": 6695 }, { "epoch": 3.033982782057091, "grad_norm": 0.2956066179481693, "learning_rate": 3.212426406437152e-05, "loss": 0.8356, "step": 6696 }, { "epoch": 3.034435885817852, "grad_norm": 0.20995871357914556, "learning_rate": 3.211185847467459e-05, "loss": 0.8411, "step": 6697 }, { "epoch": 3.0348889895786133, "grad_norm": 0.35485287291746265, "learning_rate": 3.209945367436194e-05, "loss": 0.8606, "step": 6698 }, { "epoch": 3.035342093339375, "grad_norm": 0.36738565982021776, "learning_rate": 3.208704966467495e-05, "loss": 0.8689, "step": 6699 }, { "epoch": 3.035795197100136, "grad_norm": 0.3440315792916736, "learning_rate": 3.2074646446854907e-05, "loss": 0.838, "step": 6700 }, { "epoch": 3.036248300860897, "grad_norm": 0.22129111659928558, "learning_rate": 3.2062244022143033e-05, "loss": 0.848, "step": 6701 }, { "epoch": 3.0367014046216583, "grad_norm": 0.23466351460742255, "learning_rate": 3.204984239178046e-05, "loss": 0.855, "step": 6702 }, { "epoch": 3.0371545083824194, "grad_norm": 0.2940436245533145, "learning_rate": 3.203744155700826e-05, "loss": 0.8536, "step": 6703 }, { "epoch": 3.037607612143181, "grad_norm": 0.2727515336224014, "learning_rate": 3.202504151906741e-05, "loss": 0.8566, "step": 6704 }, { "epoch": 3.038060715903942, "grad_norm": 0.27619468020234295, "learning_rate": 3.2012642279198804e-05, "loss": 0.8338, "step": 6705 }, { "epoch": 3.0385138196647032, "grad_norm": 0.2808716284342954, "learning_rate": 3.2000243838643265e-05, "loss": 0.8549, "step": 6706 }, { "epoch": 3.0389669234254644, "grad_norm": 0.2435402241130291, "learning_rate": 3.198784619864155e-05, "loss": 0.8471, "step": 6707 }, { "epoch": 3.0394200271862255, "grad_norm": 0.2754311380237125, "learning_rate": 3.1975449360434284e-05, "loss": 0.8627, "step": 6708 }, { "epoch": 3.0398731309469866, "grad_norm": 0.22682893759912204, "learning_rate": 3.196305332526207e-05, "loss": 0.8553, "step": 6709 }, { "epoch": 3.040326234707748, "grad_norm": 0.2894802700333791, "learning_rate": 3.1950658094365404e-05, "loss": 0.8581, "step": 6710 }, { "epoch": 3.0407793384685093, "grad_norm": 0.22482339373032462, "learning_rate": 3.193826366898471e-05, "loss": 0.8476, "step": 6711 }, { "epoch": 3.0412324422292705, "grad_norm": 0.24686740035866953, "learning_rate": 3.1925870050360307e-05, "loss": 0.8558, "step": 6712 }, { "epoch": 3.0416855459900316, "grad_norm": 0.26956911462859373, "learning_rate": 3.191347723973247e-05, "loss": 0.8642, "step": 6713 }, { "epoch": 3.0421386497507927, "grad_norm": 0.24170337623999372, "learning_rate": 3.1901085238341374e-05, "loss": 0.8539, "step": 6714 }, { "epoch": 3.0425917535115543, "grad_norm": 0.2777685772589853, "learning_rate": 3.1888694047427125e-05, "loss": 0.8814, "step": 6715 }, { "epoch": 3.0430448572723154, "grad_norm": 0.35260426669771816, "learning_rate": 3.187630366822971e-05, "loss": 0.8466, "step": 6716 }, { "epoch": 3.0434979610330766, "grad_norm": 0.31355527391815213, "learning_rate": 3.1863914101989086e-05, "loss": 0.8604, "step": 6717 }, { "epoch": 3.0439510647938377, "grad_norm": 0.2082094087456643, "learning_rate": 3.185152534994509e-05, "loss": 0.8241, "step": 6718 }, { "epoch": 3.044404168554599, "grad_norm": 0.31991402877285, "learning_rate": 3.18391374133375e-05, "loss": 0.8619, "step": 6719 }, { "epoch": 3.0448572723153604, "grad_norm": 0.3114451981224739, "learning_rate": 3.182675029340601e-05, "loss": 0.8461, "step": 6720 }, { "epoch": 3.0453103760761215, "grad_norm": 0.2981008010264787, "learning_rate": 3.1814363991390215e-05, "loss": 0.8678, "step": 6721 }, { "epoch": 3.0457634798368827, "grad_norm": 0.2885429885848232, "learning_rate": 3.180197850852965e-05, "loss": 0.8391, "step": 6722 }, { "epoch": 3.046216583597644, "grad_norm": 0.27965621177712885, "learning_rate": 3.1789593846063764e-05, "loss": 0.8619, "step": 6723 }, { "epoch": 3.046669687358405, "grad_norm": 0.3219173235560964, "learning_rate": 3.17772100052319e-05, "loss": 0.8565, "step": 6724 }, { "epoch": 3.0471227911191665, "grad_norm": 0.3575536573190246, "learning_rate": 3.176482698727335e-05, "loss": 0.8757, "step": 6725 }, { "epoch": 3.0475758948799276, "grad_norm": 0.28184799124299803, "learning_rate": 3.175244479342731e-05, "loss": 0.8562, "step": 6726 }, { "epoch": 3.0480289986406888, "grad_norm": 0.2548829695528201, "learning_rate": 3.17400634249329e-05, "loss": 0.8633, "step": 6727 }, { "epoch": 3.04848210240145, "grad_norm": 0.25437485008656857, "learning_rate": 3.172768288302913e-05, "loss": 0.8455, "step": 6728 }, { "epoch": 3.048935206162211, "grad_norm": 0.22989485831299555, "learning_rate": 3.171530316895497e-05, "loss": 0.8486, "step": 6729 }, { "epoch": 3.049388309922972, "grad_norm": 0.2120997398241032, "learning_rate": 3.170292428394927e-05, "loss": 0.8467, "step": 6730 }, { "epoch": 3.0498414136837337, "grad_norm": 0.23126919856427733, "learning_rate": 3.1690546229250826e-05, "loss": 0.854, "step": 6731 }, { "epoch": 3.050294517444495, "grad_norm": 0.2464752972618355, "learning_rate": 3.167816900609834e-05, "loss": 0.862, "step": 6732 }, { "epoch": 3.050747621205256, "grad_norm": 0.418183875599293, "learning_rate": 3.166579261573042e-05, "loss": 0.8873, "step": 6733 }, { "epoch": 3.051200724966017, "grad_norm": 0.27543897679918605, "learning_rate": 3.16534170593856e-05, "loss": 0.8515, "step": 6734 }, { "epoch": 3.0516538287267783, "grad_norm": 0.3094499287634318, "learning_rate": 3.164104233830234e-05, "loss": 0.8586, "step": 6735 }, { "epoch": 3.05210693248754, "grad_norm": 0.3006906349859631, "learning_rate": 3.1628668453719e-05, "loss": 0.8786, "step": 6736 }, { "epoch": 3.052560036248301, "grad_norm": 0.31103020307741064, "learning_rate": 3.161629540687385e-05, "loss": 0.869, "step": 6737 }, { "epoch": 3.053013140009062, "grad_norm": 0.34028092812530425, "learning_rate": 3.160392319900511e-05, "loss": 0.8693, "step": 6738 }, { "epoch": 3.0534662437698232, "grad_norm": 0.23226019383458948, "learning_rate": 3.159155183135089e-05, "loss": 0.8563, "step": 6739 }, { "epoch": 3.0539193475305844, "grad_norm": 0.2608670070018268, "learning_rate": 3.157918130514921e-05, "loss": 0.8493, "step": 6740 }, { "epoch": 3.054372451291346, "grad_norm": 0.2780216017497995, "learning_rate": 3.156681162163802e-05, "loss": 0.8372, "step": 6741 }, { "epoch": 3.054825555052107, "grad_norm": 0.22994454146258012, "learning_rate": 3.155444278205519e-05, "loss": 0.8551, "step": 6742 }, { "epoch": 3.055278658812868, "grad_norm": 0.2583519791449559, "learning_rate": 3.154207478763851e-05, "loss": 0.8446, "step": 6743 }, { "epoch": 3.0557317625736293, "grad_norm": 0.27932419885661325, "learning_rate": 3.1529707639625644e-05, "loss": 0.86, "step": 6744 }, { "epoch": 3.0561848663343905, "grad_norm": 0.24933296609819827, "learning_rate": 3.1517341339254215e-05, "loss": 0.8652, "step": 6745 }, { "epoch": 3.0566379700951516, "grad_norm": 0.2828550708656524, "learning_rate": 3.150497588776174e-05, "loss": 0.8697, "step": 6746 }, { "epoch": 3.057091073855913, "grad_norm": 0.3027640561456764, "learning_rate": 3.1492611286385676e-05, "loss": 0.8863, "step": 6747 }, { "epoch": 3.0575441776166743, "grad_norm": 0.2777718510182075, "learning_rate": 3.148024753636335e-05, "loss": 0.8552, "step": 6748 }, { "epoch": 3.0579972813774354, "grad_norm": 0.2819934285073144, "learning_rate": 3.146788463893206e-05, "loss": 0.8498, "step": 6749 }, { "epoch": 3.0584503851381966, "grad_norm": 0.2555559593606759, "learning_rate": 3.145552259532896e-05, "loss": 0.8687, "step": 6750 }, { "epoch": 3.0589034888989577, "grad_norm": 0.20643881102262585, "learning_rate": 3.1443161406791175e-05, "loss": 0.8478, "step": 6751 }, { "epoch": 3.0593565926597193, "grad_norm": 0.2145897089545357, "learning_rate": 3.1430801074555705e-05, "loss": 0.8599, "step": 6752 }, { "epoch": 3.0598096964204804, "grad_norm": 0.2929838101731071, "learning_rate": 3.141844159985947e-05, "loss": 0.8678, "step": 6753 }, { "epoch": 3.0602628001812415, "grad_norm": 0.2912384355690855, "learning_rate": 3.140608298393932e-05, "loss": 0.8632, "step": 6754 }, { "epoch": 3.0607159039420027, "grad_norm": 0.3096104634611178, "learning_rate": 3.1393725228032e-05, "loss": 0.8415, "step": 6755 }, { "epoch": 3.061169007702764, "grad_norm": 0.27427192819133156, "learning_rate": 3.1381368333374194e-05, "loss": 0.8689, "step": 6756 }, { "epoch": 3.0616221114635254, "grad_norm": 0.2605213714996298, "learning_rate": 3.136901230120247e-05, "loss": 0.8561, "step": 6757 }, { "epoch": 3.0620752152242865, "grad_norm": 0.28952524784455175, "learning_rate": 3.135665713275334e-05, "loss": 0.8566, "step": 6758 }, { "epoch": 3.0625283189850476, "grad_norm": 0.24324050494254784, "learning_rate": 3.134430282926319e-05, "loss": 0.8457, "step": 6759 }, { "epoch": 3.0629814227458088, "grad_norm": 0.27426662633750715, "learning_rate": 3.133194939196838e-05, "loss": 0.8414, "step": 6760 }, { "epoch": 3.06343452650657, "grad_norm": 0.3221208868004719, "learning_rate": 3.1319596822105106e-05, "loss": 0.8519, "step": 6761 }, { "epoch": 3.063887630267331, "grad_norm": 0.31570478340698305, "learning_rate": 3.130724512090954e-05, "loss": 0.8256, "step": 6762 }, { "epoch": 3.0643407340280926, "grad_norm": 0.4138573802318061, "learning_rate": 3.129489428961774e-05, "loss": 0.8394, "step": 6763 }, { "epoch": 3.0647938377888537, "grad_norm": 0.3179228650097925, "learning_rate": 3.128254432946568e-05, "loss": 0.8461, "step": 6764 }, { "epoch": 3.065246941549615, "grad_norm": 0.23869357287154605, "learning_rate": 3.1270195241689254e-05, "loss": 0.8633, "step": 6765 }, { "epoch": 3.065700045310376, "grad_norm": 0.24574437535664656, "learning_rate": 3.125784702752427e-05, "loss": 0.8336, "step": 6766 }, { "epoch": 3.066153149071137, "grad_norm": 0.3385020960604622, "learning_rate": 3.124549968820642e-05, "loss": 0.8785, "step": 6767 }, { "epoch": 3.0666062528318987, "grad_norm": 0.3567431051106734, "learning_rate": 3.123315322497135e-05, "loss": 0.8671, "step": 6768 }, { "epoch": 3.06705935659266, "grad_norm": 0.31233481678807096, "learning_rate": 3.122080763905459e-05, "loss": 0.8564, "step": 6769 }, { "epoch": 3.067512460353421, "grad_norm": 0.25905304519824823, "learning_rate": 3.120846293169159e-05, "loss": 0.863, "step": 6770 }, { "epoch": 3.067965564114182, "grad_norm": 0.22018685308736818, "learning_rate": 3.119611910411771e-05, "loss": 0.8386, "step": 6771 }, { "epoch": 3.068418667874943, "grad_norm": 0.25225141884773095, "learning_rate": 3.118377615756824e-05, "loss": 0.8623, "step": 6772 }, { "epoch": 3.0688717716357043, "grad_norm": 0.3498237967347394, "learning_rate": 3.1171434093278356e-05, "loss": 0.8579, "step": 6773 }, { "epoch": 3.069324875396466, "grad_norm": 0.3002224467124046, "learning_rate": 3.115909291248315e-05, "loss": 0.8423, "step": 6774 }, { "epoch": 3.069777979157227, "grad_norm": 0.37887500807733265, "learning_rate": 3.114675261641764e-05, "loss": 0.879, "step": 6775 }, { "epoch": 3.070231082917988, "grad_norm": 0.3008290883225651, "learning_rate": 3.113441320631676e-05, "loss": 0.8683, "step": 6776 }, { "epoch": 3.0706841866787493, "grad_norm": 0.31283638611922693, "learning_rate": 3.112207468341532e-05, "loss": 0.8577, "step": 6777 }, { "epoch": 3.0711372904395104, "grad_norm": 0.24339670338309055, "learning_rate": 3.1109737048948075e-05, "loss": 0.8649, "step": 6778 }, { "epoch": 3.071590394200272, "grad_norm": 0.2509251980511222, "learning_rate": 3.109740030414968e-05, "loss": 0.8768, "step": 6779 }, { "epoch": 3.072043497961033, "grad_norm": 0.26939573285864016, "learning_rate": 3.10850644502547e-05, "loss": 0.8518, "step": 6780 }, { "epoch": 3.0724966017217943, "grad_norm": 0.2468691587296614, "learning_rate": 3.107272948849761e-05, "loss": 0.8512, "step": 6781 }, { "epoch": 3.0729497054825554, "grad_norm": 0.25796951653497585, "learning_rate": 3.10603954201128e-05, "loss": 0.8364, "step": 6782 }, { "epoch": 3.0734028092433165, "grad_norm": 0.2553838043849567, "learning_rate": 3.104806224633459e-05, "loss": 0.8509, "step": 6783 }, { "epoch": 3.073855913004078, "grad_norm": 0.22164855934137773, "learning_rate": 3.103572996839715e-05, "loss": 0.8609, "step": 6784 }, { "epoch": 3.0743090167648393, "grad_norm": 0.1967005819914268, "learning_rate": 3.1023398587534605e-05, "loss": 0.8667, "step": 6785 }, { "epoch": 3.0747621205256004, "grad_norm": 0.23756727350501194, "learning_rate": 3.1011068104981004e-05, "loss": 0.8481, "step": 6786 }, { "epoch": 3.0752152242863615, "grad_norm": 0.21500366295615467, "learning_rate": 3.099873852197028e-05, "loss": 0.8767, "step": 6787 }, { "epoch": 3.0756683280471226, "grad_norm": 0.24221181315563442, "learning_rate": 3.0986409839736284e-05, "loss": 0.8549, "step": 6788 }, { "epoch": 3.076121431807884, "grad_norm": 0.29169695525387107, "learning_rate": 3.097408205951276e-05, "loss": 0.8621, "step": 6789 }, { "epoch": 3.0765745355686454, "grad_norm": 0.24297634300500337, "learning_rate": 3.09617551825334e-05, "loss": 0.8608, "step": 6790 }, { "epoch": 3.0770276393294065, "grad_norm": 0.2295499833289481, "learning_rate": 3.0949429210031765e-05, "loss": 0.8394, "step": 6791 }, { "epoch": 3.0774807430901676, "grad_norm": 0.26764176701325637, "learning_rate": 3.093710414324135e-05, "loss": 0.8551, "step": 6792 }, { "epoch": 3.0779338468509287, "grad_norm": 0.26037609022928004, "learning_rate": 3.092477998339555e-05, "loss": 0.841, "step": 6793 }, { "epoch": 3.07838695061169, "grad_norm": 0.2544799127875487, "learning_rate": 3.091245673172768e-05, "loss": 0.8619, "step": 6794 }, { "epoch": 3.0788400543724515, "grad_norm": 0.2933741699345733, "learning_rate": 3.090013438947094e-05, "loss": 0.8691, "step": 6795 }, { "epoch": 3.0792931581332126, "grad_norm": 0.26246135667362547, "learning_rate": 3.088781295785848e-05, "loss": 0.8606, "step": 6796 }, { "epoch": 3.0797462618939737, "grad_norm": 0.23517475096231086, "learning_rate": 3.0875492438123304e-05, "loss": 0.8609, "step": 6797 }, { "epoch": 3.080199365654735, "grad_norm": 0.2542026982249196, "learning_rate": 3.086317283149837e-05, "loss": 0.8679, "step": 6798 }, { "epoch": 3.080652469415496, "grad_norm": 0.2582107181946659, "learning_rate": 3.085085413921653e-05, "loss": 0.8689, "step": 6799 }, { "epoch": 3.0811055731762576, "grad_norm": 0.2836739579396822, "learning_rate": 3.083853636251054e-05, "loss": 0.8622, "step": 6800 }, { "epoch": 3.0815586769370187, "grad_norm": 0.2468651055031965, "learning_rate": 3.082621950261308e-05, "loss": 0.8585, "step": 6801 }, { "epoch": 3.08201178069778, "grad_norm": 0.22012911757106318, "learning_rate": 3.08139035607567e-05, "loss": 0.845, "step": 6802 }, { "epoch": 3.082464884458541, "grad_norm": 0.5246255678330417, "learning_rate": 3.08015885381739e-05, "loss": 0.8775, "step": 6803 }, { "epoch": 3.082917988219302, "grad_norm": 0.24777100531392726, "learning_rate": 3.078927443609709e-05, "loss": 0.8666, "step": 6804 }, { "epoch": 3.0833710919800636, "grad_norm": 0.3061968879589159, "learning_rate": 3.077696125575854e-05, "loss": 0.8508, "step": 6805 }, { "epoch": 3.083824195740825, "grad_norm": 0.3630811180017651, "learning_rate": 3.076464899839047e-05, "loss": 0.8585, "step": 6806 }, { "epoch": 3.084277299501586, "grad_norm": 0.37183153077711234, "learning_rate": 3.075233766522499e-05, "loss": 0.8696, "step": 6807 }, { "epoch": 3.084730403262347, "grad_norm": 0.26674514812720856, "learning_rate": 3.074002725749414e-05, "loss": 0.8521, "step": 6808 }, { "epoch": 3.085183507023108, "grad_norm": 0.2751228810837495, "learning_rate": 3.072771777642983e-05, "loss": 0.848, "step": 6809 }, { "epoch": 3.0856366107838693, "grad_norm": 0.27495850977063824, "learning_rate": 3.071540922326391e-05, "loss": 0.853, "step": 6810 }, { "epoch": 3.086089714544631, "grad_norm": 0.24144227175427105, "learning_rate": 3.070310159922813e-05, "loss": 0.8613, "step": 6811 }, { "epoch": 3.086542818305392, "grad_norm": 0.21067173599801764, "learning_rate": 3.069079490555413e-05, "loss": 0.8362, "step": 6812 }, { "epoch": 3.086995922066153, "grad_norm": 0.29096041881129625, "learning_rate": 3.0678489143473474e-05, "loss": 0.8687, "step": 6813 }, { "epoch": 3.0874490258269143, "grad_norm": 0.3146701235698847, "learning_rate": 3.066618431421762e-05, "loss": 0.8435, "step": 6814 }, { "epoch": 3.0879021295876754, "grad_norm": 0.21931139906302038, "learning_rate": 3.065388041901795e-05, "loss": 0.8553, "step": 6815 }, { "epoch": 3.088355233348437, "grad_norm": 0.258721025625552, "learning_rate": 3.064157745910574e-05, "loss": 0.8634, "step": 6816 }, { "epoch": 3.088808337109198, "grad_norm": 0.28305538590268353, "learning_rate": 3.0629275435712176e-05, "loss": 0.8667, "step": 6817 }, { "epoch": 3.0892614408699592, "grad_norm": 0.27083678563617414, "learning_rate": 3.061697435006835e-05, "loss": 0.8306, "step": 6818 }, { "epoch": 3.0897145446307204, "grad_norm": 0.33825365148540015, "learning_rate": 3.0604674203405256e-05, "loss": 0.8673, "step": 6819 }, { "epoch": 3.0901676483914815, "grad_norm": 0.3559410740126645, "learning_rate": 3.059237499695379e-05, "loss": 0.8381, "step": 6820 }, { "epoch": 3.090620752152243, "grad_norm": 0.2729379340493814, "learning_rate": 3.0580076731944795e-05, "loss": 0.8423, "step": 6821 }, { "epoch": 3.091073855913004, "grad_norm": 0.25451170716460353, "learning_rate": 3.056777940960895e-05, "loss": 0.849, "step": 6822 }, { "epoch": 3.0915269596737653, "grad_norm": 0.2646213397797145, "learning_rate": 3.055548303117689e-05, "loss": 0.8427, "step": 6823 }, { "epoch": 3.0919800634345265, "grad_norm": 0.2999813983411666, "learning_rate": 3.054318759787913e-05, "loss": 0.8666, "step": 6824 }, { "epoch": 3.0924331671952876, "grad_norm": 0.2998863125831017, "learning_rate": 3.0530893110946114e-05, "loss": 0.8473, "step": 6825 }, { "epoch": 3.0928862709560487, "grad_norm": 0.325009554497055, "learning_rate": 3.051859957160819e-05, "loss": 0.8544, "step": 6826 }, { "epoch": 3.0933393747168103, "grad_norm": 0.3134009700609845, "learning_rate": 3.0506306981095575e-05, "loss": 0.8759, "step": 6827 }, { "epoch": 3.0937924784775714, "grad_norm": 0.23141880742412904, "learning_rate": 3.0494015340638437e-05, "loss": 0.8423, "step": 6828 }, { "epoch": 3.0942455822383326, "grad_norm": 0.20133683564307778, "learning_rate": 3.0481724651466828e-05, "loss": 0.8462, "step": 6829 }, { "epoch": 3.0946986859990937, "grad_norm": 0.2946361595893816, "learning_rate": 3.0469434914810687e-05, "loss": 0.8385, "step": 6830 }, { "epoch": 3.095151789759855, "grad_norm": 0.28135088241487727, "learning_rate": 3.0457146131899892e-05, "loss": 0.8552, "step": 6831 }, { "epoch": 3.0956048935206164, "grad_norm": 0.19322318071254505, "learning_rate": 3.0444858303964203e-05, "loss": 0.8658, "step": 6832 }, { "epoch": 3.0960579972813775, "grad_norm": 0.30478736717840677, "learning_rate": 3.043257143223329e-05, "loss": 0.8568, "step": 6833 }, { "epoch": 3.0965111010421387, "grad_norm": 0.2876539465074798, "learning_rate": 3.0420285517936734e-05, "loss": 0.8558, "step": 6834 }, { "epoch": 3.0969642048029, "grad_norm": 0.2259307764323379, "learning_rate": 3.040800056230401e-05, "loss": 0.8576, "step": 6835 }, { "epoch": 3.097417308563661, "grad_norm": 0.23508030534223873, "learning_rate": 3.03957165665645e-05, "loss": 0.8471, "step": 6836 }, { "epoch": 3.0978704123244225, "grad_norm": 0.261614314219775, "learning_rate": 3.0383433531947504e-05, "loss": 0.8592, "step": 6837 }, { "epoch": 3.0983235160851836, "grad_norm": 0.25924052644070683, "learning_rate": 3.0371151459682196e-05, "loss": 0.8749, "step": 6838 }, { "epoch": 3.0987766198459448, "grad_norm": 0.26966238205579035, "learning_rate": 3.0358870350997676e-05, "loss": 0.8651, "step": 6839 }, { "epoch": 3.099229723606706, "grad_norm": 0.30405832983093195, "learning_rate": 3.034659020712295e-05, "loss": 0.84, "step": 6840 }, { "epoch": 3.099682827367467, "grad_norm": 0.31229568014451564, "learning_rate": 3.033431102928693e-05, "loss": 0.8775, "step": 6841 }, { "epoch": 3.100135931128228, "grad_norm": 0.36453552395894595, "learning_rate": 3.032203281871838e-05, "loss": 0.8511, "step": 6842 }, { "epoch": 3.1005890348889897, "grad_norm": 0.2648545757536755, "learning_rate": 3.0309755576646044e-05, "loss": 0.8502, "step": 6843 }, { "epoch": 3.101042138649751, "grad_norm": 0.2351154665291071, "learning_rate": 3.029747930429852e-05, "loss": 0.8621, "step": 6844 }, { "epoch": 3.101495242410512, "grad_norm": 0.28042150158767676, "learning_rate": 3.028520400290433e-05, "loss": 0.8486, "step": 6845 }, { "epoch": 3.101948346171273, "grad_norm": 0.27344130820164586, "learning_rate": 3.0272929673691885e-05, "loss": 0.8616, "step": 6846 }, { "epoch": 3.1024014499320343, "grad_norm": 0.24354385522192695, "learning_rate": 3.0260656317889506e-05, "loss": 0.8444, "step": 6847 }, { "epoch": 3.102854553692796, "grad_norm": 0.22697736584467126, "learning_rate": 3.0248383936725422e-05, "loss": 0.8495, "step": 6848 }, { "epoch": 3.103307657453557, "grad_norm": 0.2483619777448867, "learning_rate": 3.0236112531427763e-05, "loss": 0.8571, "step": 6849 }, { "epoch": 3.103760761214318, "grad_norm": 0.21515170443077647, "learning_rate": 3.0223842103224532e-05, "loss": 0.8566, "step": 6850 }, { "epoch": 3.1042138649750792, "grad_norm": 0.2297630026848899, "learning_rate": 3.0211572653343676e-05, "loss": 0.844, "step": 6851 }, { "epoch": 3.1046669687358404, "grad_norm": 0.27711797810041866, "learning_rate": 3.0199304183013027e-05, "loss": 0.8442, "step": 6852 }, { "epoch": 3.105120072496602, "grad_norm": 0.23807427359450223, "learning_rate": 3.0187036693460317e-05, "loss": 0.8761, "step": 6853 }, { "epoch": 3.105573176257363, "grad_norm": 0.22616391634498376, "learning_rate": 3.017477018591318e-05, "loss": 0.8514, "step": 6854 }, { "epoch": 3.106026280018124, "grad_norm": 0.22259899757082127, "learning_rate": 3.016250466159916e-05, "loss": 0.8581, "step": 6855 }, { "epoch": 3.1064793837788853, "grad_norm": 0.18718421829507786, "learning_rate": 3.0150240121745676e-05, "loss": 0.8573, "step": 6856 }, { "epoch": 3.1069324875396465, "grad_norm": 0.26374272580738145, "learning_rate": 3.0137976567580104e-05, "loss": 0.8824, "step": 6857 }, { "epoch": 3.1073855913004076, "grad_norm": 0.2465841813572801, "learning_rate": 3.0125714000329645e-05, "loss": 0.8491, "step": 6858 }, { "epoch": 3.107838695061169, "grad_norm": 0.20052353791723174, "learning_rate": 3.011345242122147e-05, "loss": 0.8408, "step": 6859 }, { "epoch": 3.1082917988219303, "grad_norm": 0.22807471644529578, "learning_rate": 3.0101191831482613e-05, "loss": 0.8591, "step": 6860 }, { "epoch": 3.1087449025826914, "grad_norm": 0.23558692342519144, "learning_rate": 3.008893223234002e-05, "loss": 0.8657, "step": 6861 }, { "epoch": 3.1091980063434526, "grad_norm": 0.2174804202556402, "learning_rate": 3.0076673625020535e-05, "loss": 0.8565, "step": 6862 }, { "epoch": 3.1096511101042137, "grad_norm": 0.23553729375584154, "learning_rate": 3.0064416010750908e-05, "loss": 0.8808, "step": 6863 }, { "epoch": 3.1101042138649753, "grad_norm": 0.25071823687762823, "learning_rate": 3.0052159390757794e-05, "loss": 0.8522, "step": 6864 }, { "epoch": 3.1105573176257364, "grad_norm": 0.30435566877349296, "learning_rate": 3.0039903766267734e-05, "loss": 0.8631, "step": 6865 }, { "epoch": 3.1110104213864975, "grad_norm": 0.22779615789935032, "learning_rate": 3.0027649138507167e-05, "loss": 0.8639, "step": 6866 }, { "epoch": 3.1114635251472587, "grad_norm": 0.23926091864834842, "learning_rate": 3.001539550870245e-05, "loss": 0.8635, "step": 6867 }, { "epoch": 3.11191662890802, "grad_norm": 0.22525340518258552, "learning_rate": 3.000314287807983e-05, "loss": 0.8782, "step": 6868 }, { "epoch": 3.1123697326687814, "grad_norm": 0.2390499834882509, "learning_rate": 2.9990891247865455e-05, "loss": 0.8666, "step": 6869 }, { "epoch": 3.1128228364295425, "grad_norm": 0.33269981002703275, "learning_rate": 2.997864061928538e-05, "loss": 0.8531, "step": 6870 }, { "epoch": 3.1132759401903036, "grad_norm": 0.2966806995440285, "learning_rate": 2.9966390993565545e-05, "loss": 0.8548, "step": 6871 }, { "epoch": 3.1137290439510648, "grad_norm": 0.2424435384779743, "learning_rate": 2.99541423719318e-05, "loss": 0.8549, "step": 6872 }, { "epoch": 3.114182147711826, "grad_norm": 0.25613306347276965, "learning_rate": 2.9941894755609905e-05, "loss": 0.8765, "step": 6873 }, { "epoch": 3.114635251472587, "grad_norm": 0.25488549787701187, "learning_rate": 2.9929648145825484e-05, "loss": 0.8395, "step": 6874 }, { "epoch": 3.1150883552333486, "grad_norm": 0.2830337551136654, "learning_rate": 2.9917402543804098e-05, "loss": 0.8624, "step": 6875 }, { "epoch": 3.1155414589941097, "grad_norm": 0.24696783743247544, "learning_rate": 2.990515795077119e-05, "loss": 0.8663, "step": 6876 }, { "epoch": 3.115994562754871, "grad_norm": 0.25941885217581184, "learning_rate": 2.9892914367952096e-05, "loss": 0.8501, "step": 6877 }, { "epoch": 3.116447666515632, "grad_norm": 0.27495249237887764, "learning_rate": 2.9880671796572075e-05, "loss": 0.8421, "step": 6878 }, { "epoch": 3.116900770276393, "grad_norm": 0.22589463669476512, "learning_rate": 2.9868430237856256e-05, "loss": 0.8612, "step": 6879 }, { "epoch": 3.1173538740371547, "grad_norm": 0.21505379017266077, "learning_rate": 2.9856189693029686e-05, "loss": 0.8422, "step": 6880 }, { "epoch": 3.117806977797916, "grad_norm": 0.2191823766057645, "learning_rate": 2.984395016331731e-05, "loss": 0.8615, "step": 6881 }, { "epoch": 3.118260081558677, "grad_norm": 0.2239178473205481, "learning_rate": 2.9831711649943953e-05, "loss": 0.8583, "step": 6882 }, { "epoch": 3.118713185319438, "grad_norm": 0.3486854842209309, "learning_rate": 2.9819474154134352e-05, "loss": 0.8496, "step": 6883 }, { "epoch": 3.119166289080199, "grad_norm": 0.26796228698285346, "learning_rate": 2.9807237677113147e-05, "loss": 0.8574, "step": 6884 }, { "epoch": 3.119619392840961, "grad_norm": 0.27903793439533897, "learning_rate": 2.9795002220104866e-05, "loss": 0.8455, "step": 6885 }, { "epoch": 3.120072496601722, "grad_norm": 0.24522835313332156, "learning_rate": 2.978276778433394e-05, "loss": 0.8482, "step": 6886 }, { "epoch": 3.120525600362483, "grad_norm": 0.20623054824976003, "learning_rate": 2.97705343710247e-05, "loss": 0.8648, "step": 6887 }, { "epoch": 3.120978704123244, "grad_norm": 0.2485859390058601, "learning_rate": 2.9758301981401367e-05, "loss": 0.852, "step": 6888 }, { "epoch": 3.1214318078840053, "grad_norm": 0.29242570109730626, "learning_rate": 2.974607061668807e-05, "loss": 0.8611, "step": 6889 }, { "epoch": 3.1218849116447664, "grad_norm": 0.32819402442404844, "learning_rate": 2.973384027810883e-05, "loss": 0.8624, "step": 6890 }, { "epoch": 3.122338015405528, "grad_norm": 0.345786086529366, "learning_rate": 2.9721610966887558e-05, "loss": 0.848, "step": 6891 }, { "epoch": 3.122791119166289, "grad_norm": 0.28457175565724324, "learning_rate": 2.9709382684248066e-05, "loss": 0.8484, "step": 6892 }, { "epoch": 3.1232442229270503, "grad_norm": 0.22986655404494996, "learning_rate": 2.9697155431414074e-05, "loss": 0.8797, "step": 6893 }, { "epoch": 3.1236973266878114, "grad_norm": 0.29900636974388084, "learning_rate": 2.9684929209609193e-05, "loss": 0.8618, "step": 6894 }, { "epoch": 3.1241504304485725, "grad_norm": 0.32103978824741414, "learning_rate": 2.9672704020056923e-05, "loss": 0.8336, "step": 6895 }, { "epoch": 3.124603534209334, "grad_norm": 0.27187504820846836, "learning_rate": 2.966047986398066e-05, "loss": 0.8398, "step": 6896 }, { "epoch": 3.1250566379700953, "grad_norm": 0.32607989827676986, "learning_rate": 2.964825674260372e-05, "loss": 0.844, "step": 6897 }, { "epoch": 3.1255097417308564, "grad_norm": 0.2948619913309637, "learning_rate": 2.9636034657149308e-05, "loss": 0.8562, "step": 6898 }, { "epoch": 3.1259628454916175, "grad_norm": 0.23166439712168285, "learning_rate": 2.962381360884048e-05, "loss": 0.8663, "step": 6899 }, { "epoch": 3.1264159492523786, "grad_norm": 0.29499682840891556, "learning_rate": 2.9611593598900236e-05, "loss": 0.867, "step": 6900 }, { "epoch": 3.1268690530131398, "grad_norm": 0.271415653970264, "learning_rate": 2.9599374628551477e-05, "loss": 0.8608, "step": 6901 }, { "epoch": 3.1273221567739014, "grad_norm": 0.29793495439502937, "learning_rate": 2.958715669901697e-05, "loss": 0.8661, "step": 6902 }, { "epoch": 3.1277752605346625, "grad_norm": 0.39064555917847926, "learning_rate": 2.9574939811519384e-05, "loss": 0.8788, "step": 6903 }, { "epoch": 3.1282283642954236, "grad_norm": 0.2440136213818834, "learning_rate": 2.956272396728131e-05, "loss": 0.8486, "step": 6904 }, { "epoch": 3.1286814680561847, "grad_norm": 0.28950066910463573, "learning_rate": 2.9550509167525193e-05, "loss": 0.8614, "step": 6905 }, { "epoch": 3.129134571816946, "grad_norm": 0.3595057212458713, "learning_rate": 2.9538295413473422e-05, "loss": 0.8609, "step": 6906 }, { "epoch": 3.1295876755777075, "grad_norm": 0.2844714152508288, "learning_rate": 2.9526082706348228e-05, "loss": 0.8395, "step": 6907 }, { "epoch": 3.1300407793384686, "grad_norm": 0.36441593804161854, "learning_rate": 2.9513871047371787e-05, "loss": 0.8532, "step": 6908 }, { "epoch": 3.1304938830992297, "grad_norm": 0.25481636413141956, "learning_rate": 2.950166043776614e-05, "loss": 0.8712, "step": 6909 }, { "epoch": 3.130946986859991, "grad_norm": 0.3449318615149107, "learning_rate": 2.9489450878753233e-05, "loss": 0.8641, "step": 6910 }, { "epoch": 3.131400090620752, "grad_norm": 0.3931281487084758, "learning_rate": 2.9477242371554893e-05, "loss": 0.8666, "step": 6911 }, { "epoch": 3.1318531943815135, "grad_norm": 0.28941581910246433, "learning_rate": 2.946503491739286e-05, "loss": 0.8432, "step": 6912 }, { "epoch": 3.1323062981422747, "grad_norm": 0.28175271491116455, "learning_rate": 2.945282851748876e-05, "loss": 0.8585, "step": 6913 }, { "epoch": 3.132759401903036, "grad_norm": 0.28444834821152676, "learning_rate": 2.944062317306412e-05, "loss": 0.8738, "step": 6914 }, { "epoch": 3.133212505663797, "grad_norm": 0.28123305741763355, "learning_rate": 2.942841888534035e-05, "loss": 0.871, "step": 6915 }, { "epoch": 3.133665609424558, "grad_norm": 0.3347582640361715, "learning_rate": 2.941621565553877e-05, "loss": 0.8479, "step": 6916 }, { "epoch": 3.1341187131853196, "grad_norm": 0.2977632777885195, "learning_rate": 2.940401348488058e-05, "loss": 0.8526, "step": 6917 }, { "epoch": 3.134571816946081, "grad_norm": 0.30161472149254587, "learning_rate": 2.9391812374586882e-05, "loss": 0.8503, "step": 6918 }, { "epoch": 3.135024920706842, "grad_norm": 0.27912543223083064, "learning_rate": 2.9379612325878662e-05, "loss": 0.8664, "step": 6919 }, { "epoch": 3.135478024467603, "grad_norm": 0.2591530178998119, "learning_rate": 2.9367413339976804e-05, "loss": 0.8628, "step": 6920 }, { "epoch": 3.135931128228364, "grad_norm": 0.29672591341778415, "learning_rate": 2.9355215418102098e-05, "loss": 0.8526, "step": 6921 }, { "epoch": 3.1363842319891253, "grad_norm": 0.27564050734136764, "learning_rate": 2.9343018561475216e-05, "loss": 0.8617, "step": 6922 }, { "epoch": 3.136837335749887, "grad_norm": 0.21856784912603283, "learning_rate": 2.9330822771316724e-05, "loss": 0.8693, "step": 6923 }, { "epoch": 3.137290439510648, "grad_norm": 0.2737573938292888, "learning_rate": 2.9318628048847076e-05, "loss": 0.8654, "step": 6924 }, { "epoch": 3.137743543271409, "grad_norm": 0.2689197703467019, "learning_rate": 2.930643439528664e-05, "loss": 0.8695, "step": 6925 }, { "epoch": 3.1381966470321703, "grad_norm": 0.23153168815552466, "learning_rate": 2.9294241811855656e-05, "loss": 0.8436, "step": 6926 }, { "epoch": 3.1386497507929314, "grad_norm": 0.23567673558970045, "learning_rate": 2.9282050299774252e-05, "loss": 0.8503, "step": 6927 }, { "epoch": 3.139102854553693, "grad_norm": 0.28350885200162024, "learning_rate": 2.9269859860262477e-05, "loss": 0.8699, "step": 6928 }, { "epoch": 3.139555958314454, "grad_norm": 0.23661025345807377, "learning_rate": 2.9257670494540248e-05, "loss": 0.8509, "step": 6929 }, { "epoch": 3.1400090620752152, "grad_norm": 0.2585846891845243, "learning_rate": 2.9245482203827387e-05, "loss": 0.8805, "step": 6930 }, { "epoch": 3.1404621658359764, "grad_norm": 0.23724655045097243, "learning_rate": 2.9233294989343593e-05, "loss": 0.8479, "step": 6931 }, { "epoch": 3.1409152695967375, "grad_norm": 0.19477421738901904, "learning_rate": 2.922110885230848e-05, "loss": 0.8741, "step": 6932 }, { "epoch": 3.1413683733574986, "grad_norm": 0.210880664408046, "learning_rate": 2.920892379394154e-05, "loss": 0.8572, "step": 6933 }, { "epoch": 3.14182147711826, "grad_norm": 0.21552241314868636, "learning_rate": 2.9196739815462164e-05, "loss": 0.8414, "step": 6934 }, { "epoch": 3.1422745808790213, "grad_norm": 0.25816937657433975, "learning_rate": 2.9184556918089614e-05, "loss": 0.847, "step": 6935 }, { "epoch": 3.1427276846397825, "grad_norm": 0.2704727956770071, "learning_rate": 2.9172375103043076e-05, "loss": 0.8533, "step": 6936 }, { "epoch": 3.1431807884005436, "grad_norm": 0.26982597681427456, "learning_rate": 2.9160194371541606e-05, "loss": 0.8529, "step": 6937 }, { "epoch": 3.1436338921613047, "grad_norm": 0.31396847840533876, "learning_rate": 2.9148014724804156e-05, "loss": 0.8951, "step": 6938 }, { "epoch": 3.1440869959220663, "grad_norm": 0.2675074864607281, "learning_rate": 2.9135836164049572e-05, "loss": 0.8399, "step": 6939 }, { "epoch": 3.1445400996828274, "grad_norm": 0.2500055891653405, "learning_rate": 2.9123658690496595e-05, "loss": 0.8678, "step": 6940 }, { "epoch": 3.1449932034435886, "grad_norm": 0.24164837679540316, "learning_rate": 2.9111482305363846e-05, "loss": 0.8274, "step": 6941 }, { "epoch": 3.1454463072043497, "grad_norm": 0.21220747766284126, "learning_rate": 2.909930700986986e-05, "loss": 0.8471, "step": 6942 }, { "epoch": 3.145899410965111, "grad_norm": 0.2860059442803462, "learning_rate": 2.908713280523302e-05, "loss": 0.8489, "step": 6943 }, { "epoch": 3.1463525147258724, "grad_norm": 0.21339461073809302, "learning_rate": 2.907495969267164e-05, "loss": 0.8428, "step": 6944 }, { "epoch": 3.1468056184866335, "grad_norm": 0.24180111813193755, "learning_rate": 2.906278767340391e-05, "loss": 0.8614, "step": 6945 }, { "epoch": 3.1472587222473947, "grad_norm": 0.20944816698338428, "learning_rate": 2.905061674864792e-05, "loss": 0.8451, "step": 6946 }, { "epoch": 3.147711826008156, "grad_norm": 0.257223002208461, "learning_rate": 2.9038446919621627e-05, "loss": 0.872, "step": 6947 }, { "epoch": 3.148164929768917, "grad_norm": 0.23714617262384313, "learning_rate": 2.9026278187542903e-05, "loss": 0.8698, "step": 6948 }, { "epoch": 3.1486180335296785, "grad_norm": 0.3600547341395043, "learning_rate": 2.9014110553629494e-05, "loss": 0.854, "step": 6949 }, { "epoch": 3.1490711372904396, "grad_norm": 0.20038186535524466, "learning_rate": 2.9001944019099056e-05, "loss": 0.8556, "step": 6950 }, { "epoch": 3.1495242410512008, "grad_norm": 0.21553708150303091, "learning_rate": 2.8989778585169115e-05, "loss": 0.8716, "step": 6951 }, { "epoch": 3.149977344811962, "grad_norm": 0.20354338069316932, "learning_rate": 2.8977614253057087e-05, "loss": 0.8606, "step": 6952 }, { "epoch": 3.150430448572723, "grad_norm": 0.2129515615292999, "learning_rate": 2.8965451023980287e-05, "loss": 0.8669, "step": 6953 }, { "epoch": 3.150883552333484, "grad_norm": 0.23672192898815003, "learning_rate": 2.8953288899155923e-05, "loss": 0.8487, "step": 6954 }, { "epoch": 3.1513366560942457, "grad_norm": 0.32002986173044246, "learning_rate": 2.8941127879801095e-05, "loss": 0.8472, "step": 6955 }, { "epoch": 3.151789759855007, "grad_norm": 0.29349377887297584, "learning_rate": 2.8928967967132763e-05, "loss": 0.8592, "step": 6956 }, { "epoch": 3.152242863615768, "grad_norm": 0.2197561690753047, "learning_rate": 2.8916809162367802e-05, "loss": 0.8619, "step": 6957 }, { "epoch": 3.152695967376529, "grad_norm": 0.2501533287832186, "learning_rate": 2.890465146672297e-05, "loss": 0.8506, "step": 6958 }, { "epoch": 3.1531490711372903, "grad_norm": 0.25305990229832165, "learning_rate": 2.8892494881414924e-05, "loss": 0.8657, "step": 6959 }, { "epoch": 3.153602174898052, "grad_norm": 0.25053185301302827, "learning_rate": 2.8880339407660202e-05, "loss": 0.8438, "step": 6960 }, { "epoch": 3.154055278658813, "grad_norm": 0.2337380977987983, "learning_rate": 2.886818504667522e-05, "loss": 0.8643, "step": 6961 }, { "epoch": 3.154508382419574, "grad_norm": 0.26697336639624547, "learning_rate": 2.885603179967631e-05, "loss": 0.8758, "step": 6962 }, { "epoch": 3.1549614861803352, "grad_norm": 0.2653631623329183, "learning_rate": 2.8843879667879647e-05, "loss": 0.8652, "step": 6963 }, { "epoch": 3.1554145899410964, "grad_norm": 0.31595259857635366, "learning_rate": 2.883172865250135e-05, "loss": 0.8525, "step": 6964 }, { "epoch": 3.1558676937018575, "grad_norm": 0.272151001220066, "learning_rate": 2.8819578754757382e-05, "loss": 0.866, "step": 6965 }, { "epoch": 3.156320797462619, "grad_norm": 0.26501770070241454, "learning_rate": 2.880742997586362e-05, "loss": 0.8606, "step": 6966 }, { "epoch": 3.15677390122338, "grad_norm": 0.2620251848304312, "learning_rate": 2.8795282317035805e-05, "loss": 0.8601, "step": 6967 }, { "epoch": 3.1572270049841413, "grad_norm": 0.23496350908713967, "learning_rate": 2.8783135779489603e-05, "loss": 0.846, "step": 6968 }, { "epoch": 3.1576801087449025, "grad_norm": 0.2691897781165931, "learning_rate": 2.8770990364440536e-05, "loss": 0.8515, "step": 6969 }, { "epoch": 3.1581332125056636, "grad_norm": 0.26145195247388653, "learning_rate": 2.8758846073104025e-05, "loss": 0.8411, "step": 6970 }, { "epoch": 3.158586316266425, "grad_norm": 0.25188435315985985, "learning_rate": 2.874670290669537e-05, "loss": 0.8461, "step": 6971 }, { "epoch": 3.1590394200271863, "grad_norm": 0.29029362508888706, "learning_rate": 2.8734560866429766e-05, "loss": 0.8626, "step": 6972 }, { "epoch": 3.1594925237879474, "grad_norm": 0.305559752540115, "learning_rate": 2.87224199535223e-05, "loss": 0.8385, "step": 6973 }, { "epoch": 3.1599456275487086, "grad_norm": 0.23643007551730935, "learning_rate": 2.8710280169187937e-05, "loss": 0.848, "step": 6974 }, { "epoch": 3.1603987313094697, "grad_norm": 0.2683996407057784, "learning_rate": 2.8698141514641533e-05, "loss": 0.8726, "step": 6975 }, { "epoch": 3.1608518350702313, "grad_norm": 0.2512738738594623, "learning_rate": 2.8686003991097837e-05, "loss": 0.8554, "step": 6976 }, { "epoch": 3.1613049388309924, "grad_norm": 0.3543045225081176, "learning_rate": 2.867386759977147e-05, "loss": 0.8508, "step": 6977 }, { "epoch": 3.1617580425917535, "grad_norm": 0.35859528033468313, "learning_rate": 2.8661732341876956e-05, "loss": 0.8687, "step": 6978 }, { "epoch": 3.1622111463525147, "grad_norm": 0.2451662461696374, "learning_rate": 2.8649598218628703e-05, "loss": 0.8692, "step": 6979 }, { "epoch": 3.162664250113276, "grad_norm": 0.19836677573227265, "learning_rate": 2.863746523124098e-05, "loss": 0.847, "step": 6980 }, { "epoch": 3.1631173538740374, "grad_norm": 0.26330612496474404, "learning_rate": 2.8625333380927974e-05, "loss": 0.864, "step": 6981 }, { "epoch": 3.1635704576347985, "grad_norm": 0.24355472989406646, "learning_rate": 2.861320266890375e-05, "loss": 0.8493, "step": 6982 }, { "epoch": 3.1640235613955596, "grad_norm": 0.22486860788771623, "learning_rate": 2.860107309638225e-05, "loss": 0.8517, "step": 6983 }, { "epoch": 3.1644766651563208, "grad_norm": 0.23825583976776646, "learning_rate": 2.858894466457731e-05, "loss": 0.8785, "step": 6984 }, { "epoch": 3.164929768917082, "grad_norm": 0.18777492874725518, "learning_rate": 2.8576817374702653e-05, "loss": 0.8546, "step": 6985 }, { "epoch": 3.165382872677843, "grad_norm": 0.23411065017618854, "learning_rate": 2.8564691227971884e-05, "loss": 0.8694, "step": 6986 }, { "epoch": 3.1658359764386046, "grad_norm": 0.23005067880792415, "learning_rate": 2.85525662255985e-05, "loss": 0.8372, "step": 6987 }, { "epoch": 3.1662890801993657, "grad_norm": 0.21280081330399217, "learning_rate": 2.854044236879586e-05, "loss": 0.8495, "step": 6988 }, { "epoch": 3.166742183960127, "grad_norm": 0.1943465904263312, "learning_rate": 2.852831965877724e-05, "loss": 0.8803, "step": 6989 }, { "epoch": 3.167195287720888, "grad_norm": 0.26762583172531196, "learning_rate": 2.851619809675578e-05, "loss": 0.8755, "step": 6990 }, { "epoch": 3.167648391481649, "grad_norm": 0.3040600929520484, "learning_rate": 2.850407768394451e-05, "loss": 0.8519, "step": 6991 }, { "epoch": 3.1681014952424107, "grad_norm": 0.36621360101468964, "learning_rate": 2.849195842155636e-05, "loss": 0.8688, "step": 6992 }, { "epoch": 3.168554599003172, "grad_norm": 0.31324142514784464, "learning_rate": 2.8479840310804124e-05, "loss": 0.8531, "step": 6993 }, { "epoch": 3.169007702763933, "grad_norm": 0.26429350522496137, "learning_rate": 2.846772335290049e-05, "loss": 0.8482, "step": 6994 }, { "epoch": 3.169460806524694, "grad_norm": 0.26932496082786567, "learning_rate": 2.8455607549058037e-05, "loss": 0.8412, "step": 6995 }, { "epoch": 3.169913910285455, "grad_norm": 0.3085647844116436, "learning_rate": 2.8443492900489204e-05, "loss": 0.8654, "step": 6996 }, { "epoch": 3.170367014046217, "grad_norm": 0.3102829402686287, "learning_rate": 2.8431379408406337e-05, "loss": 0.8563, "step": 6997 }, { "epoch": 3.170820117806978, "grad_norm": 0.30342125089317423, "learning_rate": 2.841926707402167e-05, "loss": 0.8592, "step": 6998 }, { "epoch": 3.171273221567739, "grad_norm": 0.2368366541509572, "learning_rate": 2.840715589854731e-05, "loss": 0.8349, "step": 6999 }, { "epoch": 3.1717263253285, "grad_norm": 0.2769774736435075, "learning_rate": 2.8395045883195238e-05, "loss": 0.8541, "step": 7000 }, { "epoch": 3.1721794290892613, "grad_norm": 0.2910457269467194, "learning_rate": 2.838293702917734e-05, "loss": 0.8437, "step": 7001 }, { "epoch": 3.172632532850023, "grad_norm": 0.2652133889966474, "learning_rate": 2.837082933770538e-05, "loss": 0.8553, "step": 7002 }, { "epoch": 3.173085636610784, "grad_norm": 0.32079652884052534, "learning_rate": 2.8358722809991007e-05, "loss": 0.8584, "step": 7003 }, { "epoch": 3.173538740371545, "grad_norm": 0.317786910413703, "learning_rate": 2.834661744724573e-05, "loss": 0.8658, "step": 7004 }, { "epoch": 3.1739918441323063, "grad_norm": 0.219363791031108, "learning_rate": 2.8334513250680973e-05, "loss": 0.8603, "step": 7005 }, { "epoch": 3.1744449478930674, "grad_norm": 0.263796643296055, "learning_rate": 2.8322410221508022e-05, "loss": 0.8622, "step": 7006 }, { "epoch": 3.1748980516538285, "grad_norm": 0.22883036369618734, "learning_rate": 2.831030836093806e-05, "loss": 0.8547, "step": 7007 }, { "epoch": 3.17535115541459, "grad_norm": 0.25212521325309917, "learning_rate": 2.8298207670182153e-05, "loss": 0.8701, "step": 7008 }, { "epoch": 3.1758042591753513, "grad_norm": 0.236680498493163, "learning_rate": 2.8286108150451234e-05, "loss": 0.8583, "step": 7009 }, { "epoch": 3.1762573629361124, "grad_norm": 0.2706089937281074, "learning_rate": 2.8274009802956136e-05, "loss": 0.8651, "step": 7010 }, { "epoch": 3.1767104666968735, "grad_norm": 0.2682090084112634, "learning_rate": 2.8261912628907578e-05, "loss": 0.854, "step": 7011 }, { "epoch": 3.1771635704576346, "grad_norm": 0.22745380225574663, "learning_rate": 2.824981662951614e-05, "loss": 0.8695, "step": 7012 }, { "epoch": 3.177616674218396, "grad_norm": 0.28238870156146895, "learning_rate": 2.823772180599229e-05, "loss": 0.8664, "step": 7013 }, { "epoch": 3.1780697779791574, "grad_norm": 0.3084374114390799, "learning_rate": 2.8225628159546394e-05, "loss": 0.8568, "step": 7014 }, { "epoch": 3.1785228817399185, "grad_norm": 0.20057822159127559, "learning_rate": 2.821353569138869e-05, "loss": 0.8611, "step": 7015 }, { "epoch": 3.1789759855006796, "grad_norm": 0.3667422779915181, "learning_rate": 2.8201444402729293e-05, "loss": 0.866, "step": 7016 }, { "epoch": 3.1794290892614407, "grad_norm": 0.3093548415948781, "learning_rate": 2.8189354294778212e-05, "loss": 0.8554, "step": 7017 }, { "epoch": 3.179882193022202, "grad_norm": 0.2741204728880231, "learning_rate": 2.8177265368745327e-05, "loss": 0.8462, "step": 7018 }, { "epoch": 3.1803352967829635, "grad_norm": 0.29842067569618014, "learning_rate": 2.8165177625840406e-05, "loss": 0.8406, "step": 7019 }, { "epoch": 3.1807884005437246, "grad_norm": 0.27884738592062935, "learning_rate": 2.81530910672731e-05, "loss": 0.8468, "step": 7020 }, { "epoch": 3.1812415043044857, "grad_norm": 0.30956293847066907, "learning_rate": 2.814100569425293e-05, "loss": 0.8447, "step": 7021 }, { "epoch": 3.181694608065247, "grad_norm": 0.3296505042694299, "learning_rate": 2.8128921507989323e-05, "loss": 0.8442, "step": 7022 }, { "epoch": 3.182147711826008, "grad_norm": 0.23300091433702885, "learning_rate": 2.811683850969156e-05, "loss": 0.8574, "step": 7023 }, { "epoch": 3.1826008155867695, "grad_norm": 0.28235365605314816, "learning_rate": 2.8104756700568816e-05, "loss": 0.852, "step": 7024 }, { "epoch": 3.1830539193475307, "grad_norm": 0.3494817881210834, "learning_rate": 2.8092676081830136e-05, "loss": 0.8358, "step": 7025 }, { "epoch": 3.183507023108292, "grad_norm": 0.222337021150125, "learning_rate": 2.808059665468447e-05, "loss": 0.8563, "step": 7026 }, { "epoch": 3.183960126869053, "grad_norm": 0.3136333626306609, "learning_rate": 2.8068518420340622e-05, "loss": 0.8481, "step": 7027 }, { "epoch": 3.184413230629814, "grad_norm": 0.2386599165669791, "learning_rate": 2.80564413800073e-05, "loss": 0.8278, "step": 7028 }, { "epoch": 3.1848663343905756, "grad_norm": 0.2367756893783231, "learning_rate": 2.804436553489307e-05, "loss": 0.8517, "step": 7029 }, { "epoch": 3.185319438151337, "grad_norm": 0.25630700614234747, "learning_rate": 2.8032290886206393e-05, "loss": 0.8624, "step": 7030 }, { "epoch": 3.185772541912098, "grad_norm": 0.21130421485095835, "learning_rate": 2.802021743515562e-05, "loss": 0.8588, "step": 7031 }, { "epoch": 3.186225645672859, "grad_norm": 0.22523828182828834, "learning_rate": 2.8008145182948942e-05, "loss": 0.8628, "step": 7032 }, { "epoch": 3.18667874943362, "grad_norm": 0.26174062488034017, "learning_rate": 2.7996074130794477e-05, "loss": 0.8633, "step": 7033 }, { "epoch": 3.1871318531943817, "grad_norm": 0.3162921697093732, "learning_rate": 2.7984004279900197e-05, "loss": 0.8591, "step": 7034 }, { "epoch": 3.187584956955143, "grad_norm": 0.2870343308801343, "learning_rate": 2.7971935631473953e-05, "loss": 0.8478, "step": 7035 }, { "epoch": 3.188038060715904, "grad_norm": 0.27374663209558814, "learning_rate": 2.795986818672349e-05, "loss": 0.852, "step": 7036 }, { "epoch": 3.188491164476665, "grad_norm": 0.268455970212713, "learning_rate": 2.7947801946856424e-05, "loss": 0.8574, "step": 7037 }, { "epoch": 3.1889442682374263, "grad_norm": 0.2234790067660105, "learning_rate": 2.7935736913080246e-05, "loss": 0.8525, "step": 7038 }, { "epoch": 3.1893973719981874, "grad_norm": 0.23850685726254095, "learning_rate": 2.792367308660234e-05, "loss": 0.8693, "step": 7039 }, { "epoch": 3.189850475758949, "grad_norm": 0.2940727319197081, "learning_rate": 2.7911610468629966e-05, "loss": 0.8578, "step": 7040 }, { "epoch": 3.19030357951971, "grad_norm": 0.28984228186394123, "learning_rate": 2.7899549060370237e-05, "loss": 0.8657, "step": 7041 }, { "epoch": 3.1907566832804712, "grad_norm": 0.24979944682773914, "learning_rate": 2.7887488863030175e-05, "loss": 0.8498, "step": 7042 }, { "epoch": 3.1912097870412324, "grad_norm": 0.2821219695074692, "learning_rate": 2.7875429877816673e-05, "loss": 0.8474, "step": 7043 }, { "epoch": 3.1916628908019935, "grad_norm": 0.21362498831246599, "learning_rate": 2.7863372105936498e-05, "loss": 0.8575, "step": 7044 }, { "epoch": 3.192115994562755, "grad_norm": 0.2337593810724454, "learning_rate": 2.7851315548596304e-05, "loss": 0.8324, "step": 7045 }, { "epoch": 3.192569098323516, "grad_norm": 0.2504266867308288, "learning_rate": 2.7839260207002615e-05, "loss": 0.8641, "step": 7046 }, { "epoch": 3.1930222020842773, "grad_norm": 0.25560691683081405, "learning_rate": 2.7827206082361838e-05, "loss": 0.867, "step": 7047 }, { "epoch": 3.1934753058450385, "grad_norm": 0.2472445687542034, "learning_rate": 2.781515317588026e-05, "loss": 0.8546, "step": 7048 }, { "epoch": 3.1939284096057996, "grad_norm": 0.2634672228468619, "learning_rate": 2.7803101488764036e-05, "loss": 0.8354, "step": 7049 }, { "epoch": 3.1943815133665607, "grad_norm": 0.2794459156028952, "learning_rate": 2.7791051022219204e-05, "loss": 0.8454, "step": 7050 }, { "epoch": 3.1948346171273223, "grad_norm": 0.24736827649942306, "learning_rate": 2.777900177745169e-05, "loss": 0.85, "step": 7051 }, { "epoch": 3.1952877208880834, "grad_norm": 0.29888055958353293, "learning_rate": 2.7766953755667273e-05, "loss": 0.8553, "step": 7052 }, { "epoch": 3.1957408246488446, "grad_norm": 0.2120787810199299, "learning_rate": 2.775490695807165e-05, "loss": 0.8602, "step": 7053 }, { "epoch": 3.1961939284096057, "grad_norm": 0.23805290643381005, "learning_rate": 2.7742861385870353e-05, "loss": 0.8791, "step": 7054 }, { "epoch": 3.196647032170367, "grad_norm": 0.26285644599486163, "learning_rate": 2.7730817040268823e-05, "loss": 0.869, "step": 7055 }, { "epoch": 3.1971001359311284, "grad_norm": 0.20714472492183914, "learning_rate": 2.771877392247236e-05, "loss": 0.8757, "step": 7056 }, { "epoch": 3.1975532396918895, "grad_norm": 0.18901471397869926, "learning_rate": 2.7706732033686142e-05, "loss": 0.8669, "step": 7057 }, { "epoch": 3.1980063434526507, "grad_norm": 0.20616206700534945, "learning_rate": 2.769469137511523e-05, "loss": 0.8403, "step": 7058 }, { "epoch": 3.198459447213412, "grad_norm": 0.23048555767493575, "learning_rate": 2.7682651947964557e-05, "loss": 0.8735, "step": 7059 }, { "epoch": 3.198912550974173, "grad_norm": 0.2098032422185303, "learning_rate": 2.767061375343894e-05, "loss": 0.8295, "step": 7060 }, { "epoch": 3.1993656547349345, "grad_norm": 0.2130344704931421, "learning_rate": 2.765857679274307e-05, "loss": 0.8546, "step": 7061 }, { "epoch": 3.1998187584956956, "grad_norm": 0.24023160418545203, "learning_rate": 2.7646541067081512e-05, "loss": 0.8561, "step": 7062 }, { "epoch": 3.2002718622564568, "grad_norm": 0.21352945959732111, "learning_rate": 2.7634506577658707e-05, "loss": 0.8634, "step": 7063 }, { "epoch": 3.200724966017218, "grad_norm": 0.20866335342078074, "learning_rate": 2.7622473325678982e-05, "loss": 0.8476, "step": 7064 }, { "epoch": 3.201178069777979, "grad_norm": 0.26411333724219443, "learning_rate": 2.7610441312346518e-05, "loss": 0.8649, "step": 7065 }, { "epoch": 3.2016311735387406, "grad_norm": 0.21351918560253713, "learning_rate": 2.7598410538865394e-05, "loss": 0.8475, "step": 7066 }, { "epoch": 3.2020842772995017, "grad_norm": 0.22290559504851218, "learning_rate": 2.758638100643955e-05, "loss": 0.8488, "step": 7067 }, { "epoch": 3.202537381060263, "grad_norm": 0.28738822247888174, "learning_rate": 2.757435271627282e-05, "loss": 0.8572, "step": 7068 }, { "epoch": 3.202990484821024, "grad_norm": 0.22441091625041334, "learning_rate": 2.7562325669568913e-05, "loss": 0.8561, "step": 7069 }, { "epoch": 3.203443588581785, "grad_norm": 0.27440028829546753, "learning_rate": 2.755029986753136e-05, "loss": 0.86, "step": 7070 }, { "epoch": 3.2038966923425463, "grad_norm": 0.2624121968458042, "learning_rate": 2.753827531136365e-05, "loss": 0.8418, "step": 7071 }, { "epoch": 3.204349796103308, "grad_norm": 0.2174184520924744, "learning_rate": 2.7526252002269082e-05, "loss": 0.8829, "step": 7072 }, { "epoch": 3.204802899864069, "grad_norm": 0.2345969089029024, "learning_rate": 2.751422994145088e-05, "loss": 0.8695, "step": 7073 }, { "epoch": 3.20525600362483, "grad_norm": 0.22899599559262687, "learning_rate": 2.7502209130112098e-05, "loss": 0.8774, "step": 7074 }, { "epoch": 3.2057091073855912, "grad_norm": 0.2358462116816445, "learning_rate": 2.74901895694557e-05, "loss": 0.8728, "step": 7075 }, { "epoch": 3.2061622111463524, "grad_norm": 0.2863659374258244, "learning_rate": 2.7478171260684515e-05, "loss": 0.8641, "step": 7076 }, { "epoch": 3.206615314907114, "grad_norm": 0.3207565492646311, "learning_rate": 2.7466154205001223e-05, "loss": 0.8658, "step": 7077 }, { "epoch": 3.207068418667875, "grad_norm": 0.3381163225170988, "learning_rate": 2.7454138403608407e-05, "loss": 0.8514, "step": 7078 }, { "epoch": 3.207521522428636, "grad_norm": 0.28977862074582433, "learning_rate": 2.744212385770852e-05, "loss": 0.8524, "step": 7079 }, { "epoch": 3.2079746261893973, "grad_norm": 0.27259359800007554, "learning_rate": 2.7430110568503887e-05, "loss": 0.8452, "step": 7080 }, { "epoch": 3.2084277299501585, "grad_norm": 0.2690743379302954, "learning_rate": 2.7418098537196693e-05, "loss": 0.8668, "step": 7081 }, { "epoch": 3.2088808337109196, "grad_norm": 0.3121736212503154, "learning_rate": 2.7406087764989017e-05, "loss": 0.8658, "step": 7082 }, { "epoch": 3.209333937471681, "grad_norm": 0.25579621156218596, "learning_rate": 2.7394078253082805e-05, "loss": 0.8625, "step": 7083 }, { "epoch": 3.2097870412324423, "grad_norm": 0.211634967255306, "learning_rate": 2.7382070002679885e-05, "loss": 0.8599, "step": 7084 }, { "epoch": 3.2102401449932034, "grad_norm": 0.2569856219199704, "learning_rate": 2.737006301498194e-05, "loss": 0.8578, "step": 7085 }, { "epoch": 3.2106932487539646, "grad_norm": 0.20928361642103502, "learning_rate": 2.7358057291190524e-05, "loss": 0.8592, "step": 7086 }, { "epoch": 3.2111463525147257, "grad_norm": 0.218091886477514, "learning_rate": 2.7346052832507096e-05, "loss": 0.841, "step": 7087 }, { "epoch": 3.2115994562754873, "grad_norm": 0.21128948329519798, "learning_rate": 2.7334049640132963e-05, "loss": 0.8517, "step": 7088 }, { "epoch": 3.2120525600362484, "grad_norm": 0.21197580586597436, "learning_rate": 2.7322047715269307e-05, "loss": 0.859, "step": 7089 }, { "epoch": 3.2125056637970095, "grad_norm": 0.2500489474981061, "learning_rate": 2.73100470591172e-05, "loss": 0.8583, "step": 7090 }, { "epoch": 3.2129587675577707, "grad_norm": 0.2761748437159554, "learning_rate": 2.7298047672877566e-05, "loss": 0.8554, "step": 7091 }, { "epoch": 3.213411871318532, "grad_norm": 0.20634037508616643, "learning_rate": 2.7286049557751223e-05, "loss": 0.8362, "step": 7092 }, { "epoch": 3.2138649750792934, "grad_norm": 0.27981315013896013, "learning_rate": 2.727405271493883e-05, "loss": 0.8504, "step": 7093 }, { "epoch": 3.2143180788400545, "grad_norm": 0.23497186221077546, "learning_rate": 2.7262057145640946e-05, "loss": 0.8671, "step": 7094 }, { "epoch": 3.2147711826008156, "grad_norm": 0.22479781737822138, "learning_rate": 2.7250062851057998e-05, "loss": 0.8704, "step": 7095 }, { "epoch": 3.2152242863615768, "grad_norm": 0.1955612011892064, "learning_rate": 2.723806983239028e-05, "loss": 0.8431, "step": 7096 }, { "epoch": 3.215677390122338, "grad_norm": 0.1939397604632136, "learning_rate": 2.7226078090837965e-05, "loss": 0.8461, "step": 7097 }, { "epoch": 3.2161304938830995, "grad_norm": 0.356744420772555, "learning_rate": 2.721408762760109e-05, "loss": 0.9, "step": 7098 }, { "epoch": 3.2165835976438606, "grad_norm": 0.2709971660728218, "learning_rate": 2.720209844387957e-05, "loss": 0.881, "step": 7099 }, { "epoch": 3.2170367014046217, "grad_norm": 0.2956411199361788, "learning_rate": 2.7190110540873193e-05, "loss": 0.8731, "step": 7100 }, { "epoch": 3.217489805165383, "grad_norm": 0.23593047790113222, "learning_rate": 2.7178123919781608e-05, "loss": 0.8264, "step": 7101 }, { "epoch": 3.217942908926144, "grad_norm": 0.22673329639609813, "learning_rate": 2.716613858180434e-05, "loss": 0.8794, "step": 7102 }, { "epoch": 3.218396012686905, "grad_norm": 0.25490354175076213, "learning_rate": 2.7154154528140803e-05, "loss": 0.8488, "step": 7103 }, { "epoch": 3.2188491164476667, "grad_norm": 0.23816973166532276, "learning_rate": 2.714217175999026e-05, "loss": 0.8588, "step": 7104 }, { "epoch": 3.219302220208428, "grad_norm": 0.23177071101462934, "learning_rate": 2.7130190278551853e-05, "loss": 0.8439, "step": 7105 }, { "epoch": 3.219755323969189, "grad_norm": 0.22236024238089874, "learning_rate": 2.71182100850246e-05, "loss": 0.8654, "step": 7106 }, { "epoch": 3.22020842772995, "grad_norm": 0.2068661809405282, "learning_rate": 2.7106231180607387e-05, "loss": 0.8705, "step": 7107 }, { "epoch": 3.220661531490711, "grad_norm": 0.3554660370293646, "learning_rate": 2.7094253566498966e-05, "loss": 0.8662, "step": 7108 }, { "epoch": 3.221114635251473, "grad_norm": 0.21685841893721267, "learning_rate": 2.708227724389797e-05, "loss": 0.8495, "step": 7109 }, { "epoch": 3.221567739012234, "grad_norm": 0.2621587282830676, "learning_rate": 2.707030221400289e-05, "loss": 0.8771, "step": 7110 }, { "epoch": 3.222020842772995, "grad_norm": 0.26029841340277304, "learning_rate": 2.7058328478012098e-05, "loss": 0.8625, "step": 7111 }, { "epoch": 3.222473946533756, "grad_norm": 0.22696092492780598, "learning_rate": 2.704635603712383e-05, "loss": 0.8353, "step": 7112 }, { "epoch": 3.2229270502945173, "grad_norm": 0.20963580550415364, "learning_rate": 2.7034384892536204e-05, "loss": 0.8449, "step": 7113 }, { "epoch": 3.2233801540552784, "grad_norm": 0.24651754114514687, "learning_rate": 2.702241504544719e-05, "loss": 0.8569, "step": 7114 }, { "epoch": 3.22383325781604, "grad_norm": 0.2785001085849498, "learning_rate": 2.7010446497054635e-05, "loss": 0.8585, "step": 7115 }, { "epoch": 3.224286361576801, "grad_norm": 0.2311356796234111, "learning_rate": 2.699847924855627e-05, "loss": 0.8472, "step": 7116 }, { "epoch": 3.2247394653375623, "grad_norm": 0.26986918433835927, "learning_rate": 2.698651330114969e-05, "loss": 0.846, "step": 7117 }, { "epoch": 3.2251925690983234, "grad_norm": 0.3194951825855863, "learning_rate": 2.6974548656032332e-05, "loss": 0.8415, "step": 7118 }, { "epoch": 3.2256456728590845, "grad_norm": 0.2048676095500777, "learning_rate": 2.6962585314401544e-05, "loss": 0.8649, "step": 7119 }, { "epoch": 3.226098776619846, "grad_norm": 0.3299843650947428, "learning_rate": 2.6950623277454514e-05, "loss": 0.8606, "step": 7120 }, { "epoch": 3.2265518803806073, "grad_norm": 0.3074287308445422, "learning_rate": 2.693866254638832e-05, "loss": 0.8703, "step": 7121 }, { "epoch": 3.2270049841413684, "grad_norm": 0.2423537283357825, "learning_rate": 2.6926703122399892e-05, "loss": 0.8799, "step": 7122 }, { "epoch": 3.2274580879021295, "grad_norm": 0.32693793161597373, "learning_rate": 2.691474500668604e-05, "loss": 0.87, "step": 7123 }, { "epoch": 3.2279111916628906, "grad_norm": 0.24903853526855974, "learning_rate": 2.690278820044344e-05, "loss": 0.8677, "step": 7124 }, { "epoch": 3.228364295423652, "grad_norm": 0.21978044542624975, "learning_rate": 2.6890832704868643e-05, "loss": 0.8576, "step": 7125 }, { "epoch": 3.2288173991844134, "grad_norm": 0.3144734502233083, "learning_rate": 2.6878878521158055e-05, "loss": 0.8449, "step": 7126 }, { "epoch": 3.2292705029451745, "grad_norm": 0.2789687520162681, "learning_rate": 2.6866925650507948e-05, "loss": 0.8412, "step": 7127 }, { "epoch": 3.2297236067059356, "grad_norm": 0.4117360879201466, "learning_rate": 2.6854974094114486e-05, "loss": 0.8619, "step": 7128 }, { "epoch": 3.2301767104666967, "grad_norm": 0.34491770102431446, "learning_rate": 2.68430238531737e-05, "loss": 0.884, "step": 7129 }, { "epoch": 3.2306298142274583, "grad_norm": 0.27415492317118106, "learning_rate": 2.6831074928881453e-05, "loss": 0.8518, "step": 7130 }, { "epoch": 3.2310829179882194, "grad_norm": 0.2429578738708251, "learning_rate": 2.6819127322433513e-05, "loss": 0.8491, "step": 7131 }, { "epoch": 3.2315360217489806, "grad_norm": 0.2949788007545788, "learning_rate": 2.68071810350255e-05, "loss": 0.8588, "step": 7132 }, { "epoch": 3.2319891255097417, "grad_norm": 0.22062784068329305, "learning_rate": 2.679523606785291e-05, "loss": 0.8574, "step": 7133 }, { "epoch": 3.232442229270503, "grad_norm": 0.2625813763942762, "learning_rate": 2.6783292422111107e-05, "loss": 0.8608, "step": 7134 }, { "epoch": 3.232895333031264, "grad_norm": 0.24588882462384093, "learning_rate": 2.677135009899531e-05, "loss": 0.8746, "step": 7135 }, { "epoch": 3.2333484367920255, "grad_norm": 0.26619162791697926, "learning_rate": 2.6759409099700624e-05, "loss": 0.8469, "step": 7136 }, { "epoch": 3.2338015405527867, "grad_norm": 0.3096232134308684, "learning_rate": 2.6747469425422016e-05, "loss": 0.845, "step": 7137 }, { "epoch": 3.234254644313548, "grad_norm": 0.28509776104566875, "learning_rate": 2.6735531077354294e-05, "loss": 0.8383, "step": 7138 }, { "epoch": 3.234707748074309, "grad_norm": 0.2093342609787599, "learning_rate": 2.6723594056692172e-05, "loss": 0.8532, "step": 7139 }, { "epoch": 3.23516085183507, "grad_norm": 0.20132374924875598, "learning_rate": 2.671165836463021e-05, "loss": 0.8497, "step": 7140 }, { "epoch": 3.2356139555958316, "grad_norm": 0.2360717618060155, "learning_rate": 2.6699724002362843e-05, "loss": 0.8477, "step": 7141 }, { "epoch": 3.236067059356593, "grad_norm": 0.21509934945759762, "learning_rate": 2.668779097108437e-05, "loss": 0.8435, "step": 7142 }, { "epoch": 3.236520163117354, "grad_norm": 0.19381183822273126, "learning_rate": 2.6675859271988955e-05, "loss": 0.8508, "step": 7143 }, { "epoch": 3.236973266878115, "grad_norm": 0.20174936723911022, "learning_rate": 2.666392890627063e-05, "loss": 0.8654, "step": 7144 }, { "epoch": 3.237426370638876, "grad_norm": 0.2588442635636024, "learning_rate": 2.6651999875123303e-05, "loss": 0.8791, "step": 7145 }, { "epoch": 3.2378794743996373, "grad_norm": 0.3008730424479363, "learning_rate": 2.6640072179740716e-05, "loss": 0.8495, "step": 7146 }, { "epoch": 3.238332578160399, "grad_norm": 0.30036291708247925, "learning_rate": 2.6628145821316525e-05, "loss": 0.8363, "step": 7147 }, { "epoch": 3.23878568192116, "grad_norm": 0.24250828891242815, "learning_rate": 2.6616220801044214e-05, "loss": 0.8551, "step": 7148 }, { "epoch": 3.239238785681921, "grad_norm": 0.26637181344671346, "learning_rate": 2.660429712011715e-05, "loss": 0.8623, "step": 7149 }, { "epoch": 3.2396918894426823, "grad_norm": 0.24967396117738344, "learning_rate": 2.659237477972856e-05, "loss": 0.8493, "step": 7150 }, { "epoch": 3.2401449932034434, "grad_norm": 0.25754573675957826, "learning_rate": 2.6580453781071553e-05, "loss": 0.863, "step": 7151 }, { "epoch": 3.240598096964205, "grad_norm": 0.3144946971220873, "learning_rate": 2.6568534125339075e-05, "loss": 0.8483, "step": 7152 }, { "epoch": 3.241051200724966, "grad_norm": 0.23442737681541415, "learning_rate": 2.655661581372397e-05, "loss": 0.8631, "step": 7153 }, { "epoch": 3.2415043044857272, "grad_norm": 0.23567331980799358, "learning_rate": 2.654469884741891e-05, "loss": 0.8553, "step": 7154 }, { "epoch": 3.2419574082464884, "grad_norm": 0.24021886312793794, "learning_rate": 2.6532783227616464e-05, "loss": 0.8628, "step": 7155 }, { "epoch": 3.2424105120072495, "grad_norm": 0.24045974348517865, "learning_rate": 2.6520868955509053e-05, "loss": 0.8686, "step": 7156 }, { "epoch": 3.242863615768011, "grad_norm": 0.23670928171923042, "learning_rate": 2.6508956032288967e-05, "loss": 0.86, "step": 7157 }, { "epoch": 3.243316719528772, "grad_norm": 0.2258243913698344, "learning_rate": 2.649704445914836e-05, "loss": 0.9038, "step": 7158 }, { "epoch": 3.2437698232895333, "grad_norm": 0.2698910445940504, "learning_rate": 2.6485134237279243e-05, "loss": 0.8522, "step": 7159 }, { "epoch": 3.2442229270502945, "grad_norm": 0.24939055847786418, "learning_rate": 2.6473225367873508e-05, "loss": 0.8475, "step": 7160 }, { "epoch": 3.2446760308110556, "grad_norm": 0.2392448715863759, "learning_rate": 2.646131785212291e-05, "loss": 0.864, "step": 7161 }, { "epoch": 3.245129134571817, "grad_norm": 0.22648576491120778, "learning_rate": 2.644941169121904e-05, "loss": 0.8493, "step": 7162 }, { "epoch": 3.2455822383325783, "grad_norm": 0.26148484303976666, "learning_rate": 2.643750688635339e-05, "loss": 0.8695, "step": 7163 }, { "epoch": 3.2460353420933394, "grad_norm": 0.2127155404606317, "learning_rate": 2.642560343871729e-05, "loss": 0.8476, "step": 7164 }, { "epoch": 3.2464884458541006, "grad_norm": 0.20731617113274028, "learning_rate": 2.6413701349501956e-05, "loss": 0.8504, "step": 7165 }, { "epoch": 3.2469415496148617, "grad_norm": 0.247022773513119, "learning_rate": 2.6401800619898445e-05, "loss": 0.8584, "step": 7166 }, { "epoch": 3.247394653375623, "grad_norm": 0.24934823946535742, "learning_rate": 2.6389901251097705e-05, "loss": 0.8633, "step": 7167 }, { "epoch": 3.2478477571363844, "grad_norm": 0.24848199316667768, "learning_rate": 2.637800324429052e-05, "loss": 0.8835, "step": 7168 }, { "epoch": 3.2483008608971455, "grad_norm": 0.24518389187994932, "learning_rate": 2.636610660066756e-05, "loss": 0.8429, "step": 7169 }, { "epoch": 3.2487539646579067, "grad_norm": 0.25309163816974506, "learning_rate": 2.635421132141936e-05, "loss": 0.8562, "step": 7170 }, { "epoch": 3.249207068418668, "grad_norm": 0.2752986663187818, "learning_rate": 2.6342317407736274e-05, "loss": 0.8576, "step": 7171 }, { "epoch": 3.249660172179429, "grad_norm": 0.22024359170836705, "learning_rate": 2.633042486080858e-05, "loss": 0.8518, "step": 7172 }, { "epoch": 3.2501132759401905, "grad_norm": 0.24751759939846507, "learning_rate": 2.631853368182638e-05, "loss": 0.8637, "step": 7173 }, { "epoch": 3.2505663797009516, "grad_norm": 0.27584718698095, "learning_rate": 2.6306643871979657e-05, "loss": 0.8488, "step": 7174 }, { "epoch": 3.2510194834617128, "grad_norm": 0.25431577836528924, "learning_rate": 2.6294755432458246e-05, "loss": 0.861, "step": 7175 }, { "epoch": 3.251472587222474, "grad_norm": 0.2113239760783083, "learning_rate": 2.6282868364451858e-05, "loss": 0.8502, "step": 7176 }, { "epoch": 3.251925690983235, "grad_norm": 0.2040478544630469, "learning_rate": 2.6270982669150052e-05, "loss": 0.8513, "step": 7177 }, { "epoch": 3.252378794743996, "grad_norm": 0.1892923502468876, "learning_rate": 2.6259098347742274e-05, "loss": 0.8548, "step": 7178 }, { "epoch": 3.2528318985047577, "grad_norm": 0.23492149532975357, "learning_rate": 2.6247215401417787e-05, "loss": 0.8562, "step": 7179 }, { "epoch": 3.253285002265519, "grad_norm": 0.27408100780122535, "learning_rate": 2.623533383136576e-05, "loss": 0.8677, "step": 7180 }, { "epoch": 3.25373810602628, "grad_norm": 0.24830304291748487, "learning_rate": 2.6223453638775205e-05, "loss": 0.8639, "step": 7181 }, { "epoch": 3.254191209787041, "grad_norm": 0.19779185468847726, "learning_rate": 2.6211574824835006e-05, "loss": 0.8472, "step": 7182 }, { "epoch": 3.2546443135478023, "grad_norm": 0.24561279204458167, "learning_rate": 2.6199697390733914e-05, "loss": 0.8583, "step": 7183 }, { "epoch": 3.255097417308564, "grad_norm": 0.343567222605969, "learning_rate": 2.61878213376605e-05, "loss": 0.865, "step": 7184 }, { "epoch": 3.255550521069325, "grad_norm": 0.36799817593917655, "learning_rate": 2.617594666680324e-05, "loss": 0.867, "step": 7185 }, { "epoch": 3.256003624830086, "grad_norm": 0.22660428616964476, "learning_rate": 2.616407337935047e-05, "loss": 0.8409, "step": 7186 }, { "epoch": 3.2564567285908472, "grad_norm": 0.22221375452729342, "learning_rate": 2.615220147649036e-05, "loss": 0.8422, "step": 7187 }, { "epoch": 3.2569098323516084, "grad_norm": 0.3185600861998414, "learning_rate": 2.6140330959410977e-05, "loss": 0.8732, "step": 7188 }, { "epoch": 3.2573629361123695, "grad_norm": 0.32367276857753363, "learning_rate": 2.6128461829300223e-05, "loss": 0.8636, "step": 7189 }, { "epoch": 3.257816039873131, "grad_norm": 0.2316272471002833, "learning_rate": 2.6116594087345865e-05, "loss": 0.8723, "step": 7190 }, { "epoch": 3.258269143633892, "grad_norm": 0.2832360302401134, "learning_rate": 2.610472773473553e-05, "loss": 0.8633, "step": 7191 }, { "epoch": 3.2587222473946533, "grad_norm": 0.28149863029629046, "learning_rate": 2.6092862772656724e-05, "loss": 0.8599, "step": 7192 }, { "epoch": 3.2591753511554145, "grad_norm": 0.3204253282706259, "learning_rate": 2.6080999202296795e-05, "loss": 0.8707, "step": 7193 }, { "epoch": 3.259628454916176, "grad_norm": 0.38426869170166766, "learning_rate": 2.6069137024842953e-05, "loss": 0.8623, "step": 7194 }, { "epoch": 3.260081558676937, "grad_norm": 0.3422770526897979, "learning_rate": 2.6057276241482283e-05, "loss": 0.8586, "step": 7195 }, { "epoch": 3.2605346624376983, "grad_norm": 0.2188860561881324, "learning_rate": 2.6045416853401717e-05, "loss": 0.8439, "step": 7196 }, { "epoch": 3.2609877661984594, "grad_norm": 0.2795290997071543, "learning_rate": 2.6033558861788044e-05, "loss": 0.8513, "step": 7197 }, { "epoch": 3.2614408699592206, "grad_norm": 0.2934986845567122, "learning_rate": 2.6021702267827936e-05, "loss": 0.871, "step": 7198 }, { "epoch": 3.2618939737199817, "grad_norm": 0.2664911814655664, "learning_rate": 2.6009847072707893e-05, "loss": 0.8803, "step": 7199 }, { "epoch": 3.2623470774807433, "grad_norm": 0.1994678035029635, "learning_rate": 2.5997993277614295e-05, "loss": 0.8558, "step": 7200 }, { "epoch": 3.2628001812415044, "grad_norm": 0.21720568416329, "learning_rate": 2.598614088373338e-05, "loss": 0.8483, "step": 7201 }, { "epoch": 3.2632532850022655, "grad_norm": 0.24868857423797375, "learning_rate": 2.597428989225124e-05, "loss": 0.8432, "step": 7202 }, { "epoch": 3.2637063887630267, "grad_norm": 0.3122025724780986, "learning_rate": 2.5962440304353846e-05, "loss": 0.8597, "step": 7203 }, { "epoch": 3.264159492523788, "grad_norm": 0.3256851311824766, "learning_rate": 2.5950592121226993e-05, "loss": 0.8461, "step": 7204 }, { "epoch": 3.2646125962845494, "grad_norm": 0.28138174963863405, "learning_rate": 2.593874534405637e-05, "loss": 0.8593, "step": 7205 }, { "epoch": 3.2650657000453105, "grad_norm": 0.23191202589035323, "learning_rate": 2.5926899974027514e-05, "loss": 0.8751, "step": 7206 }, { "epoch": 3.2655188038060716, "grad_norm": 0.35266431786553454, "learning_rate": 2.59150560123258e-05, "loss": 0.8516, "step": 7207 }, { "epoch": 3.2659719075668328, "grad_norm": 0.37557264336470564, "learning_rate": 2.5903213460136495e-05, "loss": 0.8747, "step": 7208 }, { "epoch": 3.266425011327594, "grad_norm": 0.3135283043271753, "learning_rate": 2.5891372318644695e-05, "loss": 0.8781, "step": 7209 }, { "epoch": 3.266878115088355, "grad_norm": 0.23312278429619437, "learning_rate": 2.587953258903539e-05, "loss": 0.8326, "step": 7210 }, { "epoch": 3.2673312188491166, "grad_norm": 0.26409276212699695, "learning_rate": 2.5867694272493396e-05, "loss": 0.8537, "step": 7211 }, { "epoch": 3.2677843226098777, "grad_norm": 0.3750506862807531, "learning_rate": 2.58558573702034e-05, "loss": 0.8634, "step": 7212 }, { "epoch": 3.268237426370639, "grad_norm": 0.3744899204965023, "learning_rate": 2.5844021883349955e-05, "loss": 0.8678, "step": 7213 }, { "epoch": 3.2686905301314, "grad_norm": 0.296380134373002, "learning_rate": 2.5832187813117475e-05, "loss": 0.8559, "step": 7214 }, { "epoch": 3.269143633892161, "grad_norm": 0.29279624098082313, "learning_rate": 2.5820355160690195e-05, "loss": 0.8609, "step": 7215 }, { "epoch": 3.2695967376529227, "grad_norm": 0.3298358013169392, "learning_rate": 2.580852392725225e-05, "loss": 0.8649, "step": 7216 }, { "epoch": 3.270049841413684, "grad_norm": 0.37426252552284106, "learning_rate": 2.5796694113987616e-05, "loss": 0.8652, "step": 7217 }, { "epoch": 3.270502945174445, "grad_norm": 0.30333336452919873, "learning_rate": 2.5784865722080136e-05, "loss": 0.8563, "step": 7218 }, { "epoch": 3.270956048935206, "grad_norm": 0.2535897056442829, "learning_rate": 2.5773038752713493e-05, "loss": 0.8579, "step": 7219 }, { "epoch": 3.271409152695967, "grad_norm": 0.21846670450886302, "learning_rate": 2.5761213207071247e-05, "loss": 0.8497, "step": 7220 }, { "epoch": 3.2718622564567283, "grad_norm": 0.23756155201473705, "learning_rate": 2.5749389086336806e-05, "loss": 0.8465, "step": 7221 }, { "epoch": 3.27231536021749, "grad_norm": 0.28719680621736027, "learning_rate": 2.573756639169344e-05, "loss": 0.8469, "step": 7222 }, { "epoch": 3.272768463978251, "grad_norm": 0.28383781164275707, "learning_rate": 2.5725745124324267e-05, "loss": 0.8544, "step": 7223 }, { "epoch": 3.273221567739012, "grad_norm": 0.18018241412860272, "learning_rate": 2.5713925285412264e-05, "loss": 0.8559, "step": 7224 }, { "epoch": 3.2736746714997733, "grad_norm": 0.23068945152325335, "learning_rate": 2.5702106876140274e-05, "loss": 0.8508, "step": 7225 }, { "epoch": 3.274127775260535, "grad_norm": 0.2549875420591406, "learning_rate": 2.5690289897691e-05, "loss": 0.8531, "step": 7226 }, { "epoch": 3.274580879021296, "grad_norm": 0.19624883415304634, "learning_rate": 2.5678474351246986e-05, "loss": 0.8551, "step": 7227 }, { "epoch": 3.275033982782057, "grad_norm": 0.40515007937621067, "learning_rate": 2.566666023799064e-05, "loss": 0.8861, "step": 7228 }, { "epoch": 3.2754870865428183, "grad_norm": 0.21433091309589739, "learning_rate": 2.565484755910423e-05, "loss": 0.8599, "step": 7229 }, { "epoch": 3.2759401903035794, "grad_norm": 0.21362446604912824, "learning_rate": 2.564303631576988e-05, "loss": 0.8522, "step": 7230 }, { "epoch": 3.2763932940643405, "grad_norm": 0.24542431775171147, "learning_rate": 2.563122650916956e-05, "loss": 0.8462, "step": 7231 }, { "epoch": 3.276846397825102, "grad_norm": 0.24790332773808382, "learning_rate": 2.5619418140485107e-05, "loss": 0.8639, "step": 7232 }, { "epoch": 3.2772995015858633, "grad_norm": 0.24382793032122768, "learning_rate": 2.5607611210898212e-05, "loss": 0.8753, "step": 7233 }, { "epoch": 3.2777526053466244, "grad_norm": 0.34740218439199283, "learning_rate": 2.559580572159043e-05, "loss": 0.8451, "step": 7234 }, { "epoch": 3.2782057091073855, "grad_norm": 0.28833032486305665, "learning_rate": 2.5584001673743153e-05, "loss": 0.8589, "step": 7235 }, { "epoch": 3.2786588128681466, "grad_norm": 0.20545573594275687, "learning_rate": 2.5572199068537637e-05, "loss": 0.8514, "step": 7236 }, { "epoch": 3.279111916628908, "grad_norm": 0.24640375571667442, "learning_rate": 2.5560397907155002e-05, "loss": 0.8421, "step": 7237 }, { "epoch": 3.2795650203896693, "grad_norm": 0.3097195042065037, "learning_rate": 2.5548598190776213e-05, "loss": 0.8632, "step": 7238 }, { "epoch": 3.2800181241504305, "grad_norm": 0.27587924453660306, "learning_rate": 2.5536799920582112e-05, "loss": 0.864, "step": 7239 }, { "epoch": 3.2804712279111916, "grad_norm": 0.3056807802708445, "learning_rate": 2.5525003097753357e-05, "loss": 0.8386, "step": 7240 }, { "epoch": 3.2809243316719527, "grad_norm": 0.23842381840340549, "learning_rate": 2.5513207723470482e-05, "loss": 0.8439, "step": 7241 }, { "epoch": 3.281377435432714, "grad_norm": 0.21981251299296709, "learning_rate": 2.5501413798913895e-05, "loss": 0.8341, "step": 7242 }, { "epoch": 3.2818305391934754, "grad_norm": 0.24260613798629221, "learning_rate": 2.5489621325263825e-05, "loss": 0.8588, "step": 7243 }, { "epoch": 3.2822836429542366, "grad_norm": 0.23499886841902215, "learning_rate": 2.5477830303700374e-05, "loss": 0.8672, "step": 7244 }, { "epoch": 3.2827367467149977, "grad_norm": 0.16514981206799567, "learning_rate": 2.5466040735403494e-05, "loss": 0.8537, "step": 7245 }, { "epoch": 3.283189850475759, "grad_norm": 0.2462235775466053, "learning_rate": 2.5454252621553e-05, "loss": 0.8593, "step": 7246 }, { "epoch": 3.28364295423652, "grad_norm": 0.23645010436424435, "learning_rate": 2.5442465963328565e-05, "loss": 0.8624, "step": 7247 }, { "epoch": 3.2840960579972815, "grad_norm": 0.19392734282767482, "learning_rate": 2.5430680761909682e-05, "loss": 0.8553, "step": 7248 }, { "epoch": 3.2845491617580427, "grad_norm": 0.23865007197395605, "learning_rate": 2.5418897018475747e-05, "loss": 0.8489, "step": 7249 }, { "epoch": 3.285002265518804, "grad_norm": 0.21239910525474617, "learning_rate": 2.540711473420598e-05, "loss": 0.8555, "step": 7250 }, { "epoch": 3.285455369279565, "grad_norm": 0.19266157617203214, "learning_rate": 2.5395333910279455e-05, "loss": 0.8685, "step": 7251 }, { "epoch": 3.285908473040326, "grad_norm": 0.1869405206417225, "learning_rate": 2.5383554547875103e-05, "loss": 0.8534, "step": 7252 }, { "epoch": 3.286361576801087, "grad_norm": 0.2200238009440733, "learning_rate": 2.537177664817172e-05, "loss": 0.8502, "step": 7253 }, { "epoch": 3.286814680561849, "grad_norm": 0.197041858415064, "learning_rate": 2.5360000212347942e-05, "loss": 0.8447, "step": 7254 }, { "epoch": 3.28726778432261, "grad_norm": 0.2075236959882449, "learning_rate": 2.5348225241582264e-05, "loss": 0.8677, "step": 7255 }, { "epoch": 3.287720888083371, "grad_norm": 0.22633821169871807, "learning_rate": 2.5336451737053037e-05, "loss": 0.8452, "step": 7256 }, { "epoch": 3.288173991844132, "grad_norm": 0.19656736467320662, "learning_rate": 2.532467969993846e-05, "loss": 0.8412, "step": 7257 }, { "epoch": 3.2886270956048937, "grad_norm": 0.2229258603324314, "learning_rate": 2.5312909131416587e-05, "loss": 0.8594, "step": 7258 }, { "epoch": 3.289080199365655, "grad_norm": 0.2116738565917769, "learning_rate": 2.530114003266534e-05, "loss": 0.8738, "step": 7259 }, { "epoch": 3.289533303126416, "grad_norm": 0.23596672716681424, "learning_rate": 2.5289372404862462e-05, "loss": 0.8624, "step": 7260 }, { "epoch": 3.289986406887177, "grad_norm": 0.23636726191264518, "learning_rate": 2.5277606249185563e-05, "loss": 0.8708, "step": 7261 }, { "epoch": 3.2904395106479383, "grad_norm": 0.22594171267112817, "learning_rate": 2.526584156681212e-05, "loss": 0.8712, "step": 7262 }, { "epoch": 3.2908926144086994, "grad_norm": 0.21737558634883253, "learning_rate": 2.5254078358919453e-05, "loss": 0.8436, "step": 7263 }, { "epoch": 3.291345718169461, "grad_norm": 0.23705622867920928, "learning_rate": 2.5242316626684723e-05, "loss": 0.8584, "step": 7264 }, { "epoch": 3.291798821930222, "grad_norm": 0.18456078033190162, "learning_rate": 2.523055637128496e-05, "loss": 0.8439, "step": 7265 }, { "epoch": 3.2922519256909832, "grad_norm": 0.22358780052373814, "learning_rate": 2.521879759389704e-05, "loss": 0.8338, "step": 7266 }, { "epoch": 3.2927050294517444, "grad_norm": 0.4225169584649395, "learning_rate": 2.5207040295697698e-05, "loss": 0.8788, "step": 7267 }, { "epoch": 3.2931581332125055, "grad_norm": 0.24283699118208696, "learning_rate": 2.51952844778635e-05, "loss": 0.8535, "step": 7268 }, { "epoch": 3.293611236973267, "grad_norm": 0.23638219427803484, "learning_rate": 2.5183530141570877e-05, "loss": 0.8576, "step": 7269 }, { "epoch": 3.294064340734028, "grad_norm": 0.21221915679687292, "learning_rate": 2.5171777287996114e-05, "loss": 0.8675, "step": 7270 }, { "epoch": 3.2945174444947893, "grad_norm": 0.22738772761684278, "learning_rate": 2.5160025918315355e-05, "loss": 0.8701, "step": 7271 }, { "epoch": 3.2949705482555505, "grad_norm": 0.25074485073619346, "learning_rate": 2.514827603370457e-05, "loss": 0.873, "step": 7272 }, { "epoch": 3.2954236520163116, "grad_norm": 0.30538543408292856, "learning_rate": 2.513652763533962e-05, "loss": 0.8618, "step": 7273 }, { "epoch": 3.2958767557770727, "grad_norm": 0.2735856007992235, "learning_rate": 2.512478072439618e-05, "loss": 0.8745, "step": 7274 }, { "epoch": 3.2963298595378343, "grad_norm": 0.20129488818197014, "learning_rate": 2.5113035302049795e-05, "loss": 0.8608, "step": 7275 }, { "epoch": 3.2967829632985954, "grad_norm": 0.32769864084209754, "learning_rate": 2.5101291369475848e-05, "loss": 0.8537, "step": 7276 }, { "epoch": 3.2972360670593566, "grad_norm": 0.3614826452096708, "learning_rate": 2.5089548927849583e-05, "loss": 0.8436, "step": 7277 }, { "epoch": 3.2976891708201177, "grad_norm": 0.25871029582240135, "learning_rate": 2.5077807978346095e-05, "loss": 0.8796, "step": 7278 }, { "epoch": 3.298142274580879, "grad_norm": 0.20091610033995297, "learning_rate": 2.5066068522140324e-05, "loss": 0.8613, "step": 7279 }, { "epoch": 3.2985953783416404, "grad_norm": 0.27956286035409905, "learning_rate": 2.5054330560407075e-05, "loss": 0.8593, "step": 7280 }, { "epoch": 3.2990484821024015, "grad_norm": 0.3629912392584932, "learning_rate": 2.5042594094320985e-05, "loss": 0.8615, "step": 7281 }, { "epoch": 3.2995015858631627, "grad_norm": 0.3363189422379541, "learning_rate": 2.5030859125056543e-05, "loss": 0.8465, "step": 7282 }, { "epoch": 3.299954689623924, "grad_norm": 0.24899583942741121, "learning_rate": 2.5019125653788114e-05, "loss": 0.8541, "step": 7283 }, { "epoch": 3.300407793384685, "grad_norm": 0.2631238120774405, "learning_rate": 2.5007393681689868e-05, "loss": 0.8546, "step": 7284 }, { "epoch": 3.300860897145446, "grad_norm": 0.19029418636298978, "learning_rate": 2.4995663209935863e-05, "loss": 0.8365, "step": 7285 }, { "epoch": 3.3013140009062076, "grad_norm": 0.28199608052682645, "learning_rate": 2.49839342397e-05, "loss": 0.8707, "step": 7286 }, { "epoch": 3.3017671046669688, "grad_norm": 0.23770002091683193, "learning_rate": 2.4972206772156008e-05, "loss": 0.827, "step": 7287 }, { "epoch": 3.30222020842773, "grad_norm": 0.26112989736900316, "learning_rate": 2.4960480808477493e-05, "loss": 0.8538, "step": 7288 }, { "epoch": 3.302673312188491, "grad_norm": 0.24614231396612563, "learning_rate": 2.4948756349837892e-05, "loss": 0.8746, "step": 7289 }, { "epoch": 3.3031264159492526, "grad_norm": 0.22967527170902055, "learning_rate": 2.4937033397410507e-05, "loss": 0.8491, "step": 7290 }, { "epoch": 3.3035795197100137, "grad_norm": 0.260260742961296, "learning_rate": 2.4925311952368484e-05, "loss": 0.847, "step": 7291 }, { "epoch": 3.304032623470775, "grad_norm": 0.2754364134127014, "learning_rate": 2.4913592015884795e-05, "loss": 0.8485, "step": 7292 }, { "epoch": 3.304485727231536, "grad_norm": 0.2802657492636461, "learning_rate": 2.490187358913229e-05, "loss": 0.8507, "step": 7293 }, { "epoch": 3.304938830992297, "grad_norm": 0.29573609424967273, "learning_rate": 2.4890156673283665e-05, "loss": 0.8562, "step": 7294 }, { "epoch": 3.3053919347530583, "grad_norm": 0.2453685811018919, "learning_rate": 2.487844126951145e-05, "loss": 0.8458, "step": 7295 }, { "epoch": 3.30584503851382, "grad_norm": 0.30381532818065077, "learning_rate": 2.486672737898804e-05, "loss": 0.8561, "step": 7296 }, { "epoch": 3.306298142274581, "grad_norm": 0.2289031195429428, "learning_rate": 2.485501500288568e-05, "loss": 0.8546, "step": 7297 }, { "epoch": 3.306751246035342, "grad_norm": 0.2167409358748046, "learning_rate": 2.4843304142376434e-05, "loss": 0.8532, "step": 7298 }, { "epoch": 3.3072043497961032, "grad_norm": 0.2605983569601727, "learning_rate": 2.4831594798632233e-05, "loss": 0.8636, "step": 7299 }, { "epoch": 3.3076574535568644, "grad_norm": 0.223142961590979, "learning_rate": 2.4819886972824866e-05, "loss": 0.8395, "step": 7300 }, { "epoch": 3.308110557317626, "grad_norm": 0.20545302150834252, "learning_rate": 2.4808180666125967e-05, "loss": 0.8553, "step": 7301 }, { "epoch": 3.308563661078387, "grad_norm": 0.27351455378633405, "learning_rate": 2.4796475879707003e-05, "loss": 0.8712, "step": 7302 }, { "epoch": 3.309016764839148, "grad_norm": 0.26828003898140057, "learning_rate": 2.4784772614739315e-05, "loss": 0.8546, "step": 7303 }, { "epoch": 3.3094698685999093, "grad_norm": 0.23713717236502688, "learning_rate": 2.477307087239405e-05, "loss": 0.8484, "step": 7304 }, { "epoch": 3.3099229723606705, "grad_norm": 0.25690029114056195, "learning_rate": 2.476137065384225e-05, "loss": 0.8629, "step": 7305 }, { "epoch": 3.3103760761214316, "grad_norm": 0.23792715634729633, "learning_rate": 2.4749671960254766e-05, "loss": 0.8623, "step": 7306 }, { "epoch": 3.310829179882193, "grad_norm": 0.21958427174103795, "learning_rate": 2.4737974792802325e-05, "loss": 0.8508, "step": 7307 }, { "epoch": 3.3112822836429543, "grad_norm": 0.23428914581718271, "learning_rate": 2.4726279152655483e-05, "loss": 0.8485, "step": 7308 }, { "epoch": 3.3117353874037154, "grad_norm": 0.1951091567443895, "learning_rate": 2.4714585040984656e-05, "loss": 0.864, "step": 7309 }, { "epoch": 3.3121884911644766, "grad_norm": 0.19189310502189602, "learning_rate": 2.470289245896009e-05, "loss": 0.842, "step": 7310 }, { "epoch": 3.3126415949252377, "grad_norm": 0.21139170327064294, "learning_rate": 2.469120140775191e-05, "loss": 0.8688, "step": 7311 }, { "epoch": 3.3130946986859993, "grad_norm": 0.23590758493769792, "learning_rate": 2.4679511888530033e-05, "loss": 0.8651, "step": 7312 }, { "epoch": 3.3135478024467604, "grad_norm": 0.23027879163009468, "learning_rate": 2.466782390246428e-05, "loss": 0.8608, "step": 7313 }, { "epoch": 3.3140009062075215, "grad_norm": 0.24732723143471813, "learning_rate": 2.4656137450724287e-05, "loss": 0.8427, "step": 7314 }, { "epoch": 3.3144540099682827, "grad_norm": 0.2599165285560934, "learning_rate": 2.464445253447954e-05, "loss": 0.8554, "step": 7315 }, { "epoch": 3.314907113729044, "grad_norm": 0.1792631276299599, "learning_rate": 2.4632769154899386e-05, "loss": 0.8698, "step": 7316 }, { "epoch": 3.315360217489805, "grad_norm": 0.22043674595942767, "learning_rate": 2.4621087313153e-05, "loss": 0.8285, "step": 7317 }, { "epoch": 3.3158133212505665, "grad_norm": 0.24689990876359413, "learning_rate": 2.4609407010409406e-05, "loss": 0.847, "step": 7318 }, { "epoch": 3.3162664250113276, "grad_norm": 0.17532785175767437, "learning_rate": 2.459772824783749e-05, "loss": 0.8733, "step": 7319 }, { "epoch": 3.3167195287720888, "grad_norm": 0.2440539559444655, "learning_rate": 2.4586051026605966e-05, "loss": 0.8566, "step": 7320 }, { "epoch": 3.31717263253285, "grad_norm": 0.24671816541080863, "learning_rate": 2.4574375347883395e-05, "loss": 0.8398, "step": 7321 }, { "epoch": 3.3176257362936115, "grad_norm": 0.246390471412382, "learning_rate": 2.4562701212838194e-05, "loss": 0.8328, "step": 7322 }, { "epoch": 3.3180788400543726, "grad_norm": 0.19270791493716322, "learning_rate": 2.4551028622638628e-05, "loss": 0.874, "step": 7323 }, { "epoch": 3.3185319438151337, "grad_norm": 0.1735406605908711, "learning_rate": 2.4539357578452786e-05, "loss": 0.8767, "step": 7324 }, { "epoch": 3.318985047575895, "grad_norm": 0.2485685784806722, "learning_rate": 2.452768808144862e-05, "loss": 0.8539, "step": 7325 }, { "epoch": 3.319438151336656, "grad_norm": 0.19103119151090733, "learning_rate": 2.451602013279393e-05, "loss": 0.8756, "step": 7326 }, { "epoch": 3.319891255097417, "grad_norm": 0.20035986582432547, "learning_rate": 2.4504353733656347e-05, "loss": 0.8482, "step": 7327 }, { "epoch": 3.3203443588581787, "grad_norm": 0.2355464877526527, "learning_rate": 2.4492688885203366e-05, "loss": 0.8419, "step": 7328 }, { "epoch": 3.32079746261894, "grad_norm": 0.23113767623790377, "learning_rate": 2.4481025588602297e-05, "loss": 0.8463, "step": 7329 }, { "epoch": 3.321250566379701, "grad_norm": 0.23474146483279446, "learning_rate": 2.4469363845020318e-05, "loss": 0.8588, "step": 7330 }, { "epoch": 3.321703670140462, "grad_norm": 0.1916264078539388, "learning_rate": 2.445770365562445e-05, "loss": 0.8801, "step": 7331 }, { "epoch": 3.322156773901223, "grad_norm": 0.1896333647694634, "learning_rate": 2.444604502158156e-05, "loss": 0.892, "step": 7332 }, { "epoch": 3.322609877661985, "grad_norm": 0.2212923081370578, "learning_rate": 2.443438794405834e-05, "loss": 0.8543, "step": 7333 }, { "epoch": 3.323062981422746, "grad_norm": 0.22253472426587623, "learning_rate": 2.442273242422135e-05, "loss": 0.857, "step": 7334 }, { "epoch": 3.323516085183507, "grad_norm": 0.28236295363242186, "learning_rate": 2.441107846323698e-05, "loss": 0.8496, "step": 7335 }, { "epoch": 3.323969188944268, "grad_norm": 0.25522297138580546, "learning_rate": 2.439942606227149e-05, "loss": 0.8751, "step": 7336 }, { "epoch": 3.3244222927050293, "grad_norm": 0.2043168042905987, "learning_rate": 2.4387775222490926e-05, "loss": 0.868, "step": 7337 }, { "epoch": 3.3248753964657904, "grad_norm": 0.2736554212733009, "learning_rate": 2.4376125945061233e-05, "loss": 0.8517, "step": 7338 }, { "epoch": 3.325328500226552, "grad_norm": 0.2329469409185666, "learning_rate": 2.4364478231148176e-05, "loss": 0.8516, "step": 7339 }, { "epoch": 3.325781603987313, "grad_norm": 0.22614859187409103, "learning_rate": 2.435283208191738e-05, "loss": 0.8658, "step": 7340 }, { "epoch": 3.3262347077480743, "grad_norm": 0.22474358777152162, "learning_rate": 2.4341187498534287e-05, "loss": 0.8606, "step": 7341 }, { "epoch": 3.3266878115088354, "grad_norm": 0.30935785389060305, "learning_rate": 2.4329544482164203e-05, "loss": 0.8511, "step": 7342 }, { "epoch": 3.3271409152695965, "grad_norm": 0.2540403768560898, "learning_rate": 2.431790303397227e-05, "loss": 0.8561, "step": 7343 }, { "epoch": 3.327594019030358, "grad_norm": 0.19897332829429298, "learning_rate": 2.4306263155123486e-05, "loss": 0.8334, "step": 7344 }, { "epoch": 3.3280471227911192, "grad_norm": 0.21455251336465292, "learning_rate": 2.429462484678266e-05, "loss": 0.848, "step": 7345 }, { "epoch": 3.3285002265518804, "grad_norm": 0.20687052019192842, "learning_rate": 2.4282988110114475e-05, "loss": 0.8734, "step": 7346 }, { "epoch": 3.3289533303126415, "grad_norm": 0.2877043512235351, "learning_rate": 2.4271352946283445e-05, "loss": 0.8464, "step": 7347 }, { "epoch": 3.3294064340734026, "grad_norm": 0.2555276866604074, "learning_rate": 2.425971935645392e-05, "loss": 0.839, "step": 7348 }, { "epoch": 3.3298595378341638, "grad_norm": 0.21774932283702403, "learning_rate": 2.4248087341790117e-05, "loss": 0.8741, "step": 7349 }, { "epoch": 3.3303126415949253, "grad_norm": 0.28435131471096253, "learning_rate": 2.4236456903456062e-05, "loss": 0.8501, "step": 7350 }, { "epoch": 3.3307657453556865, "grad_norm": 0.31696648427439333, "learning_rate": 2.4224828042615643e-05, "loss": 0.8409, "step": 7351 }, { "epoch": 3.3312188491164476, "grad_norm": 0.33114374720083045, "learning_rate": 2.4213200760432597e-05, "loss": 0.8659, "step": 7352 }, { "epoch": 3.3316719528772087, "grad_norm": 0.22899438798698196, "learning_rate": 2.4201575058070475e-05, "loss": 0.8702, "step": 7353 }, { "epoch": 3.3321250566379703, "grad_norm": 0.23946859035385554, "learning_rate": 2.4189950936692713e-05, "loss": 0.8351, "step": 7354 }, { "epoch": 3.3325781603987314, "grad_norm": 0.30396621371553073, "learning_rate": 2.4178328397462528e-05, "loss": 0.8599, "step": 7355 }, { "epoch": 3.3330312641594926, "grad_norm": 0.3514363439669115, "learning_rate": 2.416670744154305e-05, "loss": 0.8596, "step": 7356 }, { "epoch": 3.3334843679202537, "grad_norm": 0.2848415885394628, "learning_rate": 2.4155088070097186e-05, "loss": 0.8663, "step": 7357 }, { "epoch": 3.333937471681015, "grad_norm": 0.2734413673871564, "learning_rate": 2.4143470284287728e-05, "loss": 0.8645, "step": 7358 }, { "epoch": 3.334390575441776, "grad_norm": 0.2764190444596464, "learning_rate": 2.413185408527728e-05, "loss": 0.8521, "step": 7359 }, { "epoch": 3.3348436792025375, "grad_norm": 0.33107594471146035, "learning_rate": 2.412023947422832e-05, "loss": 0.8821, "step": 7360 }, { "epoch": 3.3352967829632987, "grad_norm": 0.32197531833648896, "learning_rate": 2.4108626452303135e-05, "loss": 0.8451, "step": 7361 }, { "epoch": 3.33574988672406, "grad_norm": 0.22331616667832582, "learning_rate": 2.4097015020663878e-05, "loss": 0.8657, "step": 7362 }, { "epoch": 3.336202990484821, "grad_norm": 0.2648890162610852, "learning_rate": 2.4085405180472528e-05, "loss": 0.8559, "step": 7363 }, { "epoch": 3.336656094245582, "grad_norm": 0.2895018160276969, "learning_rate": 2.407379693289091e-05, "loss": 0.8278, "step": 7364 }, { "epoch": 3.3371091980063436, "grad_norm": 0.24627750582107977, "learning_rate": 2.4062190279080674e-05, "loss": 0.8591, "step": 7365 }, { "epoch": 3.3375623017671048, "grad_norm": 0.26119954697339415, "learning_rate": 2.4050585220203334e-05, "loss": 0.8707, "step": 7366 }, { "epoch": 3.338015405527866, "grad_norm": 0.1788149813406437, "learning_rate": 2.4038981757420235e-05, "loss": 0.8589, "step": 7367 }, { "epoch": 3.338468509288627, "grad_norm": 0.23315524887830594, "learning_rate": 2.4027379891892563e-05, "loss": 0.8719, "step": 7368 }, { "epoch": 3.338921613049388, "grad_norm": 0.22671170260060722, "learning_rate": 2.4015779624781338e-05, "loss": 0.8496, "step": 7369 }, { "epoch": 3.3393747168101493, "grad_norm": 0.18344286378076585, "learning_rate": 2.4004180957247433e-05, "loss": 0.8591, "step": 7370 }, { "epoch": 3.339827820570911, "grad_norm": 0.1942025634782635, "learning_rate": 2.3992583890451543e-05, "loss": 0.8844, "step": 7371 }, { "epoch": 3.340280924331672, "grad_norm": 0.23016244435686364, "learning_rate": 2.3980988425554232e-05, "loss": 0.8556, "step": 7372 }, { "epoch": 3.340734028092433, "grad_norm": 0.17318944966107527, "learning_rate": 2.3969394563715857e-05, "loss": 0.8626, "step": 7373 }, { "epoch": 3.3411871318531943, "grad_norm": 0.21470293266669285, "learning_rate": 2.3957802306096662e-05, "loss": 0.8507, "step": 7374 }, { "epoch": 3.341640235613956, "grad_norm": 0.21207966325282676, "learning_rate": 2.3946211653856707e-05, "loss": 0.862, "step": 7375 }, { "epoch": 3.342093339374717, "grad_norm": 0.18895003499188015, "learning_rate": 2.3934622608155883e-05, "loss": 0.8664, "step": 7376 }, { "epoch": 3.342546443135478, "grad_norm": 0.2355336405613857, "learning_rate": 2.392303517015395e-05, "loss": 0.8478, "step": 7377 }, { "epoch": 3.3429995468962392, "grad_norm": 0.17656256023167938, "learning_rate": 2.391144934101048e-05, "loss": 0.8535, "step": 7378 }, { "epoch": 3.3434526506570004, "grad_norm": 0.21095585409802547, "learning_rate": 2.3899865121884892e-05, "loss": 0.8513, "step": 7379 }, { "epoch": 3.3439057544177615, "grad_norm": 0.20841848756563244, "learning_rate": 2.388828251393646e-05, "loss": 0.852, "step": 7380 }, { "epoch": 3.344358858178523, "grad_norm": 0.21257321043599783, "learning_rate": 2.3876701518324255e-05, "loss": 0.8479, "step": 7381 }, { "epoch": 3.344811961939284, "grad_norm": 0.22241462831281786, "learning_rate": 2.3865122136207228e-05, "loss": 0.87, "step": 7382 }, { "epoch": 3.3452650657000453, "grad_norm": 0.20814867464868336, "learning_rate": 2.3853544368744164e-05, "loss": 0.8527, "step": 7383 }, { "epoch": 3.3457181694608065, "grad_norm": 0.17732287642843367, "learning_rate": 2.384196821709366e-05, "loss": 0.8623, "step": 7384 }, { "epoch": 3.3461712732215676, "grad_norm": 0.19280908119737486, "learning_rate": 2.3830393682414173e-05, "loss": 0.8724, "step": 7385 }, { "epoch": 3.346624376982329, "grad_norm": 0.2401251367004501, "learning_rate": 2.3818820765863997e-05, "loss": 0.8552, "step": 7386 }, { "epoch": 3.3470774807430903, "grad_norm": 0.20558863342176267, "learning_rate": 2.3807249468601265e-05, "loss": 0.8585, "step": 7387 }, { "epoch": 3.3475305845038514, "grad_norm": 0.21692039717441433, "learning_rate": 2.3795679791783925e-05, "loss": 0.8418, "step": 7388 }, { "epoch": 3.3479836882646126, "grad_norm": 0.21128227405641278, "learning_rate": 2.3784111736569803e-05, "loss": 0.8436, "step": 7389 }, { "epoch": 3.3484367920253737, "grad_norm": 0.18875006338412104, "learning_rate": 2.3772545304116524e-05, "loss": 0.867, "step": 7390 }, { "epoch": 3.348889895786135, "grad_norm": 0.20740067025208847, "learning_rate": 2.3760980495581573e-05, "loss": 0.8567, "step": 7391 }, { "epoch": 3.3493429995468964, "grad_norm": 0.2117278902494259, "learning_rate": 2.3749417312122263e-05, "loss": 0.8549, "step": 7392 }, { "epoch": 3.3497961033076575, "grad_norm": 0.19138132731452087, "learning_rate": 2.373785575489575e-05, "loss": 0.8503, "step": 7393 }, { "epoch": 3.3502492070684187, "grad_norm": 0.2072785755084952, "learning_rate": 2.3726295825059032e-05, "loss": 0.8568, "step": 7394 }, { "epoch": 3.35070231082918, "grad_norm": 0.18474716031613375, "learning_rate": 2.3714737523768928e-05, "loss": 0.8694, "step": 7395 }, { "epoch": 3.351155414589941, "grad_norm": 0.18208063791042658, "learning_rate": 2.3703180852182114e-05, "loss": 0.8666, "step": 7396 }, { "epoch": 3.3516085183507025, "grad_norm": 0.1975234207060734, "learning_rate": 2.369162581145509e-05, "loss": 0.8684, "step": 7397 }, { "epoch": 3.3520616221114636, "grad_norm": 0.18285980231420848, "learning_rate": 2.3680072402744184e-05, "loss": 0.8505, "step": 7398 }, { "epoch": 3.3525147258722248, "grad_norm": 0.20906477728824055, "learning_rate": 2.3668520627205577e-05, "loss": 0.864, "step": 7399 }, { "epoch": 3.352967829632986, "grad_norm": 0.18964148600344907, "learning_rate": 2.365697048599529e-05, "loss": 0.8547, "step": 7400 }, { "epoch": 3.353420933393747, "grad_norm": 0.1801969786667389, "learning_rate": 2.364542198026916e-05, "loss": 0.8513, "step": 7401 }, { "epoch": 3.353874037154508, "grad_norm": 0.18935249634305246, "learning_rate": 2.363387511118288e-05, "loss": 0.8532, "step": 7402 }, { "epoch": 3.3543271409152697, "grad_norm": 0.18808796063072336, "learning_rate": 2.362232987989197e-05, "loss": 0.8566, "step": 7403 }, { "epoch": 3.354780244676031, "grad_norm": 0.18599567060619274, "learning_rate": 2.3610786287551784e-05, "loss": 0.857, "step": 7404 }, { "epoch": 3.355233348436792, "grad_norm": 0.16765674999559416, "learning_rate": 2.3599244335317528e-05, "loss": 0.8267, "step": 7405 }, { "epoch": 3.355686452197553, "grad_norm": 0.2326785964729258, "learning_rate": 2.3587704024344217e-05, "loss": 0.8447, "step": 7406 }, { "epoch": 3.3561395559583147, "grad_norm": 0.2319168146969691, "learning_rate": 2.3576165355786717e-05, "loss": 0.8511, "step": 7407 }, { "epoch": 3.356592659719076, "grad_norm": 0.2023869569099263, "learning_rate": 2.3564628330799737e-05, "loss": 0.841, "step": 7408 }, { "epoch": 3.357045763479837, "grad_norm": 0.2519916084457495, "learning_rate": 2.3553092950537807e-05, "loss": 0.8367, "step": 7409 }, { "epoch": 3.357498867240598, "grad_norm": 0.2590306068630774, "learning_rate": 2.3541559216155303e-05, "loss": 0.8763, "step": 7410 }, { "epoch": 3.3579519710013592, "grad_norm": 0.21667583911200358, "learning_rate": 2.3530027128806444e-05, "loss": 0.8535, "step": 7411 }, { "epoch": 3.3584050747621204, "grad_norm": 0.27885708889735544, "learning_rate": 2.3518496689645245e-05, "loss": 0.864, "step": 7412 }, { "epoch": 3.358858178522882, "grad_norm": 0.22569403732158616, "learning_rate": 2.3506967899825597e-05, "loss": 0.8681, "step": 7413 }, { "epoch": 3.359311282283643, "grad_norm": 0.21879963600920896, "learning_rate": 2.3495440760501208e-05, "loss": 0.8441, "step": 7414 }, { "epoch": 3.359764386044404, "grad_norm": 0.23102303321044818, "learning_rate": 2.348391527282563e-05, "loss": 0.8639, "step": 7415 }, { "epoch": 3.3602174898051653, "grad_norm": 0.21413193630792351, "learning_rate": 2.3472391437952247e-05, "loss": 0.8517, "step": 7416 }, { "epoch": 3.3606705935659265, "grad_norm": 0.23218461534235624, "learning_rate": 2.346086925703428e-05, "loss": 0.8674, "step": 7417 }, { "epoch": 3.361123697326688, "grad_norm": 0.20510626209134186, "learning_rate": 2.344934873122476e-05, "loss": 0.8537, "step": 7418 }, { "epoch": 3.361576801087449, "grad_norm": 0.22781357787771536, "learning_rate": 2.343782986167659e-05, "loss": 0.8551, "step": 7419 }, { "epoch": 3.3620299048482103, "grad_norm": 0.22261505235035656, "learning_rate": 2.3426312649542478e-05, "loss": 0.8558, "step": 7420 }, { "epoch": 3.3624830086089714, "grad_norm": 0.21959581790542684, "learning_rate": 2.341479709597499e-05, "loss": 0.8583, "step": 7421 }, { "epoch": 3.3629361123697326, "grad_norm": 0.21294226800896787, "learning_rate": 2.3403283202126503e-05, "loss": 0.841, "step": 7422 }, { "epoch": 3.3633892161304937, "grad_norm": 0.21321874833140161, "learning_rate": 2.3391770969149247e-05, "loss": 0.8617, "step": 7423 }, { "epoch": 3.3638423198912553, "grad_norm": 0.2479739633333841, "learning_rate": 2.338026039819527e-05, "loss": 0.8353, "step": 7424 }, { "epoch": 3.3642954236520164, "grad_norm": 0.20982434637857117, "learning_rate": 2.336875149041647e-05, "loss": 0.8344, "step": 7425 }, { "epoch": 3.3647485274127775, "grad_norm": 0.19519170829828605, "learning_rate": 2.335724424696456e-05, "loss": 0.8777, "step": 7426 }, { "epoch": 3.3652016311735387, "grad_norm": 0.23099964186526842, "learning_rate": 2.3345738668991084e-05, "loss": 0.8634, "step": 7427 }, { "epoch": 3.3656547349343, "grad_norm": 0.21181965122485363, "learning_rate": 2.333423475764747e-05, "loss": 0.8447, "step": 7428 }, { "epoch": 3.3661078386950614, "grad_norm": 0.3296701321683772, "learning_rate": 2.3322732514084913e-05, "loss": 0.8639, "step": 7429 }, { "epoch": 3.3665609424558225, "grad_norm": 0.2560931729221973, "learning_rate": 2.3311231939454463e-05, "loss": 0.8833, "step": 7430 }, { "epoch": 3.3670140462165836, "grad_norm": 0.2413126432252204, "learning_rate": 2.3299733034907028e-05, "loss": 0.876, "step": 7431 }, { "epoch": 3.3674671499773448, "grad_norm": 0.34924977351694236, "learning_rate": 2.3288235801593313e-05, "loss": 0.8443, "step": 7432 }, { "epoch": 3.367920253738106, "grad_norm": 0.27297961993546754, "learning_rate": 2.327674024066389e-05, "loss": 0.848, "step": 7433 }, { "epoch": 3.368373357498867, "grad_norm": 0.28240963908646877, "learning_rate": 2.3265246353269122e-05, "loss": 0.8584, "step": 7434 }, { "epoch": 3.3688264612596286, "grad_norm": 0.21382678855967777, "learning_rate": 2.3253754140559258e-05, "loss": 0.8463, "step": 7435 }, { "epoch": 3.3692795650203897, "grad_norm": 0.20347327611575142, "learning_rate": 2.324226360368432e-05, "loss": 0.8346, "step": 7436 }, { "epoch": 3.369732668781151, "grad_norm": 0.19917284229203458, "learning_rate": 2.3230774743794217e-05, "loss": 0.8633, "step": 7437 }, { "epoch": 3.370185772541912, "grad_norm": 0.23333770135781204, "learning_rate": 2.3219287562038667e-05, "loss": 0.8545, "step": 7438 }, { "epoch": 3.3706388763026736, "grad_norm": 0.2233402908284768, "learning_rate": 2.3207802059567176e-05, "loss": 0.8596, "step": 7439 }, { "epoch": 3.3710919800634347, "grad_norm": 0.26284902105401176, "learning_rate": 2.319631823752918e-05, "loss": 0.8357, "step": 7440 }, { "epoch": 3.371545083824196, "grad_norm": 0.3799581221320321, "learning_rate": 2.3184836097073845e-05, "loss": 0.8455, "step": 7441 }, { "epoch": 3.371998187584957, "grad_norm": 0.29178352559633175, "learning_rate": 2.317335563935025e-05, "loss": 0.8703, "step": 7442 }, { "epoch": 3.372451291345718, "grad_norm": 0.19795642944873654, "learning_rate": 2.3161876865507248e-05, "loss": 0.8614, "step": 7443 }, { "epoch": 3.372904395106479, "grad_norm": 0.21552180771572324, "learning_rate": 2.3150399776693564e-05, "loss": 0.8466, "step": 7444 }, { "epoch": 3.373357498867241, "grad_norm": 0.26978440906546514, "learning_rate": 2.3138924374057727e-05, "loss": 0.879, "step": 7445 }, { "epoch": 3.373810602628002, "grad_norm": 0.2247120237200685, "learning_rate": 2.3127450658748096e-05, "loss": 0.8732, "step": 7446 }, { "epoch": 3.374263706388763, "grad_norm": 0.24971763733151756, "learning_rate": 2.311597863191289e-05, "loss": 0.8681, "step": 7447 }, { "epoch": 3.374716810149524, "grad_norm": 0.24128804613295382, "learning_rate": 2.3104508294700124e-05, "loss": 0.8775, "step": 7448 }, { "epoch": 3.3751699139102853, "grad_norm": 0.2965956674099863, "learning_rate": 2.3093039648257676e-05, "loss": 0.8672, "step": 7449 }, { "epoch": 3.375623017671047, "grad_norm": 0.3067874617302245, "learning_rate": 2.3081572693733224e-05, "loss": 0.8655, "step": 7450 }, { "epoch": 3.376076121431808, "grad_norm": 0.20576637344706558, "learning_rate": 2.3070107432274317e-05, "loss": 0.8675, "step": 7451 }, { "epoch": 3.376529225192569, "grad_norm": 0.2654188015129961, "learning_rate": 2.305864386502828e-05, "loss": 0.8553, "step": 7452 }, { "epoch": 3.3769823289533303, "grad_norm": 0.2618042643325438, "learning_rate": 2.3047181993142332e-05, "loss": 0.8655, "step": 7453 }, { "epoch": 3.3774354327140914, "grad_norm": 0.19377632479942772, "learning_rate": 2.303572181776346e-05, "loss": 0.8424, "step": 7454 }, { "epoch": 3.3778885364748525, "grad_norm": 0.26274738203521014, "learning_rate": 2.302426334003851e-05, "loss": 0.8521, "step": 7455 }, { "epoch": 3.378341640235614, "grad_norm": 0.2938269419016066, "learning_rate": 2.301280656111418e-05, "loss": 0.8547, "step": 7456 }, { "epoch": 3.3787947439963752, "grad_norm": 0.33564390690589696, "learning_rate": 2.300135148213695e-05, "loss": 0.8612, "step": 7457 }, { "epoch": 3.3792478477571364, "grad_norm": 0.24340414209717154, "learning_rate": 2.298989810425318e-05, "loss": 0.8584, "step": 7458 }, { "epoch": 3.3797009515178975, "grad_norm": 0.2380469985021485, "learning_rate": 2.297844642860901e-05, "loss": 0.8547, "step": 7459 }, { "epoch": 3.3801540552786586, "grad_norm": 0.35271850920646897, "learning_rate": 2.2966996456350467e-05, "loss": 0.8401, "step": 7460 }, { "epoch": 3.38060715903942, "grad_norm": 0.29273099670398833, "learning_rate": 2.2955548188623347e-05, "loss": 0.8652, "step": 7461 }, { "epoch": 3.3810602628001813, "grad_norm": 0.24518301163010978, "learning_rate": 2.2944101626573323e-05, "loss": 0.8554, "step": 7462 }, { "epoch": 3.3815133665609425, "grad_norm": 0.27629321056441547, "learning_rate": 2.2932656771345876e-05, "loss": 0.8538, "step": 7463 }, { "epoch": 3.3819664703217036, "grad_norm": 0.23887793381941372, "learning_rate": 2.2921213624086295e-05, "loss": 0.8662, "step": 7464 }, { "epoch": 3.3824195740824647, "grad_norm": 0.2613866202144253, "learning_rate": 2.2909772185939757e-05, "loss": 0.8518, "step": 7465 }, { "epoch": 3.382872677843226, "grad_norm": 0.18072634086919345, "learning_rate": 2.28983324580512e-05, "loss": 0.8683, "step": 7466 }, { "epoch": 3.3833257816039874, "grad_norm": 0.24932174172770977, "learning_rate": 2.288689444156546e-05, "loss": 0.8686, "step": 7467 }, { "epoch": 3.3837788853647486, "grad_norm": 0.24307575332041237, "learning_rate": 2.2875458137627143e-05, "loss": 0.8461, "step": 7468 }, { "epoch": 3.3842319891255097, "grad_norm": 0.27155939926906014, "learning_rate": 2.286402354738071e-05, "loss": 0.8697, "step": 7469 }, { "epoch": 3.384685092886271, "grad_norm": 0.20885182948198577, "learning_rate": 2.285259067197043e-05, "loss": 0.8418, "step": 7470 }, { "epoch": 3.3851381966470324, "grad_norm": 0.2710991044667358, "learning_rate": 2.284115951254044e-05, "loss": 0.8408, "step": 7471 }, { "epoch": 3.3855913004077935, "grad_norm": 0.28921784886390006, "learning_rate": 2.2829730070234666e-05, "loss": 0.852, "step": 7472 }, { "epoch": 3.3860444041685547, "grad_norm": 0.22023231789398173, "learning_rate": 2.2818302346196894e-05, "loss": 0.8586, "step": 7473 }, { "epoch": 3.386497507929316, "grad_norm": 0.20484891679623893, "learning_rate": 2.2806876341570723e-05, "loss": 0.8566, "step": 7474 }, { "epoch": 3.386950611690077, "grad_norm": 0.24380214511254544, "learning_rate": 2.279545205749955e-05, "loss": 0.8807, "step": 7475 }, { "epoch": 3.387403715450838, "grad_norm": 0.20455815457287504, "learning_rate": 2.278402949512667e-05, "loss": 0.8366, "step": 7476 }, { "epoch": 3.3878568192115996, "grad_norm": 0.24951563656646672, "learning_rate": 2.2772608655595124e-05, "loss": 0.8461, "step": 7477 }, { "epoch": 3.3883099229723608, "grad_norm": 0.21453538615463186, "learning_rate": 2.276118954004786e-05, "loss": 0.8505, "step": 7478 }, { "epoch": 3.388763026733122, "grad_norm": 0.22274619614719285, "learning_rate": 2.2749772149627578e-05, "loss": 0.8721, "step": 7479 }, { "epoch": 3.389216130493883, "grad_norm": 0.28429819658452354, "learning_rate": 2.2738356485476878e-05, "loss": 0.8703, "step": 7480 }, { "epoch": 3.389669234254644, "grad_norm": 0.22830263080310514, "learning_rate": 2.2726942548738124e-05, "loss": 0.8623, "step": 7481 }, { "epoch": 3.3901223380154057, "grad_norm": 0.25719948666620795, "learning_rate": 2.2715530340553555e-05, "loss": 0.8394, "step": 7482 }, { "epoch": 3.390575441776167, "grad_norm": 0.26321712136097764, "learning_rate": 2.2704119862065214e-05, "loss": 0.8495, "step": 7483 }, { "epoch": 3.391028545536928, "grad_norm": 0.22470498158491534, "learning_rate": 2.269271111441494e-05, "loss": 0.8722, "step": 7484 }, { "epoch": 3.391481649297689, "grad_norm": 0.23189252123256024, "learning_rate": 2.2681304098744485e-05, "loss": 0.838, "step": 7485 }, { "epoch": 3.3919347530584503, "grad_norm": 0.23095505348604006, "learning_rate": 2.266989881619533e-05, "loss": 0.8518, "step": 7486 }, { "epoch": 3.3923878568192114, "grad_norm": 0.21448481734086805, "learning_rate": 2.265849526790886e-05, "loss": 0.858, "step": 7487 }, { "epoch": 3.392840960579973, "grad_norm": 0.24816998813795438, "learning_rate": 2.2647093455026224e-05, "loss": 0.8662, "step": 7488 }, { "epoch": 3.393294064340734, "grad_norm": 0.21896121972005178, "learning_rate": 2.2635693378688463e-05, "loss": 0.8581, "step": 7489 }, { "epoch": 3.3937471681014952, "grad_norm": 0.2489088579310893, "learning_rate": 2.262429504003639e-05, "loss": 0.8429, "step": 7490 }, { "epoch": 3.3942002718622564, "grad_norm": 0.22620660899102787, "learning_rate": 2.2612898440210645e-05, "loss": 0.8781, "step": 7491 }, { "epoch": 3.3946533756230175, "grad_norm": 0.24691259896620424, "learning_rate": 2.2601503580351744e-05, "loss": 0.8587, "step": 7492 }, { "epoch": 3.395106479383779, "grad_norm": 0.19533625154652948, "learning_rate": 2.259011046159997e-05, "loss": 0.8625, "step": 7493 }, { "epoch": 3.39555958314454, "grad_norm": 0.23825563617412926, "learning_rate": 2.257871908509548e-05, "loss": 0.8594, "step": 7494 }, { "epoch": 3.3960126869053013, "grad_norm": 0.2523955313306957, "learning_rate": 2.256732945197821e-05, "loss": 0.8169, "step": 7495 }, { "epoch": 3.3964657906660625, "grad_norm": 0.19957499678646393, "learning_rate": 2.2555941563387974e-05, "loss": 0.8353, "step": 7496 }, { "epoch": 3.3969188944268236, "grad_norm": 0.23129718245302058, "learning_rate": 2.2544555420464364e-05, "loss": 0.8463, "step": 7497 }, { "epoch": 3.3973719981875847, "grad_norm": 0.19363073110603152, "learning_rate": 2.253317102434683e-05, "loss": 0.8614, "step": 7498 }, { "epoch": 3.3978251019483463, "grad_norm": 0.23341068372214338, "learning_rate": 2.252178837617464e-05, "loss": 0.8437, "step": 7499 }, { "epoch": 3.3982782057091074, "grad_norm": 0.23661347877240127, "learning_rate": 2.2510407477086845e-05, "loss": 0.8479, "step": 7500 }, { "epoch": 3.3987313094698686, "grad_norm": 0.21373629923004445, "learning_rate": 2.2499028328222403e-05, "loss": 0.8794, "step": 7501 }, { "epoch": 3.3991844132306297, "grad_norm": 0.21116110030130503, "learning_rate": 2.2487650930720014e-05, "loss": 0.8692, "step": 7502 }, { "epoch": 3.3996375169913913, "grad_norm": 0.19783452842250748, "learning_rate": 2.247627528571828e-05, "loss": 0.8499, "step": 7503 }, { "epoch": 3.4000906207521524, "grad_norm": 0.2043701394429158, "learning_rate": 2.2464901394355537e-05, "loss": 0.8543, "step": 7504 }, { "epoch": 3.4005437245129135, "grad_norm": 0.23887358585708354, "learning_rate": 2.245352925777005e-05, "loss": 0.8658, "step": 7505 }, { "epoch": 3.4009968282736747, "grad_norm": 0.21868781399102738, "learning_rate": 2.2442158877099824e-05, "loss": 0.8544, "step": 7506 }, { "epoch": 3.401449932034436, "grad_norm": 0.21472511005927133, "learning_rate": 2.2430790253482712e-05, "loss": 0.8553, "step": 7507 }, { "epoch": 3.401903035795197, "grad_norm": 0.1937296440689607, "learning_rate": 2.241942338805642e-05, "loss": 0.8474, "step": 7508 }, { "epoch": 3.4023561395559585, "grad_norm": 0.2171257512962372, "learning_rate": 2.240805828195843e-05, "loss": 0.8458, "step": 7509 }, { "epoch": 3.4028092433167196, "grad_norm": 0.20184382264249504, "learning_rate": 2.2396694936326106e-05, "loss": 0.841, "step": 7510 }, { "epoch": 3.4032623470774808, "grad_norm": 0.22858582272970665, "learning_rate": 2.2385333352296576e-05, "loss": 0.8553, "step": 7511 }, { "epoch": 3.403715450838242, "grad_norm": 0.23396099773345103, "learning_rate": 2.237397353100684e-05, "loss": 0.836, "step": 7512 }, { "epoch": 3.404168554599003, "grad_norm": 0.21450739382852393, "learning_rate": 2.2362615473593684e-05, "loss": 0.8829, "step": 7513 }, { "epoch": 3.4046216583597646, "grad_norm": 0.2542182993905758, "learning_rate": 2.2351259181193753e-05, "loss": 0.8552, "step": 7514 }, { "epoch": 3.4050747621205257, "grad_norm": 0.22079669498822035, "learning_rate": 2.233990465494349e-05, "loss": 0.8517, "step": 7515 }, { "epoch": 3.405527865881287, "grad_norm": 0.2806510448191875, "learning_rate": 2.232855189597915e-05, "loss": 0.8612, "step": 7516 }, { "epoch": 3.405980969642048, "grad_norm": 0.20787126964713692, "learning_rate": 2.231720090543686e-05, "loss": 0.8711, "step": 7517 }, { "epoch": 3.406434073402809, "grad_norm": 0.20029544668615196, "learning_rate": 2.2305851684452503e-05, "loss": 0.8638, "step": 7518 }, { "epoch": 3.4068871771635703, "grad_norm": 0.25489501327503883, "learning_rate": 2.2294504234161868e-05, "loss": 0.8641, "step": 7519 }, { "epoch": 3.407340280924332, "grad_norm": 0.29384880166169025, "learning_rate": 2.2283158555700473e-05, "loss": 0.854, "step": 7520 }, { "epoch": 3.407793384685093, "grad_norm": 0.293483439129025, "learning_rate": 2.2271814650203748e-05, "loss": 0.8618, "step": 7521 }, { "epoch": 3.408246488445854, "grad_norm": 0.2279885024598758, "learning_rate": 2.2260472518806862e-05, "loss": 0.8709, "step": 7522 }, { "epoch": 3.4086995922066152, "grad_norm": 0.2568020927307055, "learning_rate": 2.2249132162644892e-05, "loss": 0.8583, "step": 7523 }, { "epoch": 3.4091526959673764, "grad_norm": 0.25515295538363697, "learning_rate": 2.223779358285266e-05, "loss": 0.8522, "step": 7524 }, { "epoch": 3.409605799728138, "grad_norm": 0.251889253256516, "learning_rate": 2.2226456780564857e-05, "loss": 0.8504, "step": 7525 }, { "epoch": 3.410058903488899, "grad_norm": 0.18867535494765836, "learning_rate": 2.2215121756915985e-05, "loss": 0.8632, "step": 7526 }, { "epoch": 3.41051200724966, "grad_norm": 0.27121310631895784, "learning_rate": 2.2203788513040337e-05, "loss": 0.8784, "step": 7527 }, { "epoch": 3.4109651110104213, "grad_norm": 0.2604638351704252, "learning_rate": 2.21924570500721e-05, "loss": 0.8632, "step": 7528 }, { "epoch": 3.4114182147711825, "grad_norm": 0.24879081630884287, "learning_rate": 2.2181127369145202e-05, "loss": 0.8414, "step": 7529 }, { "epoch": 3.4118713185319436, "grad_norm": 0.2823299977781845, "learning_rate": 2.2169799471393466e-05, "loss": 0.8604, "step": 7530 }, { "epoch": 3.412324422292705, "grad_norm": 0.2871572713375418, "learning_rate": 2.2158473357950463e-05, "loss": 0.8443, "step": 7531 }, { "epoch": 3.4127775260534663, "grad_norm": 0.26906074824380466, "learning_rate": 2.214714902994965e-05, "loss": 0.871, "step": 7532 }, { "epoch": 3.4132306298142274, "grad_norm": 0.19795218707860462, "learning_rate": 2.2135826488524264e-05, "loss": 0.8497, "step": 7533 }, { "epoch": 3.4136837335749886, "grad_norm": 0.23095721641193892, "learning_rate": 2.2124505734807392e-05, "loss": 0.8581, "step": 7534 }, { "epoch": 3.41413683733575, "grad_norm": 0.28334601217094196, "learning_rate": 2.211318676993192e-05, "loss": 0.8636, "step": 7535 }, { "epoch": 3.4145899410965113, "grad_norm": 0.22280385314436937, "learning_rate": 2.2101869595030553e-05, "loss": 0.856, "step": 7536 }, { "epoch": 3.4150430448572724, "grad_norm": 0.24101588619034955, "learning_rate": 2.209055421123584e-05, "loss": 0.8685, "step": 7537 }, { "epoch": 3.4154961486180335, "grad_norm": 0.2754151787284054, "learning_rate": 2.2079240619680115e-05, "loss": 0.8401, "step": 7538 }, { "epoch": 3.4159492523787947, "grad_norm": 0.3006657925968915, "learning_rate": 2.2067928821495598e-05, "loss": 0.853, "step": 7539 }, { "epoch": 3.416402356139556, "grad_norm": 0.2433898195790782, "learning_rate": 2.2056618817814238e-05, "loss": 0.8842, "step": 7540 }, { "epoch": 3.4168554599003174, "grad_norm": 0.23786784289265375, "learning_rate": 2.204531060976789e-05, "loss": 0.8411, "step": 7541 }, { "epoch": 3.4173085636610785, "grad_norm": 0.23254985113062984, "learning_rate": 2.2034004198488175e-05, "loss": 0.8577, "step": 7542 }, { "epoch": 3.4177616674218396, "grad_norm": 0.2686852343073672, "learning_rate": 2.2022699585106544e-05, "loss": 0.8434, "step": 7543 }, { "epoch": 3.4182147711826008, "grad_norm": 0.19903054753951815, "learning_rate": 2.2011396770754296e-05, "loss": 0.8592, "step": 7544 }, { "epoch": 3.418667874943362, "grad_norm": 0.19599953124518787, "learning_rate": 2.2000095756562506e-05, "loss": 0.852, "step": 7545 }, { "epoch": 3.4191209787041235, "grad_norm": 0.21357987289423286, "learning_rate": 2.1988796543662115e-05, "loss": 0.8621, "step": 7546 }, { "epoch": 3.4195740824648846, "grad_norm": 0.21943863100491462, "learning_rate": 2.1977499133183837e-05, "loss": 0.8648, "step": 7547 }, { "epoch": 3.4200271862256457, "grad_norm": 0.1905528659851096, "learning_rate": 2.1966203526258258e-05, "loss": 0.8535, "step": 7548 }, { "epoch": 3.420480289986407, "grad_norm": 0.2273785149262698, "learning_rate": 2.1954909724015723e-05, "loss": 0.8508, "step": 7549 }, { "epoch": 3.420933393747168, "grad_norm": 0.24056145313778393, "learning_rate": 2.1943617727586463e-05, "loss": 0.8579, "step": 7550 }, { "epoch": 3.421386497507929, "grad_norm": 0.2274683490684438, "learning_rate": 2.1932327538100476e-05, "loss": 0.8678, "step": 7551 }, { "epoch": 3.4218396012686907, "grad_norm": 0.24832922179167016, "learning_rate": 2.1921039156687582e-05, "loss": 0.8594, "step": 7552 }, { "epoch": 3.422292705029452, "grad_norm": 0.2149211575489883, "learning_rate": 2.1909752584477464e-05, "loss": 0.848, "step": 7553 }, { "epoch": 3.422745808790213, "grad_norm": 0.235205350750718, "learning_rate": 2.1898467822599567e-05, "loss": 0.8719, "step": 7554 }, { "epoch": 3.423198912550974, "grad_norm": 0.26138909081499584, "learning_rate": 2.188718487218321e-05, "loss": 0.8577, "step": 7555 }, { "epoch": 3.423652016311735, "grad_norm": 0.21256347857135957, "learning_rate": 2.1875903734357477e-05, "loss": 0.859, "step": 7556 }, { "epoch": 3.424105120072497, "grad_norm": 0.21413723005294208, "learning_rate": 2.1864624410251327e-05, "loss": 0.8382, "step": 7557 }, { "epoch": 3.424558223833258, "grad_norm": 0.2779631676072218, "learning_rate": 2.1853346900993478e-05, "loss": 0.858, "step": 7558 }, { "epoch": 3.425011327594019, "grad_norm": 0.23074996647467425, "learning_rate": 2.1842071207712516e-05, "loss": 0.841, "step": 7559 }, { "epoch": 3.42546443135478, "grad_norm": 0.20708104156800894, "learning_rate": 2.183079733153683e-05, "loss": 0.839, "step": 7560 }, { "epoch": 3.4259175351155413, "grad_norm": 0.2771310705009886, "learning_rate": 2.1819525273594585e-05, "loss": 0.8554, "step": 7561 }, { "epoch": 3.4263706388763024, "grad_norm": 0.2532686115270148, "learning_rate": 2.180825503501385e-05, "loss": 0.8525, "step": 7562 }, { "epoch": 3.426823742637064, "grad_norm": 0.24777124755597818, "learning_rate": 2.1796986616922424e-05, "loss": 0.8469, "step": 7563 }, { "epoch": 3.427276846397825, "grad_norm": 0.20214894299207387, "learning_rate": 2.178572002044799e-05, "loss": 0.8703, "step": 7564 }, { "epoch": 3.4277299501585863, "grad_norm": 0.17931817626303692, "learning_rate": 2.1774455246718004e-05, "loss": 0.8476, "step": 7565 }, { "epoch": 3.4281830539193474, "grad_norm": 0.2160532948663639, "learning_rate": 2.176319229685978e-05, "loss": 0.8566, "step": 7566 }, { "epoch": 3.428636157680109, "grad_norm": 0.20429385307079512, "learning_rate": 2.175193117200041e-05, "loss": 0.8496, "step": 7567 }, { "epoch": 3.42908926144087, "grad_norm": 0.21440225542585808, "learning_rate": 2.174067187326681e-05, "loss": 0.8601, "step": 7568 }, { "epoch": 3.4295423652016312, "grad_norm": 0.269407594761201, "learning_rate": 2.1729414401785753e-05, "loss": 0.8518, "step": 7569 }, { "epoch": 3.4299954689623924, "grad_norm": 0.18137344603901207, "learning_rate": 2.1718158758683762e-05, "loss": 0.857, "step": 7570 }, { "epoch": 3.4304485727231535, "grad_norm": 0.27136721372442735, "learning_rate": 2.1706904945087258e-05, "loss": 0.8624, "step": 7571 }, { "epoch": 3.4309016764839146, "grad_norm": 0.18523521772913135, "learning_rate": 2.1695652962122392e-05, "loss": 0.838, "step": 7572 }, { "epoch": 3.431354780244676, "grad_norm": 0.2769160470315618, "learning_rate": 2.1684402810915223e-05, "loss": 0.8682, "step": 7573 }, { "epoch": 3.4318078840054373, "grad_norm": 0.22054950997860387, "learning_rate": 2.167315449259153e-05, "loss": 0.8522, "step": 7574 }, { "epoch": 3.4322609877661985, "grad_norm": 0.24962642249272654, "learning_rate": 2.1661908008277e-05, "loss": 0.8553, "step": 7575 }, { "epoch": 3.4327140915269596, "grad_norm": 0.2468277176081092, "learning_rate": 2.1650663359097078e-05, "loss": 0.8654, "step": 7576 }, { "epoch": 3.4331671952877207, "grad_norm": 0.22272634975136776, "learning_rate": 2.1639420546177022e-05, "loss": 0.8442, "step": 7577 }, { "epoch": 3.4336202990484823, "grad_norm": 0.2046908898503734, "learning_rate": 2.162817957064196e-05, "loss": 0.8598, "step": 7578 }, { "epoch": 3.4340734028092434, "grad_norm": 0.21063656242859916, "learning_rate": 2.1616940433616768e-05, "loss": 0.8482, "step": 7579 }, { "epoch": 3.4345265065700046, "grad_norm": 0.2507827162709439, "learning_rate": 2.160570313622621e-05, "loss": 0.8731, "step": 7580 }, { "epoch": 3.4349796103307657, "grad_norm": 0.203114880554659, "learning_rate": 2.159446767959479e-05, "loss": 0.8733, "step": 7581 }, { "epoch": 3.435432714091527, "grad_norm": 0.21650081575498378, "learning_rate": 2.158323406484691e-05, "loss": 0.8284, "step": 7582 }, { "epoch": 3.435885817852288, "grad_norm": 0.20795484576674675, "learning_rate": 2.157200229310669e-05, "loss": 0.8489, "step": 7583 }, { "epoch": 3.4363389216130495, "grad_norm": 0.21730944784007578, "learning_rate": 2.156077236549816e-05, "loss": 0.8674, "step": 7584 }, { "epoch": 3.4367920253738107, "grad_norm": 0.241369894572405, "learning_rate": 2.1549544283145094e-05, "loss": 0.8399, "step": 7585 }, { "epoch": 3.437245129134572, "grad_norm": 0.17737390328085287, "learning_rate": 2.1538318047171142e-05, "loss": 0.8444, "step": 7586 }, { "epoch": 3.437698232895333, "grad_norm": 0.2060426707765262, "learning_rate": 2.1527093658699725e-05, "loss": 0.8501, "step": 7587 }, { "epoch": 3.438151336656094, "grad_norm": 0.1962524418656349, "learning_rate": 2.151587111885408e-05, "loss": 0.8509, "step": 7588 }, { "epoch": 3.4386044404168556, "grad_norm": 0.23587964328758976, "learning_rate": 2.1504650428757298e-05, "loss": 0.8422, "step": 7589 }, { "epoch": 3.4390575441776168, "grad_norm": 0.17276419113249233, "learning_rate": 2.1493431589532232e-05, "loss": 0.8614, "step": 7590 }, { "epoch": 3.439510647938378, "grad_norm": 0.22580232121635097, "learning_rate": 2.1482214602301602e-05, "loss": 0.8609, "step": 7591 }, { "epoch": 3.439963751699139, "grad_norm": 0.22303925165422936, "learning_rate": 2.1470999468187896e-05, "loss": 0.8535, "step": 7592 }, { "epoch": 3.4404168554599, "grad_norm": 0.22633821814693078, "learning_rate": 2.145978618831346e-05, "loss": 0.8713, "step": 7593 }, { "epoch": 3.4408699592206613, "grad_norm": 0.23648601070684933, "learning_rate": 2.1448574763800405e-05, "loss": 0.8641, "step": 7594 }, { "epoch": 3.441323062981423, "grad_norm": 0.24273594923139444, "learning_rate": 2.1437365195770717e-05, "loss": 0.8454, "step": 7595 }, { "epoch": 3.441776166742184, "grad_norm": 0.19491876476781306, "learning_rate": 2.142615748534615e-05, "loss": 0.844, "step": 7596 }, { "epoch": 3.442229270502945, "grad_norm": 0.18548984679415453, "learning_rate": 2.1414951633648258e-05, "loss": 0.8508, "step": 7597 }, { "epoch": 3.4426823742637063, "grad_norm": 0.20160531267506399, "learning_rate": 2.1403747641798475e-05, "loss": 0.8651, "step": 7598 }, { "epoch": 3.443135478024468, "grad_norm": 0.2086398214098943, "learning_rate": 2.139254551091798e-05, "loss": 0.8629, "step": 7599 }, { "epoch": 3.443588581785229, "grad_norm": 0.20882882344558668, "learning_rate": 2.138134524212782e-05, "loss": 0.8624, "step": 7600 }, { "epoch": 3.44404168554599, "grad_norm": 0.19823539475871818, "learning_rate": 2.137014683654881e-05, "loss": 0.8492, "step": 7601 }, { "epoch": 3.4444947893067512, "grad_norm": 0.183460508024303, "learning_rate": 2.1358950295301625e-05, "loss": 0.8555, "step": 7602 }, { "epoch": 3.4449478930675124, "grad_norm": 0.22759211558802406, "learning_rate": 2.1347755619506714e-05, "loss": 0.8634, "step": 7603 }, { "epoch": 3.4454009968282735, "grad_norm": 0.1590740084194823, "learning_rate": 2.133656281028434e-05, "loss": 0.8606, "step": 7604 }, { "epoch": 3.445854100589035, "grad_norm": 0.228957729111086, "learning_rate": 2.132537186875462e-05, "loss": 0.8527, "step": 7605 }, { "epoch": 3.446307204349796, "grad_norm": 0.20622544185418068, "learning_rate": 2.1314182796037424e-05, "loss": 0.852, "step": 7606 }, { "epoch": 3.4467603081105573, "grad_norm": 0.19963715888671515, "learning_rate": 2.130299559325251e-05, "loss": 0.8299, "step": 7607 }, { "epoch": 3.4472134118713185, "grad_norm": 0.1890130885370368, "learning_rate": 2.1291810261519365e-05, "loss": 0.8455, "step": 7608 }, { "epoch": 3.4476665156320796, "grad_norm": 0.17410771753426307, "learning_rate": 2.1280626801957363e-05, "loss": 0.8523, "step": 7609 }, { "epoch": 3.448119619392841, "grad_norm": 0.19732025444767665, "learning_rate": 2.1269445215685637e-05, "loss": 0.8738, "step": 7610 }, { "epoch": 3.4485727231536023, "grad_norm": 0.21606013293567625, "learning_rate": 2.1258265503823166e-05, "loss": 0.8532, "step": 7611 }, { "epoch": 3.4490258269143634, "grad_norm": 0.18969718035053917, "learning_rate": 2.1247087667488735e-05, "loss": 0.8559, "step": 7612 }, { "epoch": 3.4494789306751246, "grad_norm": 0.22276013795737773, "learning_rate": 2.1235911707800908e-05, "loss": 0.8585, "step": 7613 }, { "epoch": 3.4499320344358857, "grad_norm": 0.23632912412934498, "learning_rate": 2.122473762587812e-05, "loss": 0.8639, "step": 7614 }, { "epoch": 3.450385138196647, "grad_norm": 0.21788409392854405, "learning_rate": 2.1213565422838563e-05, "loss": 0.8481, "step": 7615 }, { "epoch": 3.4508382419574084, "grad_norm": 0.18617723858231236, "learning_rate": 2.1202395099800287e-05, "loss": 0.8477, "step": 7616 }, { "epoch": 3.4512913457181695, "grad_norm": 0.2500763501874297, "learning_rate": 2.11912266578811e-05, "loss": 0.8689, "step": 7617 }, { "epoch": 3.4517444494789307, "grad_norm": 0.18817321028640652, "learning_rate": 2.1180060098198693e-05, "loss": 0.8531, "step": 7618 }, { "epoch": 3.452197553239692, "grad_norm": 0.21537705425328937, "learning_rate": 2.11688954218705e-05, "loss": 0.8715, "step": 7619 }, { "epoch": 3.452650657000453, "grad_norm": 0.24720028407311892, "learning_rate": 2.115773263001381e-05, "loss": 0.8475, "step": 7620 }, { "epoch": 3.4531037607612145, "grad_norm": 0.2455617146417371, "learning_rate": 2.1146571723745705e-05, "loss": 0.8509, "step": 7621 }, { "epoch": 3.4535568645219756, "grad_norm": 0.20391152937047555, "learning_rate": 2.1135412704183067e-05, "loss": 0.834, "step": 7622 }, { "epoch": 3.4540099682827368, "grad_norm": 0.22297297423898782, "learning_rate": 2.1124255572442634e-05, "loss": 0.8603, "step": 7623 }, { "epoch": 3.454463072043498, "grad_norm": 0.21434643147757032, "learning_rate": 2.11131003296409e-05, "loss": 0.8525, "step": 7624 }, { "epoch": 3.454916175804259, "grad_norm": 0.21075778849898705, "learning_rate": 2.110194697689421e-05, "loss": 0.8523, "step": 7625 }, { "epoch": 3.45536927956502, "grad_norm": 0.1953896355881209, "learning_rate": 2.1090795515318695e-05, "loss": 0.8909, "step": 7626 }, { "epoch": 3.4558223833257817, "grad_norm": 0.19255533658048324, "learning_rate": 2.1079645946030323e-05, "loss": 0.8619, "step": 7627 }, { "epoch": 3.456275487086543, "grad_norm": 0.2126615457802581, "learning_rate": 2.106849827014485e-05, "loss": 0.849, "step": 7628 }, { "epoch": 3.456728590847304, "grad_norm": 0.19678654353382707, "learning_rate": 2.1057352488777826e-05, "loss": 0.8735, "step": 7629 }, { "epoch": 3.457181694608065, "grad_norm": 0.18258238117939174, "learning_rate": 2.1046208603044677e-05, "loss": 0.8481, "step": 7630 }, { "epoch": 3.4576347983688267, "grad_norm": 0.22635888864992085, "learning_rate": 2.103506661406056e-05, "loss": 0.8481, "step": 7631 }, { "epoch": 3.458087902129588, "grad_norm": 0.24465167395363704, "learning_rate": 2.10239265229405e-05, "loss": 0.8777, "step": 7632 }, { "epoch": 3.458541005890349, "grad_norm": 0.24446057411448702, "learning_rate": 2.10127883307993e-05, "loss": 0.8527, "step": 7633 }, { "epoch": 3.45899410965111, "grad_norm": 0.27504109549324823, "learning_rate": 2.10016520387516e-05, "loss": 0.8756, "step": 7634 }, { "epoch": 3.4594472134118712, "grad_norm": 0.1875503683677984, "learning_rate": 2.099051764791181e-05, "loss": 0.8357, "step": 7635 }, { "epoch": 3.4599003171726324, "grad_norm": 0.23162340435363188, "learning_rate": 2.0979385159394203e-05, "loss": 0.856, "step": 7636 }, { "epoch": 3.460353420933394, "grad_norm": 0.19179050988454316, "learning_rate": 2.0968254574312815e-05, "loss": 0.8465, "step": 7637 }, { "epoch": 3.460806524694155, "grad_norm": 0.23016627426263964, "learning_rate": 2.0957125893781497e-05, "loss": 0.8661, "step": 7638 }, { "epoch": 3.461259628454916, "grad_norm": 0.18389789038460727, "learning_rate": 2.0945999118913952e-05, "loss": 0.8446, "step": 7639 }, { "epoch": 3.4617127322156773, "grad_norm": 0.21343261512410983, "learning_rate": 2.093487425082364e-05, "loss": 0.856, "step": 7640 }, { "epoch": 3.4621658359764385, "grad_norm": 0.18540917701457782, "learning_rate": 2.0923751290623862e-05, "loss": 0.8523, "step": 7641 }, { "epoch": 3.4626189397372, "grad_norm": 0.24804036377176789, "learning_rate": 2.0912630239427695e-05, "loss": 0.8574, "step": 7642 }, { "epoch": 3.463072043497961, "grad_norm": 0.2743500188614412, "learning_rate": 2.0901511098348087e-05, "loss": 0.8419, "step": 7643 }, { "epoch": 3.4635251472587223, "grad_norm": 0.2663549651875696, "learning_rate": 2.089039386849771e-05, "loss": 0.8593, "step": 7644 }, { "epoch": 3.4639782510194834, "grad_norm": 0.35819142474086796, "learning_rate": 2.0879278550989137e-05, "loss": 0.8677, "step": 7645 }, { "epoch": 3.4644313547802446, "grad_norm": 0.22502806249580465, "learning_rate": 2.0868165146934665e-05, "loss": 0.8642, "step": 7646 }, { "epoch": 3.4648844585410057, "grad_norm": 0.4360242772814925, "learning_rate": 2.0857053657446467e-05, "loss": 0.8629, "step": 7647 }, { "epoch": 3.4653375623017673, "grad_norm": 0.3215836157904443, "learning_rate": 2.084594408363648e-05, "loss": 0.8799, "step": 7648 }, { "epoch": 3.4657906660625284, "grad_norm": 0.27914336251304905, "learning_rate": 2.0834836426616456e-05, "loss": 0.861, "step": 7649 }, { "epoch": 3.4662437698232895, "grad_norm": 0.2875735974263299, "learning_rate": 2.0823730687497985e-05, "loss": 0.8602, "step": 7650 }, { "epoch": 3.4666968735840507, "grad_norm": 0.26740728133385583, "learning_rate": 2.0812626867392423e-05, "loss": 0.8543, "step": 7651 }, { "epoch": 3.467149977344812, "grad_norm": 0.22815992998040657, "learning_rate": 2.0801524967410972e-05, "loss": 0.8881, "step": 7652 }, { "epoch": 3.4676030811055734, "grad_norm": 0.191489683647731, "learning_rate": 2.0790424988664607e-05, "loss": 0.8618, "step": 7653 }, { "epoch": 3.4680561848663345, "grad_norm": 0.3677101345463394, "learning_rate": 2.0779326932264147e-05, "loss": 0.8743, "step": 7654 }, { "epoch": 3.4685092886270956, "grad_norm": 0.24593056629414767, "learning_rate": 2.076823079932018e-05, "loss": 0.8457, "step": 7655 }, { "epoch": 3.4689623923878568, "grad_norm": 0.22491445317015435, "learning_rate": 2.0757136590943144e-05, "loss": 0.8706, "step": 7656 }, { "epoch": 3.469415496148618, "grad_norm": 0.1884911476058863, "learning_rate": 2.0746044308243246e-05, "loss": 0.8411, "step": 7657 }, { "epoch": 3.469868599909379, "grad_norm": 0.2021221178655566, "learning_rate": 2.0734953952330512e-05, "loss": 0.8712, "step": 7658 }, { "epoch": 3.4703217036701406, "grad_norm": 0.20645779211925738, "learning_rate": 2.072386552431479e-05, "loss": 0.855, "step": 7659 }, { "epoch": 3.4707748074309017, "grad_norm": 0.18548992432263878, "learning_rate": 2.0712779025305714e-05, "loss": 0.8693, "step": 7660 }, { "epoch": 3.471227911191663, "grad_norm": 0.21006952340951182, "learning_rate": 2.070169445641276e-05, "loss": 0.841, "step": 7661 }, { "epoch": 3.471681014952424, "grad_norm": 0.16202796190356675, "learning_rate": 2.0690611818745142e-05, "loss": 0.8646, "step": 7662 }, { "epoch": 3.4721341187131856, "grad_norm": 0.18789176824741166, "learning_rate": 2.0679531113411974e-05, "loss": 0.8467, "step": 7663 }, { "epoch": 3.4725872224739467, "grad_norm": 0.19328451975226876, "learning_rate": 2.06684523415221e-05, "loss": 0.8687, "step": 7664 }, { "epoch": 3.473040326234708, "grad_norm": 0.1832715139478923, "learning_rate": 2.065737550418419e-05, "loss": 0.8469, "step": 7665 }, { "epoch": 3.473493429995469, "grad_norm": 0.1790931567764358, "learning_rate": 2.064630060250675e-05, "loss": 0.8538, "step": 7666 }, { "epoch": 3.47394653375623, "grad_norm": 0.169588554677636, "learning_rate": 2.0635227637598053e-05, "loss": 0.8452, "step": 7667 }, { "epoch": 3.474399637516991, "grad_norm": 0.2148255781460299, "learning_rate": 2.062415661056621e-05, "loss": 0.8402, "step": 7668 }, { "epoch": 3.474852741277753, "grad_norm": 0.18568428706278095, "learning_rate": 2.061308752251911e-05, "loss": 0.8432, "step": 7669 }, { "epoch": 3.475305845038514, "grad_norm": 0.23716208365654728, "learning_rate": 2.0602020374564483e-05, "loss": 0.8704, "step": 7670 }, { "epoch": 3.475758948799275, "grad_norm": 0.25386583578579464, "learning_rate": 2.059095516780981e-05, "loss": 0.8384, "step": 7671 }, { "epoch": 3.476212052560036, "grad_norm": 0.19273571014349244, "learning_rate": 2.057989190336245e-05, "loss": 0.84, "step": 7672 }, { "epoch": 3.4766651563207973, "grad_norm": 0.23375029105150646, "learning_rate": 2.056883058232951e-05, "loss": 0.8686, "step": 7673 }, { "epoch": 3.477118260081559, "grad_norm": 0.19236021146762067, "learning_rate": 2.055777120581791e-05, "loss": 0.8608, "step": 7674 }, { "epoch": 3.47757136384232, "grad_norm": 0.22422578782215355, "learning_rate": 2.054671377493441e-05, "loss": 0.8649, "step": 7675 }, { "epoch": 3.478024467603081, "grad_norm": 0.1834169157204344, "learning_rate": 2.0535658290785525e-05, "loss": 0.8461, "step": 7676 }, { "epoch": 3.4784775713638423, "grad_norm": 0.2753079073133639, "learning_rate": 2.052460475447764e-05, "loss": 0.8462, "step": 7677 }, { "epoch": 3.4789306751246034, "grad_norm": 0.20686631525113797, "learning_rate": 2.0513553167116864e-05, "loss": 0.8434, "step": 7678 }, { "epoch": 3.4793837788853645, "grad_norm": 0.19695120798924082, "learning_rate": 2.0502503529809195e-05, "loss": 0.8553, "step": 7679 }, { "epoch": 3.479836882646126, "grad_norm": 0.21027115173535266, "learning_rate": 2.049145584366036e-05, "loss": 0.8615, "step": 7680 }, { "epoch": 3.4802899864068872, "grad_norm": 0.17280375592329666, "learning_rate": 2.0480410109775963e-05, "loss": 0.8432, "step": 7681 }, { "epoch": 3.4807430901676484, "grad_norm": 0.19995144898048245, "learning_rate": 2.0469366329261352e-05, "loss": 0.837, "step": 7682 }, { "epoch": 3.4811961939284095, "grad_norm": 0.18568141821303374, "learning_rate": 2.0458324503221697e-05, "loss": 0.8661, "step": 7683 }, { "epoch": 3.4816492976891706, "grad_norm": 0.17117660795182574, "learning_rate": 2.0447284632762e-05, "loss": 0.8434, "step": 7684 }, { "epoch": 3.482102401449932, "grad_norm": 0.18769206611340503, "learning_rate": 2.043624671898702e-05, "loss": 0.8464, "step": 7685 }, { "epoch": 3.4825555052106933, "grad_norm": 0.18760190740125762, "learning_rate": 2.0425210763001376e-05, "loss": 0.857, "step": 7686 }, { "epoch": 3.4830086089714545, "grad_norm": 0.19833747104049215, "learning_rate": 2.0414176765909433e-05, "loss": 0.8584, "step": 7687 }, { "epoch": 3.4834617127322156, "grad_norm": 0.17874289230016818, "learning_rate": 2.0403144728815417e-05, "loss": 0.8553, "step": 7688 }, { "epoch": 3.4839148164929767, "grad_norm": 0.2028396791296418, "learning_rate": 2.0392114652823308e-05, "loss": 0.8547, "step": 7689 }, { "epoch": 3.484367920253738, "grad_norm": 0.18621093568295283, "learning_rate": 2.0381086539036906e-05, "loss": 0.863, "step": 7690 }, { "epoch": 3.4848210240144994, "grad_norm": 0.18703354066686512, "learning_rate": 2.037006038855984e-05, "loss": 0.843, "step": 7691 }, { "epoch": 3.4852741277752606, "grad_norm": 0.18742040506645719, "learning_rate": 2.035903620249551e-05, "loss": 0.8464, "step": 7692 }, { "epoch": 3.4857272315360217, "grad_norm": 0.18291343104076707, "learning_rate": 2.0348013981947137e-05, "loss": 0.8552, "step": 7693 }, { "epoch": 3.486180335296783, "grad_norm": 0.16217824983871615, "learning_rate": 2.0336993728017723e-05, "loss": 0.8583, "step": 7694 }, { "epoch": 3.4866334390575444, "grad_norm": 0.1966074713885507, "learning_rate": 2.0325975441810115e-05, "loss": 0.8625, "step": 7695 }, { "epoch": 3.4870865428183055, "grad_norm": 0.19629053946691963, "learning_rate": 2.031495912442691e-05, "loss": 0.8505, "step": 7696 }, { "epoch": 3.4875396465790667, "grad_norm": 0.19912571370819804, "learning_rate": 2.0303944776970587e-05, "loss": 0.8902, "step": 7697 }, { "epoch": 3.487992750339828, "grad_norm": 0.22407718610592775, "learning_rate": 2.029293240054331e-05, "loss": 0.8464, "step": 7698 }, { "epoch": 3.488445854100589, "grad_norm": 0.18258342349360154, "learning_rate": 2.028192199624716e-05, "loss": 0.8564, "step": 7699 }, { "epoch": 3.48889895786135, "grad_norm": 0.22930791963224764, "learning_rate": 2.0270913565183942e-05, "loss": 0.8564, "step": 7700 }, { "epoch": 3.4893520616221116, "grad_norm": 0.19017732558755554, "learning_rate": 2.025990710845532e-05, "loss": 0.8493, "step": 7701 }, { "epoch": 3.4898051653828728, "grad_norm": 0.22020885871998164, "learning_rate": 2.024890262716273e-05, "loss": 0.8493, "step": 7702 }, { "epoch": 3.490258269143634, "grad_norm": 0.20312944385613363, "learning_rate": 2.0237900122407392e-05, "loss": 0.8422, "step": 7703 }, { "epoch": 3.490711372904395, "grad_norm": 0.194818400987498, "learning_rate": 2.0226899595290393e-05, "loss": 0.8597, "step": 7704 }, { "epoch": 3.491164476665156, "grad_norm": 0.17140529055366954, "learning_rate": 2.0215901046912537e-05, "loss": 0.8654, "step": 7705 }, { "epoch": 3.4916175804259177, "grad_norm": 0.19634936959121593, "learning_rate": 2.020490447837451e-05, "loss": 0.8626, "step": 7706 }, { "epoch": 3.492070684186679, "grad_norm": 0.1800134128435048, "learning_rate": 2.019390989077674e-05, "loss": 0.8757, "step": 7707 }, { "epoch": 3.49252378794744, "grad_norm": 0.16196528645271832, "learning_rate": 2.0182917285219493e-05, "loss": 0.844, "step": 7708 }, { "epoch": 3.492976891708201, "grad_norm": 0.16091018853478994, "learning_rate": 2.0171926662802827e-05, "loss": 0.8631, "step": 7709 }, { "epoch": 3.4934299954689623, "grad_norm": 0.20757211113237403, "learning_rate": 2.0160938024626574e-05, "loss": 0.8378, "step": 7710 }, { "epoch": 3.4938830992297234, "grad_norm": 0.2189408307794617, "learning_rate": 2.014995137179043e-05, "loss": 0.8496, "step": 7711 }, { "epoch": 3.494336202990485, "grad_norm": 0.18163822243389793, "learning_rate": 2.0138966705393814e-05, "loss": 0.8579, "step": 7712 }, { "epoch": 3.494789306751246, "grad_norm": 0.27083844187083056, "learning_rate": 2.012798402653602e-05, "loss": 0.8625, "step": 7713 }, { "epoch": 3.4952424105120072, "grad_norm": 0.25560020046948534, "learning_rate": 2.0117003336316084e-05, "loss": 0.8401, "step": 7714 }, { "epoch": 3.4956955142727684, "grad_norm": 0.21385303394048275, "learning_rate": 2.0106024635832898e-05, "loss": 0.8581, "step": 7715 }, { "epoch": 3.4961486180335295, "grad_norm": 0.1862659108318867, "learning_rate": 2.0095047926185095e-05, "loss": 0.8563, "step": 7716 }, { "epoch": 3.496601721794291, "grad_norm": 0.23691182184069962, "learning_rate": 2.0084073208471162e-05, "loss": 0.8592, "step": 7717 }, { "epoch": 3.497054825555052, "grad_norm": 0.25752612826203314, "learning_rate": 2.0073100483789362e-05, "loss": 0.8446, "step": 7718 }, { "epoch": 3.4975079293158133, "grad_norm": 0.1700287746874578, "learning_rate": 2.0062129753237728e-05, "loss": 0.848, "step": 7719 }, { "epoch": 3.4979610330765745, "grad_norm": 0.18659480448389829, "learning_rate": 2.0051161017914173e-05, "loss": 0.8619, "step": 7720 }, { "epoch": 3.4984141368373356, "grad_norm": 0.23897850580053812, "learning_rate": 2.004019427891633e-05, "loss": 0.851, "step": 7721 }, { "epoch": 3.4988672405980967, "grad_norm": 0.22391230469867454, "learning_rate": 2.002922953734168e-05, "loss": 0.8492, "step": 7722 }, { "epoch": 3.4993203443588583, "grad_norm": 0.21167476816720268, "learning_rate": 2.0018266794287485e-05, "loss": 0.8566, "step": 7723 }, { "epoch": 3.4997734481196194, "grad_norm": 0.18863221544996095, "learning_rate": 2.0007306050850824e-05, "loss": 0.8661, "step": 7724 }, { "epoch": 3.5002265518803806, "grad_norm": 0.19618471269666307, "learning_rate": 1.9996347308128552e-05, "loss": 0.8517, "step": 7725 }, { "epoch": 3.5006796556411417, "grad_norm": 0.15907882019534003, "learning_rate": 1.998539056721732e-05, "loss": 0.845, "step": 7726 }, { "epoch": 3.5011327594019033, "grad_norm": 0.18563242842039362, "learning_rate": 1.9974435829213626e-05, "loss": 0.8488, "step": 7727 }, { "epoch": 3.5015858631626644, "grad_norm": 0.20579710567439585, "learning_rate": 1.9963483095213706e-05, "loss": 0.8648, "step": 7728 }, { "epoch": 3.5020389669234255, "grad_norm": 0.1989906498172126, "learning_rate": 1.995253236631366e-05, "loss": 0.8496, "step": 7729 }, { "epoch": 3.5024920706841867, "grad_norm": 0.20070599082296892, "learning_rate": 1.9941583643609307e-05, "loss": 0.8433, "step": 7730 }, { "epoch": 3.502945174444948, "grad_norm": 0.2628320965730324, "learning_rate": 1.993063692819635e-05, "loss": 0.8416, "step": 7731 }, { "epoch": 3.503398278205709, "grad_norm": 0.2674576299041253, "learning_rate": 1.9919692221170223e-05, "loss": 0.8514, "step": 7732 }, { "epoch": 3.50385138196647, "grad_norm": 0.17876107051582776, "learning_rate": 1.9908749523626222e-05, "loss": 0.8558, "step": 7733 }, { "epoch": 3.5043044857272316, "grad_norm": 0.2832503981154127, "learning_rate": 1.9897808836659382e-05, "loss": 0.8509, "step": 7734 }, { "epoch": 3.5047575894879928, "grad_norm": 0.31266564090930093, "learning_rate": 1.9886870161364556e-05, "loss": 0.8706, "step": 7735 }, { "epoch": 3.505210693248754, "grad_norm": 0.230153414019902, "learning_rate": 1.9875933498836423e-05, "loss": 0.8747, "step": 7736 }, { "epoch": 3.505663797009515, "grad_norm": 0.17797814276553703, "learning_rate": 1.986499885016942e-05, "loss": 0.862, "step": 7737 }, { "epoch": 3.5061169007702766, "grad_norm": 0.293046278562541, "learning_rate": 1.9854066216457823e-05, "loss": 0.8572, "step": 7738 }, { "epoch": 3.5065700045310377, "grad_norm": 0.21094500797789753, "learning_rate": 1.9843135598795667e-05, "loss": 0.8621, "step": 7739 }, { "epoch": 3.507023108291799, "grad_norm": 0.2128080293536721, "learning_rate": 1.983220699827682e-05, "loss": 0.864, "step": 7740 }, { "epoch": 3.50747621205256, "grad_norm": 0.2616969118644563, "learning_rate": 1.9821280415994914e-05, "loss": 0.8626, "step": 7741 }, { "epoch": 3.507929315813321, "grad_norm": 0.26216009613511004, "learning_rate": 1.9810355853043418e-05, "loss": 0.8536, "step": 7742 }, { "epoch": 3.5083824195740823, "grad_norm": 0.21444523151611664, "learning_rate": 1.979943331051557e-05, "loss": 0.8743, "step": 7743 }, { "epoch": 3.508835523334844, "grad_norm": 0.16241455586582637, "learning_rate": 1.9788512789504392e-05, "loss": 0.8536, "step": 7744 }, { "epoch": 3.509288627095605, "grad_norm": 0.22607121027096902, "learning_rate": 1.9777594291102756e-05, "loss": 0.8151, "step": 7745 }, { "epoch": 3.509741730856366, "grad_norm": 0.21555216834456878, "learning_rate": 1.9766677816403283e-05, "loss": 0.8449, "step": 7746 }, { "epoch": 3.5101948346171272, "grad_norm": 0.13674894308879917, "learning_rate": 1.975576336649842e-05, "loss": 0.8474, "step": 7747 }, { "epoch": 3.510647938377889, "grad_norm": 0.23914996202800617, "learning_rate": 1.974485094248039e-05, "loss": 0.8619, "step": 7748 }, { "epoch": 3.51110104213865, "grad_norm": 0.3227276357472399, "learning_rate": 1.973394054544124e-05, "loss": 0.858, "step": 7749 }, { "epoch": 3.511554145899411, "grad_norm": 0.23409328292367773, "learning_rate": 1.9723032176472787e-05, "loss": 0.8443, "step": 7750 }, { "epoch": 3.512007249660172, "grad_norm": 0.19653402998587, "learning_rate": 1.9712125836666645e-05, "loss": 0.8607, "step": 7751 }, { "epoch": 3.5124603534209333, "grad_norm": 0.32033007355264287, "learning_rate": 1.9701221527114262e-05, "loss": 0.8443, "step": 7752 }, { "epoch": 3.5129134571816945, "grad_norm": 0.3015853432204533, "learning_rate": 1.9690319248906834e-05, "loss": 0.8743, "step": 7753 }, { "epoch": 3.5133665609424556, "grad_norm": 0.2007684218564718, "learning_rate": 1.9679419003135413e-05, "loss": 0.8629, "step": 7754 }, { "epoch": 3.513819664703217, "grad_norm": 0.31740792916734606, "learning_rate": 1.9668520790890758e-05, "loss": 0.8428, "step": 7755 }, { "epoch": 3.5142727684639783, "grad_norm": 0.3039201310236514, "learning_rate": 1.9657624613263518e-05, "loss": 0.8558, "step": 7756 }, { "epoch": 3.5147258722247394, "grad_norm": 0.1889544350210027, "learning_rate": 1.9646730471344065e-05, "loss": 0.8348, "step": 7757 }, { "epoch": 3.5151789759855006, "grad_norm": 0.20356764325618815, "learning_rate": 1.9635838366222643e-05, "loss": 0.851, "step": 7758 }, { "epoch": 3.515632079746262, "grad_norm": 0.24648689302349602, "learning_rate": 1.9624948298989214e-05, "loss": 0.8654, "step": 7759 }, { "epoch": 3.5160851835070233, "grad_norm": 0.2297515474167099, "learning_rate": 1.9614060270733593e-05, "loss": 0.8493, "step": 7760 }, { "epoch": 3.5165382872677844, "grad_norm": 0.17815666878320727, "learning_rate": 1.960317428254537e-05, "loss": 0.8425, "step": 7761 }, { "epoch": 3.5169913910285455, "grad_norm": 0.21300385903689706, "learning_rate": 1.9592290335513902e-05, "loss": 0.8638, "step": 7762 }, { "epoch": 3.5174444947893067, "grad_norm": 0.2233983861305638, "learning_rate": 1.958140843072841e-05, "loss": 0.8287, "step": 7763 }, { "epoch": 3.517897598550068, "grad_norm": 0.17650376983936422, "learning_rate": 1.957052856927784e-05, "loss": 0.8786, "step": 7764 }, { "epoch": 3.518350702310829, "grad_norm": 0.19627857101523355, "learning_rate": 1.955965075225099e-05, "loss": 0.852, "step": 7765 }, { "epoch": 3.5188038060715905, "grad_norm": 0.18623661281145937, "learning_rate": 1.9548774980736397e-05, "loss": 0.8593, "step": 7766 }, { "epoch": 3.5192569098323516, "grad_norm": 0.24658477685581554, "learning_rate": 1.9537901255822453e-05, "loss": 0.8432, "step": 7767 }, { "epoch": 3.5197100135931128, "grad_norm": 0.16800605558388343, "learning_rate": 1.9527029578597297e-05, "loss": 0.8576, "step": 7768 }, { "epoch": 3.520163117353874, "grad_norm": 0.22423316382615413, "learning_rate": 1.9516159950148904e-05, "loss": 0.8547, "step": 7769 }, { "epoch": 3.5206162211146355, "grad_norm": 0.2067820691497513, "learning_rate": 1.9505292371565008e-05, "loss": 0.8378, "step": 7770 }, { "epoch": 3.5210693248753966, "grad_norm": 0.19492888447223497, "learning_rate": 1.949442684393314e-05, "loss": 0.8597, "step": 7771 }, { "epoch": 3.5215224286361577, "grad_norm": 0.226605788838416, "learning_rate": 1.9483563368340666e-05, "loss": 0.8586, "step": 7772 }, { "epoch": 3.521975532396919, "grad_norm": 0.20641667396237046, "learning_rate": 1.947270194587469e-05, "loss": 0.8694, "step": 7773 }, { "epoch": 3.52242863615768, "grad_norm": 0.18310348072131277, "learning_rate": 1.9461842577622167e-05, "loss": 0.8509, "step": 7774 }, { "epoch": 3.522881739918441, "grad_norm": 0.17354678243888366, "learning_rate": 1.9450985264669793e-05, "loss": 0.861, "step": 7775 }, { "epoch": 3.5233348436792027, "grad_norm": 0.2797394750583422, "learning_rate": 1.944013000810411e-05, "loss": 0.8524, "step": 7776 }, { "epoch": 3.523787947439964, "grad_norm": 0.18300656457251388, "learning_rate": 1.94292768090114e-05, "loss": 0.8701, "step": 7777 }, { "epoch": 3.524241051200725, "grad_norm": 0.15027074029377835, "learning_rate": 1.9418425668477798e-05, "loss": 0.8539, "step": 7778 }, { "epoch": 3.524694154961486, "grad_norm": 0.19413441410240836, "learning_rate": 1.9407576587589183e-05, "loss": 0.8685, "step": 7779 }, { "epoch": 3.5251472587222477, "grad_norm": 0.1557007410242591, "learning_rate": 1.9396729567431237e-05, "loss": 0.8559, "step": 7780 }, { "epoch": 3.525600362483009, "grad_norm": 0.2042652857789066, "learning_rate": 1.9385884609089473e-05, "loss": 0.8433, "step": 7781 }, { "epoch": 3.52605346624377, "grad_norm": 0.1728979260014999, "learning_rate": 1.9375041713649147e-05, "loss": 0.8682, "step": 7782 }, { "epoch": 3.526506570004531, "grad_norm": 0.2072613404887748, "learning_rate": 1.9364200882195355e-05, "loss": 0.8565, "step": 7783 }, { "epoch": 3.526959673765292, "grad_norm": 0.1848554740552151, "learning_rate": 1.9353362115812934e-05, "loss": 0.8581, "step": 7784 }, { "epoch": 3.5274127775260533, "grad_norm": 0.1952391290581268, "learning_rate": 1.934252541558658e-05, "loss": 0.876, "step": 7785 }, { "epoch": 3.5278658812868144, "grad_norm": 0.1785773509440729, "learning_rate": 1.9331690782600722e-05, "loss": 0.8693, "step": 7786 }, { "epoch": 3.528318985047576, "grad_norm": 0.21536459756354676, "learning_rate": 1.93208582179396e-05, "loss": 0.836, "step": 7787 }, { "epoch": 3.528772088808337, "grad_norm": 0.19463347875681802, "learning_rate": 1.931002772268728e-05, "loss": 0.8542, "step": 7788 }, { "epoch": 3.5292251925690983, "grad_norm": 0.19877982039470465, "learning_rate": 1.9299199297927563e-05, "loss": 0.8503, "step": 7789 }, { "epoch": 3.5296782963298594, "grad_norm": 0.18808652468786574, "learning_rate": 1.928837294474411e-05, "loss": 0.856, "step": 7790 }, { "epoch": 3.530131400090621, "grad_norm": 0.1906006563691474, "learning_rate": 1.9277548664220297e-05, "loss": 0.8536, "step": 7791 }, { "epoch": 3.530584503851382, "grad_norm": 0.23210661744243685, "learning_rate": 1.9266726457439375e-05, "loss": 0.8721, "step": 7792 }, { "epoch": 3.5310376076121432, "grad_norm": 0.23079372819727786, "learning_rate": 1.9255906325484315e-05, "loss": 0.8473, "step": 7793 }, { "epoch": 3.5314907113729044, "grad_norm": 0.20182473870038622, "learning_rate": 1.924508826943794e-05, "loss": 0.8515, "step": 7794 }, { "epoch": 3.5319438151336655, "grad_norm": 0.19320388511894415, "learning_rate": 1.9234272290382827e-05, "loss": 0.8607, "step": 7795 }, { "epoch": 3.5323969188944266, "grad_norm": 0.17992451095139167, "learning_rate": 1.922345838940133e-05, "loss": 0.8639, "step": 7796 }, { "epoch": 3.5328500226551878, "grad_norm": 0.22603599474722877, "learning_rate": 1.9212646567575662e-05, "loss": 0.831, "step": 7797 }, { "epoch": 3.5333031264159493, "grad_norm": 0.20109344638701204, "learning_rate": 1.920183682598775e-05, "loss": 0.8541, "step": 7798 }, { "epoch": 3.5337562301767105, "grad_norm": 0.19377001446819594, "learning_rate": 1.9191029165719386e-05, "loss": 0.857, "step": 7799 }, { "epoch": 3.5342093339374716, "grad_norm": 0.16902106275066736, "learning_rate": 1.9180223587852076e-05, "loss": 0.8666, "step": 7800 }, { "epoch": 3.5346624376982327, "grad_norm": 0.20018873559193062, "learning_rate": 1.91694200934672e-05, "loss": 0.8496, "step": 7801 }, { "epoch": 3.5351155414589943, "grad_norm": 0.1385625564279342, "learning_rate": 1.915861868364587e-05, "loss": 0.843, "step": 7802 }, { "epoch": 3.5355686452197554, "grad_norm": 0.19738344036989797, "learning_rate": 1.9147819359468987e-05, "loss": 0.8312, "step": 7803 }, { "epoch": 3.5360217489805166, "grad_norm": 0.1912701032889425, "learning_rate": 1.9137022122017305e-05, "loss": 0.8421, "step": 7804 }, { "epoch": 3.5364748527412777, "grad_norm": 0.1753275670932422, "learning_rate": 1.912622697237129e-05, "loss": 0.8639, "step": 7805 }, { "epoch": 3.536927956502039, "grad_norm": 0.1720212378798853, "learning_rate": 1.9115433911611268e-05, "loss": 0.8305, "step": 7806 }, { "epoch": 3.5373810602628, "grad_norm": 0.16696240408268628, "learning_rate": 1.9104642940817294e-05, "loss": 0.8615, "step": 7807 }, { "epoch": 3.5378341640235615, "grad_norm": 0.1836939250549781, "learning_rate": 1.9093854061069278e-05, "loss": 0.8587, "step": 7808 }, { "epoch": 3.5382872677843227, "grad_norm": 0.16118606040535471, "learning_rate": 1.9083067273446857e-05, "loss": 0.8546, "step": 7809 }, { "epoch": 3.538740371545084, "grad_norm": 0.18199860169171617, "learning_rate": 1.9072282579029518e-05, "loss": 0.8591, "step": 7810 }, { "epoch": 3.539193475305845, "grad_norm": 0.17653738713021572, "learning_rate": 1.90614999788965e-05, "loss": 0.8699, "step": 7811 }, { "epoch": 3.5396465790666065, "grad_norm": 0.21980423259209458, "learning_rate": 1.9050719474126837e-05, "loss": 0.8536, "step": 7812 }, { "epoch": 3.5400996828273676, "grad_norm": 0.22546230378448123, "learning_rate": 1.9039941065799338e-05, "loss": 0.8684, "step": 7813 }, { "epoch": 3.5405527865881288, "grad_norm": 0.18128172255533093, "learning_rate": 1.902916475499267e-05, "loss": 0.8486, "step": 7814 }, { "epoch": 3.54100589034889, "grad_norm": 0.2098337346854023, "learning_rate": 1.901839054278521e-05, "loss": 0.8526, "step": 7815 }, { "epoch": 3.541458994109651, "grad_norm": 0.18033708257940018, "learning_rate": 1.900761843025516e-05, "loss": 0.8531, "step": 7816 }, { "epoch": 3.541912097870412, "grad_norm": 0.1926634803715893, "learning_rate": 1.8996848418480526e-05, "loss": 0.8575, "step": 7817 }, { "epoch": 3.5423652016311733, "grad_norm": 0.20008268275612542, "learning_rate": 1.898608050853906e-05, "loss": 0.8412, "step": 7818 }, { "epoch": 3.542818305391935, "grad_norm": 0.16867664051983433, "learning_rate": 1.8975314701508366e-05, "loss": 0.8556, "step": 7819 }, { "epoch": 3.543271409152696, "grad_norm": 0.17918863411776084, "learning_rate": 1.8964550998465766e-05, "loss": 0.8523, "step": 7820 }, { "epoch": 3.543724512913457, "grad_norm": 0.21390374939925294, "learning_rate": 1.8953789400488448e-05, "loss": 0.8608, "step": 7821 }, { "epoch": 3.5441776166742183, "grad_norm": 0.21817454821561416, "learning_rate": 1.8943029908653328e-05, "loss": 0.8594, "step": 7822 }, { "epoch": 3.54463072043498, "grad_norm": 0.20996748144085917, "learning_rate": 1.8932272524037114e-05, "loss": 0.8573, "step": 7823 }, { "epoch": 3.545083824195741, "grad_norm": 0.1981180108237735, "learning_rate": 1.8921517247716362e-05, "loss": 0.8541, "step": 7824 }, { "epoch": 3.545536927956502, "grad_norm": 0.2088490580395011, "learning_rate": 1.8910764080767343e-05, "loss": 0.8518, "step": 7825 }, { "epoch": 3.5459900317172632, "grad_norm": 0.19480828137102255, "learning_rate": 1.8900013024266173e-05, "loss": 0.8458, "step": 7826 }, { "epoch": 3.5464431354780244, "grad_norm": 0.2086059763024022, "learning_rate": 1.888926407928872e-05, "loss": 0.8532, "step": 7827 }, { "epoch": 3.5468962392387855, "grad_norm": 0.20631700014933063, "learning_rate": 1.8878517246910672e-05, "loss": 0.868, "step": 7828 }, { "epoch": 3.5473493429995466, "grad_norm": 0.17291993763338437, "learning_rate": 1.8867772528207465e-05, "loss": 0.8555, "step": 7829 }, { "epoch": 3.547802446760308, "grad_norm": 0.19108947980149737, "learning_rate": 1.8857029924254378e-05, "loss": 0.84, "step": 7830 }, { "epoch": 3.5482555505210693, "grad_norm": 0.14451287612713806, "learning_rate": 1.8846289436126432e-05, "loss": 0.8602, "step": 7831 }, { "epoch": 3.5487086542818305, "grad_norm": 0.1824227310139155, "learning_rate": 1.8835551064898434e-05, "loss": 0.8563, "step": 7832 }, { "epoch": 3.5491617580425916, "grad_norm": 0.16260051543358287, "learning_rate": 1.882481481164503e-05, "loss": 0.8642, "step": 7833 }, { "epoch": 3.549614861803353, "grad_norm": 0.22147919136220862, "learning_rate": 1.881408067744059e-05, "loss": 0.8473, "step": 7834 }, { "epoch": 3.5500679655641143, "grad_norm": 0.14560199602886778, "learning_rate": 1.8803348663359333e-05, "loss": 0.8386, "step": 7835 }, { "epoch": 3.5505210693248754, "grad_norm": 0.22964703313969217, "learning_rate": 1.8792618770475212e-05, "loss": 0.8466, "step": 7836 }, { "epoch": 3.5509741730856366, "grad_norm": 0.19220186477756582, "learning_rate": 1.878189099986201e-05, "loss": 0.8532, "step": 7837 }, { "epoch": 3.5514272768463977, "grad_norm": 0.19274157529774705, "learning_rate": 1.8771165352593265e-05, "loss": 0.8359, "step": 7838 }, { "epoch": 3.551880380607159, "grad_norm": 0.1727271262693973, "learning_rate": 1.876044182974233e-05, "loss": 0.8789, "step": 7839 }, { "epoch": 3.5523334843679204, "grad_norm": 0.18542662123466536, "learning_rate": 1.8749720432382333e-05, "loss": 0.8395, "step": 7840 }, { "epoch": 3.5527865881286815, "grad_norm": 0.18730106032661284, "learning_rate": 1.8739001161586162e-05, "loss": 0.8436, "step": 7841 }, { "epoch": 3.5532396918894427, "grad_norm": 0.19478754449643965, "learning_rate": 1.8728284018426556e-05, "loss": 0.8438, "step": 7842 }, { "epoch": 3.553692795650204, "grad_norm": 0.15668823196860177, "learning_rate": 1.8717569003975966e-05, "loss": 0.8621, "step": 7843 }, { "epoch": 3.5541458994109654, "grad_norm": 0.18999644745066, "learning_rate": 1.8706856119306707e-05, "loss": 0.8461, "step": 7844 }, { "epoch": 3.5545990031717265, "grad_norm": 0.15178411174281792, "learning_rate": 1.8696145365490802e-05, "loss": 0.8338, "step": 7845 }, { "epoch": 3.5550521069324876, "grad_norm": 0.18050710575965245, "learning_rate": 1.8685436743600146e-05, "loss": 0.8588, "step": 7846 }, { "epoch": 3.5555052106932488, "grad_norm": 0.16449103069306023, "learning_rate": 1.8674730254706342e-05, "loss": 0.8581, "step": 7847 }, { "epoch": 3.55595831445401, "grad_norm": 0.21172256341507992, "learning_rate": 1.866402589988081e-05, "loss": 0.8495, "step": 7848 }, { "epoch": 3.556411418214771, "grad_norm": 0.16240059972811122, "learning_rate": 1.8653323680194785e-05, "loss": 0.8636, "step": 7849 }, { "epoch": 3.556864521975532, "grad_norm": 0.21225776768491508, "learning_rate": 1.864262359671923e-05, "loss": 0.8607, "step": 7850 }, { "epoch": 3.5573176257362937, "grad_norm": 0.18693506354886705, "learning_rate": 1.863192565052496e-05, "loss": 0.8545, "step": 7851 }, { "epoch": 3.557770729497055, "grad_norm": 0.16557376079978037, "learning_rate": 1.8621229842682518e-05, "loss": 0.8408, "step": 7852 }, { "epoch": 3.558223833257816, "grad_norm": 0.1726207046149816, "learning_rate": 1.8610536174262276e-05, "loss": 0.8571, "step": 7853 }, { "epoch": 3.558676937018577, "grad_norm": 0.18670167800356938, "learning_rate": 1.8599844646334353e-05, "loss": 0.8629, "step": 7854 }, { "epoch": 3.5591300407793387, "grad_norm": 0.20205214037332653, "learning_rate": 1.8589155259968704e-05, "loss": 0.8512, "step": 7855 }, { "epoch": 3.5595831445401, "grad_norm": 0.1846288808843656, "learning_rate": 1.8578468016235014e-05, "loss": 0.8464, "step": 7856 }, { "epoch": 3.560036248300861, "grad_norm": 0.17208210554339093, "learning_rate": 1.8567782916202782e-05, "loss": 0.8732, "step": 7857 }, { "epoch": 3.560489352061622, "grad_norm": 0.18456639192784144, "learning_rate": 1.8557099960941304e-05, "loss": 0.8544, "step": 7858 }, { "epoch": 3.5609424558223832, "grad_norm": 0.1703133595406315, "learning_rate": 1.8546419151519633e-05, "loss": 0.8716, "step": 7859 }, { "epoch": 3.5613955595831444, "grad_norm": 0.20782418999812233, "learning_rate": 1.8535740489006637e-05, "loss": 0.8847, "step": 7860 }, { "epoch": 3.5618486633439055, "grad_norm": 0.1540736324587328, "learning_rate": 1.8525063974470937e-05, "loss": 0.8386, "step": 7861 }, { "epoch": 3.562301767104667, "grad_norm": 0.18389935087017859, "learning_rate": 1.8514389608980975e-05, "loss": 0.8553, "step": 7862 }, { "epoch": 3.562754870865428, "grad_norm": 0.16348382587077662, "learning_rate": 1.8503717393604955e-05, "loss": 0.8487, "step": 7863 }, { "epoch": 3.5632079746261893, "grad_norm": 0.1628140730573764, "learning_rate": 1.8493047329410848e-05, "loss": 0.8592, "step": 7864 }, { "epoch": 3.5636610783869505, "grad_norm": 0.17112898503310361, "learning_rate": 1.8482379417466463e-05, "loss": 0.8618, "step": 7865 }, { "epoch": 3.564114182147712, "grad_norm": 0.16029763477272319, "learning_rate": 1.847171365883933e-05, "loss": 0.8577, "step": 7866 }, { "epoch": 3.564567285908473, "grad_norm": 0.17220239817068342, "learning_rate": 1.846105005459683e-05, "loss": 0.8455, "step": 7867 }, { "epoch": 3.5650203896692343, "grad_norm": 0.13912149532149135, "learning_rate": 1.8450388605806074e-05, "loss": 0.8466, "step": 7868 }, { "epoch": 3.5654734934299954, "grad_norm": 0.1696147469373846, "learning_rate": 1.8439729313533984e-05, "loss": 0.8595, "step": 7869 }, { "epoch": 3.5659265971907566, "grad_norm": 0.16110923620161693, "learning_rate": 1.842907217884724e-05, "loss": 0.8476, "step": 7870 }, { "epoch": 3.5663797009515177, "grad_norm": 0.184623385811977, "learning_rate": 1.841841720281236e-05, "loss": 0.8654, "step": 7871 }, { "epoch": 3.5668328047122793, "grad_norm": 0.17706299561588631, "learning_rate": 1.8407764386495572e-05, "loss": 0.8671, "step": 7872 }, { "epoch": 3.5672859084730404, "grad_norm": 0.1715338654546571, "learning_rate": 1.8397113730962965e-05, "loss": 0.8659, "step": 7873 }, { "epoch": 3.5677390122338015, "grad_norm": 0.18270997779634007, "learning_rate": 1.8386465237280348e-05, "loss": 0.8644, "step": 7874 }, { "epoch": 3.5681921159945627, "grad_norm": 0.1847444548075782, "learning_rate": 1.8375818906513366e-05, "loss": 0.8428, "step": 7875 }, { "epoch": 3.5686452197553242, "grad_norm": 0.21219679036563996, "learning_rate": 1.8365174739727402e-05, "loss": 0.8665, "step": 7876 }, { "epoch": 3.5690983235160854, "grad_norm": 0.1535752696515413, "learning_rate": 1.8354532737987633e-05, "loss": 0.8496, "step": 7877 }, { "epoch": 3.5695514272768465, "grad_norm": 0.2054582475683959, "learning_rate": 1.8343892902359056e-05, "loss": 0.8654, "step": 7878 }, { "epoch": 3.5700045310376076, "grad_norm": 0.19983038007679188, "learning_rate": 1.8333255233906393e-05, "loss": 0.8433, "step": 7879 }, { "epoch": 3.5704576347983688, "grad_norm": 0.1968642048562203, "learning_rate": 1.832261973369421e-05, "loss": 0.8495, "step": 7880 }, { "epoch": 3.57091073855913, "grad_norm": 0.17490533572254435, "learning_rate": 1.83119864027868e-05, "loss": 0.8338, "step": 7881 }, { "epoch": 3.571363842319891, "grad_norm": 0.15464331342666116, "learning_rate": 1.830135524224829e-05, "loss": 0.8577, "step": 7882 }, { "epoch": 3.5718169460806526, "grad_norm": 0.1712288553052261, "learning_rate": 1.8290726253142547e-05, "loss": 0.844, "step": 7883 }, { "epoch": 3.5722700498414137, "grad_norm": 0.15099984686251094, "learning_rate": 1.828009943653323e-05, "loss": 0.8449, "step": 7884 }, { "epoch": 3.572723153602175, "grad_norm": 0.19601687944588306, "learning_rate": 1.8269474793483816e-05, "loss": 0.8603, "step": 7885 }, { "epoch": 3.573176257362936, "grad_norm": 0.1683671145646774, "learning_rate": 1.8258852325057504e-05, "loss": 0.8581, "step": 7886 }, { "epoch": 3.5736293611236976, "grad_norm": 0.19581924074520096, "learning_rate": 1.8248232032317346e-05, "loss": 0.8622, "step": 7887 }, { "epoch": 3.5740824648844587, "grad_norm": 0.204559717891536, "learning_rate": 1.82376139163261e-05, "loss": 0.8562, "step": 7888 }, { "epoch": 3.57453556864522, "grad_norm": 0.18566715116446003, "learning_rate": 1.822699797814638e-05, "loss": 0.8615, "step": 7889 }, { "epoch": 3.574988672405981, "grad_norm": 0.16860936328859855, "learning_rate": 1.821638421884051e-05, "loss": 0.8648, "step": 7890 }, { "epoch": 3.575441776166742, "grad_norm": 0.17985314857957865, "learning_rate": 1.8205772639470676e-05, "loss": 0.86, "step": 7891 }, { "epoch": 3.575894879927503, "grad_norm": 0.20837793891525722, "learning_rate": 1.8195163241098777e-05, "loss": 0.8913, "step": 7892 }, { "epoch": 3.5763479836882643, "grad_norm": 0.24268488386779896, "learning_rate": 1.8184556024786513e-05, "loss": 0.8393, "step": 7893 }, { "epoch": 3.576801087449026, "grad_norm": 0.1794389432612748, "learning_rate": 1.81739509915954e-05, "loss": 0.864, "step": 7894 }, { "epoch": 3.577254191209787, "grad_norm": 0.2086141761409848, "learning_rate": 1.8163348142586667e-05, "loss": 0.848, "step": 7895 }, { "epoch": 3.577707294970548, "grad_norm": 0.16851535070077353, "learning_rate": 1.815274747882141e-05, "loss": 0.8513, "step": 7896 }, { "epoch": 3.5781603987313093, "grad_norm": 0.20613792627987268, "learning_rate": 1.8142149001360424e-05, "loss": 0.8632, "step": 7897 }, { "epoch": 3.578613502492071, "grad_norm": 0.1492118658597635, "learning_rate": 1.8131552711264357e-05, "loss": 0.856, "step": 7898 }, { "epoch": 3.579066606252832, "grad_norm": 0.1968852872370156, "learning_rate": 1.812095860959357e-05, "loss": 0.8565, "step": 7899 }, { "epoch": 3.579519710013593, "grad_norm": 0.1639528638979256, "learning_rate": 1.811036669740827e-05, "loss": 0.8512, "step": 7900 }, { "epoch": 3.5799728137743543, "grad_norm": 0.15218063071056634, "learning_rate": 1.8099776975768396e-05, "loss": 0.845, "step": 7901 }, { "epoch": 3.5804259175351154, "grad_norm": 0.16846019683898616, "learning_rate": 1.8089189445733676e-05, "loss": 0.8547, "step": 7902 }, { "epoch": 3.5808790212958765, "grad_norm": 0.17288165843083242, "learning_rate": 1.8078604108363657e-05, "loss": 0.8546, "step": 7903 }, { "epoch": 3.581332125056638, "grad_norm": 0.15825035016315653, "learning_rate": 1.8068020964717603e-05, "loss": 0.8486, "step": 7904 }, { "epoch": 3.5817852288173992, "grad_norm": 0.1961942606120653, "learning_rate": 1.8057440015854628e-05, "loss": 0.8456, "step": 7905 }, { "epoch": 3.5822383325781604, "grad_norm": 0.18280288750520354, "learning_rate": 1.8046861262833563e-05, "loss": 0.8734, "step": 7906 }, { "epoch": 3.5826914363389215, "grad_norm": 0.194783207188888, "learning_rate": 1.803628470671307e-05, "loss": 0.8511, "step": 7907 }, { "epoch": 3.583144540099683, "grad_norm": 0.19625577617744627, "learning_rate": 1.802571034855156e-05, "loss": 0.8531, "step": 7908 }, { "epoch": 3.583597643860444, "grad_norm": 0.1717847561111245, "learning_rate": 1.8015138189407223e-05, "loss": 0.8646, "step": 7909 }, { "epoch": 3.5840507476212053, "grad_norm": 0.16686293478642697, "learning_rate": 1.8004568230338063e-05, "loss": 0.8635, "step": 7910 }, { "epoch": 3.5845038513819665, "grad_norm": 0.18402736494005686, "learning_rate": 1.7994000472401802e-05, "loss": 0.8399, "step": 7911 }, { "epoch": 3.5849569551427276, "grad_norm": 0.18544893269727653, "learning_rate": 1.7983434916656028e-05, "loss": 0.8659, "step": 7912 }, { "epoch": 3.5854100589034887, "grad_norm": 0.1718732631801253, "learning_rate": 1.7972871564158013e-05, "loss": 0.8508, "step": 7913 }, { "epoch": 3.58586316266425, "grad_norm": 0.19165170367792966, "learning_rate": 1.7962310415964893e-05, "loss": 0.8772, "step": 7914 }, { "epoch": 3.5863162664250114, "grad_norm": 0.1623824843559078, "learning_rate": 1.7951751473133518e-05, "loss": 0.8768, "step": 7915 }, { "epoch": 3.5867693701857726, "grad_norm": 0.17016301703173667, "learning_rate": 1.7941194736720574e-05, "loss": 0.8903, "step": 7916 }, { "epoch": 3.5872224739465337, "grad_norm": 0.175298497293858, "learning_rate": 1.7930640207782478e-05, "loss": 0.8583, "step": 7917 }, { "epoch": 3.587675577707295, "grad_norm": 0.1893721697437588, "learning_rate": 1.7920087887375445e-05, "loss": 0.8306, "step": 7918 }, { "epoch": 3.5881286814680564, "grad_norm": 0.16900412475947849, "learning_rate": 1.7909537776555476e-05, "loss": 0.8599, "step": 7919 }, { "epoch": 3.5885817852288175, "grad_norm": 0.23247459096867476, "learning_rate": 1.789898987637834e-05, "loss": 0.8478, "step": 7920 }, { "epoch": 3.5890348889895787, "grad_norm": 0.18242242154914767, "learning_rate": 1.78884441878996e-05, "loss": 0.8603, "step": 7921 }, { "epoch": 3.58948799275034, "grad_norm": 0.1916086395481729, "learning_rate": 1.787790071217457e-05, "loss": 0.8599, "step": 7922 }, { "epoch": 3.589941096511101, "grad_norm": 0.17474108114002454, "learning_rate": 1.7867359450258373e-05, "loss": 0.8551, "step": 7923 }, { "epoch": 3.590394200271862, "grad_norm": 0.20249167975437282, "learning_rate": 1.7856820403205907e-05, "loss": 0.8394, "step": 7924 }, { "epoch": 3.590847304032623, "grad_norm": 0.19935073309473372, "learning_rate": 1.7846283572071815e-05, "loss": 0.8462, "step": 7925 }, { "epoch": 3.5913004077933848, "grad_norm": 0.17355375989401953, "learning_rate": 1.783574895791054e-05, "loss": 0.8694, "step": 7926 }, { "epoch": 3.591753511554146, "grad_norm": 0.20700359010057803, "learning_rate": 1.7825216561776334e-05, "loss": 0.8664, "step": 7927 }, { "epoch": 3.592206615314907, "grad_norm": 0.172500482832728, "learning_rate": 1.7814686384723173e-05, "loss": 0.8597, "step": 7928 }, { "epoch": 3.592659719075668, "grad_norm": 0.19499206949677325, "learning_rate": 1.7804158427804835e-05, "loss": 0.8618, "step": 7929 }, { "epoch": 3.5931128228364297, "grad_norm": 0.1508355302465444, "learning_rate": 1.7793632692074893e-05, "loss": 0.8311, "step": 7930 }, { "epoch": 3.593565926597191, "grad_norm": 0.2035601506768486, "learning_rate": 1.778310917858666e-05, "loss": 0.8556, "step": 7931 }, { "epoch": 3.594019030357952, "grad_norm": 0.1905515781339536, "learning_rate": 1.7772587888393276e-05, "loss": 0.8402, "step": 7932 }, { "epoch": 3.594472134118713, "grad_norm": 0.18767344543929396, "learning_rate": 1.7762068822547603e-05, "loss": 0.8533, "step": 7933 }, { "epoch": 3.5949252378794743, "grad_norm": 0.20435238429757932, "learning_rate": 1.775155198210233e-05, "loss": 0.8581, "step": 7934 }, { "epoch": 3.5953783416402354, "grad_norm": 0.19840729239651564, "learning_rate": 1.774103736810988e-05, "loss": 0.8551, "step": 7935 }, { "epoch": 3.595831445400997, "grad_norm": 0.18695261644369146, "learning_rate": 1.7730524981622496e-05, "loss": 0.8545, "step": 7936 }, { "epoch": 3.596284549161758, "grad_norm": 0.23469139833635425, "learning_rate": 1.772001482369217e-05, "loss": 0.8572, "step": 7937 }, { "epoch": 3.5967376529225192, "grad_norm": 0.18724796674187388, "learning_rate": 1.7709506895370658e-05, "loss": 0.8532, "step": 7938 }, { "epoch": 3.5971907566832804, "grad_norm": 0.15999157033264325, "learning_rate": 1.7699001197709545e-05, "loss": 0.8552, "step": 7939 }, { "epoch": 3.597643860444042, "grad_norm": 0.15934760443358764, "learning_rate": 1.7688497731760122e-05, "loss": 0.8577, "step": 7940 }, { "epoch": 3.598096964204803, "grad_norm": 0.16321266520144873, "learning_rate": 1.7677996498573535e-05, "loss": 0.8386, "step": 7941 }, { "epoch": 3.598550067965564, "grad_norm": 0.18608310509855477, "learning_rate": 1.766749749920063e-05, "loss": 0.8505, "step": 7942 }, { "epoch": 3.5990031717263253, "grad_norm": 0.18855124756334057, "learning_rate": 1.76570007346921e-05, "loss": 0.8498, "step": 7943 }, { "epoch": 3.5994562754870865, "grad_norm": 0.1797406519572638, "learning_rate": 1.7646506206098357e-05, "loss": 0.8481, "step": 7944 }, { "epoch": 3.5999093792478476, "grad_norm": 0.1533168691518076, "learning_rate": 1.7636013914469608e-05, "loss": 0.8543, "step": 7945 }, { "epoch": 3.6003624830086087, "grad_norm": 0.20645783677540017, "learning_rate": 1.7625523860855863e-05, "loss": 0.8594, "step": 7946 }, { "epoch": 3.6008155867693703, "grad_norm": 0.20725316556782786, "learning_rate": 1.7615036046306856e-05, "loss": 0.8475, "step": 7947 }, { "epoch": 3.6012686905301314, "grad_norm": 0.22638400646803913, "learning_rate": 1.7604550471872162e-05, "loss": 0.8543, "step": 7948 }, { "epoch": 3.6017217942908926, "grad_norm": 0.20489839479981548, "learning_rate": 1.759406713860106e-05, "loss": 0.8631, "step": 7949 }, { "epoch": 3.6021748980516537, "grad_norm": 0.1913455287984867, "learning_rate": 1.758358604754268e-05, "loss": 0.8715, "step": 7950 }, { "epoch": 3.6026280018124153, "grad_norm": 0.24827280395498175, "learning_rate": 1.7573107199745848e-05, "loss": 0.8502, "step": 7951 }, { "epoch": 3.6030811055731764, "grad_norm": 0.21767432896908287, "learning_rate": 1.756263059625924e-05, "loss": 0.8241, "step": 7952 }, { "epoch": 3.6035342093339375, "grad_norm": 0.2137417648585404, "learning_rate": 1.755215623813126e-05, "loss": 0.8658, "step": 7953 }, { "epoch": 3.6039873130946987, "grad_norm": 0.22563067976344636, "learning_rate": 1.7541684126410087e-05, "loss": 0.8609, "step": 7954 }, { "epoch": 3.60444041685546, "grad_norm": 0.16765404639406226, "learning_rate": 1.7531214262143718e-05, "loss": 0.8783, "step": 7955 }, { "epoch": 3.604893520616221, "grad_norm": 0.21945277181155892, "learning_rate": 1.752074664637986e-05, "loss": 0.8707, "step": 7956 }, { "epoch": 3.605346624376982, "grad_norm": 0.1536050727113246, "learning_rate": 1.751028128016607e-05, "loss": 0.8739, "step": 7957 }, { "epoch": 3.6057997281377436, "grad_norm": 0.18329339334134812, "learning_rate": 1.749981816454961e-05, "loss": 0.8553, "step": 7958 }, { "epoch": 3.6062528318985048, "grad_norm": 0.16538951788453057, "learning_rate": 1.748935730057757e-05, "loss": 0.8417, "step": 7959 }, { "epoch": 3.606705935659266, "grad_norm": 0.21274070581772117, "learning_rate": 1.7478898689296772e-05, "loss": 0.8576, "step": 7960 }, { "epoch": 3.607159039420027, "grad_norm": 0.21056672940951815, "learning_rate": 1.7468442331753852e-05, "loss": 0.8577, "step": 7961 }, { "epoch": 3.6076121431807886, "grad_norm": 0.22720099625715826, "learning_rate": 1.74579882289952e-05, "loss": 0.872, "step": 7962 }, { "epoch": 3.6080652469415497, "grad_norm": 0.24704340514640072, "learning_rate": 1.7447536382066957e-05, "loss": 0.8711, "step": 7963 }, { "epoch": 3.608518350702311, "grad_norm": 0.21501529837784628, "learning_rate": 1.7437086792015096e-05, "loss": 0.857, "step": 7964 }, { "epoch": 3.608971454463072, "grad_norm": 0.19361966055465765, "learning_rate": 1.74266394598853e-05, "loss": 0.8628, "step": 7965 }, { "epoch": 3.609424558223833, "grad_norm": 0.21048135792174508, "learning_rate": 1.741619438672309e-05, "loss": 0.8811, "step": 7966 }, { "epoch": 3.6098776619845943, "grad_norm": 0.2304033340142909, "learning_rate": 1.74057515735737e-05, "loss": 0.8539, "step": 7967 }, { "epoch": 3.610330765745356, "grad_norm": 0.151801219283284, "learning_rate": 1.739531102148219e-05, "loss": 0.8565, "step": 7968 }, { "epoch": 3.610783869506117, "grad_norm": 0.2269925239653014, "learning_rate": 1.7384872731493364e-05, "loss": 0.8747, "step": 7969 }, { "epoch": 3.611236973266878, "grad_norm": 0.1792382550954444, "learning_rate": 1.737443670465178e-05, "loss": 0.8587, "step": 7970 }, { "epoch": 3.6116900770276392, "grad_norm": 0.2192996063118868, "learning_rate": 1.7364002942001828e-05, "loss": 0.8422, "step": 7971 }, { "epoch": 3.612143180788401, "grad_norm": 0.18971962530278275, "learning_rate": 1.735357144458761e-05, "loss": 0.8509, "step": 7972 }, { "epoch": 3.612596284549162, "grad_norm": 0.21274713582953883, "learning_rate": 1.734314221345306e-05, "loss": 0.8707, "step": 7973 }, { "epoch": 3.613049388309923, "grad_norm": 0.18805714240406232, "learning_rate": 1.7332715249641828e-05, "loss": 0.8472, "step": 7974 }, { "epoch": 3.613502492070684, "grad_norm": 0.1740843693252281, "learning_rate": 1.7322290554197385e-05, "loss": 0.8491, "step": 7975 }, { "epoch": 3.6139555958314453, "grad_norm": 0.20837791777644787, "learning_rate": 1.731186812816293e-05, "loss": 0.8651, "step": 7976 }, { "epoch": 3.6144086995922065, "grad_norm": 0.20453717111153816, "learning_rate": 1.730144797258149e-05, "loss": 0.8727, "step": 7977 }, { "epoch": 3.6148618033529676, "grad_norm": 0.1964500241987209, "learning_rate": 1.7291030088495813e-05, "loss": 0.8706, "step": 7978 }, { "epoch": 3.615314907113729, "grad_norm": 0.16292463837111876, "learning_rate": 1.7280614476948434e-05, "loss": 0.8595, "step": 7979 }, { "epoch": 3.6157680108744903, "grad_norm": 0.17057916263837924, "learning_rate": 1.7270201138981693e-05, "loss": 0.8448, "step": 7980 }, { "epoch": 3.6162211146352514, "grad_norm": 0.1481042802437913, "learning_rate": 1.725979007563764e-05, "loss": 0.8515, "step": 7981 }, { "epoch": 3.6166742183960126, "grad_norm": 0.1555559124591059, "learning_rate": 1.7249381287958187e-05, "loss": 0.8477, "step": 7982 }, { "epoch": 3.617127322156774, "grad_norm": 0.17266567774381925, "learning_rate": 1.7238974776984906e-05, "loss": 0.8665, "step": 7983 }, { "epoch": 3.6175804259175353, "grad_norm": 0.16188558076514473, "learning_rate": 1.722857054375924e-05, "loss": 0.8655, "step": 7984 }, { "epoch": 3.6180335296782964, "grad_norm": 0.17127875836682566, "learning_rate": 1.7218168589322336e-05, "loss": 0.8505, "step": 7985 }, { "epoch": 3.6184866334390575, "grad_norm": 0.19158129996333853, "learning_rate": 1.7207768914715166e-05, "loss": 0.8618, "step": 7986 }, { "epoch": 3.6189397371998187, "grad_norm": 0.15789906679708637, "learning_rate": 1.719737152097843e-05, "loss": 0.8506, "step": 7987 }, { "epoch": 3.61939284096058, "grad_norm": 0.19423456141201, "learning_rate": 1.7186976409152642e-05, "loss": 0.8697, "step": 7988 }, { "epoch": 3.619845944721341, "grad_norm": 0.1735826383285651, "learning_rate": 1.717658358027805e-05, "loss": 0.872, "step": 7989 }, { "epoch": 3.6202990484821025, "grad_norm": 0.1965062870649474, "learning_rate": 1.7166193035394677e-05, "loss": 0.847, "step": 7990 }, { "epoch": 3.6207521522428636, "grad_norm": 0.18145983922438558, "learning_rate": 1.715580477554235e-05, "loss": 0.8561, "step": 7991 }, { "epoch": 3.6212052560036248, "grad_norm": 0.183598694145561, "learning_rate": 1.7145418801760625e-05, "loss": 0.8676, "step": 7992 }, { "epoch": 3.621658359764386, "grad_norm": 0.16624826936808018, "learning_rate": 1.7135035115088877e-05, "loss": 0.8462, "step": 7993 }, { "epoch": 3.6221114635251475, "grad_norm": 0.16944482124085258, "learning_rate": 1.7124653716566187e-05, "loss": 0.8527, "step": 7994 }, { "epoch": 3.6225645672859086, "grad_norm": 0.1940602647683241, "learning_rate": 1.711427460723149e-05, "loss": 0.8382, "step": 7995 }, { "epoch": 3.6230176710466697, "grad_norm": 0.17865491320438467, "learning_rate": 1.7103897788123407e-05, "loss": 0.845, "step": 7996 }, { "epoch": 3.623470774807431, "grad_norm": 0.2132037594318177, "learning_rate": 1.7093523260280402e-05, "loss": 0.8838, "step": 7997 }, { "epoch": 3.623923878568192, "grad_norm": 0.16791980668931716, "learning_rate": 1.7083151024740665e-05, "loss": 0.8461, "step": 7998 }, { "epoch": 3.624376982328953, "grad_norm": 0.1698564392792099, "learning_rate": 1.7072781082542153e-05, "loss": 0.8489, "step": 7999 }, { "epoch": 3.6248300860897147, "grad_norm": 0.17357481310490128, "learning_rate": 1.7062413434722635e-05, "loss": 0.8609, "step": 8000 }, { "epoch": 3.625283189850476, "grad_norm": 0.1768487843822683, "learning_rate": 1.7052048082319602e-05, "loss": 0.8277, "step": 8001 }, { "epoch": 3.625736293611237, "grad_norm": 0.1677908559404881, "learning_rate": 1.7041685026370367e-05, "loss": 0.8425, "step": 8002 }, { "epoch": 3.626189397371998, "grad_norm": 0.20962070306389619, "learning_rate": 1.703132426791195e-05, "loss": 0.8382, "step": 8003 }, { "epoch": 3.6266425011327597, "grad_norm": 0.18321739805578405, "learning_rate": 1.7020965807981208e-05, "loss": 0.8517, "step": 8004 }, { "epoch": 3.627095604893521, "grad_norm": 0.19679799373197285, "learning_rate": 1.7010609647614728e-05, "loss": 0.8652, "step": 8005 }, { "epoch": 3.627548708654282, "grad_norm": 0.19665600254265858, "learning_rate": 1.700025578784885e-05, "loss": 0.8681, "step": 8006 }, { "epoch": 3.628001812415043, "grad_norm": 0.18659015167796414, "learning_rate": 1.6989904229719732e-05, "loss": 0.8469, "step": 8007 }, { "epoch": 3.628454916175804, "grad_norm": 0.1750627310128107, "learning_rate": 1.697955497426327e-05, "loss": 0.857, "step": 8008 }, { "epoch": 3.6289080199365653, "grad_norm": 0.19005239451622266, "learning_rate": 1.6969208022515147e-05, "loss": 0.857, "step": 8009 }, { "epoch": 3.6293611236973264, "grad_norm": 0.15479874506779995, "learning_rate": 1.695886337551079e-05, "loss": 0.8464, "step": 8010 }, { "epoch": 3.629814227458088, "grad_norm": 0.19310030322953034, "learning_rate": 1.6948521034285433e-05, "loss": 0.8644, "step": 8011 }, { "epoch": 3.630267331218849, "grad_norm": 0.1653192241796836, "learning_rate": 1.6938180999874027e-05, "loss": 0.8546, "step": 8012 }, { "epoch": 3.6307204349796103, "grad_norm": 0.19145593309036635, "learning_rate": 1.6927843273311357e-05, "loss": 0.8648, "step": 8013 }, { "epoch": 3.6311735387403714, "grad_norm": 0.16914518281919844, "learning_rate": 1.691750785563193e-05, "loss": 0.8632, "step": 8014 }, { "epoch": 3.631626642501133, "grad_norm": 0.1997107172105663, "learning_rate": 1.6907174747870013e-05, "loss": 0.855, "step": 8015 }, { "epoch": 3.632079746261894, "grad_norm": 0.17806549040654063, "learning_rate": 1.6896843951059698e-05, "loss": 0.8529, "step": 8016 }, { "epoch": 3.6325328500226552, "grad_norm": 0.1525489465558621, "learning_rate": 1.6886515466234777e-05, "loss": 0.8406, "step": 8017 }, { "epoch": 3.6329859537834164, "grad_norm": 0.17822406118094497, "learning_rate": 1.6876189294428882e-05, "loss": 0.8514, "step": 8018 }, { "epoch": 3.6334390575441775, "grad_norm": 0.15272065244196342, "learning_rate": 1.6865865436675345e-05, "loss": 0.8635, "step": 8019 }, { "epoch": 3.6338921613049386, "grad_norm": 0.1790533018115583, "learning_rate": 1.685554389400732e-05, "loss": 0.8408, "step": 8020 }, { "epoch": 3.6343452650656998, "grad_norm": 0.17387619883870536, "learning_rate": 1.6845224667457702e-05, "loss": 0.8554, "step": 8021 }, { "epoch": 3.6347983688264613, "grad_norm": 0.18633517575455882, "learning_rate": 1.683490775805914e-05, "loss": 0.8595, "step": 8022 }, { "epoch": 3.6352514725872225, "grad_norm": 0.18706393044734798, "learning_rate": 1.6824593166844097e-05, "loss": 0.8561, "step": 8023 }, { "epoch": 3.6357045763479836, "grad_norm": 0.1742645342120996, "learning_rate": 1.6814280894844753e-05, "loss": 0.8414, "step": 8024 }, { "epoch": 3.6361576801087447, "grad_norm": 0.16887666198615628, "learning_rate": 1.6803970943093112e-05, "loss": 0.8466, "step": 8025 }, { "epoch": 3.6366107838695063, "grad_norm": 0.3100516010796533, "learning_rate": 1.679366331262088e-05, "loss": 0.8943, "step": 8026 }, { "epoch": 3.6370638876302674, "grad_norm": 0.17333784350241718, "learning_rate": 1.67833580044596e-05, "loss": 0.8578, "step": 8027 }, { "epoch": 3.6375169913910286, "grad_norm": 0.17531441280733057, "learning_rate": 1.677305501964051e-05, "loss": 0.8324, "step": 8028 }, { "epoch": 3.6379700951517897, "grad_norm": 0.16561795132080076, "learning_rate": 1.6762754359194685e-05, "loss": 0.8717, "step": 8029 }, { "epoch": 3.638423198912551, "grad_norm": 0.20204812332479558, "learning_rate": 1.6752456024152927e-05, "loss": 0.8441, "step": 8030 }, { "epoch": 3.638876302673312, "grad_norm": 0.17139999391801095, "learning_rate": 1.6742160015545797e-05, "loss": 0.8504, "step": 8031 }, { "epoch": 3.6393294064340735, "grad_norm": 0.6908065895121862, "learning_rate": 1.6731866334403663e-05, "loss": 0.8733, "step": 8032 }, { "epoch": 3.6397825101948347, "grad_norm": 0.1778430954065627, "learning_rate": 1.6721574981756615e-05, "loss": 0.823, "step": 8033 }, { "epoch": 3.640235613955596, "grad_norm": 0.17191314390332937, "learning_rate": 1.671128595863456e-05, "loss": 0.8704, "step": 8034 }, { "epoch": 3.640688717716357, "grad_norm": 0.16603314360284183, "learning_rate": 1.670099926606711e-05, "loss": 0.8539, "step": 8035 }, { "epoch": 3.6411418214771185, "grad_norm": 0.19003775012657603, "learning_rate": 1.669071490508371e-05, "loss": 0.843, "step": 8036 }, { "epoch": 3.6415949252378796, "grad_norm": 0.2220886551362204, "learning_rate": 1.668043287671351e-05, "loss": 0.8651, "step": 8037 }, { "epoch": 3.6420480289986408, "grad_norm": 0.18775604734345128, "learning_rate": 1.6670153181985486e-05, "loss": 0.8385, "step": 8038 }, { "epoch": 3.642501132759402, "grad_norm": 0.19385575367735652, "learning_rate": 1.665987582192834e-05, "loss": 0.849, "step": 8039 }, { "epoch": 3.642954236520163, "grad_norm": 0.1782795746215909, "learning_rate": 1.6649600797570544e-05, "loss": 0.8596, "step": 8040 }, { "epoch": 3.643407340280924, "grad_norm": 0.15175891006936118, "learning_rate": 1.663932810994034e-05, "loss": 0.8502, "step": 8041 }, { "epoch": 3.6438604440416853, "grad_norm": 0.17881120733040562, "learning_rate": 1.6629057760065736e-05, "loss": 0.8598, "step": 8042 }, { "epoch": 3.644313547802447, "grad_norm": 0.16840760903835647, "learning_rate": 1.6618789748974532e-05, "loss": 0.8663, "step": 8043 }, { "epoch": 3.644766651563208, "grad_norm": 0.19807508746136035, "learning_rate": 1.6608524077694244e-05, "loss": 0.8713, "step": 8044 }, { "epoch": 3.645219755323969, "grad_norm": 0.15619045925329758, "learning_rate": 1.6598260747252206e-05, "loss": 0.8757, "step": 8045 }, { "epoch": 3.6456728590847303, "grad_norm": 0.18742059956057544, "learning_rate": 1.6587999758675466e-05, "loss": 0.8626, "step": 8046 }, { "epoch": 3.646125962845492, "grad_norm": 0.1526489957095354, "learning_rate": 1.6577741112990896e-05, "loss": 0.8759, "step": 8047 }, { "epoch": 3.646579066606253, "grad_norm": 0.1687705202432552, "learning_rate": 1.6567484811225066e-05, "loss": 0.8435, "step": 8048 }, { "epoch": 3.647032170367014, "grad_norm": 0.17547169922641834, "learning_rate": 1.655723085440439e-05, "loss": 0.852, "step": 8049 }, { "epoch": 3.6474852741277752, "grad_norm": 0.1599804010966413, "learning_rate": 1.6546979243554978e-05, "loss": 0.8637, "step": 8050 }, { "epoch": 3.6479383778885364, "grad_norm": 0.1939087906936247, "learning_rate": 1.6536729979702718e-05, "loss": 0.8574, "step": 8051 }, { "epoch": 3.6483914816492975, "grad_norm": 0.18221843187345446, "learning_rate": 1.6526483063873313e-05, "loss": 0.8506, "step": 8052 }, { "epoch": 3.6488445854100586, "grad_norm": 0.18375284577697246, "learning_rate": 1.6516238497092165e-05, "loss": 0.832, "step": 8053 }, { "epoch": 3.64929768917082, "grad_norm": 0.18435184788820247, "learning_rate": 1.650599628038449e-05, "loss": 0.8672, "step": 8054 }, { "epoch": 3.6497507929315813, "grad_norm": 0.21000454444585143, "learning_rate": 1.649575641477523e-05, "loss": 0.8626, "step": 8055 }, { "epoch": 3.6502038966923425, "grad_norm": 0.20730434455114868, "learning_rate": 1.6485518901289147e-05, "loss": 0.8566, "step": 8056 }, { "epoch": 3.6506570004531036, "grad_norm": 0.1737622466201369, "learning_rate": 1.647528374095069e-05, "loss": 0.8506, "step": 8057 }, { "epoch": 3.651110104213865, "grad_norm": 0.2289125436134594, "learning_rate": 1.6465050934784154e-05, "loss": 0.85, "step": 8058 }, { "epoch": 3.6515632079746263, "grad_norm": 0.21214256277004218, "learning_rate": 1.6454820483813536e-05, "loss": 0.8701, "step": 8059 }, { "epoch": 3.6520163117353874, "grad_norm": 0.159171076362068, "learning_rate": 1.6444592389062616e-05, "loss": 0.8351, "step": 8060 }, { "epoch": 3.6524694154961486, "grad_norm": 0.18617913732962152, "learning_rate": 1.643436665155496e-05, "loss": 0.8401, "step": 8061 }, { "epoch": 3.6529225192569097, "grad_norm": 0.19636940790970986, "learning_rate": 1.6424143272313857e-05, "loss": 0.8418, "step": 8062 }, { "epoch": 3.653375623017671, "grad_norm": 0.19264241312437555, "learning_rate": 1.6413922252362417e-05, "loss": 0.8687, "step": 8063 }, { "epoch": 3.6538287267784324, "grad_norm": 0.21140333154659502, "learning_rate": 1.6403703592723446e-05, "loss": 0.8478, "step": 8064 }, { "epoch": 3.6542818305391935, "grad_norm": 0.20818504726877135, "learning_rate": 1.6393487294419582e-05, "loss": 0.8489, "step": 8065 }, { "epoch": 3.6547349342999547, "grad_norm": 0.13416664890483934, "learning_rate": 1.6383273358473175e-05, "loss": 0.8514, "step": 8066 }, { "epoch": 3.655188038060716, "grad_norm": 0.22893538215930204, "learning_rate": 1.6373061785906344e-05, "loss": 0.8532, "step": 8067 }, { "epoch": 3.6556411418214774, "grad_norm": 0.20203991708760394, "learning_rate": 1.6362852577741018e-05, "loss": 0.8443, "step": 8068 }, { "epoch": 3.6560942455822385, "grad_norm": 0.16963422422735366, "learning_rate": 1.6352645734998816e-05, "loss": 0.8676, "step": 8069 }, { "epoch": 3.6565473493429996, "grad_norm": 0.20692151988360838, "learning_rate": 1.6342441258701196e-05, "loss": 0.8526, "step": 8070 }, { "epoch": 3.6570004531037608, "grad_norm": 0.19063035491961297, "learning_rate": 1.6332239149869313e-05, "loss": 0.8574, "step": 8071 }, { "epoch": 3.657453556864522, "grad_norm": 0.18780686503587765, "learning_rate": 1.6322039409524147e-05, "loss": 0.8743, "step": 8072 }, { "epoch": 3.657906660625283, "grad_norm": 0.17261453498773485, "learning_rate": 1.6311842038686373e-05, "loss": 0.8602, "step": 8073 }, { "epoch": 3.658359764386044, "grad_norm": 0.21410720751688428, "learning_rate": 1.6301647038376503e-05, "loss": 0.8612, "step": 8074 }, { "epoch": 3.6588128681468057, "grad_norm": 0.16408513301680958, "learning_rate": 1.6291454409614754e-05, "loss": 0.8532, "step": 8075 }, { "epoch": 3.659265971907567, "grad_norm": 0.16481115890256887, "learning_rate": 1.6281264153421114e-05, "loss": 0.8515, "step": 8076 }, { "epoch": 3.659719075668328, "grad_norm": 0.2167298340610933, "learning_rate": 1.6271076270815368e-05, "loss": 0.8739, "step": 8077 }, { "epoch": 3.660172179429089, "grad_norm": 0.16707495133084824, "learning_rate": 1.626089076281702e-05, "loss": 0.8549, "step": 8078 }, { "epoch": 3.6606252831898507, "grad_norm": 0.16633607807749454, "learning_rate": 1.625070763044538e-05, "loss": 0.8521, "step": 8079 }, { "epoch": 3.661078386950612, "grad_norm": 0.1869711780413498, "learning_rate": 1.624052687471947e-05, "loss": 0.8452, "step": 8080 }, { "epoch": 3.661531490711373, "grad_norm": 0.17632963716970435, "learning_rate": 1.6230348496658125e-05, "loss": 0.8832, "step": 8081 }, { "epoch": 3.661984594472134, "grad_norm": 0.17042965135656032, "learning_rate": 1.622017249727992e-05, "loss": 0.8625, "step": 8082 }, { "epoch": 3.6624376982328952, "grad_norm": 0.18173960260458935, "learning_rate": 1.6209998877603156e-05, "loss": 0.8698, "step": 8083 }, { "epoch": 3.6628908019936564, "grad_norm": 0.15673189879350927, "learning_rate": 1.6199827638645974e-05, "loss": 0.8691, "step": 8084 }, { "epoch": 3.6633439057544175, "grad_norm": 0.18323551421374462, "learning_rate": 1.6189658781426197e-05, "loss": 0.8501, "step": 8085 }, { "epoch": 3.663797009515179, "grad_norm": 0.21121170050173346, "learning_rate": 1.6179492306961475e-05, "loss": 0.8615, "step": 8086 }, { "epoch": 3.66425011327594, "grad_norm": 0.1697806719183866, "learning_rate": 1.6169328216269166e-05, "loss": 0.8558, "step": 8087 }, { "epoch": 3.6647032170367013, "grad_norm": 0.19947388402256955, "learning_rate": 1.6159166510366438e-05, "loss": 0.85, "step": 8088 }, { "epoch": 3.6651563207974625, "grad_norm": 0.2125125437783954, "learning_rate": 1.6149007190270166e-05, "loss": 0.8236, "step": 8089 }, { "epoch": 3.665609424558224, "grad_norm": 0.20577125965543464, "learning_rate": 1.613885025699705e-05, "loss": 0.8484, "step": 8090 }, { "epoch": 3.666062528318985, "grad_norm": 0.20166104016253317, "learning_rate": 1.61286957115635e-05, "loss": 0.8526, "step": 8091 }, { "epoch": 3.6665156320797463, "grad_norm": 0.19787265292099218, "learning_rate": 1.611854355498569e-05, "loss": 0.8635, "step": 8092 }, { "epoch": 3.6669687358405074, "grad_norm": 0.16021330905696216, "learning_rate": 1.61083937882796e-05, "loss": 0.8646, "step": 8093 }, { "epoch": 3.6674218396012686, "grad_norm": 0.18078538182414117, "learning_rate": 1.609824641246091e-05, "loss": 0.8478, "step": 8094 }, { "epoch": 3.6678749433620297, "grad_norm": 0.14969386728103518, "learning_rate": 1.6088101428545114e-05, "loss": 0.8542, "step": 8095 }, { "epoch": 3.6683280471227913, "grad_norm": 0.17868421247049524, "learning_rate": 1.6077958837547445e-05, "loss": 0.8589, "step": 8096 }, { "epoch": 3.6687811508835524, "grad_norm": 0.16401696412653663, "learning_rate": 1.606781864048288e-05, "loss": 0.8413, "step": 8097 }, { "epoch": 3.6692342546443135, "grad_norm": 0.21799551999894543, "learning_rate": 1.6057680838366163e-05, "loss": 0.8522, "step": 8098 }, { "epoch": 3.6696873584050747, "grad_norm": 0.1557865491581747, "learning_rate": 1.6047545432211834e-05, "loss": 0.8121, "step": 8099 }, { "epoch": 3.6701404621658362, "grad_norm": 0.23040714094911535, "learning_rate": 1.6037412423034142e-05, "loss": 0.861, "step": 8100 }, { "epoch": 3.6705935659265974, "grad_norm": 0.18987723497545772, "learning_rate": 1.6027281811847144e-05, "loss": 0.826, "step": 8101 }, { "epoch": 3.6710466696873585, "grad_norm": 0.1993715269459263, "learning_rate": 1.6017153599664627e-05, "loss": 0.8632, "step": 8102 }, { "epoch": 3.6714997734481196, "grad_norm": 0.19221661283731073, "learning_rate": 1.6007027787500112e-05, "loss": 0.8451, "step": 8103 }, { "epoch": 3.6719528772088808, "grad_norm": 0.19766648861321484, "learning_rate": 1.599690437636696e-05, "loss": 0.8508, "step": 8104 }, { "epoch": 3.672405980969642, "grad_norm": 0.20125941323329663, "learning_rate": 1.598678336727821e-05, "loss": 0.8317, "step": 8105 }, { "epoch": 3.672859084730403, "grad_norm": 0.21419413415422767, "learning_rate": 1.5976664761246715e-05, "loss": 0.8483, "step": 8106 }, { "epoch": 3.6733121884911646, "grad_norm": 0.1893226040852088, "learning_rate": 1.5966548559285045e-05, "loss": 0.8451, "step": 8107 }, { "epoch": 3.6737652922519257, "grad_norm": 0.15696094316412743, "learning_rate": 1.5956434762405582e-05, "loss": 0.8513, "step": 8108 }, { "epoch": 3.674218396012687, "grad_norm": 0.17701353636277695, "learning_rate": 1.59463233716204e-05, "loss": 0.8762, "step": 8109 }, { "epoch": 3.674671499773448, "grad_norm": 0.15897867049342654, "learning_rate": 1.5936214387941405e-05, "loss": 0.8579, "step": 8110 }, { "epoch": 3.6751246035342096, "grad_norm": 0.15677434079750566, "learning_rate": 1.5926107812380212e-05, "loss": 0.8486, "step": 8111 }, { "epoch": 3.6755777072949707, "grad_norm": 0.16128537876934113, "learning_rate": 1.5916003645948186e-05, "loss": 0.8501, "step": 8112 }, { "epoch": 3.676030811055732, "grad_norm": 0.19198046736679888, "learning_rate": 1.5905901889656515e-05, "loss": 0.8703, "step": 8113 }, { "epoch": 3.676483914816493, "grad_norm": 0.16251990263670504, "learning_rate": 1.5895802544516062e-05, "loss": 0.8391, "step": 8114 }, { "epoch": 3.676937018577254, "grad_norm": 0.1448455679114384, "learning_rate": 1.5885705611537537e-05, "loss": 0.8548, "step": 8115 }, { "epoch": 3.677390122338015, "grad_norm": 0.1659773609472324, "learning_rate": 1.587561109173132e-05, "loss": 0.8558, "step": 8116 }, { "epoch": 3.677843226098777, "grad_norm": 0.17578759517425852, "learning_rate": 1.5865518986107624e-05, "loss": 0.8796, "step": 8117 }, { "epoch": 3.678296329859538, "grad_norm": 0.17985244966127903, "learning_rate": 1.5855429295676366e-05, "loss": 0.8604, "step": 8118 }, { "epoch": 3.678749433620299, "grad_norm": 0.17319487503759945, "learning_rate": 1.5845342021447264e-05, "loss": 0.8316, "step": 8119 }, { "epoch": 3.67920253738106, "grad_norm": 0.1991455441678537, "learning_rate": 1.5835257164429773e-05, "loss": 0.8637, "step": 8120 }, { "epoch": 3.6796556411418218, "grad_norm": 0.1561956794167701, "learning_rate": 1.5825174725633083e-05, "loss": 0.8629, "step": 8121 }, { "epoch": 3.680108744902583, "grad_norm": 0.271323416330209, "learning_rate": 1.58150947060662e-05, "loss": 0.8614, "step": 8122 }, { "epoch": 3.680561848663344, "grad_norm": 0.1758307314762869, "learning_rate": 1.5805017106737822e-05, "loss": 0.8534, "step": 8123 }, { "epoch": 3.681014952424105, "grad_norm": 0.20147591323678396, "learning_rate": 1.5794941928656466e-05, "loss": 0.8566, "step": 8124 }, { "epoch": 3.6814680561848663, "grad_norm": 0.15514365750320272, "learning_rate": 1.5784869172830358e-05, "loss": 0.8604, "step": 8125 }, { "epoch": 3.6819211599456274, "grad_norm": 0.22057102593733668, "learning_rate": 1.5774798840267517e-05, "loss": 0.8438, "step": 8126 }, { "epoch": 3.6823742637063885, "grad_norm": 0.24293598191963325, "learning_rate": 1.57647309319757e-05, "loss": 0.8534, "step": 8127 }, { "epoch": 3.68282736746715, "grad_norm": 0.15829019976038605, "learning_rate": 1.5754665448962408e-05, "loss": 0.8458, "step": 8128 }, { "epoch": 3.6832804712279112, "grad_norm": 0.2847987685398027, "learning_rate": 1.574460239223494e-05, "loss": 0.8452, "step": 8129 }, { "epoch": 3.6837335749886724, "grad_norm": 0.24057661837613434, "learning_rate": 1.573454176280031e-05, "loss": 0.8787, "step": 8130 }, { "epoch": 3.6841866787494335, "grad_norm": 0.21174876643260013, "learning_rate": 1.5724483561665333e-05, "loss": 0.8576, "step": 8131 }, { "epoch": 3.684639782510195, "grad_norm": 0.3065422787440296, "learning_rate": 1.5714427789836526e-05, "loss": 0.8677, "step": 8132 }, { "epoch": 3.685092886270956, "grad_norm": 0.2573481793866333, "learning_rate": 1.570437444832022e-05, "loss": 0.8528, "step": 8133 }, { "epoch": 3.6855459900317173, "grad_norm": 0.16957017245468664, "learning_rate": 1.569432353812245e-05, "loss": 0.8659, "step": 8134 }, { "epoch": 3.6859990937924785, "grad_norm": 0.28986029444917444, "learning_rate": 1.568427506024906e-05, "loss": 0.8521, "step": 8135 }, { "epoch": 3.6864521975532396, "grad_norm": 0.21187534784122902, "learning_rate": 1.5674229015705616e-05, "loss": 0.8442, "step": 8136 }, { "epoch": 3.6869053013140007, "grad_norm": 0.23195835784175045, "learning_rate": 1.5664185405497424e-05, "loss": 0.8616, "step": 8137 }, { "epoch": 3.687358405074762, "grad_norm": 0.28167991704965584, "learning_rate": 1.565414423062961e-05, "loss": 0.8739, "step": 8138 }, { "epoch": 3.6878115088355234, "grad_norm": 0.21107949932111067, "learning_rate": 1.5644105492106974e-05, "loss": 0.859, "step": 8139 }, { "epoch": 3.6882646125962846, "grad_norm": 0.2148202897993927, "learning_rate": 1.5634069190934164e-05, "loss": 0.8609, "step": 8140 }, { "epoch": 3.6887177163570457, "grad_norm": 0.26373518344308666, "learning_rate": 1.5624035328115496e-05, "loss": 0.8439, "step": 8141 }, { "epoch": 3.689170820117807, "grad_norm": 0.20019664116150349, "learning_rate": 1.5614003904655103e-05, "loss": 0.862, "step": 8142 }, { "epoch": 3.6896239238785684, "grad_norm": 0.2351382110641363, "learning_rate": 1.560397492155685e-05, "loss": 0.851, "step": 8143 }, { "epoch": 3.6900770276393295, "grad_norm": 0.2245355239884779, "learning_rate": 1.5593948379824344e-05, "loss": 0.8594, "step": 8144 }, { "epoch": 3.6905301314000907, "grad_norm": 0.17473330061752826, "learning_rate": 1.558392428046099e-05, "loss": 0.8299, "step": 8145 }, { "epoch": 3.690983235160852, "grad_norm": 0.2019428041508029, "learning_rate": 1.5573902624469886e-05, "loss": 0.8461, "step": 8146 }, { "epoch": 3.691436338921613, "grad_norm": 0.18789195584087812, "learning_rate": 1.5563883412853967e-05, "loss": 0.8618, "step": 8147 }, { "epoch": 3.691889442682374, "grad_norm": 0.1800194237875754, "learning_rate": 1.5553866646615835e-05, "loss": 0.8619, "step": 8148 }, { "epoch": 3.6923425464431356, "grad_norm": 0.197375763237344, "learning_rate": 1.554385232675793e-05, "loss": 0.8656, "step": 8149 }, { "epoch": 3.6927956502038968, "grad_norm": 0.19201867369362677, "learning_rate": 1.553384045428237e-05, "loss": 0.8401, "step": 8150 }, { "epoch": 3.693248753964658, "grad_norm": 0.21442092082340597, "learning_rate": 1.55238310301911e-05, "loss": 0.8368, "step": 8151 }, { "epoch": 3.693701857725419, "grad_norm": 0.26754564162169137, "learning_rate": 1.5513824055485774e-05, "loss": 0.8563, "step": 8152 }, { "epoch": 3.6941549614861806, "grad_norm": 0.18511926815306856, "learning_rate": 1.5503819531167806e-05, "loss": 0.8532, "step": 8153 }, { "epoch": 3.6946080652469417, "grad_norm": 0.245695410008071, "learning_rate": 1.549381745823836e-05, "loss": 0.8586, "step": 8154 }, { "epoch": 3.695061169007703, "grad_norm": 0.2328950618239662, "learning_rate": 1.5483817837698392e-05, "loss": 0.8528, "step": 8155 }, { "epoch": 3.695514272768464, "grad_norm": 0.19567786486313926, "learning_rate": 1.5473820670548587e-05, "loss": 0.861, "step": 8156 }, { "epoch": 3.695967376529225, "grad_norm": 0.23537320307559978, "learning_rate": 1.5463825957789347e-05, "loss": 0.8521, "step": 8157 }, { "epoch": 3.6964204802899863, "grad_norm": 0.18772752131871095, "learning_rate": 1.545383370042091e-05, "loss": 0.8603, "step": 8158 }, { "epoch": 3.6968735840507474, "grad_norm": 0.20519008100251193, "learning_rate": 1.5443843899443193e-05, "loss": 0.8588, "step": 8159 }, { "epoch": 3.697326687811509, "grad_norm": 0.20376973473003623, "learning_rate": 1.5433856555855923e-05, "loss": 0.8397, "step": 8160 }, { "epoch": 3.69777979157227, "grad_norm": 0.1947067791975058, "learning_rate": 1.542387167065853e-05, "loss": 0.8485, "step": 8161 }, { "epoch": 3.6982328953330312, "grad_norm": 0.16443074443376177, "learning_rate": 1.5413889244850254e-05, "loss": 0.8364, "step": 8162 }, { "epoch": 3.6986859990937924, "grad_norm": 0.22986035678892378, "learning_rate": 1.540390927943004e-05, "loss": 0.8671, "step": 8163 }, { "epoch": 3.699139102854554, "grad_norm": 0.17281409520196844, "learning_rate": 1.53939317753966e-05, "loss": 0.8549, "step": 8164 }, { "epoch": 3.699592206615315, "grad_norm": 0.21561559393324534, "learning_rate": 1.5383956733748423e-05, "loss": 0.8571, "step": 8165 }, { "epoch": 3.700045310376076, "grad_norm": 0.18165175438038675, "learning_rate": 1.5373984155483714e-05, "loss": 0.8592, "step": 8166 }, { "epoch": 3.7004984141368373, "grad_norm": 0.1708058499059133, "learning_rate": 1.5364014041600474e-05, "loss": 0.8257, "step": 8167 }, { "epoch": 3.7009515178975985, "grad_norm": 0.17984841673037497, "learning_rate": 1.5354046393096415e-05, "loss": 0.8651, "step": 8168 }, { "epoch": 3.7014046216583596, "grad_norm": 0.17731980656030258, "learning_rate": 1.5344081210969045e-05, "loss": 0.8423, "step": 8169 }, { "epoch": 3.7018577254191207, "grad_norm": 0.18972778821349906, "learning_rate": 1.5334118496215572e-05, "loss": 0.8714, "step": 8170 }, { "epoch": 3.7023108291798823, "grad_norm": 0.18433698703105408, "learning_rate": 1.532415824983302e-05, "loss": 0.8629, "step": 8171 }, { "epoch": 3.7027639329406434, "grad_norm": 0.19626901429157648, "learning_rate": 1.531420047281812e-05, "loss": 0.8572, "step": 8172 }, { "epoch": 3.7032170367014046, "grad_norm": 0.1584894269840551, "learning_rate": 1.5304245166167354e-05, "loss": 0.8419, "step": 8173 }, { "epoch": 3.7036701404621657, "grad_norm": 0.20440417416552872, "learning_rate": 1.5294292330876994e-05, "loss": 0.8335, "step": 8174 }, { "epoch": 3.7041232442229273, "grad_norm": 0.14596608424061963, "learning_rate": 1.5284341967943025e-05, "loss": 0.8617, "step": 8175 }, { "epoch": 3.7045763479836884, "grad_norm": 0.20275488093371155, "learning_rate": 1.5274394078361225e-05, "loss": 0.852, "step": 8176 }, { "epoch": 3.7050294517444495, "grad_norm": 0.18151593627341456, "learning_rate": 1.5264448663127075e-05, "loss": 0.8342, "step": 8177 }, { "epoch": 3.7054825555052107, "grad_norm": 0.18912569639617563, "learning_rate": 1.5254505723235861e-05, "loss": 0.8731, "step": 8178 }, { "epoch": 3.705935659265972, "grad_norm": 0.22085770040029176, "learning_rate": 1.5244565259682574e-05, "loss": 0.8439, "step": 8179 }, { "epoch": 3.706388763026733, "grad_norm": 0.18199951039286938, "learning_rate": 1.5234627273462002e-05, "loss": 0.8556, "step": 8180 }, { "epoch": 3.7068418667874945, "grad_norm": 0.18139517030121638, "learning_rate": 1.522469176556865e-05, "loss": 0.8571, "step": 8181 }, { "epoch": 3.7072949705482556, "grad_norm": 0.2278845746346806, "learning_rate": 1.5214758736996764e-05, "loss": 0.8652, "step": 8182 }, { "epoch": 3.7077480743090168, "grad_norm": 0.14464671078060753, "learning_rate": 1.5204828188740407e-05, "loss": 0.8494, "step": 8183 }, { "epoch": 3.708201178069778, "grad_norm": 0.22075969187391617, "learning_rate": 1.5194900121793317e-05, "loss": 0.8723, "step": 8184 }, { "epoch": 3.7086542818305395, "grad_norm": 0.1864548230577428, "learning_rate": 1.5184974537149044e-05, "loss": 0.8446, "step": 8185 }, { "epoch": 3.7091073855913006, "grad_norm": 0.1539645970854302, "learning_rate": 1.5175051435800839e-05, "loss": 0.8372, "step": 8186 }, { "epoch": 3.7095604893520617, "grad_norm": 0.2050718986288208, "learning_rate": 1.5165130818741754e-05, "loss": 0.8476, "step": 8187 }, { "epoch": 3.710013593112823, "grad_norm": 0.1463191964255394, "learning_rate": 1.515521268696456e-05, "loss": 0.8355, "step": 8188 }, { "epoch": 3.710466696873584, "grad_norm": 0.1529699839780098, "learning_rate": 1.5145297041461765e-05, "loss": 0.8555, "step": 8189 }, { "epoch": 3.710919800634345, "grad_norm": 0.17017599720778104, "learning_rate": 1.5135383883225685e-05, "loss": 0.8673, "step": 8190 }, { "epoch": 3.7113729043951063, "grad_norm": 0.15857449277175337, "learning_rate": 1.5125473213248319e-05, "loss": 0.8456, "step": 8191 }, { "epoch": 3.711826008155868, "grad_norm": 0.17080980361292186, "learning_rate": 1.5115565032521482e-05, "loss": 0.8483, "step": 8192 }, { "epoch": 3.712279111916629, "grad_norm": 0.17331403730999112, "learning_rate": 1.5105659342036681e-05, "loss": 0.8365, "step": 8193 }, { "epoch": 3.71273221567739, "grad_norm": 0.16338576987667794, "learning_rate": 1.5095756142785227e-05, "loss": 0.8412, "step": 8194 }, { "epoch": 3.713185319438151, "grad_norm": 0.17643021161706, "learning_rate": 1.5085855435758125e-05, "loss": 0.8431, "step": 8195 }, { "epoch": 3.713638423198913, "grad_norm": 0.14710778095575758, "learning_rate": 1.5075957221946195e-05, "loss": 0.8605, "step": 8196 }, { "epoch": 3.714091526959674, "grad_norm": 0.15824038484925473, "learning_rate": 1.5066061502339953e-05, "loss": 0.8457, "step": 8197 }, { "epoch": 3.714544630720435, "grad_norm": 0.20071118817906414, "learning_rate": 1.505616827792968e-05, "loss": 0.853, "step": 8198 }, { "epoch": 3.714997734481196, "grad_norm": 0.2117699663187654, "learning_rate": 1.5046277549705437e-05, "loss": 0.8403, "step": 8199 }, { "epoch": 3.7154508382419573, "grad_norm": 0.16645330798789207, "learning_rate": 1.503638931865698e-05, "loss": 0.8579, "step": 8200 }, { "epoch": 3.7159039420027185, "grad_norm": 0.24085607602688744, "learning_rate": 1.5026503585773884e-05, "loss": 0.8917, "step": 8201 }, { "epoch": 3.7163570457634796, "grad_norm": 0.16930625182178902, "learning_rate": 1.5016620352045403e-05, "loss": 0.8529, "step": 8202 }, { "epoch": 3.716810149524241, "grad_norm": 0.2201687677046315, "learning_rate": 1.5006739618460601e-05, "loss": 0.8604, "step": 8203 }, { "epoch": 3.7172632532850023, "grad_norm": 0.19265707290288872, "learning_rate": 1.4996861386008258e-05, "loss": 0.8688, "step": 8204 }, { "epoch": 3.7177163570457634, "grad_norm": 0.22178958525250855, "learning_rate": 1.498698565567689e-05, "loss": 0.8598, "step": 8205 }, { "epoch": 3.7181694608065246, "grad_norm": 0.1866994791131662, "learning_rate": 1.4977112428454813e-05, "loss": 0.8562, "step": 8206 }, { "epoch": 3.718622564567286, "grad_norm": 0.21013336093666668, "learning_rate": 1.4967241705330043e-05, "loss": 0.8584, "step": 8207 }, { "epoch": 3.7190756683280473, "grad_norm": 0.21302530997516175, "learning_rate": 1.4957373487290383e-05, "loss": 0.8351, "step": 8208 }, { "epoch": 3.7195287720888084, "grad_norm": 0.18959258181076916, "learning_rate": 1.494750777532335e-05, "loss": 0.8457, "step": 8209 }, { "epoch": 3.7199818758495695, "grad_norm": 0.20351271146519082, "learning_rate": 1.493764457041626e-05, "loss": 0.8624, "step": 8210 }, { "epoch": 3.7204349796103307, "grad_norm": 0.21595717979013732, "learning_rate": 1.4927783873556098e-05, "loss": 0.8538, "step": 8211 }, { "epoch": 3.720888083371092, "grad_norm": 0.18877176939268783, "learning_rate": 1.4917925685729687e-05, "loss": 0.8676, "step": 8212 }, { "epoch": 3.7213411871318534, "grad_norm": 0.19198596349885622, "learning_rate": 1.4908070007923527e-05, "loss": 0.8685, "step": 8213 }, { "epoch": 3.7217942908926145, "grad_norm": 0.21179428355207755, "learning_rate": 1.4898216841123931e-05, "loss": 0.8551, "step": 8214 }, { "epoch": 3.7222473946533756, "grad_norm": 0.1623971119378309, "learning_rate": 1.4888366186316896e-05, "loss": 0.8356, "step": 8215 }, { "epoch": 3.7227004984141367, "grad_norm": 0.1769254430206729, "learning_rate": 1.4878518044488233e-05, "loss": 0.8487, "step": 8216 }, { "epoch": 3.7231536021748983, "grad_norm": 0.16467995513761202, "learning_rate": 1.4868672416623442e-05, "loss": 0.8418, "step": 8217 }, { "epoch": 3.7236067059356595, "grad_norm": 0.17239438634353438, "learning_rate": 1.4858829303707799e-05, "loss": 0.8705, "step": 8218 }, { "epoch": 3.7240598096964206, "grad_norm": 0.17411439475561388, "learning_rate": 1.484898870672634e-05, "loss": 0.8441, "step": 8219 }, { "epoch": 3.7245129134571817, "grad_norm": 0.1803407286078281, "learning_rate": 1.4839150626663821e-05, "loss": 0.8431, "step": 8220 }, { "epoch": 3.724966017217943, "grad_norm": 0.16855338053903993, "learning_rate": 1.482931506450478e-05, "loss": 0.8514, "step": 8221 }, { "epoch": 3.725419120978704, "grad_norm": 0.19372999092364845, "learning_rate": 1.4819482021233458e-05, "loss": 0.8484, "step": 8222 }, { "epoch": 3.725872224739465, "grad_norm": 0.19627396140592018, "learning_rate": 1.4809651497833897e-05, "loss": 0.8441, "step": 8223 }, { "epoch": 3.7263253285002267, "grad_norm": 0.2733086434579735, "learning_rate": 1.479982349528985e-05, "loss": 0.8754, "step": 8224 }, { "epoch": 3.726778432260988, "grad_norm": 0.16259083893747148, "learning_rate": 1.4789998014584815e-05, "loss": 0.8635, "step": 8225 }, { "epoch": 3.727231536021749, "grad_norm": 0.2656689515543522, "learning_rate": 1.478017505670207e-05, "loss": 0.8673, "step": 8226 }, { "epoch": 3.72768463978251, "grad_norm": 0.16756994412408951, "learning_rate": 1.4770354622624598e-05, "loss": 0.8375, "step": 8227 }, { "epoch": 3.7281377435432717, "grad_norm": 0.25955722116752883, "learning_rate": 1.476053671333518e-05, "loss": 0.8765, "step": 8228 }, { "epoch": 3.728590847304033, "grad_norm": 0.23076330327355443, "learning_rate": 1.4750721329816288e-05, "loss": 0.8323, "step": 8229 }, { "epoch": 3.729043951064794, "grad_norm": 0.20403251167958483, "learning_rate": 1.47409084730502e-05, "loss": 0.8543, "step": 8230 }, { "epoch": 3.729497054825555, "grad_norm": 0.188282644327036, "learning_rate": 1.4731098144018879e-05, "loss": 0.8615, "step": 8231 }, { "epoch": 3.729950158586316, "grad_norm": 0.2501198855191333, "learning_rate": 1.4721290343704095e-05, "loss": 0.8577, "step": 8232 }, { "epoch": 3.7304032623470773, "grad_norm": 0.1571400379590629, "learning_rate": 1.4711485073087327e-05, "loss": 0.8364, "step": 8233 }, { "epoch": 3.7308563661078384, "grad_norm": 0.2257809445162661, "learning_rate": 1.4701682333149796e-05, "loss": 0.8483, "step": 8234 }, { "epoch": 3.7313094698686, "grad_norm": 0.18361661708747667, "learning_rate": 1.4691882124872506e-05, "loss": 0.851, "step": 8235 }, { "epoch": 3.731762573629361, "grad_norm": 0.19577670994374768, "learning_rate": 1.4682084449236164e-05, "loss": 0.8726, "step": 8236 }, { "epoch": 3.7322156773901223, "grad_norm": 0.22538897625173174, "learning_rate": 1.467228930722128e-05, "loss": 0.8488, "step": 8237 }, { "epoch": 3.7326687811508834, "grad_norm": 0.17263398637050154, "learning_rate": 1.4662496699808033e-05, "loss": 0.8478, "step": 8238 }, { "epoch": 3.733121884911645, "grad_norm": 0.20046291170892458, "learning_rate": 1.465270662797643e-05, "loss": 0.8667, "step": 8239 }, { "epoch": 3.733574988672406, "grad_norm": 0.178219325370002, "learning_rate": 1.4642919092706169e-05, "loss": 0.8697, "step": 8240 }, { "epoch": 3.7340280924331672, "grad_norm": 0.16865450738234986, "learning_rate": 1.4633134094976695e-05, "loss": 0.8594, "step": 8241 }, { "epoch": 3.7344811961939284, "grad_norm": 0.15239863545002963, "learning_rate": 1.4623351635767247e-05, "loss": 0.8426, "step": 8242 }, { "epoch": 3.7349342999546895, "grad_norm": 0.20019776066575254, "learning_rate": 1.4613571716056752e-05, "loss": 0.8723, "step": 8243 }, { "epoch": 3.7353874037154506, "grad_norm": 0.14517023951505478, "learning_rate": 1.4603794336823928e-05, "loss": 0.8451, "step": 8244 }, { "epoch": 3.735840507476212, "grad_norm": 0.19851781860339385, "learning_rate": 1.4594019499047196e-05, "loss": 0.8558, "step": 8245 }, { "epoch": 3.7362936112369733, "grad_norm": 0.17282741607046367, "learning_rate": 1.4584247203704775e-05, "loss": 0.8448, "step": 8246 }, { "epoch": 3.7367467149977345, "grad_norm": 0.2105141088305298, "learning_rate": 1.4574477451774578e-05, "loss": 0.8699, "step": 8247 }, { "epoch": 3.7371998187584956, "grad_norm": 0.16392089227196785, "learning_rate": 1.4564710244234302e-05, "loss": 0.8398, "step": 8248 }, { "epoch": 3.737652922519257, "grad_norm": 0.18700015873335066, "learning_rate": 1.4554945582061368e-05, "loss": 0.8605, "step": 8249 }, { "epoch": 3.7381060262800183, "grad_norm": 0.15311038197462432, "learning_rate": 1.4545183466232936e-05, "loss": 0.8622, "step": 8250 }, { "epoch": 3.7385591300407794, "grad_norm": 0.16694578083097286, "learning_rate": 1.4535423897725944e-05, "loss": 0.8571, "step": 8251 }, { "epoch": 3.7390122338015406, "grad_norm": 0.18925173277948001, "learning_rate": 1.4525666877517034e-05, "loss": 0.8746, "step": 8252 }, { "epoch": 3.7394653375623017, "grad_norm": 0.13563794283117073, "learning_rate": 1.4515912406582632e-05, "loss": 0.8425, "step": 8253 }, { "epoch": 3.739918441323063, "grad_norm": 0.1719085117432022, "learning_rate": 1.4506160485898875e-05, "loss": 0.8428, "step": 8254 }, { "epoch": 3.740371545083824, "grad_norm": 0.15289854073363832, "learning_rate": 1.449641111644167e-05, "loss": 0.8425, "step": 8255 }, { "epoch": 3.7408246488445855, "grad_norm": 0.16307435827615646, "learning_rate": 1.448666429918665e-05, "loss": 0.8565, "step": 8256 }, { "epoch": 3.7412777526053467, "grad_norm": 0.15037061858123554, "learning_rate": 1.4476920035109214e-05, "loss": 0.8522, "step": 8257 }, { "epoch": 3.741730856366108, "grad_norm": 0.1498578859624203, "learning_rate": 1.4467178325184486e-05, "loss": 0.8668, "step": 8258 }, { "epoch": 3.742183960126869, "grad_norm": 0.16912740603086945, "learning_rate": 1.4457439170387328e-05, "loss": 0.8442, "step": 8259 }, { "epoch": 3.7426370638876305, "grad_norm": 0.14929905526599924, "learning_rate": 1.4447702571692385e-05, "loss": 0.8594, "step": 8260 }, { "epoch": 3.7430901676483916, "grad_norm": 0.1643570216825187, "learning_rate": 1.443796853007399e-05, "loss": 0.8528, "step": 8261 }, { "epoch": 3.7435432714091528, "grad_norm": 0.16218493040827212, "learning_rate": 1.4428237046506284e-05, "loss": 0.8599, "step": 8262 }, { "epoch": 3.743996375169914, "grad_norm": 0.19038146291995706, "learning_rate": 1.4418508121963086e-05, "loss": 0.8663, "step": 8263 }, { "epoch": 3.744449478930675, "grad_norm": 0.13686348463819095, "learning_rate": 1.4408781757418022e-05, "loss": 0.8597, "step": 8264 }, { "epoch": 3.744902582691436, "grad_norm": 0.1835305156791907, "learning_rate": 1.4399057953844415e-05, "loss": 0.8694, "step": 8265 }, { "epoch": 3.7453556864521973, "grad_norm": 0.1431762994063905, "learning_rate": 1.4389336712215341e-05, "loss": 0.8363, "step": 8266 }, { "epoch": 3.745808790212959, "grad_norm": 0.13647109982094136, "learning_rate": 1.4379618033503651e-05, "loss": 0.8532, "step": 8267 }, { "epoch": 3.74626189397372, "grad_norm": 0.17240534037234556, "learning_rate": 1.4369901918681896e-05, "loss": 0.8365, "step": 8268 }, { "epoch": 3.746714997734481, "grad_norm": 0.1479425534768405, "learning_rate": 1.4360188368722394e-05, "loss": 0.863, "step": 8269 }, { "epoch": 3.7471681014952423, "grad_norm": 0.20357507683105855, "learning_rate": 1.4350477384597191e-05, "loss": 0.8663, "step": 8270 }, { "epoch": 3.747621205256004, "grad_norm": 0.1723613119107191, "learning_rate": 1.434076896727811e-05, "loss": 0.8705, "step": 8271 }, { "epoch": 3.748074309016765, "grad_norm": 0.16931758335343688, "learning_rate": 1.433106311773667e-05, "loss": 0.8438, "step": 8272 }, { "epoch": 3.748527412777526, "grad_norm": 0.16087964546477684, "learning_rate": 1.4321359836944186e-05, "loss": 0.846, "step": 8273 }, { "epoch": 3.7489805165382872, "grad_norm": 0.14656389236667125, "learning_rate": 1.4311659125871656e-05, "loss": 0.8553, "step": 8274 }, { "epoch": 3.7494336202990484, "grad_norm": 0.14989994034576887, "learning_rate": 1.4301960985489883e-05, "loss": 0.8611, "step": 8275 }, { "epoch": 3.7498867240598095, "grad_norm": 0.1381762704742944, "learning_rate": 1.4292265416769353e-05, "loss": 0.8718, "step": 8276 }, { "epoch": 3.750339827820571, "grad_norm": 0.15258416836806218, "learning_rate": 1.428257242068035e-05, "loss": 0.8474, "step": 8277 }, { "epoch": 3.750792931581332, "grad_norm": 0.14795057439606243, "learning_rate": 1.4272881998192865e-05, "loss": 0.8399, "step": 8278 }, { "epoch": 3.7512460353420933, "grad_norm": 0.160973123480901, "learning_rate": 1.4263194150276621e-05, "loss": 0.8578, "step": 8279 }, { "epoch": 3.7516991391028545, "grad_norm": 0.1489657644626548, "learning_rate": 1.425350887790113e-05, "loss": 0.8685, "step": 8280 }, { "epoch": 3.752152242863616, "grad_norm": 0.1586056286696044, "learning_rate": 1.42438261820356e-05, "loss": 0.8557, "step": 8281 }, { "epoch": 3.752605346624377, "grad_norm": 0.15570250810238878, "learning_rate": 1.4234146063649021e-05, "loss": 0.865, "step": 8282 }, { "epoch": 3.7530584503851383, "grad_norm": 0.18679646512275055, "learning_rate": 1.422446852371008e-05, "loss": 0.8622, "step": 8283 }, { "epoch": 3.7535115541458994, "grad_norm": 0.163153165151269, "learning_rate": 1.4214793563187254e-05, "loss": 0.8511, "step": 8284 }, { "epoch": 3.7539646579066606, "grad_norm": 0.2936754047693289, "learning_rate": 1.420512118304873e-05, "loss": 0.841, "step": 8285 }, { "epoch": 3.7544177616674217, "grad_norm": 0.1693320743222294, "learning_rate": 1.4195451384262428e-05, "loss": 0.8596, "step": 8286 }, { "epoch": 3.754870865428183, "grad_norm": 0.1826963980976861, "learning_rate": 1.4185784167796053e-05, "loss": 0.8558, "step": 8287 }, { "epoch": 3.7553239691889444, "grad_norm": 0.16137603284172297, "learning_rate": 1.4176119534617003e-05, "loss": 0.8738, "step": 8288 }, { "epoch": 3.7557770729497055, "grad_norm": 0.22179262362186405, "learning_rate": 1.4166457485692463e-05, "loss": 0.8676, "step": 8289 }, { "epoch": 3.7562301767104667, "grad_norm": 0.14648726371264112, "learning_rate": 1.4156798021989308e-05, "loss": 0.8506, "step": 8290 }, { "epoch": 3.756683280471228, "grad_norm": 0.21288490954952255, "learning_rate": 1.4147141144474215e-05, "loss": 0.8637, "step": 8291 }, { "epoch": 3.7571363842319894, "grad_norm": 0.1557457474111895, "learning_rate": 1.4137486854113536e-05, "loss": 0.8587, "step": 8292 }, { "epoch": 3.7575894879927505, "grad_norm": 0.18979686714317867, "learning_rate": 1.4127835151873432e-05, "loss": 0.8625, "step": 8293 }, { "epoch": 3.7580425917535116, "grad_norm": 0.16141820989166838, "learning_rate": 1.411818603871975e-05, "loss": 0.8466, "step": 8294 }, { "epoch": 3.7584956955142728, "grad_norm": 0.17966705552922516, "learning_rate": 1.410853951561809e-05, "loss": 0.8368, "step": 8295 }, { "epoch": 3.758948799275034, "grad_norm": 0.16806180421221312, "learning_rate": 1.409889558353383e-05, "loss": 0.8699, "step": 8296 }, { "epoch": 3.759401903035795, "grad_norm": 0.1600626808229934, "learning_rate": 1.4089254243432024e-05, "loss": 0.8537, "step": 8297 }, { "epoch": 3.759855006796556, "grad_norm": 0.16189268358924255, "learning_rate": 1.4079615496277543e-05, "loss": 0.8613, "step": 8298 }, { "epoch": 3.7603081105573177, "grad_norm": 0.20064297402105052, "learning_rate": 1.4069979343034922e-05, "loss": 0.8443, "step": 8299 }, { "epoch": 3.760761214318079, "grad_norm": 0.16795777767063, "learning_rate": 1.4060345784668501e-05, "loss": 0.8428, "step": 8300 }, { "epoch": 3.76121431807884, "grad_norm": 0.1894577097443553, "learning_rate": 1.4050714822142322e-05, "loss": 0.8537, "step": 8301 }, { "epoch": 3.761667421839601, "grad_norm": 0.19596072652312155, "learning_rate": 1.4041086456420158e-05, "loss": 0.852, "step": 8302 }, { "epoch": 3.7621205256003627, "grad_norm": 0.1720692893504025, "learning_rate": 1.403146068846558e-05, "loss": 0.8464, "step": 8303 }, { "epoch": 3.762573629361124, "grad_norm": 0.1949791755434125, "learning_rate": 1.4021837519241822e-05, "loss": 0.8806, "step": 8304 }, { "epoch": 3.763026733121885, "grad_norm": 0.17650103097132197, "learning_rate": 1.4012216949711923e-05, "loss": 0.8538, "step": 8305 }, { "epoch": 3.763479836882646, "grad_norm": 0.1753993202899679, "learning_rate": 1.4002598980838618e-05, "loss": 0.8648, "step": 8306 }, { "epoch": 3.763932940643407, "grad_norm": 0.188081363853381, "learning_rate": 1.3992983613584419e-05, "loss": 0.8505, "step": 8307 }, { "epoch": 3.7643860444041684, "grad_norm": 0.15697937014500585, "learning_rate": 1.3983370848911531e-05, "loss": 0.855, "step": 8308 }, { "epoch": 3.76483914816493, "grad_norm": 0.18362888930832505, "learning_rate": 1.3973760687781956e-05, "loss": 0.8625, "step": 8309 }, { "epoch": 3.765292251925691, "grad_norm": 0.1642955868823864, "learning_rate": 1.396415313115739e-05, "loss": 0.842, "step": 8310 }, { "epoch": 3.765745355686452, "grad_norm": 0.17271576633684588, "learning_rate": 1.3954548179999266e-05, "loss": 0.835, "step": 8311 }, { "epoch": 3.7661984594472133, "grad_norm": 0.22501674588669693, "learning_rate": 1.3944945835268802e-05, "loss": 0.8722, "step": 8312 }, { "epoch": 3.766651563207975, "grad_norm": 0.18290921686018674, "learning_rate": 1.3935346097926901e-05, "loss": 0.8369, "step": 8313 }, { "epoch": 3.767104666968736, "grad_norm": 0.16963809637405866, "learning_rate": 1.3925748968934255e-05, "loss": 0.8651, "step": 8314 }, { "epoch": 3.767557770729497, "grad_norm": 0.16881781853673283, "learning_rate": 1.3916154449251247e-05, "loss": 0.8443, "step": 8315 }, { "epoch": 3.7680108744902583, "grad_norm": 0.16967756005121892, "learning_rate": 1.3906562539838047e-05, "loss": 0.8471, "step": 8316 }, { "epoch": 3.7684639782510194, "grad_norm": 0.1761678785746007, "learning_rate": 1.3896973241654514e-05, "loss": 0.8664, "step": 8317 }, { "epoch": 3.7689170820117806, "grad_norm": 0.15703963747222704, "learning_rate": 1.3887386555660297e-05, "loss": 0.8494, "step": 8318 }, { "epoch": 3.7693701857725417, "grad_norm": 0.16032190535046242, "learning_rate": 1.3877802482814745e-05, "loss": 0.8513, "step": 8319 }, { "epoch": 3.7698232895333033, "grad_norm": 0.15441546696418318, "learning_rate": 1.386822102407694e-05, "loss": 0.8685, "step": 8320 }, { "epoch": 3.7702763932940644, "grad_norm": 0.18360165534669148, "learning_rate": 1.3858642180405752e-05, "loss": 0.8476, "step": 8321 }, { "epoch": 3.7707294970548255, "grad_norm": 0.14133305954501507, "learning_rate": 1.384906595275973e-05, "loss": 0.8654, "step": 8322 }, { "epoch": 3.7711826008155866, "grad_norm": 0.18275739522738782, "learning_rate": 1.3839492342097214e-05, "loss": 0.8738, "step": 8323 }, { "epoch": 3.7716357045763482, "grad_norm": 0.1491905271219199, "learning_rate": 1.3829921349376246e-05, "loss": 0.8606, "step": 8324 }, { "epoch": 3.7720888083371094, "grad_norm": 0.15526907581777585, "learning_rate": 1.3820352975554618e-05, "loss": 0.8657, "step": 8325 }, { "epoch": 3.7725419120978705, "grad_norm": 0.16341037664224173, "learning_rate": 1.3810787221589839e-05, "loss": 0.8658, "step": 8326 }, { "epoch": 3.7729950158586316, "grad_norm": 0.15742116784412152, "learning_rate": 1.3801224088439211e-05, "loss": 0.8695, "step": 8327 }, { "epoch": 3.7734481196193927, "grad_norm": 0.15956570484322957, "learning_rate": 1.3791663577059704e-05, "loss": 0.8722, "step": 8328 }, { "epoch": 3.773901223380154, "grad_norm": 0.21184680676463385, "learning_rate": 1.378210568840809e-05, "loss": 0.8649, "step": 8329 }, { "epoch": 3.774354327140915, "grad_norm": 0.15438133907393692, "learning_rate": 1.3772550423440838e-05, "loss": 0.8614, "step": 8330 }, { "epoch": 3.7748074309016766, "grad_norm": 0.26166368982180804, "learning_rate": 1.3762997783114149e-05, "loss": 0.8656, "step": 8331 }, { "epoch": 3.7752605346624377, "grad_norm": 0.15431498387843184, "learning_rate": 1.3753447768384005e-05, "loss": 0.8604, "step": 8332 }, { "epoch": 3.775713638423199, "grad_norm": 0.2225829133502041, "learning_rate": 1.3743900380206072e-05, "loss": 0.8614, "step": 8333 }, { "epoch": 3.77616674218396, "grad_norm": 0.14979845326213273, "learning_rate": 1.3734355619535804e-05, "loss": 0.865, "step": 8334 }, { "epoch": 3.7766198459447216, "grad_norm": 0.20380512860617384, "learning_rate": 1.372481348732834e-05, "loss": 0.8369, "step": 8335 }, { "epoch": 3.7770729497054827, "grad_norm": 0.15664440620260292, "learning_rate": 1.371527398453861e-05, "loss": 0.8455, "step": 8336 }, { "epoch": 3.777526053466244, "grad_norm": 0.18112211958423483, "learning_rate": 1.3705737112121224e-05, "loss": 0.8426, "step": 8337 }, { "epoch": 3.777979157227005, "grad_norm": 0.17815067635046516, "learning_rate": 1.3696202871030595e-05, "loss": 0.8452, "step": 8338 }, { "epoch": 3.778432260987766, "grad_norm": 0.19275763317925115, "learning_rate": 1.3686671262220812e-05, "loss": 0.8467, "step": 8339 }, { "epoch": 3.778885364748527, "grad_norm": 0.15245815206592034, "learning_rate": 1.3677142286645717e-05, "loss": 0.8697, "step": 8340 }, { "epoch": 3.779338468509289, "grad_norm": 0.1820612182717892, "learning_rate": 1.366761594525892e-05, "loss": 0.8639, "step": 8341 }, { "epoch": 3.77979157227005, "grad_norm": 0.16490210178942719, "learning_rate": 1.3658092239013718e-05, "loss": 0.8586, "step": 8342 }, { "epoch": 3.780244676030811, "grad_norm": 0.16655366604604427, "learning_rate": 1.36485711688632e-05, "loss": 0.8545, "step": 8343 }, { "epoch": 3.780697779791572, "grad_norm": 0.1811773971856876, "learning_rate": 1.363905273576013e-05, "loss": 0.8623, "step": 8344 }, { "epoch": 3.7811508835523338, "grad_norm": 0.15032276516277912, "learning_rate": 1.3629536940657068e-05, "loss": 0.8661, "step": 8345 }, { "epoch": 3.781603987313095, "grad_norm": 0.20200872473481143, "learning_rate": 1.3620023784506269e-05, "loss": 0.8436, "step": 8346 }, { "epoch": 3.782057091073856, "grad_norm": 0.1679440309772418, "learning_rate": 1.3610513268259724e-05, "loss": 0.8519, "step": 8347 }, { "epoch": 3.782510194834617, "grad_norm": 0.19520114111962103, "learning_rate": 1.3601005392869193e-05, "loss": 0.8533, "step": 8348 }, { "epoch": 3.7829632985953783, "grad_norm": 0.18567555517692014, "learning_rate": 1.3591500159286125e-05, "loss": 0.8524, "step": 8349 }, { "epoch": 3.7834164023561394, "grad_norm": 0.15472116985610207, "learning_rate": 1.358199756846176e-05, "loss": 0.8645, "step": 8350 }, { "epoch": 3.7838695061169005, "grad_norm": 0.17327055542712208, "learning_rate": 1.357249762134702e-05, "loss": 0.8615, "step": 8351 }, { "epoch": 3.784322609877662, "grad_norm": 0.16310164300988783, "learning_rate": 1.3563000318892608e-05, "loss": 0.8574, "step": 8352 }, { "epoch": 3.7847757136384232, "grad_norm": 0.17384599910981732, "learning_rate": 1.3553505662048916e-05, "loss": 0.865, "step": 8353 }, { "epoch": 3.7852288173991844, "grad_norm": 0.21799649961741988, "learning_rate": 1.3544013651766118e-05, "loss": 0.8408, "step": 8354 }, { "epoch": 3.7856819211599455, "grad_norm": 0.13770245148752874, "learning_rate": 1.3534524288994097e-05, "loss": 0.858, "step": 8355 }, { "epoch": 3.786135024920707, "grad_norm": 0.23532777653306447, "learning_rate": 1.3525037574682456e-05, "loss": 0.8699, "step": 8356 }, { "epoch": 3.786588128681468, "grad_norm": 0.25166566167369553, "learning_rate": 1.3515553509780577e-05, "loss": 0.8616, "step": 8357 }, { "epoch": 3.7870412324422293, "grad_norm": 0.16458684814103744, "learning_rate": 1.3506072095237533e-05, "loss": 0.8549, "step": 8358 }, { "epoch": 3.7874943362029905, "grad_norm": 0.17783271178671406, "learning_rate": 1.3496593332002163e-05, "loss": 0.8662, "step": 8359 }, { "epoch": 3.7879474399637516, "grad_norm": 0.12710143176798683, "learning_rate": 1.3487117221023019e-05, "loss": 0.8459, "step": 8360 }, { "epoch": 3.7884005437245127, "grad_norm": 0.17898737226913985, "learning_rate": 1.3477643763248414e-05, "loss": 0.853, "step": 8361 }, { "epoch": 3.788853647485274, "grad_norm": 0.17710652053711373, "learning_rate": 1.3468172959626365e-05, "loss": 0.8586, "step": 8362 }, { "epoch": 3.7893067512460354, "grad_norm": 0.14956894134550106, "learning_rate": 1.3458704811104623e-05, "loss": 0.8474, "step": 8363 }, { "epoch": 3.7897598550067966, "grad_norm": 0.1941457174400899, "learning_rate": 1.3449239318630718e-05, "loss": 0.8533, "step": 8364 }, { "epoch": 3.7902129587675577, "grad_norm": 0.14650250593174846, "learning_rate": 1.3439776483151858e-05, "loss": 0.862, "step": 8365 }, { "epoch": 3.790666062528319, "grad_norm": 0.1458430534924156, "learning_rate": 1.343031630561503e-05, "loss": 0.8688, "step": 8366 }, { "epoch": 3.7911191662890804, "grad_norm": 0.16175803681887263, "learning_rate": 1.3420858786966914e-05, "loss": 0.8556, "step": 8367 }, { "epoch": 3.7915722700498415, "grad_norm": 0.15046680151097966, "learning_rate": 1.341140392815397e-05, "loss": 0.8708, "step": 8368 }, { "epoch": 3.7920253738106027, "grad_norm": 0.14958125959198357, "learning_rate": 1.3401951730122345e-05, "loss": 0.8649, "step": 8369 }, { "epoch": 3.792478477571364, "grad_norm": 0.15228375356764984, "learning_rate": 1.3392502193817962e-05, "loss": 0.8527, "step": 8370 }, { "epoch": 3.792931581332125, "grad_norm": 0.1606019818250051, "learning_rate": 1.3383055320186444e-05, "loss": 0.8609, "step": 8371 }, { "epoch": 3.793384685092886, "grad_norm": 0.16354233607033963, "learning_rate": 1.3373611110173159e-05, "loss": 0.8513, "step": 8372 }, { "epoch": 3.7938377888536476, "grad_norm": 0.14890454172028353, "learning_rate": 1.3364169564723222e-05, "loss": 0.8572, "step": 8373 }, { "epoch": 3.7942908926144088, "grad_norm": 0.1370848765155295, "learning_rate": 1.3354730684781453e-05, "loss": 0.8479, "step": 8374 }, { "epoch": 3.79474399637517, "grad_norm": 0.18107483725171017, "learning_rate": 1.3345294471292452e-05, "loss": 0.843, "step": 8375 }, { "epoch": 3.795197100135931, "grad_norm": 0.1423251816466162, "learning_rate": 1.3335860925200486e-05, "loss": 0.8604, "step": 8376 }, { "epoch": 3.7956502038966926, "grad_norm": 0.16185335920493688, "learning_rate": 1.3326430047449619e-05, "loss": 0.8363, "step": 8377 }, { "epoch": 3.7961033076574537, "grad_norm": 0.17281494211744736, "learning_rate": 1.3317001838983599e-05, "loss": 0.8435, "step": 8378 }, { "epoch": 3.796556411418215, "grad_norm": 0.13523771177014485, "learning_rate": 1.3307576300745955e-05, "loss": 0.8417, "step": 8379 }, { "epoch": 3.797009515178976, "grad_norm": 0.18007980468559587, "learning_rate": 1.3298153433679906e-05, "loss": 0.8589, "step": 8380 }, { "epoch": 3.797462618939737, "grad_norm": 0.16007830189970373, "learning_rate": 1.3288733238728409e-05, "loss": 0.857, "step": 8381 }, { "epoch": 3.7979157227004983, "grad_norm": 0.17921217006396328, "learning_rate": 1.3279315716834198e-05, "loss": 0.8487, "step": 8382 }, { "epoch": 3.7983688264612594, "grad_norm": 0.15163885552407338, "learning_rate": 1.3269900868939662e-05, "loss": 0.84, "step": 8383 }, { "epoch": 3.798821930222021, "grad_norm": 0.151709569710286, "learning_rate": 1.3260488695987e-05, "loss": 0.849, "step": 8384 }, { "epoch": 3.799275033982782, "grad_norm": 0.1794539945385164, "learning_rate": 1.3251079198918081e-05, "loss": 0.8548, "step": 8385 }, { "epoch": 3.7997281377435432, "grad_norm": 0.12843883480990417, "learning_rate": 1.3241672378674565e-05, "loss": 0.8357, "step": 8386 }, { "epoch": 3.8001812415043044, "grad_norm": 0.15684918514703394, "learning_rate": 1.3232268236197791e-05, "loss": 0.8691, "step": 8387 }, { "epoch": 3.800634345265066, "grad_norm": 0.15120267120084557, "learning_rate": 1.3222866772428872e-05, "loss": 0.854, "step": 8388 }, { "epoch": 3.801087449025827, "grad_norm": 0.14664939965852844, "learning_rate": 1.3213467988308613e-05, "loss": 0.8637, "step": 8389 }, { "epoch": 3.801540552786588, "grad_norm": 0.16449775365863134, "learning_rate": 1.3204071884777591e-05, "loss": 0.8713, "step": 8390 }, { "epoch": 3.8019936565473493, "grad_norm": 0.1706743105523921, "learning_rate": 1.3194678462776094e-05, "loss": 0.8587, "step": 8391 }, { "epoch": 3.8024467603081105, "grad_norm": 0.1552357234463801, "learning_rate": 1.3185287723244122e-05, "loss": 0.8522, "step": 8392 }, { "epoch": 3.8028998640688716, "grad_norm": 0.21741412335699833, "learning_rate": 1.317589966712145e-05, "loss": 0.8616, "step": 8393 }, { "epoch": 3.8033529678296327, "grad_norm": 0.1433638522967412, "learning_rate": 1.3166514295347548e-05, "loss": 0.856, "step": 8394 }, { "epoch": 3.8038060715903943, "grad_norm": 0.20077872268228317, "learning_rate": 1.3157131608861646e-05, "loss": 0.8491, "step": 8395 }, { "epoch": 3.8042591753511554, "grad_norm": 0.14456840046722233, "learning_rate": 1.314775160860267e-05, "loss": 0.854, "step": 8396 }, { "epoch": 3.8047122791119166, "grad_norm": 0.21236041991582194, "learning_rate": 1.3138374295509326e-05, "loss": 0.8719, "step": 8397 }, { "epoch": 3.8051653828726777, "grad_norm": 0.15914332534312614, "learning_rate": 1.3128999670520007e-05, "loss": 0.874, "step": 8398 }, { "epoch": 3.8056184866334393, "grad_norm": 0.18039255700927387, "learning_rate": 1.311962773457284e-05, "loss": 0.8648, "step": 8399 }, { "epoch": 3.8060715903942004, "grad_norm": 0.17313608724158905, "learning_rate": 1.311025848860572e-05, "loss": 0.8466, "step": 8400 }, { "epoch": 3.8065246941549615, "grad_norm": 0.19146529213716207, "learning_rate": 1.3100891933556228e-05, "loss": 0.8464, "step": 8401 }, { "epoch": 3.8069777979157227, "grad_norm": 0.16919881293610073, "learning_rate": 1.3091528070361719e-05, "loss": 0.8403, "step": 8402 }, { "epoch": 3.807430901676484, "grad_norm": 0.16883571850192977, "learning_rate": 1.308216689995923e-05, "loss": 0.8474, "step": 8403 }, { "epoch": 3.807884005437245, "grad_norm": 0.165366597968754, "learning_rate": 1.3072808423285586e-05, "loss": 0.8695, "step": 8404 }, { "epoch": 3.8083371091980065, "grad_norm": 0.1598537065489675, "learning_rate": 1.3063452641277277e-05, "loss": 0.8445, "step": 8405 }, { "epoch": 3.8087902129587676, "grad_norm": 0.1559569897861675, "learning_rate": 1.3054099554870585e-05, "loss": 0.8616, "step": 8406 }, { "epoch": 3.8092433167195288, "grad_norm": 0.16082370103714877, "learning_rate": 1.3044749165001488e-05, "loss": 0.8785, "step": 8407 }, { "epoch": 3.80969642048029, "grad_norm": 0.16589623047855434, "learning_rate": 1.3035401472605682e-05, "loss": 0.8767, "step": 8408 }, { "epoch": 3.8101495242410515, "grad_norm": 0.18923636073750674, "learning_rate": 1.3026056478618637e-05, "loss": 0.868, "step": 8409 }, { "epoch": 3.8106026280018126, "grad_norm": 0.1604676303252083, "learning_rate": 1.3016714183975507e-05, "loss": 0.8193, "step": 8410 }, { "epoch": 3.8110557317625737, "grad_norm": 0.17973320638722137, "learning_rate": 1.300737458961122e-05, "loss": 0.8633, "step": 8411 }, { "epoch": 3.811508835523335, "grad_norm": 0.1816033361471804, "learning_rate": 1.2998037696460379e-05, "loss": 0.8507, "step": 8412 }, { "epoch": 3.811961939284096, "grad_norm": 0.186240127342365, "learning_rate": 1.2988703505457387e-05, "loss": 0.8596, "step": 8413 }, { "epoch": 3.812415043044857, "grad_norm": 0.1869561409173765, "learning_rate": 1.29793720175363e-05, "loss": 0.858, "step": 8414 }, { "epoch": 3.8128681468056183, "grad_norm": 0.19482003207568074, "learning_rate": 1.2970043233630971e-05, "loss": 0.8527, "step": 8415 }, { "epoch": 3.81332125056638, "grad_norm": 0.16382623171970573, "learning_rate": 1.2960717154674943e-05, "loss": 0.8797, "step": 8416 }, { "epoch": 3.813774354327141, "grad_norm": 0.14696193967828283, "learning_rate": 1.295139378160148e-05, "loss": 0.8579, "step": 8417 }, { "epoch": 3.814227458087902, "grad_norm": 0.18258558109974507, "learning_rate": 1.2942073115343621e-05, "loss": 0.8594, "step": 8418 }, { "epoch": 3.814680561848663, "grad_norm": 0.14529598702651766, "learning_rate": 1.293275515683408e-05, "loss": 0.8663, "step": 8419 }, { "epoch": 3.815133665609425, "grad_norm": 0.1822581537492411, "learning_rate": 1.2923439907005352e-05, "loss": 0.8674, "step": 8420 }, { "epoch": 3.815586769370186, "grad_norm": 0.14608052876345917, "learning_rate": 1.2914127366789613e-05, "loss": 0.8443, "step": 8421 }, { "epoch": 3.816039873130947, "grad_norm": 0.15293573504263727, "learning_rate": 1.2904817537118812e-05, "loss": 0.8744, "step": 8422 }, { "epoch": 3.816492976891708, "grad_norm": 0.14111233613124421, "learning_rate": 1.2895510418924592e-05, "loss": 0.8387, "step": 8423 }, { "epoch": 3.8169460806524693, "grad_norm": 0.14318062303781504, "learning_rate": 1.2886206013138325e-05, "loss": 0.875, "step": 8424 }, { "epoch": 3.8173991844132305, "grad_norm": 0.17486060868349265, "learning_rate": 1.2876904320691153e-05, "loss": 0.856, "step": 8425 }, { "epoch": 3.8178522881739916, "grad_norm": 0.16163419490438002, "learning_rate": 1.2867605342513887e-05, "loss": 0.8302, "step": 8426 }, { "epoch": 3.818305391934753, "grad_norm": 0.18326680649824392, "learning_rate": 1.2858309079537125e-05, "loss": 0.8562, "step": 8427 }, { "epoch": 3.8187584956955143, "grad_norm": 0.1518702550715421, "learning_rate": 1.2849015532691143e-05, "loss": 0.8609, "step": 8428 }, { "epoch": 3.8192115994562754, "grad_norm": 0.25004841262109495, "learning_rate": 1.2839724702905984e-05, "loss": 0.8797, "step": 8429 }, { "epoch": 3.8196647032170365, "grad_norm": 0.14803968018822178, "learning_rate": 1.2830436591111389e-05, "loss": 0.8654, "step": 8430 }, { "epoch": 3.820117806977798, "grad_norm": 0.2540398918475422, "learning_rate": 1.2821151198236855e-05, "loss": 0.8724, "step": 8431 }, { "epoch": 3.8205709107385593, "grad_norm": 0.21137532389554073, "learning_rate": 1.2811868525211582e-05, "loss": 0.8578, "step": 8432 }, { "epoch": 3.8210240144993204, "grad_norm": 0.16300565856660873, "learning_rate": 1.2802588572964499e-05, "loss": 0.849, "step": 8433 }, { "epoch": 3.8214771182600815, "grad_norm": 0.19476139105065293, "learning_rate": 1.2793311342424297e-05, "loss": 0.8587, "step": 8434 }, { "epoch": 3.8219302220208426, "grad_norm": 0.16438451668310486, "learning_rate": 1.2784036834519342e-05, "loss": 0.8637, "step": 8435 }, { "epoch": 3.822383325781604, "grad_norm": 0.15190301460568584, "learning_rate": 1.277476505017778e-05, "loss": 0.8502, "step": 8436 }, { "epoch": 3.8228364295423654, "grad_norm": 0.19795118720791277, "learning_rate": 1.2765495990327437e-05, "loss": 0.873, "step": 8437 }, { "epoch": 3.8232895333031265, "grad_norm": 0.14178522979838815, "learning_rate": 1.2756229655895913e-05, "loss": 0.8549, "step": 8438 }, { "epoch": 3.8237426370638876, "grad_norm": 0.22071471462932776, "learning_rate": 1.2746966047810499e-05, "loss": 0.8699, "step": 8439 }, { "epoch": 3.8241957408246487, "grad_norm": 0.14167075345953756, "learning_rate": 1.2737705166998224e-05, "loss": 0.8437, "step": 8440 }, { "epoch": 3.8246488445854103, "grad_norm": 0.19791631377221142, "learning_rate": 1.2728447014385833e-05, "loss": 0.8572, "step": 8441 }, { "epoch": 3.8251019483461715, "grad_norm": 0.1720035647944826, "learning_rate": 1.2719191590899839e-05, "loss": 0.8592, "step": 8442 }, { "epoch": 3.8255550521069326, "grad_norm": 0.14453347628035365, "learning_rate": 1.2709938897466438e-05, "loss": 0.8471, "step": 8443 }, { "epoch": 3.8260081558676937, "grad_norm": 0.1608979865555715, "learning_rate": 1.2700688935011551e-05, "loss": 0.857, "step": 8444 }, { "epoch": 3.826461259628455, "grad_norm": 0.14552460400781295, "learning_rate": 1.2691441704460879e-05, "loss": 0.854, "step": 8445 }, { "epoch": 3.826914363389216, "grad_norm": 0.16763317067339106, "learning_rate": 1.2682197206739777e-05, "loss": 0.8581, "step": 8446 }, { "epoch": 3.827367467149977, "grad_norm": 0.1353061989145462, "learning_rate": 1.267295544277339e-05, "loss": 0.8742, "step": 8447 }, { "epoch": 3.8278205709107387, "grad_norm": 0.17022612983934107, "learning_rate": 1.2663716413486547e-05, "loss": 0.861, "step": 8448 }, { "epoch": 3.8282736746715, "grad_norm": 0.1512400185805383, "learning_rate": 1.2654480119803831e-05, "loss": 0.8458, "step": 8449 }, { "epoch": 3.828726778432261, "grad_norm": 0.16278344890620217, "learning_rate": 1.2645246562649524e-05, "loss": 0.8656, "step": 8450 }, { "epoch": 3.829179882193022, "grad_norm": 0.15738840987698424, "learning_rate": 1.2636015742947665e-05, "loss": 0.8581, "step": 8451 }, { "epoch": 3.8296329859537837, "grad_norm": 0.15670037738938247, "learning_rate": 1.2626787661621997e-05, "loss": 0.8612, "step": 8452 }, { "epoch": 3.830086089714545, "grad_norm": 0.18825053664086563, "learning_rate": 1.2617562319595984e-05, "loss": 0.8285, "step": 8453 }, { "epoch": 3.830539193475306, "grad_norm": 0.17084536869445685, "learning_rate": 1.2608339717792846e-05, "loss": 0.8415, "step": 8454 }, { "epoch": 3.830992297236067, "grad_norm": 0.16660721795270955, "learning_rate": 1.2599119857135489e-05, "loss": 0.8573, "step": 8455 }, { "epoch": 3.831445400996828, "grad_norm": 0.15876617629273684, "learning_rate": 1.2589902738546589e-05, "loss": 0.8599, "step": 8456 }, { "epoch": 3.8318985047575893, "grad_norm": 0.14602262305655023, "learning_rate": 1.2580688362948497e-05, "loss": 0.8871, "step": 8457 }, { "epoch": 3.8323516085183504, "grad_norm": 0.19154072727053578, "learning_rate": 1.2571476731263346e-05, "loss": 0.8681, "step": 8458 }, { "epoch": 3.832804712279112, "grad_norm": 0.16243818605436286, "learning_rate": 1.256226784441295e-05, "loss": 0.8774, "step": 8459 }, { "epoch": 3.833257816039873, "grad_norm": 0.1640181031900831, "learning_rate": 1.2553061703318851e-05, "loss": 0.8473, "step": 8460 }, { "epoch": 3.8337109198006343, "grad_norm": 0.17798847674769036, "learning_rate": 1.2543858308902355e-05, "loss": 0.8684, "step": 8461 }, { "epoch": 3.8341640235613954, "grad_norm": 0.13430007621186604, "learning_rate": 1.2534657662084438e-05, "loss": 0.8592, "step": 8462 }, { "epoch": 3.834617127322157, "grad_norm": 0.15708910789488245, "learning_rate": 1.2525459763785851e-05, "loss": 0.8498, "step": 8463 }, { "epoch": 3.835070231082918, "grad_norm": 0.14208745440695614, "learning_rate": 1.2516264614927037e-05, "loss": 0.8536, "step": 8464 }, { "epoch": 3.8355233348436792, "grad_norm": 0.15945047221074657, "learning_rate": 1.2507072216428191e-05, "loss": 0.8753, "step": 8465 }, { "epoch": 3.8359764386044404, "grad_norm": 0.1374288080130318, "learning_rate": 1.2497882569209195e-05, "loss": 0.8661, "step": 8466 }, { "epoch": 3.8364295423652015, "grad_norm": 0.16542271026830543, "learning_rate": 1.2488695674189702e-05, "loss": 0.8528, "step": 8467 }, { "epoch": 3.8368826461259626, "grad_norm": 0.16109948694718748, "learning_rate": 1.2479511532289048e-05, "loss": 0.8719, "step": 8468 }, { "epoch": 3.837335749886724, "grad_norm": 0.16316717481443394, "learning_rate": 1.2470330144426307e-05, "loss": 0.8628, "step": 8469 }, { "epoch": 3.8377888536474853, "grad_norm": 0.1631413269325801, "learning_rate": 1.2461151511520306e-05, "loss": 0.8717, "step": 8470 }, { "epoch": 3.8382419574082465, "grad_norm": 0.198953106947128, "learning_rate": 1.245197563448954e-05, "loss": 0.8674, "step": 8471 }, { "epoch": 3.8386950611690076, "grad_norm": 0.16280101022614146, "learning_rate": 1.244280251425229e-05, "loss": 0.8483, "step": 8472 }, { "epoch": 3.839148164929769, "grad_norm": 0.2084580858250063, "learning_rate": 1.24336321517265e-05, "loss": 0.8494, "step": 8473 }, { "epoch": 3.8396012686905303, "grad_norm": 0.14947169027488208, "learning_rate": 1.2424464547829901e-05, "loss": 0.8541, "step": 8474 }, { "epoch": 3.8400543724512914, "grad_norm": 0.17211985599301075, "learning_rate": 1.241529970347989e-05, "loss": 0.8529, "step": 8475 }, { "epoch": 3.8405074762120526, "grad_norm": 0.14853849057763174, "learning_rate": 1.240613761959364e-05, "loss": 0.878, "step": 8476 }, { "epoch": 3.8409605799728137, "grad_norm": 0.18170945532146432, "learning_rate": 1.2396978297088e-05, "loss": 0.8322, "step": 8477 }, { "epoch": 3.841413683733575, "grad_norm": 0.14701797141062756, "learning_rate": 1.2387821736879561e-05, "loss": 0.8422, "step": 8478 }, { "epoch": 3.841866787494336, "grad_norm": 0.29550850117567645, "learning_rate": 1.2378667939884665e-05, "loss": 0.8574, "step": 8479 }, { "epoch": 3.8423198912550975, "grad_norm": 0.13801129567694415, "learning_rate": 1.236951690701932e-05, "loss": 0.8461, "step": 8480 }, { "epoch": 3.8427729950158587, "grad_norm": 0.181307930991119, "learning_rate": 1.236036863919933e-05, "loss": 0.8629, "step": 8481 }, { "epoch": 3.84322609877662, "grad_norm": 0.15925180684057175, "learning_rate": 1.2351223137340146e-05, "loss": 0.8697, "step": 8482 }, { "epoch": 3.843679202537381, "grad_norm": 0.15320690513175625, "learning_rate": 1.2342080402357008e-05, "loss": 0.841, "step": 8483 }, { "epoch": 3.8441323062981425, "grad_norm": 0.16296091941580257, "learning_rate": 1.2332940435164838e-05, "loss": 0.8625, "step": 8484 }, { "epoch": 3.8445854100589036, "grad_norm": 0.16872065056384344, "learning_rate": 1.2323803236678282e-05, "loss": 0.8637, "step": 8485 }, { "epoch": 3.8450385138196648, "grad_norm": 0.1605506626769784, "learning_rate": 1.231466880781174e-05, "loss": 0.8305, "step": 8486 }, { "epoch": 3.845491617580426, "grad_norm": 0.18631673917918523, "learning_rate": 1.2305537149479298e-05, "loss": 0.8543, "step": 8487 }, { "epoch": 3.845944721341187, "grad_norm": 0.1746444345372834, "learning_rate": 1.2296408262594803e-05, "loss": 0.8345, "step": 8488 }, { "epoch": 3.846397825101948, "grad_norm": 0.1515568373258212, "learning_rate": 1.2287282148071777e-05, "loss": 0.8468, "step": 8489 }, { "epoch": 3.8468509288627093, "grad_norm": 0.14062555582507655, "learning_rate": 1.2278158806823516e-05, "loss": 0.8354, "step": 8490 }, { "epoch": 3.847304032623471, "grad_norm": 0.1555124107590674, "learning_rate": 1.2269038239762994e-05, "loss": 0.8663, "step": 8491 }, { "epoch": 3.847757136384232, "grad_norm": 0.1493643486429064, "learning_rate": 1.2259920447802943e-05, "loss": 0.859, "step": 8492 }, { "epoch": 3.848210240144993, "grad_norm": 0.14931220040668275, "learning_rate": 1.22508054318558e-05, "loss": 0.8262, "step": 8493 }, { "epoch": 3.8486633439057543, "grad_norm": 0.16162981658281317, "learning_rate": 1.2241693192833699e-05, "loss": 0.8596, "step": 8494 }, { "epoch": 3.849116447666516, "grad_norm": 0.1583210094708569, "learning_rate": 1.223258373164856e-05, "loss": 0.8607, "step": 8495 }, { "epoch": 3.849569551427277, "grad_norm": 0.1362324821546752, "learning_rate": 1.2223477049211972e-05, "loss": 0.8501, "step": 8496 }, { "epoch": 3.850022655188038, "grad_norm": 0.16453480437848983, "learning_rate": 1.2214373146435255e-05, "loss": 0.8549, "step": 8497 }, { "epoch": 3.8504757589487992, "grad_norm": 0.13643064283596723, "learning_rate": 1.2205272024229449e-05, "loss": 0.8692, "step": 8498 }, { "epoch": 3.8509288627095604, "grad_norm": 0.1719406892746586, "learning_rate": 1.219617368350535e-05, "loss": 0.8572, "step": 8499 }, { "epoch": 3.8513819664703215, "grad_norm": 0.13747436036862187, "learning_rate": 1.2187078125173427e-05, "loss": 0.8558, "step": 8500 }, { "epoch": 3.851835070231083, "grad_norm": 0.16697946144512285, "learning_rate": 1.2177985350143913e-05, "loss": 0.8696, "step": 8501 }, { "epoch": 3.852288173991844, "grad_norm": 0.14495959131821343, "learning_rate": 1.2168895359326719e-05, "loss": 0.8592, "step": 8502 }, { "epoch": 3.8527412777526053, "grad_norm": 0.1619995663147737, "learning_rate": 1.2159808153631528e-05, "loss": 0.8747, "step": 8503 }, { "epoch": 3.8531943815133665, "grad_norm": 0.1496704832173916, "learning_rate": 1.2150723733967701e-05, "loss": 0.8579, "step": 8504 }, { "epoch": 3.853647485274128, "grad_norm": 0.14533776306500404, "learning_rate": 1.214164210124433e-05, "loss": 0.8524, "step": 8505 }, { "epoch": 3.854100589034889, "grad_norm": 0.171577464727123, "learning_rate": 1.2132563256370253e-05, "loss": 0.8581, "step": 8506 }, { "epoch": 3.8545536927956503, "grad_norm": 0.13851182179828145, "learning_rate": 1.2123487200253994e-05, "loss": 0.8533, "step": 8507 }, { "epoch": 3.8550067965564114, "grad_norm": 0.21925515331509032, "learning_rate": 1.2114413933803828e-05, "loss": 0.8618, "step": 8508 }, { "epoch": 3.8554599003171726, "grad_norm": 0.1582054351237611, "learning_rate": 1.2105343457927723e-05, "loss": 0.87, "step": 8509 }, { "epoch": 3.8559130040779337, "grad_norm": 0.18619840218884037, "learning_rate": 1.2096275773533406e-05, "loss": 0.8544, "step": 8510 }, { "epoch": 3.856366107838695, "grad_norm": 0.15741195835035945, "learning_rate": 1.208721088152827e-05, "loss": 0.8508, "step": 8511 }, { "epoch": 3.8568192115994564, "grad_norm": 0.1549684175123197, "learning_rate": 1.207814878281949e-05, "loss": 0.8672, "step": 8512 }, { "epoch": 3.8572723153602175, "grad_norm": 0.13720927186906937, "learning_rate": 1.206908947831391e-05, "loss": 0.8606, "step": 8513 }, { "epoch": 3.8577254191209787, "grad_norm": 0.16016208051878872, "learning_rate": 1.2060032968918116e-05, "loss": 0.8455, "step": 8514 }, { "epoch": 3.85817852288174, "grad_norm": 0.13972971474881715, "learning_rate": 1.2050979255538428e-05, "loss": 0.8348, "step": 8515 }, { "epoch": 3.8586316266425014, "grad_norm": 0.20240452592936983, "learning_rate": 1.2041928339080849e-05, "loss": 0.8644, "step": 8516 }, { "epoch": 3.8590847304032625, "grad_norm": 0.1696832071605433, "learning_rate": 1.2032880220451148e-05, "loss": 0.8619, "step": 8517 }, { "epoch": 3.8595378341640236, "grad_norm": 0.17004621763811612, "learning_rate": 1.2023834900554773e-05, "loss": 0.8537, "step": 8518 }, { "epoch": 3.8599909379247848, "grad_norm": 0.1512689142224882, "learning_rate": 1.201479238029693e-05, "loss": 0.8675, "step": 8519 }, { "epoch": 3.860444041685546, "grad_norm": 0.15493161910919903, "learning_rate": 1.2005752660582508e-05, "loss": 0.8639, "step": 8520 }, { "epoch": 3.860897145446307, "grad_norm": 0.1436251865792388, "learning_rate": 1.199671574231613e-05, "loss": 0.8562, "step": 8521 }, { "epoch": 3.861350249207068, "grad_norm": 0.16501612451488662, "learning_rate": 1.1987681626402154e-05, "loss": 0.8621, "step": 8522 }, { "epoch": 3.8618033529678297, "grad_norm": 0.155915525646624, "learning_rate": 1.1978650313744628e-05, "loss": 0.8655, "step": 8523 }, { "epoch": 3.862256456728591, "grad_norm": 0.14359658764484706, "learning_rate": 1.1969621805247362e-05, "loss": 0.8672, "step": 8524 }, { "epoch": 3.862709560489352, "grad_norm": 0.17653819829759199, "learning_rate": 1.1960596101813829e-05, "loss": 0.8362, "step": 8525 }, { "epoch": 3.863162664250113, "grad_norm": 0.136508779421055, "learning_rate": 1.1951573204347278e-05, "loss": 0.8417, "step": 8526 }, { "epoch": 3.8636157680108747, "grad_norm": 0.16839843168959054, "learning_rate": 1.1942553113750624e-05, "loss": 0.874, "step": 8527 }, { "epoch": 3.864068871771636, "grad_norm": 0.15302353443037317, "learning_rate": 1.1933535830926557e-05, "loss": 0.8502, "step": 8528 }, { "epoch": 3.864521975532397, "grad_norm": 0.1569843220784525, "learning_rate": 1.1924521356777445e-05, "loss": 0.8482, "step": 8529 }, { "epoch": 3.864975079293158, "grad_norm": 0.12789008060158213, "learning_rate": 1.1915509692205371e-05, "loss": 0.8571, "step": 8530 }, { "epoch": 3.865428183053919, "grad_norm": 0.14008228928078048, "learning_rate": 1.1906500838112183e-05, "loss": 0.8591, "step": 8531 }, { "epoch": 3.8658812868146804, "grad_norm": 0.11246423172765299, "learning_rate": 1.1897494795399385e-05, "loss": 0.8308, "step": 8532 }, { "epoch": 3.866334390575442, "grad_norm": 0.12856964439996632, "learning_rate": 1.188849156496826e-05, "loss": 0.857, "step": 8533 }, { "epoch": 3.866787494336203, "grad_norm": 0.13025739424283342, "learning_rate": 1.1879491147719762e-05, "loss": 0.8493, "step": 8534 }, { "epoch": 3.867240598096964, "grad_norm": 0.12583828733339972, "learning_rate": 1.1870493544554602e-05, "loss": 0.8488, "step": 8535 }, { "epoch": 3.8676937018577253, "grad_norm": 0.13530599183409017, "learning_rate": 1.1861498756373164e-05, "loss": 0.8669, "step": 8536 }, { "epoch": 3.868146805618487, "grad_norm": 0.11721552681137029, "learning_rate": 1.1852506784075613e-05, "loss": 0.8417, "step": 8537 }, { "epoch": 3.868599909379248, "grad_norm": 0.12744669618833812, "learning_rate": 1.1843517628561769e-05, "loss": 0.8662, "step": 8538 }, { "epoch": 3.869053013140009, "grad_norm": 0.13890516319954102, "learning_rate": 1.1834531290731194e-05, "loss": 0.865, "step": 8539 }, { "epoch": 3.8695061169007703, "grad_norm": 0.1432245362228614, "learning_rate": 1.1825547771483196e-05, "loss": 0.8534, "step": 8540 }, { "epoch": 3.8699592206615314, "grad_norm": 0.15678756565234217, "learning_rate": 1.1816567071716754e-05, "loss": 0.8551, "step": 8541 }, { "epoch": 3.8704123244222925, "grad_norm": 0.16018444462670833, "learning_rate": 1.18075891923306e-05, "loss": 0.8396, "step": 8542 }, { "epoch": 3.8708654281830537, "grad_norm": 0.1501558325833406, "learning_rate": 1.1798614134223154e-05, "loss": 0.8461, "step": 8543 }, { "epoch": 3.8713185319438153, "grad_norm": 0.1709752019967061, "learning_rate": 1.1789641898292601e-05, "loss": 0.8688, "step": 8544 }, { "epoch": 3.8717716357045764, "grad_norm": 0.16676132849226596, "learning_rate": 1.178067248543679e-05, "loss": 0.8677, "step": 8545 }, { "epoch": 3.8722247394653375, "grad_norm": 0.16478990550183403, "learning_rate": 1.1771705896553302e-05, "loss": 0.8508, "step": 8546 }, { "epoch": 3.8726778432260986, "grad_norm": 0.1557864063260858, "learning_rate": 1.1762742132539469e-05, "loss": 0.8639, "step": 8547 }, { "epoch": 3.8731309469868602, "grad_norm": 0.15630408608816704, "learning_rate": 1.1753781194292296e-05, "loss": 0.8483, "step": 8548 }, { "epoch": 3.8735840507476214, "grad_norm": 0.1553800136617688, "learning_rate": 1.1744823082708537e-05, "loss": 0.8693, "step": 8549 }, { "epoch": 3.8740371545083825, "grad_norm": 0.14973850407172065, "learning_rate": 1.1735867798684639e-05, "loss": 0.8357, "step": 8550 }, { "epoch": 3.8744902582691436, "grad_norm": 0.16297763873650842, "learning_rate": 1.1726915343116793e-05, "loss": 0.8206, "step": 8551 }, { "epoch": 3.8749433620299047, "grad_norm": 0.17159553445859277, "learning_rate": 1.171796571690087e-05, "loss": 0.8553, "step": 8552 }, { "epoch": 3.875396465790666, "grad_norm": 0.1529539333619934, "learning_rate": 1.170901892093252e-05, "loss": 0.8543, "step": 8553 }, { "epoch": 3.875849569551427, "grad_norm": 0.21499587419933416, "learning_rate": 1.170007495610701e-05, "loss": 0.8601, "step": 8554 }, { "epoch": 3.8763026733121886, "grad_norm": 0.1535538683270342, "learning_rate": 1.1691133823319434e-05, "loss": 0.8512, "step": 8555 }, { "epoch": 3.8767557770729497, "grad_norm": 0.22959528676100321, "learning_rate": 1.1682195523464514e-05, "loss": 0.8579, "step": 8556 }, { "epoch": 3.877208880833711, "grad_norm": 0.15762192903925, "learning_rate": 1.167326005743676e-05, "loss": 0.8499, "step": 8557 }, { "epoch": 3.877661984594472, "grad_norm": 0.21367731927430894, "learning_rate": 1.1664327426130346e-05, "loss": 0.8774, "step": 8558 }, { "epoch": 3.8781150883552336, "grad_norm": 0.17134559412469139, "learning_rate": 1.1655397630439174e-05, "loss": 0.848, "step": 8559 }, { "epoch": 3.8785681921159947, "grad_norm": 0.18592635981765934, "learning_rate": 1.1646470671256886e-05, "loss": 0.8538, "step": 8560 }, { "epoch": 3.879021295876756, "grad_norm": 0.18209236537791887, "learning_rate": 1.1637546549476806e-05, "loss": 0.8445, "step": 8561 }, { "epoch": 3.879474399637517, "grad_norm": 0.17520362637851264, "learning_rate": 1.1628625265992009e-05, "loss": 0.8506, "step": 8562 }, { "epoch": 3.879927503398278, "grad_norm": 0.20361952127034266, "learning_rate": 1.1619706821695247e-05, "loss": 0.8603, "step": 8563 }, { "epoch": 3.880380607159039, "grad_norm": 0.14299463519104283, "learning_rate": 1.1610791217479037e-05, "loss": 0.8714, "step": 8564 }, { "epoch": 3.880833710919801, "grad_norm": 0.21247244808499988, "learning_rate": 1.1601878454235566e-05, "loss": 0.8705, "step": 8565 }, { "epoch": 3.881286814680562, "grad_norm": 0.13813819809484423, "learning_rate": 1.1592968532856745e-05, "loss": 0.8543, "step": 8566 }, { "epoch": 3.881739918441323, "grad_norm": 0.19104812882855815, "learning_rate": 1.1584061454234235e-05, "loss": 0.858, "step": 8567 }, { "epoch": 3.882193022202084, "grad_norm": 0.13322818835269673, "learning_rate": 1.157515721925936e-05, "loss": 0.8523, "step": 8568 }, { "epoch": 3.8826461259628458, "grad_norm": 0.17016267249506767, "learning_rate": 1.1566255828823216e-05, "loss": 0.8708, "step": 8569 }, { "epoch": 3.883099229723607, "grad_norm": 0.1353957528199625, "learning_rate": 1.1557357283816559e-05, "loss": 0.8631, "step": 8570 }, { "epoch": 3.883552333484368, "grad_norm": 0.1383421332690397, "learning_rate": 1.1548461585129909e-05, "loss": 0.8618, "step": 8571 }, { "epoch": 3.884005437245129, "grad_norm": 0.1517368962764463, "learning_rate": 1.1539568733653455e-05, "loss": 0.8782, "step": 8572 }, { "epoch": 3.8844585410058903, "grad_norm": 0.13278562434790944, "learning_rate": 1.1530678730277151e-05, "loss": 0.8556, "step": 8573 }, { "epoch": 3.8849116447666514, "grad_norm": 0.1583849398069538, "learning_rate": 1.152179157589063e-05, "loss": 0.8609, "step": 8574 }, { "epoch": 3.8853647485274125, "grad_norm": 0.13450335773236907, "learning_rate": 1.1512907271383234e-05, "loss": 0.8518, "step": 8575 }, { "epoch": 3.885817852288174, "grad_norm": 0.15255964968925534, "learning_rate": 1.1504025817644057e-05, "loss": 0.8535, "step": 8576 }, { "epoch": 3.8862709560489352, "grad_norm": 0.1503606111691112, "learning_rate": 1.1495147215561872e-05, "loss": 0.8569, "step": 8577 }, { "epoch": 3.8867240598096964, "grad_norm": 0.1356978369022614, "learning_rate": 1.1486271466025194e-05, "loss": 0.8556, "step": 8578 }, { "epoch": 3.8871771635704575, "grad_norm": 0.15701675139441343, "learning_rate": 1.1477398569922223e-05, "loss": 0.8788, "step": 8579 }, { "epoch": 3.887630267331219, "grad_norm": 0.1362356399344266, "learning_rate": 1.1468528528140906e-05, "loss": 0.8479, "step": 8580 }, { "epoch": 3.88808337109198, "grad_norm": 0.14360876417055196, "learning_rate": 1.1459661341568883e-05, "loss": 0.8501, "step": 8581 }, { "epoch": 3.8885364748527413, "grad_norm": 0.13842812225309511, "learning_rate": 1.1450797011093501e-05, "loss": 0.8938, "step": 8582 }, { "epoch": 3.8889895786135025, "grad_norm": 0.15869399174503535, "learning_rate": 1.1441935537601858e-05, "loss": 0.8527, "step": 8583 }, { "epoch": 3.8894426823742636, "grad_norm": 0.1420899761515793, "learning_rate": 1.1433076921980715e-05, "loss": 0.8587, "step": 8584 }, { "epoch": 3.8898957861350247, "grad_norm": 0.18015763689528244, "learning_rate": 1.1424221165116602e-05, "loss": 0.8578, "step": 8585 }, { "epoch": 3.890348889895786, "grad_norm": 0.14938234964164995, "learning_rate": 1.1415368267895706e-05, "loss": 0.8259, "step": 8586 }, { "epoch": 3.8908019936565474, "grad_norm": 0.1662288038848901, "learning_rate": 1.1406518231203983e-05, "loss": 0.8518, "step": 8587 }, { "epoch": 3.8912550974173086, "grad_norm": 0.1848572167151051, "learning_rate": 1.139767105592705e-05, "loss": 0.8614, "step": 8588 }, { "epoch": 3.8917082011780697, "grad_norm": 0.1656226825572055, "learning_rate": 1.138882674295029e-05, "loss": 0.8731, "step": 8589 }, { "epoch": 3.892161304938831, "grad_norm": 0.14859198721324388, "learning_rate": 1.1379985293158762e-05, "loss": 0.8546, "step": 8590 }, { "epoch": 3.8926144086995924, "grad_norm": 0.16957055091673595, "learning_rate": 1.1371146707437238e-05, "loss": 0.8729, "step": 8591 }, { "epoch": 3.8930675124603535, "grad_norm": 0.1485041304323603, "learning_rate": 1.136231098667024e-05, "loss": 0.8541, "step": 8592 }, { "epoch": 3.8935206162211147, "grad_norm": 0.14765482342724603, "learning_rate": 1.135347813174195e-05, "loss": 0.8518, "step": 8593 }, { "epoch": 3.893973719981876, "grad_norm": 0.12089258893856272, "learning_rate": 1.134464814353632e-05, "loss": 0.8733, "step": 8594 }, { "epoch": 3.894426823742637, "grad_norm": 0.13610646863779355, "learning_rate": 1.1335821022936964e-05, "loss": 0.8463, "step": 8595 }, { "epoch": 3.894879927503398, "grad_norm": 0.13715654272716793, "learning_rate": 1.1326996770827257e-05, "loss": 0.8605, "step": 8596 }, { "epoch": 3.8953330312641596, "grad_norm": 0.13701769432066946, "learning_rate": 1.1318175388090231e-05, "loss": 0.8575, "step": 8597 }, { "epoch": 3.8957861350249208, "grad_norm": 0.13845091159885553, "learning_rate": 1.1309356875608693e-05, "loss": 0.8691, "step": 8598 }, { "epoch": 3.896239238785682, "grad_norm": 0.13464884158929, "learning_rate": 1.1300541234265112e-05, "loss": 0.861, "step": 8599 }, { "epoch": 3.896692342546443, "grad_norm": 0.14241185030924747, "learning_rate": 1.1291728464941688e-05, "loss": 0.8669, "step": 8600 }, { "epoch": 3.8971454463072046, "grad_norm": 0.13841021423087402, "learning_rate": 1.1282918568520351e-05, "loss": 0.8417, "step": 8601 }, { "epoch": 3.8975985500679657, "grad_norm": 0.13663471769002034, "learning_rate": 1.1274111545882702e-05, "loss": 0.8397, "step": 8602 }, { "epoch": 3.898051653828727, "grad_norm": 0.1453777364281865, "learning_rate": 1.1265307397910115e-05, "loss": 0.8506, "step": 8603 }, { "epoch": 3.898504757589488, "grad_norm": 0.13446349910830394, "learning_rate": 1.1256506125483604e-05, "loss": 0.8444, "step": 8604 }, { "epoch": 3.898957861350249, "grad_norm": 0.15333122019201073, "learning_rate": 1.1247707729483963e-05, "loss": 0.8282, "step": 8605 }, { "epoch": 3.8994109651110103, "grad_norm": 0.13282914753472844, "learning_rate": 1.1238912210791653e-05, "loss": 0.8359, "step": 8606 }, { "epoch": 3.8998640688717714, "grad_norm": 0.15685892605084487, "learning_rate": 1.1230119570286853e-05, "loss": 0.8674, "step": 8607 }, { "epoch": 3.900317172632533, "grad_norm": 0.13880978819241305, "learning_rate": 1.1221329808849481e-05, "loss": 0.8598, "step": 8608 }, { "epoch": 3.900770276393294, "grad_norm": 0.16598238767386794, "learning_rate": 1.121254292735913e-05, "loss": 0.8473, "step": 8609 }, { "epoch": 3.9012233801540552, "grad_norm": 0.1471464657719892, "learning_rate": 1.1203758926695159e-05, "loss": 0.8627, "step": 8610 }, { "epoch": 3.9016764839148164, "grad_norm": 0.15677713925014392, "learning_rate": 1.1194977807736552e-05, "loss": 0.8334, "step": 8611 }, { "epoch": 3.902129587675578, "grad_norm": 0.14367639118795786, "learning_rate": 1.1186199571362089e-05, "loss": 0.8612, "step": 8612 }, { "epoch": 3.902582691436339, "grad_norm": 0.15745238245836812, "learning_rate": 1.117742421845021e-05, "loss": 0.8384, "step": 8613 }, { "epoch": 3.9030357951971, "grad_norm": 0.1813753456887977, "learning_rate": 1.1168651749879106e-05, "loss": 0.8613, "step": 8614 }, { "epoch": 3.9034888989578613, "grad_norm": 0.14290057914105803, "learning_rate": 1.1159882166526637e-05, "loss": 0.8536, "step": 8615 }, { "epoch": 3.9039420027186225, "grad_norm": 0.1744271492596852, "learning_rate": 1.1151115469270412e-05, "loss": 0.8797, "step": 8616 }, { "epoch": 3.9043951064793836, "grad_norm": 0.153137180196252, "learning_rate": 1.114235165898773e-05, "loss": 0.8441, "step": 8617 }, { "epoch": 3.9048482102401447, "grad_norm": 0.152502498681608, "learning_rate": 1.1133590736555586e-05, "loss": 0.8524, "step": 8618 }, { "epoch": 3.9053013140009063, "grad_norm": 0.15416647342128278, "learning_rate": 1.1124832702850736e-05, "loss": 0.8386, "step": 8619 }, { "epoch": 3.9057544177616674, "grad_norm": 0.14628994511737853, "learning_rate": 1.1116077558749589e-05, "loss": 0.8355, "step": 8620 }, { "epoch": 3.9062075215224286, "grad_norm": 0.14773523393884908, "learning_rate": 1.110732530512832e-05, "loss": 0.8486, "step": 8621 }, { "epoch": 3.9066606252831897, "grad_norm": 0.15027418148000785, "learning_rate": 1.1098575942862757e-05, "loss": 0.8615, "step": 8622 }, { "epoch": 3.9071137290439513, "grad_norm": 0.15995304216617431, "learning_rate": 1.1089829472828497e-05, "loss": 0.857, "step": 8623 }, { "epoch": 3.9075668328047124, "grad_norm": 0.16796952476145993, "learning_rate": 1.1081085895900792e-05, "loss": 0.8663, "step": 8624 }, { "epoch": 3.9080199365654735, "grad_norm": 0.16827557461007, "learning_rate": 1.1072345212954661e-05, "loss": 0.8589, "step": 8625 }, { "epoch": 3.9084730403262347, "grad_norm": 0.18978049111790468, "learning_rate": 1.1063607424864787e-05, "loss": 0.8786, "step": 8626 }, { "epoch": 3.908926144086996, "grad_norm": 0.170314780728482, "learning_rate": 1.1054872532505567e-05, "loss": 0.8474, "step": 8627 }, { "epoch": 3.909379247847757, "grad_norm": 0.21147833915462103, "learning_rate": 1.1046140536751148e-05, "loss": 0.8659, "step": 8628 }, { "epoch": 3.9098323516085185, "grad_norm": 0.19404959268022312, "learning_rate": 1.1037411438475344e-05, "loss": 0.8679, "step": 8629 }, { "epoch": 3.9102854553692796, "grad_norm": 0.2420981715161454, "learning_rate": 1.1028685238551704e-05, "loss": 0.8618, "step": 8630 }, { "epoch": 3.9107385591300408, "grad_norm": 0.19771850138489344, "learning_rate": 1.1019961937853468e-05, "loss": 0.8596, "step": 8631 }, { "epoch": 3.911191662890802, "grad_norm": 0.21960101738854249, "learning_rate": 1.1011241537253614e-05, "loss": 0.8609, "step": 8632 }, { "epoch": 3.9116447666515635, "grad_norm": 0.22115616030996474, "learning_rate": 1.1002524037624793e-05, "loss": 0.8356, "step": 8633 }, { "epoch": 3.9120978704123246, "grad_norm": 0.1812366434565941, "learning_rate": 1.0993809439839404e-05, "loss": 0.8458, "step": 8634 }, { "epoch": 3.9125509741730857, "grad_norm": 0.21685531448703993, "learning_rate": 1.0985097744769529e-05, "loss": 0.8509, "step": 8635 }, { "epoch": 3.913004077933847, "grad_norm": 0.14175194256042553, "learning_rate": 1.0976388953286952e-05, "loss": 0.8559, "step": 8636 }, { "epoch": 3.913457181694608, "grad_norm": 0.2072203668640189, "learning_rate": 1.0967683066263208e-05, "loss": 0.8589, "step": 8637 }, { "epoch": 3.913910285455369, "grad_norm": 0.13632360011625297, "learning_rate": 1.095898008456949e-05, "loss": 0.8545, "step": 8638 }, { "epoch": 3.9143633892161303, "grad_norm": 0.16785298519063913, "learning_rate": 1.0950280009076745e-05, "loss": 0.8595, "step": 8639 }, { "epoch": 3.914816492976892, "grad_norm": 0.14459816712513976, "learning_rate": 1.0941582840655593e-05, "loss": 0.8489, "step": 8640 }, { "epoch": 3.915269596737653, "grad_norm": 0.18789318232640478, "learning_rate": 1.09328885801764e-05, "loss": 0.8479, "step": 8641 }, { "epoch": 3.915722700498414, "grad_norm": 0.1513918088760998, "learning_rate": 1.0924197228509206e-05, "loss": 0.8492, "step": 8642 }, { "epoch": 3.916175804259175, "grad_norm": 0.18910934471118754, "learning_rate": 1.0915508786523766e-05, "loss": 0.8567, "step": 8643 }, { "epoch": 3.916628908019937, "grad_norm": 0.1408799050870736, "learning_rate": 1.0906823255089574e-05, "loss": 0.8518, "step": 8644 }, { "epoch": 3.917082011780698, "grad_norm": 0.17709341779434454, "learning_rate": 1.0898140635075789e-05, "loss": 0.8492, "step": 8645 }, { "epoch": 3.917535115541459, "grad_norm": 0.1353320791829695, "learning_rate": 1.088946092735132e-05, "loss": 0.8406, "step": 8646 }, { "epoch": 3.91798821930222, "grad_norm": 0.1623889943478645, "learning_rate": 1.0880784132784744e-05, "loss": 0.8525, "step": 8647 }, { "epoch": 3.9184413230629813, "grad_norm": 0.16664302356896996, "learning_rate": 1.0872110252244395e-05, "loss": 0.8692, "step": 8648 }, { "epoch": 3.9188944268237424, "grad_norm": 0.13512902956418515, "learning_rate": 1.0863439286598258e-05, "loss": 0.854, "step": 8649 }, { "epoch": 3.9193475305845036, "grad_norm": 0.22790653652670595, "learning_rate": 1.0854771236714083e-05, "loss": 0.8706, "step": 8650 }, { "epoch": 3.919800634345265, "grad_norm": 0.1439991916632017, "learning_rate": 1.084610610345929e-05, "loss": 0.8432, "step": 8651 }, { "epoch": 3.9202537381060263, "grad_norm": 0.15289321401564418, "learning_rate": 1.083744388770101e-05, "loss": 0.8337, "step": 8652 }, { "epoch": 3.9207068418667874, "grad_norm": 0.15885110509812844, "learning_rate": 1.0828784590306106e-05, "loss": 0.853, "step": 8653 }, { "epoch": 3.9211599456275485, "grad_norm": 0.1498981988751178, "learning_rate": 1.0820128212141116e-05, "loss": 0.8594, "step": 8654 }, { "epoch": 3.92161304938831, "grad_norm": 0.16082229222012243, "learning_rate": 1.0811474754072323e-05, "loss": 0.8578, "step": 8655 }, { "epoch": 3.9220661531490713, "grad_norm": 0.14676266485102255, "learning_rate": 1.0802824216965684e-05, "loss": 0.8435, "step": 8656 }, { "epoch": 3.9225192569098324, "grad_norm": 0.14820473383709984, "learning_rate": 1.079417660168689e-05, "loss": 0.8531, "step": 8657 }, { "epoch": 3.9229723606705935, "grad_norm": 0.12936875597084438, "learning_rate": 1.0785531909101321e-05, "loss": 0.8663, "step": 8658 }, { "epoch": 3.9234254644313546, "grad_norm": 0.1536776563510941, "learning_rate": 1.0776890140074058e-05, "loss": 0.8516, "step": 8659 }, { "epoch": 3.923878568192116, "grad_norm": 0.13833583959877643, "learning_rate": 1.0768251295469926e-05, "loss": 0.8649, "step": 8660 }, { "epoch": 3.9243316719528774, "grad_norm": 0.18049658186269577, "learning_rate": 1.0759615376153412e-05, "loss": 0.8713, "step": 8661 }, { "epoch": 3.9247847757136385, "grad_norm": 0.12783291933485982, "learning_rate": 1.075098238298876e-05, "loss": 0.842, "step": 8662 }, { "epoch": 3.9252378794743996, "grad_norm": 0.16817535339106301, "learning_rate": 1.0742352316839856e-05, "loss": 0.8649, "step": 8663 }, { "epoch": 3.9256909832351607, "grad_norm": 0.14045468079191747, "learning_rate": 1.0733725178570368e-05, "loss": 0.8583, "step": 8664 }, { "epoch": 3.9261440869959223, "grad_norm": 0.14235300868779202, "learning_rate": 1.0725100969043605e-05, "loss": 0.872, "step": 8665 }, { "epoch": 3.9265971907566835, "grad_norm": 0.1411340940821661, "learning_rate": 1.071647968912263e-05, "loss": 0.8399, "step": 8666 }, { "epoch": 3.9270502945174446, "grad_norm": 0.17009029646797932, "learning_rate": 1.070786133967019e-05, "loss": 0.8468, "step": 8667 }, { "epoch": 3.9275033982782057, "grad_norm": 0.12842962680866046, "learning_rate": 1.0699245921548735e-05, "loss": 0.857, "step": 8668 }, { "epoch": 3.927956502038967, "grad_norm": 0.14901200105230844, "learning_rate": 1.0690633435620428e-05, "loss": 0.8694, "step": 8669 }, { "epoch": 3.928409605799728, "grad_norm": 0.1408537649148267, "learning_rate": 1.0682023882747155e-05, "loss": 0.8524, "step": 8670 }, { "epoch": 3.928862709560489, "grad_norm": 0.14059759701324395, "learning_rate": 1.067341726379048e-05, "loss": 0.8492, "step": 8671 }, { "epoch": 3.9293158133212507, "grad_norm": 0.1505652427872954, "learning_rate": 1.0664813579611683e-05, "loss": 0.8427, "step": 8672 }, { "epoch": 3.929768917082012, "grad_norm": 0.12159703334802538, "learning_rate": 1.0656212831071774e-05, "loss": 0.8607, "step": 8673 }, { "epoch": 3.930222020842773, "grad_norm": 0.16968109360299566, "learning_rate": 1.064761501903143e-05, "loss": 0.8464, "step": 8674 }, { "epoch": 3.930675124603534, "grad_norm": 0.125477885017927, "learning_rate": 1.0639020144351071e-05, "loss": 0.8716, "step": 8675 }, { "epoch": 3.9311282283642957, "grad_norm": 0.14945847737021048, "learning_rate": 1.0630428207890783e-05, "loss": 0.8538, "step": 8676 }, { "epoch": 3.931581332125057, "grad_norm": 0.14710306869111237, "learning_rate": 1.0621839210510405e-05, "loss": 0.8549, "step": 8677 }, { "epoch": 3.932034435885818, "grad_norm": 0.16281025317433204, "learning_rate": 1.0613253153069452e-05, "loss": 0.8421, "step": 8678 }, { "epoch": 3.932487539646579, "grad_norm": 0.13461703282896784, "learning_rate": 1.060467003642713e-05, "loss": 0.8555, "step": 8679 }, { "epoch": 3.93294064340734, "grad_norm": 0.16375184442864618, "learning_rate": 1.0596089861442396e-05, "loss": 0.8638, "step": 8680 }, { "epoch": 3.9333937471681013, "grad_norm": 0.17148811118757312, "learning_rate": 1.0587512628973865e-05, "loss": 0.8585, "step": 8681 }, { "epoch": 3.9338468509288624, "grad_norm": 0.1451199638361889, "learning_rate": 1.0578938339879907e-05, "loss": 0.8551, "step": 8682 }, { "epoch": 3.934299954689624, "grad_norm": 0.18139655492635, "learning_rate": 1.0570366995018544e-05, "loss": 0.8551, "step": 8683 }, { "epoch": 3.934753058450385, "grad_norm": 0.14110945163475244, "learning_rate": 1.0561798595247552e-05, "loss": 0.8553, "step": 8684 }, { "epoch": 3.9352061622111463, "grad_norm": 0.20342479864706106, "learning_rate": 1.0553233141424371e-05, "loss": 0.8427, "step": 8685 }, { "epoch": 3.9356592659719074, "grad_norm": 0.12738367586587257, "learning_rate": 1.0544670634406189e-05, "loss": 0.8516, "step": 8686 }, { "epoch": 3.936112369732669, "grad_norm": 0.18511169468995287, "learning_rate": 1.0536111075049856e-05, "loss": 0.8456, "step": 8687 }, { "epoch": 3.93656547349343, "grad_norm": 0.13600170047235857, "learning_rate": 1.0527554464211943e-05, "loss": 0.8572, "step": 8688 }, { "epoch": 3.9370185772541912, "grad_norm": 0.16542832847888878, "learning_rate": 1.051900080274875e-05, "loss": 0.8397, "step": 8689 }, { "epoch": 3.9374716810149524, "grad_norm": 0.14102459973980505, "learning_rate": 1.0510450091516242e-05, "loss": 0.8309, "step": 8690 }, { "epoch": 3.9379247847757135, "grad_norm": 0.14616137100679058, "learning_rate": 1.0501902331370127e-05, "loss": 0.8472, "step": 8691 }, { "epoch": 3.9383778885364746, "grad_norm": 0.14776928422191063, "learning_rate": 1.0493357523165772e-05, "loss": 0.8488, "step": 8692 }, { "epoch": 3.938830992297236, "grad_norm": 0.14133463321183218, "learning_rate": 1.0484815667758306e-05, "loss": 0.8765, "step": 8693 }, { "epoch": 3.9392840960579973, "grad_norm": 0.14258827073318114, "learning_rate": 1.0476276766002504e-05, "loss": 0.8196, "step": 8694 }, { "epoch": 3.9397371998187585, "grad_norm": 0.13445131710197306, "learning_rate": 1.0467740818752898e-05, "loss": 0.8596, "step": 8695 }, { "epoch": 3.9401903035795196, "grad_norm": 0.15955895690688432, "learning_rate": 1.0459207826863692e-05, "loss": 0.8618, "step": 8696 }, { "epoch": 3.940643407340281, "grad_norm": 0.15272641517840374, "learning_rate": 1.0450677791188783e-05, "loss": 0.8554, "step": 8697 }, { "epoch": 3.9410965111010423, "grad_norm": 0.12832113175798723, "learning_rate": 1.044215071258182e-05, "loss": 0.8636, "step": 8698 }, { "epoch": 3.9415496148618034, "grad_norm": 0.14463048580349183, "learning_rate": 1.0433626591896102e-05, "loss": 0.8287, "step": 8699 }, { "epoch": 3.9420027186225646, "grad_norm": 0.13816153492364724, "learning_rate": 1.0425105429984681e-05, "loss": 0.8463, "step": 8700 }, { "epoch": 3.9424558223833257, "grad_norm": 0.16026672217034615, "learning_rate": 1.0416587227700261e-05, "loss": 0.8636, "step": 8701 }, { "epoch": 3.942908926144087, "grad_norm": 0.14115637116129529, "learning_rate": 1.0408071985895312e-05, "loss": 0.8623, "step": 8702 }, { "epoch": 3.943362029904848, "grad_norm": 0.16961230045937448, "learning_rate": 1.0399559705421951e-05, "loss": 0.8424, "step": 8703 }, { "epoch": 3.9438151336656095, "grad_norm": 0.12968288126830813, "learning_rate": 1.0391050387132014e-05, "loss": 0.8512, "step": 8704 }, { "epoch": 3.9442682374263707, "grad_norm": 0.1438267763339451, "learning_rate": 1.0382544031877072e-05, "loss": 0.8491, "step": 8705 }, { "epoch": 3.944721341187132, "grad_norm": 0.13091128041111474, "learning_rate": 1.037404064050835e-05, "loss": 0.8596, "step": 8706 }, { "epoch": 3.945174444947893, "grad_norm": 0.13589962312363735, "learning_rate": 1.0365540213876826e-05, "loss": 0.8577, "step": 8707 }, { "epoch": 3.9456275487086545, "grad_norm": 0.1375483429414981, "learning_rate": 1.0357042752833135e-05, "loss": 0.8626, "step": 8708 }, { "epoch": 3.9460806524694156, "grad_norm": 0.13002318679026262, "learning_rate": 1.034854825822766e-05, "loss": 0.8603, "step": 8709 }, { "epoch": 3.9465337562301768, "grad_norm": 0.13667342662826326, "learning_rate": 1.0340056730910435e-05, "loss": 0.8465, "step": 8710 }, { "epoch": 3.946986859990938, "grad_norm": 0.13728624810366424, "learning_rate": 1.0331568171731261e-05, "loss": 0.8749, "step": 8711 }, { "epoch": 3.947439963751699, "grad_norm": 0.12763537281954357, "learning_rate": 1.0323082581539591e-05, "loss": 0.8553, "step": 8712 }, { "epoch": 3.94789306751246, "grad_norm": 0.1471880489573456, "learning_rate": 1.0314599961184579e-05, "loss": 0.8716, "step": 8713 }, { "epoch": 3.9483461712732213, "grad_norm": 0.13295873530938618, "learning_rate": 1.0306120311515136e-05, "loss": 0.8641, "step": 8714 }, { "epoch": 3.948799275033983, "grad_norm": 0.1864063487278123, "learning_rate": 1.0297643633379803e-05, "loss": 0.8725, "step": 8715 }, { "epoch": 3.949252378794744, "grad_norm": 0.14398508941070984, "learning_rate": 1.0289169927626897e-05, "loss": 0.8412, "step": 8716 }, { "epoch": 3.949705482555505, "grad_norm": 0.21994473036321507, "learning_rate": 1.0280699195104367e-05, "loss": 0.8261, "step": 8717 }, { "epoch": 3.9501585863162663, "grad_norm": 0.18567936646271002, "learning_rate": 1.0272231436659927e-05, "loss": 0.8656, "step": 8718 }, { "epoch": 3.950611690077028, "grad_norm": 0.16544553434508397, "learning_rate": 1.0263766653140955e-05, "loss": 0.8536, "step": 8719 }, { "epoch": 3.951064793837789, "grad_norm": 0.2015225407478143, "learning_rate": 1.025530484539452e-05, "loss": 0.8574, "step": 8720 }, { "epoch": 3.95151789759855, "grad_norm": 0.14131286639549273, "learning_rate": 1.024684601426745e-05, "loss": 0.8607, "step": 8721 }, { "epoch": 3.9519710013593112, "grad_norm": 0.16203241295563256, "learning_rate": 1.0238390160606211e-05, "loss": 0.8396, "step": 8722 }, { "epoch": 3.9524241051200724, "grad_norm": 0.16660502237172278, "learning_rate": 1.0229937285257022e-05, "loss": 0.8545, "step": 8723 }, { "epoch": 3.9528772088808335, "grad_norm": 0.12455659287529115, "learning_rate": 1.0221487389065765e-05, "loss": 0.8584, "step": 8724 }, { "epoch": 3.953330312641595, "grad_norm": 0.15478342780881765, "learning_rate": 1.021304047287805e-05, "loss": 0.8512, "step": 8725 }, { "epoch": 3.953783416402356, "grad_norm": 0.15131259729651644, "learning_rate": 1.0204596537539163e-05, "loss": 0.8587, "step": 8726 }, { "epoch": 3.9542365201631173, "grad_norm": 0.14556388923316876, "learning_rate": 1.019615558389413e-05, "loss": 0.8607, "step": 8727 }, { "epoch": 3.9546896239238785, "grad_norm": 0.1497622196029836, "learning_rate": 1.0187717612787633e-05, "loss": 0.8554, "step": 8728 }, { "epoch": 3.95514272768464, "grad_norm": 0.1458467218187921, "learning_rate": 1.0179282625064108e-05, "loss": 0.8616, "step": 8729 }, { "epoch": 3.955595831445401, "grad_norm": 0.1701986781481152, "learning_rate": 1.0170850621567631e-05, "loss": 0.8714, "step": 8730 }, { "epoch": 3.9560489352061623, "grad_norm": 0.17054925972036342, "learning_rate": 1.0162421603142043e-05, "loss": 0.8559, "step": 8731 }, { "epoch": 3.9565020389669234, "grad_norm": 0.13491836272598642, "learning_rate": 1.015399557063084e-05, "loss": 0.8617, "step": 8732 }, { "epoch": 3.9569551427276846, "grad_norm": 0.1801315915418446, "learning_rate": 1.0145572524877223e-05, "loss": 0.8523, "step": 8733 }, { "epoch": 3.9574082464884457, "grad_norm": 0.15563575455353826, "learning_rate": 1.0137152466724128e-05, "loss": 0.8584, "step": 8734 }, { "epoch": 3.957861350249207, "grad_norm": 0.16792312697682965, "learning_rate": 1.0128735397014142e-05, "loss": 0.8473, "step": 8735 }, { "epoch": 3.9583144540099684, "grad_norm": 0.17426493162647028, "learning_rate": 1.0120321316589612e-05, "loss": 0.8526, "step": 8736 }, { "epoch": 3.9587675577707295, "grad_norm": 0.15554135931468513, "learning_rate": 1.0111910226292529e-05, "loss": 0.8721, "step": 8737 }, { "epoch": 3.9592206615314907, "grad_norm": 0.18376000325542288, "learning_rate": 1.0103502126964631e-05, "loss": 0.8524, "step": 8738 }, { "epoch": 3.959673765292252, "grad_norm": 0.13617173556256232, "learning_rate": 1.0095097019447322e-05, "loss": 0.8742, "step": 8739 }, { "epoch": 3.9601268690530134, "grad_norm": 0.21259499705790136, "learning_rate": 1.0086694904581714e-05, "loss": 0.8578, "step": 8740 }, { "epoch": 3.9605799728137745, "grad_norm": 0.14867979604642723, "learning_rate": 1.0078295783208643e-05, "loss": 0.8698, "step": 8741 }, { "epoch": 3.9610330765745356, "grad_norm": 0.15817305965418468, "learning_rate": 1.0069899656168607e-05, "loss": 0.8487, "step": 8742 }, { "epoch": 3.9614861803352968, "grad_norm": 0.17547675165376198, "learning_rate": 1.006150652430185e-05, "loss": 0.8538, "step": 8743 }, { "epoch": 3.961939284096058, "grad_norm": 0.12910566795384532, "learning_rate": 1.0053116388448272e-05, "loss": 0.8549, "step": 8744 }, { "epoch": 3.962392387856819, "grad_norm": 0.2003323814581074, "learning_rate": 1.004472924944751e-05, "loss": 0.8525, "step": 8745 }, { "epoch": 3.96284549161758, "grad_norm": 0.1506044172632504, "learning_rate": 1.0036345108138862e-05, "loss": 0.8597, "step": 8746 }, { "epoch": 3.9632985953783417, "grad_norm": 0.15204629713185985, "learning_rate": 1.0027963965361373e-05, "loss": 0.8679, "step": 8747 }, { "epoch": 3.963751699139103, "grad_norm": 0.15887484188284318, "learning_rate": 1.0019585821953752e-05, "loss": 0.8598, "step": 8748 }, { "epoch": 3.964204802899864, "grad_norm": 0.1324020218314922, "learning_rate": 1.0011210678754403e-05, "loss": 0.8526, "step": 8749 }, { "epoch": 3.964657906660625, "grad_norm": 0.17734606205684839, "learning_rate": 1.000283853660148e-05, "loss": 0.8609, "step": 8750 }, { "epoch": 3.9651110104213867, "grad_norm": 0.12982731644176615, "learning_rate": 9.994469396332765e-06, "loss": 0.8386, "step": 8751 }, { "epoch": 3.965564114182148, "grad_norm": 0.15501689911744213, "learning_rate": 9.986103258785808e-06, "loss": 0.8504, "step": 8752 }, { "epoch": 3.966017217942909, "grad_norm": 0.13005879125410205, "learning_rate": 9.977740124797801e-06, "loss": 0.8658, "step": 8753 }, { "epoch": 3.96647032170367, "grad_norm": 0.1418626633039595, "learning_rate": 9.96937999520569e-06, "loss": 0.8588, "step": 8754 }, { "epoch": 3.966923425464431, "grad_norm": 0.12978904791199028, "learning_rate": 9.961022870846064e-06, "loss": 0.8609, "step": 8755 }, { "epoch": 3.9673765292251923, "grad_norm": 0.1394311030941313, "learning_rate": 9.952668752555263e-06, "loss": 0.8441, "step": 8756 }, { "epoch": 3.967829632985954, "grad_norm": 0.13018030882969267, "learning_rate": 9.9443176411693e-06, "loss": 0.8585, "step": 8757 }, { "epoch": 3.968282736746715, "grad_norm": 0.14929666298043778, "learning_rate": 9.935969537523862e-06, "loss": 0.8546, "step": 8758 }, { "epoch": 3.968735840507476, "grad_norm": 0.12108219442036522, "learning_rate": 9.927624442454396e-06, "loss": 0.8729, "step": 8759 }, { "epoch": 3.9691889442682373, "grad_norm": 0.134072619295781, "learning_rate": 9.919282356795992e-06, "loss": 0.861, "step": 8760 }, { "epoch": 3.969642048028999, "grad_norm": 0.13971772384243136, "learning_rate": 9.91094328138348e-06, "loss": 0.8481, "step": 8761 }, { "epoch": 3.97009515178976, "grad_norm": 0.13113987442698957, "learning_rate": 9.90260721705135e-06, "loss": 0.8709, "step": 8762 }, { "epoch": 3.970548255550521, "grad_norm": 0.14841760740565574, "learning_rate": 9.894274164633835e-06, "loss": 0.8657, "step": 8763 }, { "epoch": 3.9710013593112823, "grad_norm": 0.13255494467127615, "learning_rate": 9.885944124964832e-06, "loss": 0.848, "step": 8764 }, { "epoch": 3.9714544630720434, "grad_norm": 0.16420127860103256, "learning_rate": 9.87761709887793e-06, "loss": 0.8612, "step": 8765 }, { "epoch": 3.9719075668328045, "grad_norm": 0.12352707325613753, "learning_rate": 9.86929308720646e-06, "loss": 0.8379, "step": 8766 }, { "epoch": 3.9723606705935657, "grad_norm": 0.15640934106011448, "learning_rate": 9.860972090783405e-06, "loss": 0.8336, "step": 8767 }, { "epoch": 3.9728137743543273, "grad_norm": 0.13405583075530514, "learning_rate": 9.852654110441482e-06, "loss": 0.8464, "step": 8768 }, { "epoch": 3.9732668781150884, "grad_norm": 0.14469557882755343, "learning_rate": 9.84433914701307e-06, "loss": 0.8615, "step": 8769 }, { "epoch": 3.9737199818758495, "grad_norm": 0.13145936699626057, "learning_rate": 9.836027201330291e-06, "loss": 0.8416, "step": 8770 }, { "epoch": 3.9741730856366106, "grad_norm": 0.1222093065826357, "learning_rate": 9.82771827422492e-06, "loss": 0.8477, "step": 8771 }, { "epoch": 3.9746261893973722, "grad_norm": 0.1455877170335531, "learning_rate": 9.819412366528467e-06, "loss": 0.8276, "step": 8772 }, { "epoch": 3.9750792931581334, "grad_norm": 0.14121935213594283, "learning_rate": 9.811109479072112e-06, "loss": 0.8464, "step": 8773 }, { "epoch": 3.9755323969188945, "grad_norm": 0.16461484112050898, "learning_rate": 9.802809612686737e-06, "loss": 0.8514, "step": 8774 }, { "epoch": 3.9759855006796556, "grad_norm": 0.12322715859096488, "learning_rate": 9.794512768202949e-06, "loss": 0.8751, "step": 8775 }, { "epoch": 3.9764386044404167, "grad_norm": 0.1428934490167226, "learning_rate": 9.786218946451007e-06, "loss": 0.8533, "step": 8776 }, { "epoch": 3.976891708201178, "grad_norm": 0.1437487300572671, "learning_rate": 9.77792814826092e-06, "loss": 0.8532, "step": 8777 }, { "epoch": 3.977344811961939, "grad_norm": 0.1297698916086517, "learning_rate": 9.769640374462344e-06, "loss": 0.8576, "step": 8778 }, { "epoch": 3.9777979157227006, "grad_norm": 0.14309528604831068, "learning_rate": 9.761355625884672e-06, "loss": 0.8405, "step": 8779 }, { "epoch": 3.9782510194834617, "grad_norm": 0.12664704027933066, "learning_rate": 9.753073903356972e-06, "loss": 0.8487, "step": 8780 }, { "epoch": 3.978704123244223, "grad_norm": 0.15834760279282473, "learning_rate": 9.744795207708013e-06, "loss": 0.852, "step": 8781 }, { "epoch": 3.979157227004984, "grad_norm": 0.1365437015532402, "learning_rate": 9.736519539766251e-06, "loss": 0.8873, "step": 8782 }, { "epoch": 3.9796103307657456, "grad_norm": 0.15984632904401447, "learning_rate": 9.728246900359877e-06, "loss": 0.8431, "step": 8783 }, { "epoch": 3.9800634345265067, "grad_norm": 0.13132743679656858, "learning_rate": 9.719977290316738e-06, "loss": 0.8578, "step": 8784 }, { "epoch": 3.980516538287268, "grad_norm": 0.16644875648750548, "learning_rate": 9.711710710464381e-06, "loss": 0.8511, "step": 8785 }, { "epoch": 3.980969642048029, "grad_norm": 0.2870201064397205, "learning_rate": 9.703447161630088e-06, "loss": 0.8566, "step": 8786 }, { "epoch": 3.98142274580879, "grad_norm": 0.1381672164196009, "learning_rate": 9.695186644640784e-06, "loss": 0.88, "step": 8787 }, { "epoch": 3.981875849569551, "grad_norm": 0.1442166022859317, "learning_rate": 9.686929160323143e-06, "loss": 0.8585, "step": 8788 }, { "epoch": 3.982328953330313, "grad_norm": 0.14243402403851327, "learning_rate": 9.67867470950349e-06, "loss": 0.8702, "step": 8789 }, { "epoch": 3.982782057091074, "grad_norm": 0.1493997975958885, "learning_rate": 9.670423293007883e-06, "loss": 0.865, "step": 8790 }, { "epoch": 3.983235160851835, "grad_norm": 0.1547888756659604, "learning_rate": 9.662174911662042e-06, "loss": 0.8504, "step": 8791 }, { "epoch": 3.983688264612596, "grad_norm": 0.14011278408433167, "learning_rate": 9.653929566291417e-06, "loss": 0.8592, "step": 8792 }, { "epoch": 3.9841413683733577, "grad_norm": 0.15942030485422892, "learning_rate": 9.645687257721139e-06, "loss": 0.8492, "step": 8793 }, { "epoch": 3.984594472134119, "grad_norm": 0.14658243749225833, "learning_rate": 9.637447986776016e-06, "loss": 0.8556, "step": 8794 }, { "epoch": 3.98504757589488, "grad_norm": 0.15624193777936105, "learning_rate": 9.629211754280594e-06, "loss": 0.8484, "step": 8795 }, { "epoch": 3.985500679655641, "grad_norm": 0.14933708134659035, "learning_rate": 9.620978561059067e-06, "loss": 0.8567, "step": 8796 }, { "epoch": 3.9859537834164023, "grad_norm": 0.16589898287170102, "learning_rate": 9.612748407935375e-06, "loss": 0.8641, "step": 8797 }, { "epoch": 3.9864068871771634, "grad_norm": 0.1709190357537583, "learning_rate": 9.604521295733105e-06, "loss": 0.8255, "step": 8798 }, { "epoch": 3.9868599909379245, "grad_norm": 0.14947337151308884, "learning_rate": 9.596297225275588e-06, "loss": 0.8365, "step": 8799 }, { "epoch": 3.987313094698686, "grad_norm": 0.14136819558149952, "learning_rate": 9.588076197385807e-06, "loss": 0.8573, "step": 8800 }, { "epoch": 3.9877661984594472, "grad_norm": 0.1384754369223182, "learning_rate": 9.579858212886456e-06, "loss": 0.8688, "step": 8801 }, { "epoch": 3.9882193022202084, "grad_norm": 0.13194763008854868, "learning_rate": 9.57164327259994e-06, "loss": 0.8487, "step": 8802 }, { "epoch": 3.9886724059809695, "grad_norm": 0.1329820413806333, "learning_rate": 9.563431377348338e-06, "loss": 0.8535, "step": 8803 }, { "epoch": 3.989125509741731, "grad_norm": 0.1449260141538113, "learning_rate": 9.55522252795344e-06, "loss": 0.868, "step": 8804 }, { "epoch": 3.989578613502492, "grad_norm": 0.1516785116702068, "learning_rate": 9.547016725236715e-06, "loss": 0.8659, "step": 8805 }, { "epoch": 3.9900317172632533, "grad_norm": 0.13810416869482436, "learning_rate": 9.53881397001935e-06, "loss": 0.846, "step": 8806 }, { "epoch": 3.9904848210240145, "grad_norm": 0.15166815780785742, "learning_rate": 9.530614263122198e-06, "loss": 0.8499, "step": 8807 }, { "epoch": 3.9909379247847756, "grad_norm": 0.12760621782790668, "learning_rate": 9.522417605365839e-06, "loss": 0.8385, "step": 8808 }, { "epoch": 3.9913910285455367, "grad_norm": 0.1622667799239956, "learning_rate": 9.51422399757052e-06, "loss": 0.8475, "step": 8809 }, { "epoch": 3.991844132306298, "grad_norm": 0.1147675311848901, "learning_rate": 9.506033440556188e-06, "loss": 0.847, "step": 8810 }, { "epoch": 3.9922972360670594, "grad_norm": 0.14576349557386334, "learning_rate": 9.49784593514251e-06, "loss": 0.8442, "step": 8811 }, { "epoch": 3.9927503398278206, "grad_norm": 0.12291245675503419, "learning_rate": 9.4896614821488e-06, "loss": 0.8453, "step": 8812 }, { "epoch": 3.9932034435885817, "grad_norm": 0.14234835495694048, "learning_rate": 9.48148008239413e-06, "loss": 0.8538, "step": 8813 }, { "epoch": 3.993656547349343, "grad_norm": 0.1257835757842282, "learning_rate": 9.473301736697196e-06, "loss": 0.8586, "step": 8814 }, { "epoch": 3.9941096511101044, "grad_norm": 0.1629008451534394, "learning_rate": 9.465126445876458e-06, "loss": 0.8535, "step": 8815 }, { "epoch": 3.9945627548708655, "grad_norm": 0.15315915469706617, "learning_rate": 9.456954210750005e-06, "loss": 0.8772, "step": 8816 }, { "epoch": 3.9950158586316267, "grad_norm": 0.29400274776589597, "learning_rate": 9.448785032135675e-06, "loss": 0.8537, "step": 8817 }, { "epoch": 3.995468962392388, "grad_norm": 0.1308580534233549, "learning_rate": 9.440618910850964e-06, "loss": 0.8626, "step": 8818 }, { "epoch": 3.995922066153149, "grad_norm": 0.1503112450050845, "learning_rate": 9.43245584771307e-06, "loss": 0.8345, "step": 8819 }, { "epoch": 3.99637516991391, "grad_norm": 0.16427206231710248, "learning_rate": 9.42429584353891e-06, "loss": 0.8503, "step": 8820 }, { "epoch": 3.9968282736746716, "grad_norm": 0.14840953245002145, "learning_rate": 9.416138899145042e-06, "loss": 0.8598, "step": 8821 }, { "epoch": 3.9972813774354328, "grad_norm": 0.1363519280954702, "learning_rate": 9.407985015347782e-06, "loss": 0.8608, "step": 8822 }, { "epoch": 3.997734481196194, "grad_norm": 0.18060696591116854, "learning_rate": 9.399834192963081e-06, "loss": 0.8534, "step": 8823 }, { "epoch": 3.998187584956955, "grad_norm": 0.12325750543620921, "learning_rate": 9.391686432806635e-06, "loss": 0.8622, "step": 8824 }, { "epoch": 3.9986406887177166, "grad_norm": 0.1933707942083629, "learning_rate": 9.383541735693798e-06, "loss": 0.8644, "step": 8825 }, { "epoch": 3.9990937924784777, "grad_norm": 0.16530650039506362, "learning_rate": 9.375400102439616e-06, "loss": 0.8389, "step": 8826 }, { "epoch": 3.999546896239239, "grad_norm": 0.16309128628791286, "learning_rate": 9.367261533858865e-06, "loss": 0.8672, "step": 8827 }, { "epoch": 4.0, "grad_norm": 0.17333826513015055, "learning_rate": 9.359126030765964e-06, "loss": 0.8434, "step": 8828 }, { "epoch": 4.000453103760761, "grad_norm": 0.12445161792114366, "learning_rate": 9.350993593975075e-06, "loss": 0.8327, "step": 8829 }, { "epoch": 4.000906207521522, "grad_norm": 0.16670783486869606, "learning_rate": 9.342864224300006e-06, "loss": 0.8508, "step": 8830 }, { "epoch": 4.001359311282283, "grad_norm": 0.14926808883453407, "learning_rate": 9.334737922554308e-06, "loss": 0.8356, "step": 8831 }, { "epoch": 4.0018124150430445, "grad_norm": 0.1448690484034359, "learning_rate": 9.326614689551174e-06, "loss": 0.8585, "step": 8832 }, { "epoch": 4.002265518803806, "grad_norm": 0.15399890756505655, "learning_rate": 9.318494526103538e-06, "loss": 0.8518, "step": 8833 }, { "epoch": 4.002718622564568, "grad_norm": 0.15142708495624962, "learning_rate": 9.310377433023992e-06, "loss": 0.876, "step": 8834 }, { "epoch": 4.003171726325329, "grad_norm": 0.132314547261074, "learning_rate": 9.302263411124816e-06, "loss": 0.8514, "step": 8835 }, { "epoch": 4.00362483008609, "grad_norm": 0.1660998026915115, "learning_rate": 9.294152461218027e-06, "loss": 0.8562, "step": 8836 }, { "epoch": 4.004077933846851, "grad_norm": 0.1315284702597734, "learning_rate": 9.286044584115278e-06, "loss": 0.8449, "step": 8837 }, { "epoch": 4.004531037607612, "grad_norm": 0.1549778506759883, "learning_rate": 9.277939780627983e-06, "loss": 0.8618, "step": 8838 }, { "epoch": 4.004984141368373, "grad_norm": 0.15312569291556455, "learning_rate": 9.269838051567159e-06, "loss": 0.8242, "step": 8839 }, { "epoch": 4.0054372451291345, "grad_norm": 0.15662221923312572, "learning_rate": 9.2617393977436e-06, "loss": 0.8447, "step": 8840 }, { "epoch": 4.005890348889896, "grad_norm": 0.15254291248471966, "learning_rate": 9.25364381996773e-06, "loss": 0.844, "step": 8841 }, { "epoch": 4.006343452650657, "grad_norm": 0.15752076833774223, "learning_rate": 9.245551319049721e-06, "loss": 0.8327, "step": 8842 }, { "epoch": 4.006796556411418, "grad_norm": 0.13856300321265771, "learning_rate": 9.23746189579938e-06, "loss": 0.8478, "step": 8843 }, { "epoch": 4.00724966017218, "grad_norm": 0.14168940382571013, "learning_rate": 9.229375551026262e-06, "loss": 0.8234, "step": 8844 }, { "epoch": 4.007702763932941, "grad_norm": 0.13015917302170146, "learning_rate": 9.221292285539568e-06, "loss": 0.8306, "step": 8845 }, { "epoch": 4.008155867693702, "grad_norm": 0.1554151543197663, "learning_rate": 9.213212100148202e-06, "loss": 0.8587, "step": 8846 }, { "epoch": 4.008608971454463, "grad_norm": 0.13276278060819413, "learning_rate": 9.20513499566079e-06, "loss": 0.8501, "step": 8847 }, { "epoch": 4.009062075215224, "grad_norm": 0.14864462909127948, "learning_rate": 9.197060972885602e-06, "loss": 0.8429, "step": 8848 }, { "epoch": 4.0095151789759855, "grad_norm": 0.12934145565407534, "learning_rate": 9.188990032630642e-06, "loss": 0.8404, "step": 8849 }, { "epoch": 4.009968282736747, "grad_norm": 0.1383307763980896, "learning_rate": 9.18092217570357e-06, "loss": 0.8383, "step": 8850 }, { "epoch": 4.010421386497508, "grad_norm": 0.18248044832550434, "learning_rate": 9.17285740291178e-06, "loss": 0.8538, "step": 8851 }, { "epoch": 4.010874490258269, "grad_norm": 0.12174809742726042, "learning_rate": 9.164795715062302e-06, "loss": 0.8366, "step": 8852 }, { "epoch": 4.01132759401903, "grad_norm": 0.2099830957241385, "learning_rate": 9.156737112961913e-06, "loss": 0.8503, "step": 8853 }, { "epoch": 4.011780697779791, "grad_norm": 0.14062331854777124, "learning_rate": 9.148681597417046e-06, "loss": 0.8349, "step": 8854 }, { "epoch": 4.012233801540553, "grad_norm": 0.19329516601897498, "learning_rate": 9.140629169233826e-06, "loss": 0.8522, "step": 8855 }, { "epoch": 4.012686905301314, "grad_norm": 0.1523138881010568, "learning_rate": 9.132579829218095e-06, "loss": 0.854, "step": 8856 }, { "epoch": 4.0131400090620755, "grad_norm": 0.13707913344527017, "learning_rate": 9.124533578175346e-06, "loss": 0.8403, "step": 8857 }, { "epoch": 4.013593112822837, "grad_norm": 0.13200758589584216, "learning_rate": 9.116490416910815e-06, "loss": 0.8337, "step": 8858 }, { "epoch": 4.014046216583598, "grad_norm": 0.1360034298361964, "learning_rate": 9.108450346229367e-06, "loss": 0.8427, "step": 8859 }, { "epoch": 4.014499320344359, "grad_norm": 0.12769220491920288, "learning_rate": 9.100413366935626e-06, "loss": 0.8529, "step": 8860 }, { "epoch": 4.01495242410512, "grad_norm": 0.12365092889433026, "learning_rate": 9.092379479833843e-06, "loss": 0.8466, "step": 8861 }, { "epoch": 4.015405527865881, "grad_norm": 0.14654738832911546, "learning_rate": 9.084348685727993e-06, "loss": 0.8476, "step": 8862 }, { "epoch": 4.015858631626642, "grad_norm": 0.12060861152566384, "learning_rate": 9.076320985421745e-06, "loss": 0.8488, "step": 8863 }, { "epoch": 4.016311735387403, "grad_norm": 0.1387366964924681, "learning_rate": 9.068296379718431e-06, "loss": 0.8479, "step": 8864 }, { "epoch": 4.0167648391481645, "grad_norm": 0.13317843599295787, "learning_rate": 9.06027486942112e-06, "loss": 0.8374, "step": 8865 }, { "epoch": 4.0172179429089265, "grad_norm": 0.13303953513515004, "learning_rate": 9.052256455332511e-06, "loss": 0.8407, "step": 8866 }, { "epoch": 4.017671046669688, "grad_norm": 0.15632082727127422, "learning_rate": 9.044241138255051e-06, "loss": 0.8495, "step": 8867 }, { "epoch": 4.018124150430449, "grad_norm": 0.1282127381405766, "learning_rate": 9.03622891899083e-06, "loss": 0.8502, "step": 8868 }, { "epoch": 4.01857725419121, "grad_norm": 0.12289863804084893, "learning_rate": 9.028219798341675e-06, "loss": 0.8417, "step": 8869 }, { "epoch": 4.019030357951971, "grad_norm": 0.13245296397454498, "learning_rate": 9.020213777109057e-06, "loss": 0.8431, "step": 8870 }, { "epoch": 4.019483461712732, "grad_norm": 0.12782069682990507, "learning_rate": 9.012210856094152e-06, "loss": 0.8387, "step": 8871 }, { "epoch": 4.019936565473493, "grad_norm": 0.12904707285428335, "learning_rate": 9.00421103609785e-06, "loss": 0.85, "step": 8872 }, { "epoch": 4.0203896692342544, "grad_norm": 0.12055360532078378, "learning_rate": 8.996214317920691e-06, "loss": 0.8583, "step": 8873 }, { "epoch": 4.020842772995016, "grad_norm": 0.13836983818421503, "learning_rate": 8.988220702362941e-06, "loss": 0.8418, "step": 8874 }, { "epoch": 4.021295876755777, "grad_norm": 0.12599883931465955, "learning_rate": 8.980230190224528e-06, "loss": 0.8342, "step": 8875 }, { "epoch": 4.021748980516539, "grad_norm": 0.13331696418142341, "learning_rate": 8.972242782305093e-06, "loss": 0.8662, "step": 8876 }, { "epoch": 4.0222020842773, "grad_norm": 0.16494668999181236, "learning_rate": 8.964258479403951e-06, "loss": 0.8434, "step": 8877 }, { "epoch": 4.022655188038061, "grad_norm": 0.11713611748343628, "learning_rate": 8.956277282320091e-06, "loss": 0.8428, "step": 8878 }, { "epoch": 4.023108291798822, "grad_norm": 0.13391245669411408, "learning_rate": 8.948299191852237e-06, "loss": 0.8614, "step": 8879 }, { "epoch": 4.023561395559583, "grad_norm": 0.16730134242745534, "learning_rate": 8.940324208798748e-06, "loss": 0.8526, "step": 8880 }, { "epoch": 4.024014499320344, "grad_norm": 0.12933235567037427, "learning_rate": 8.932352333957723e-06, "loss": 0.8583, "step": 8881 }, { "epoch": 4.0244676030811055, "grad_norm": 0.1824381736531991, "learning_rate": 8.924383568126904e-06, "loss": 0.8336, "step": 8882 }, { "epoch": 4.024920706841867, "grad_norm": 0.12481638181777596, "learning_rate": 8.916417912103768e-06, "loss": 0.8564, "step": 8883 }, { "epoch": 4.025373810602628, "grad_norm": 0.15117457024231515, "learning_rate": 8.908455366685431e-06, "loss": 0.8412, "step": 8884 }, { "epoch": 4.025826914363389, "grad_norm": 0.1390402801923857, "learning_rate": 8.900495932668742e-06, "loss": 0.8388, "step": 8885 }, { "epoch": 4.02628001812415, "grad_norm": 0.13723669675375536, "learning_rate": 8.892539610850216e-06, "loss": 0.856, "step": 8886 }, { "epoch": 4.026733121884912, "grad_norm": 0.163877703387031, "learning_rate": 8.884586402026043e-06, "loss": 0.8537, "step": 8887 }, { "epoch": 4.027186225645673, "grad_norm": 0.13929238737705307, "learning_rate": 8.876636306992146e-06, "loss": 0.8357, "step": 8888 }, { "epoch": 4.027639329406434, "grad_norm": 0.14735798863097077, "learning_rate": 8.868689326544081e-06, "loss": 0.8536, "step": 8889 }, { "epoch": 4.0280924331671955, "grad_norm": 0.13881759597773427, "learning_rate": 8.860745461477149e-06, "loss": 0.8445, "step": 8890 }, { "epoch": 4.028545536927957, "grad_norm": 0.14889286705725097, "learning_rate": 8.85280471258629e-06, "loss": 0.8526, "step": 8891 }, { "epoch": 4.028998640688718, "grad_norm": 0.12944364315105378, "learning_rate": 8.844867080666163e-06, "loss": 0.8476, "step": 8892 }, { "epoch": 4.029451744449479, "grad_norm": 0.131447175356184, "learning_rate": 8.83693256651109e-06, "loss": 0.8423, "step": 8893 }, { "epoch": 4.02990484821024, "grad_norm": 0.13083903727822443, "learning_rate": 8.829001170915124e-06, "loss": 0.846, "step": 8894 }, { "epoch": 4.030357951971001, "grad_norm": 0.135402470155557, "learning_rate": 8.821072894671957e-06, "loss": 0.8334, "step": 8895 }, { "epoch": 4.030811055731762, "grad_norm": 0.16296728941813587, "learning_rate": 8.813147738574992e-06, "loss": 0.8383, "step": 8896 }, { "epoch": 4.031264159492523, "grad_norm": 0.12771540958811495, "learning_rate": 8.805225703417322e-06, "loss": 0.8253, "step": 8897 }, { "epoch": 4.031717263253285, "grad_norm": 0.1481370986507368, "learning_rate": 8.797306789991707e-06, "loss": 0.8337, "step": 8898 }, { "epoch": 4.0321703670140465, "grad_norm": 0.12465007931143315, "learning_rate": 8.789390999090637e-06, "loss": 0.8553, "step": 8899 }, { "epoch": 4.032623470774808, "grad_norm": 0.1355041662587558, "learning_rate": 8.78147833150624e-06, "loss": 0.8489, "step": 8900 }, { "epoch": 4.033076574535569, "grad_norm": 0.11181400774939063, "learning_rate": 8.77356878803037e-06, "loss": 0.8291, "step": 8901 }, { "epoch": 4.03352967829633, "grad_norm": 0.14017878421749355, "learning_rate": 8.765662369454539e-06, "loss": 0.8475, "step": 8902 }, { "epoch": 4.033982782057091, "grad_norm": 0.11822817481965538, "learning_rate": 8.757759076569985e-06, "loss": 0.8407, "step": 8903 }, { "epoch": 4.034435885817852, "grad_norm": 0.12181001016770741, "learning_rate": 8.749858910167575e-06, "loss": 0.8327, "step": 8904 }, { "epoch": 4.034888989578613, "grad_norm": 0.12717392361672006, "learning_rate": 8.741961871037929e-06, "loss": 0.8443, "step": 8905 }, { "epoch": 4.035342093339374, "grad_norm": 0.11290020122370571, "learning_rate": 8.734067959971306e-06, "loss": 0.8554, "step": 8906 }, { "epoch": 4.035795197100136, "grad_norm": 0.1314688946015283, "learning_rate": 8.726177177757655e-06, "loss": 0.8451, "step": 8907 }, { "epoch": 4.036248300860898, "grad_norm": 0.12507845552182367, "learning_rate": 8.718289525186652e-06, "loss": 0.8491, "step": 8908 }, { "epoch": 4.036701404621659, "grad_norm": 0.12965956197449555, "learning_rate": 8.710405003047607e-06, "loss": 0.8548, "step": 8909 }, { "epoch": 4.03715450838242, "grad_norm": 0.11373241309366225, "learning_rate": 8.702523612129564e-06, "loss": 0.8408, "step": 8910 }, { "epoch": 4.037607612143181, "grad_norm": 0.1339628717768926, "learning_rate": 8.694645353221207e-06, "loss": 0.845, "step": 8911 }, { "epoch": 4.038060715903942, "grad_norm": 0.12427656988055119, "learning_rate": 8.68677022711096e-06, "loss": 0.826, "step": 8912 }, { "epoch": 4.038513819664703, "grad_norm": 0.13521811692052027, "learning_rate": 8.67889823458688e-06, "loss": 0.8484, "step": 8913 }, { "epoch": 4.038966923425464, "grad_norm": 0.13313075290939058, "learning_rate": 8.671029376436753e-06, "loss": 0.8568, "step": 8914 }, { "epoch": 4.0394200271862255, "grad_norm": 0.15425599039312052, "learning_rate": 8.663163653448028e-06, "loss": 0.8477, "step": 8915 }, { "epoch": 4.039873130946987, "grad_norm": 0.1267932087228106, "learning_rate": 8.655301066407831e-06, "loss": 0.8369, "step": 8916 }, { "epoch": 4.040326234707748, "grad_norm": 0.16131854601549062, "learning_rate": 8.647441616103016e-06, "loss": 0.8532, "step": 8917 }, { "epoch": 4.040779338468509, "grad_norm": 0.12169528202531608, "learning_rate": 8.639585303320071e-06, "loss": 0.849, "step": 8918 }, { "epoch": 4.041232442229271, "grad_norm": 0.14698247026612363, "learning_rate": 8.631732128845214e-06, "loss": 0.8528, "step": 8919 }, { "epoch": 4.041685545990032, "grad_norm": 0.15053057629283068, "learning_rate": 8.62388209346431e-06, "loss": 0.8583, "step": 8920 }, { "epoch": 4.042138649750793, "grad_norm": 0.12545802942015352, "learning_rate": 8.616035197962956e-06, "loss": 0.8281, "step": 8921 }, { "epoch": 4.042591753511554, "grad_norm": 0.17530457870145366, "learning_rate": 8.608191443126395e-06, "loss": 0.8362, "step": 8922 }, { "epoch": 4.043044857272315, "grad_norm": 0.14067951053142194, "learning_rate": 8.600350829739556e-06, "loss": 0.8339, "step": 8923 }, { "epoch": 4.043497961033077, "grad_norm": 0.15000862992045635, "learning_rate": 8.592513358587094e-06, "loss": 0.8516, "step": 8924 }, { "epoch": 4.043951064793838, "grad_norm": 0.147703075590296, "learning_rate": 8.584679030453293e-06, "loss": 0.8343, "step": 8925 }, { "epoch": 4.044404168554599, "grad_norm": 0.13416044319117332, "learning_rate": 8.576847846122183e-06, "loss": 0.8543, "step": 8926 }, { "epoch": 4.04485727231536, "grad_norm": 0.18281105060747407, "learning_rate": 8.569019806377418e-06, "loss": 0.8451, "step": 8927 }, { "epoch": 4.045310376076121, "grad_norm": 0.14823345132804314, "learning_rate": 8.561194912002398e-06, "loss": 0.8323, "step": 8928 }, { "epoch": 4.045763479836882, "grad_norm": 0.14165002059090728, "learning_rate": 8.553373163780146e-06, "loss": 0.838, "step": 8929 }, { "epoch": 4.046216583597644, "grad_norm": 0.14201157296810057, "learning_rate": 8.545554562493437e-06, "loss": 0.8654, "step": 8930 }, { "epoch": 4.046669687358405, "grad_norm": 0.14963179086595446, "learning_rate": 8.537739108924672e-06, "loss": 0.8468, "step": 8931 }, { "epoch": 4.0471227911191665, "grad_norm": 0.14561436677328535, "learning_rate": 8.529926803855958e-06, "loss": 0.8462, "step": 8932 }, { "epoch": 4.047575894879928, "grad_norm": 0.13835634267328378, "learning_rate": 8.522117648069113e-06, "loss": 0.8354, "step": 8933 }, { "epoch": 4.048028998640689, "grad_norm": 0.13814703513923804, "learning_rate": 8.514311642345592e-06, "loss": 0.8479, "step": 8934 }, { "epoch": 4.04848210240145, "grad_norm": 0.13023841876083905, "learning_rate": 8.506508787466581e-06, "loss": 0.8355, "step": 8935 }, { "epoch": 4.048935206162211, "grad_norm": 0.12509346698114537, "learning_rate": 8.498709084212914e-06, "loss": 0.8312, "step": 8936 }, { "epoch": 4.049388309922972, "grad_norm": 0.1283998406161009, "learning_rate": 8.490912533365137e-06, "loss": 0.8601, "step": 8937 }, { "epoch": 4.049841413683733, "grad_norm": 0.13392762146514475, "learning_rate": 8.483119135703459e-06, "loss": 0.8395, "step": 8938 }, { "epoch": 4.050294517444494, "grad_norm": 0.1284417432958856, "learning_rate": 8.475328892007781e-06, "loss": 0.8628, "step": 8939 }, { "epoch": 4.050747621205256, "grad_norm": 0.1306808587985899, "learning_rate": 8.467541803057706e-06, "loss": 0.8558, "step": 8940 }, { "epoch": 4.051200724966018, "grad_norm": 0.13395854661146933, "learning_rate": 8.459757869632481e-06, "loss": 0.8399, "step": 8941 }, { "epoch": 4.051653828726779, "grad_norm": 0.13162925266624104, "learning_rate": 8.451977092511092e-06, "loss": 0.841, "step": 8942 }, { "epoch": 4.05210693248754, "grad_norm": 0.13644752815711678, "learning_rate": 8.444199472472152e-06, "loss": 0.8478, "step": 8943 }, { "epoch": 4.052560036248301, "grad_norm": 0.12655658107633094, "learning_rate": 8.436425010294008e-06, "loss": 0.8513, "step": 8944 }, { "epoch": 4.053013140009062, "grad_norm": 0.14976485537101014, "learning_rate": 8.428653706754644e-06, "loss": 0.8437, "step": 8945 }, { "epoch": 4.053466243769823, "grad_norm": 0.12398200718478937, "learning_rate": 8.420885562631777e-06, "loss": 0.8361, "step": 8946 }, { "epoch": 4.053919347530584, "grad_norm": 0.12477105558273793, "learning_rate": 8.41312057870277e-06, "loss": 0.8305, "step": 8947 }, { "epoch": 4.0543724512913455, "grad_norm": 0.12873100549119215, "learning_rate": 8.405358755744672e-06, "loss": 0.8454, "step": 8948 }, { "epoch": 4.054825555052107, "grad_norm": 0.11747522629265662, "learning_rate": 8.397600094534248e-06, "loss": 0.8528, "step": 8949 }, { "epoch": 4.055278658812868, "grad_norm": 0.14556641092451245, "learning_rate": 8.389844595847908e-06, "loss": 0.8509, "step": 8950 }, { "epoch": 4.05573176257363, "grad_norm": 0.14684726126595699, "learning_rate": 8.382092260461778e-06, "loss": 0.8332, "step": 8951 }, { "epoch": 4.056184866334391, "grad_norm": 0.13516902065703065, "learning_rate": 8.374343089151642e-06, "loss": 0.8445, "step": 8952 }, { "epoch": 4.056637970095152, "grad_norm": 0.12570281619368226, "learning_rate": 8.366597082692976e-06, "loss": 0.8283, "step": 8953 }, { "epoch": 4.057091073855913, "grad_norm": 0.16047428715414816, "learning_rate": 8.358854241860932e-06, "loss": 0.8654, "step": 8954 }, { "epoch": 4.057544177616674, "grad_norm": 0.11905675522413363, "learning_rate": 8.35111456743038e-06, "loss": 0.827, "step": 8955 }, { "epoch": 4.057997281377435, "grad_norm": 0.1360039117113899, "learning_rate": 8.343378060175818e-06, "loss": 0.8406, "step": 8956 }, { "epoch": 4.058450385138197, "grad_norm": 0.1262819424222326, "learning_rate": 8.335644720871477e-06, "loss": 0.8442, "step": 8957 }, { "epoch": 4.058903488898958, "grad_norm": 0.11801228029862691, "learning_rate": 8.327914550291245e-06, "loss": 0.8461, "step": 8958 }, { "epoch": 4.059356592659719, "grad_norm": 0.1531478467425587, "learning_rate": 8.320187549208682e-06, "loss": 0.8203, "step": 8959 }, { "epoch": 4.05980969642048, "grad_norm": 0.11497814741859595, "learning_rate": 8.312463718397068e-06, "loss": 0.839, "step": 8960 }, { "epoch": 4.060262800181241, "grad_norm": 0.13876965893028217, "learning_rate": 8.30474305862933e-06, "loss": 0.856, "step": 8961 }, { "epoch": 4.060715903942003, "grad_norm": 0.1307434191868257, "learning_rate": 8.297025570678103e-06, "loss": 0.8347, "step": 8962 }, { "epoch": 4.061169007702764, "grad_norm": 0.16539078828341172, "learning_rate": 8.289311255315682e-06, "loss": 0.8577, "step": 8963 }, { "epoch": 4.061622111463525, "grad_norm": 0.12530847699929457, "learning_rate": 8.281600113314069e-06, "loss": 0.8493, "step": 8964 }, { "epoch": 4.0620752152242865, "grad_norm": 0.14690555211939846, "learning_rate": 8.273892145444922e-06, "loss": 0.8502, "step": 8965 }, { "epoch": 4.062528318985048, "grad_norm": 0.13112164090484416, "learning_rate": 8.266187352479612e-06, "loss": 0.8365, "step": 8966 }, { "epoch": 4.062981422745809, "grad_norm": 0.15125268276584553, "learning_rate": 8.258485735189161e-06, "loss": 0.8281, "step": 8967 }, { "epoch": 4.06343452650657, "grad_norm": 0.11760567938414848, "learning_rate": 8.25078729434428e-06, "loss": 0.8445, "step": 8968 }, { "epoch": 4.063887630267331, "grad_norm": 0.13973159171185864, "learning_rate": 8.243092030715396e-06, "loss": 0.8286, "step": 8969 }, { "epoch": 4.064340734028092, "grad_norm": 0.12579624466384312, "learning_rate": 8.235399945072564e-06, "loss": 0.8347, "step": 8970 }, { "epoch": 4.064793837788853, "grad_norm": 0.1308119159876165, "learning_rate": 8.22771103818557e-06, "loss": 0.8291, "step": 8971 }, { "epoch": 4.065246941549615, "grad_norm": 0.12535331106252082, "learning_rate": 8.220025310823842e-06, "loss": 0.8504, "step": 8972 }, { "epoch": 4.065700045310376, "grad_norm": 0.14505831583158577, "learning_rate": 8.212342763756526e-06, "loss": 0.8526, "step": 8973 }, { "epoch": 4.066153149071138, "grad_norm": 0.1286513814064682, "learning_rate": 8.204663397752415e-06, "loss": 0.8416, "step": 8974 }, { "epoch": 4.066606252831899, "grad_norm": 0.1375879649857786, "learning_rate": 8.196987213580017e-06, "loss": 0.8544, "step": 8975 }, { "epoch": 4.06705935659266, "grad_norm": 0.16409906234811597, "learning_rate": 8.1893142120075e-06, "loss": 0.8471, "step": 8976 }, { "epoch": 4.067512460353421, "grad_norm": 0.16195912589048847, "learning_rate": 8.181644393802704e-06, "loss": 0.85, "step": 8977 }, { "epoch": 4.067965564114182, "grad_norm": 0.12972659130433792, "learning_rate": 8.173977759733187e-06, "loss": 0.8435, "step": 8978 }, { "epoch": 4.068418667874943, "grad_norm": 0.13849279358478184, "learning_rate": 8.166314310566146e-06, "loss": 0.8403, "step": 8979 }, { "epoch": 4.068871771635704, "grad_norm": 0.12827507537281765, "learning_rate": 8.158654047068504e-06, "loss": 0.8398, "step": 8980 }, { "epoch": 4.0693248753964655, "grad_norm": 0.15067653232804254, "learning_rate": 8.150996970006813e-06, "loss": 0.8473, "step": 8981 }, { "epoch": 4.069777979157227, "grad_norm": 0.14436455535141712, "learning_rate": 8.143343080147362e-06, "loss": 0.8453, "step": 8982 }, { "epoch": 4.070231082917989, "grad_norm": 0.14526850028048935, "learning_rate": 8.13569237825608e-06, "loss": 0.8239, "step": 8983 }, { "epoch": 4.07068418667875, "grad_norm": 0.16625763529652432, "learning_rate": 8.128044865098577e-06, "loss": 0.8696, "step": 8984 }, { "epoch": 4.071137290439511, "grad_norm": 0.132096438605392, "learning_rate": 8.120400541440183e-06, "loss": 0.856, "step": 8985 }, { "epoch": 4.071590394200272, "grad_norm": 0.12526497352805932, "learning_rate": 8.112759408045856e-06, "loss": 0.8428, "step": 8986 }, { "epoch": 4.072043497961033, "grad_norm": 0.14899210916764122, "learning_rate": 8.10512146568029e-06, "loss": 0.8525, "step": 8987 }, { "epoch": 4.072496601721794, "grad_norm": 0.15961779412970395, "learning_rate": 8.097486715107803e-06, "loss": 0.8794, "step": 8988 }, { "epoch": 4.072949705482555, "grad_norm": 0.15946902329750232, "learning_rate": 8.089855157092445e-06, "loss": 0.8389, "step": 8989 }, { "epoch": 4.0734028092433165, "grad_norm": 0.16618708297435456, "learning_rate": 8.082226792397904e-06, "loss": 0.8592, "step": 8990 }, { "epoch": 4.073855913004078, "grad_norm": 0.1565533399276409, "learning_rate": 8.074601621787593e-06, "loss": 0.8478, "step": 8991 }, { "epoch": 4.074309016764839, "grad_norm": 0.15852728919868422, "learning_rate": 8.06697964602456e-06, "loss": 0.8617, "step": 8992 }, { "epoch": 4.0747621205256, "grad_norm": 0.1419936748723508, "learning_rate": 8.059360865871553e-06, "loss": 0.8611, "step": 8993 }, { "epoch": 4.075215224286362, "grad_norm": 0.13874966939114033, "learning_rate": 8.05174528209101e-06, "loss": 0.8467, "step": 8994 }, { "epoch": 4.075668328047123, "grad_norm": 0.13371080814301664, "learning_rate": 8.04413289544503e-06, "loss": 0.8413, "step": 8995 }, { "epoch": 4.076121431807884, "grad_norm": 0.14036678839658814, "learning_rate": 8.036523706695419e-06, "loss": 0.8397, "step": 8996 }, { "epoch": 4.076574535568645, "grad_norm": 0.1286758912317606, "learning_rate": 8.028917716603621e-06, "loss": 0.8284, "step": 8997 }, { "epoch": 4.0770276393294065, "grad_norm": 0.11712905237418081, "learning_rate": 8.021314925930807e-06, "loss": 0.8315, "step": 8998 }, { "epoch": 4.077480743090168, "grad_norm": 0.122876026904278, "learning_rate": 8.0137153354378e-06, "loss": 0.8518, "step": 8999 }, { "epoch": 4.077933846850929, "grad_norm": 0.11693305482137183, "learning_rate": 8.006118945885096e-06, "loss": 0.832, "step": 9000 }, { "epoch": 4.07838695061169, "grad_norm": 0.12999726563090921, "learning_rate": 7.9985257580329e-06, "loss": 0.8585, "step": 9001 }, { "epoch": 4.078840054372451, "grad_norm": 0.1095858991058873, "learning_rate": 7.990935772641065e-06, "loss": 0.8416, "step": 9002 }, { "epoch": 4.079293158133212, "grad_norm": 0.11846465353631379, "learning_rate": 7.983348990469152e-06, "loss": 0.8622, "step": 9003 }, { "epoch": 4.079746261893973, "grad_norm": 0.1189468019089531, "learning_rate": 7.975765412276368e-06, "loss": 0.8415, "step": 9004 }, { "epoch": 4.080199365654735, "grad_norm": 0.11988089693215145, "learning_rate": 7.968185038821646e-06, "loss": 0.8317, "step": 9005 }, { "epoch": 4.080652469415496, "grad_norm": 0.1115090782989817, "learning_rate": 7.960607870863545e-06, "loss": 0.8306, "step": 9006 }, { "epoch": 4.0811055731762576, "grad_norm": 0.12430040515760214, "learning_rate": 7.953033909160353e-06, "loss": 0.8514, "step": 9007 }, { "epoch": 4.081558676937019, "grad_norm": 0.11646471279421368, "learning_rate": 7.945463154469996e-06, "loss": 0.8448, "step": 9008 }, { "epoch": 4.08201178069778, "grad_norm": 0.11290592405795655, "learning_rate": 7.937895607550108e-06, "loss": 0.8554, "step": 9009 }, { "epoch": 4.082464884458541, "grad_norm": 0.12004151740471158, "learning_rate": 7.930331269157974e-06, "loss": 0.8612, "step": 9010 }, { "epoch": 4.082917988219302, "grad_norm": 0.11265422230796375, "learning_rate": 7.922770140050594e-06, "loss": 0.8622, "step": 9011 }, { "epoch": 4.083371091980063, "grad_norm": 0.11649693387452224, "learning_rate": 7.915212220984619e-06, "loss": 0.8335, "step": 9012 }, { "epoch": 4.083824195740824, "grad_norm": 0.12055571310298817, "learning_rate": 7.907657512716378e-06, "loss": 0.8299, "step": 9013 }, { "epoch": 4.0842772995015855, "grad_norm": 0.11866877221938224, "learning_rate": 7.900106016001906e-06, "loss": 0.8367, "step": 9014 }, { "epoch": 4.0847304032623475, "grad_norm": 0.11253878481683474, "learning_rate": 7.892557731596882e-06, "loss": 0.8433, "step": 9015 }, { "epoch": 4.085183507023109, "grad_norm": 0.13496524985707947, "learning_rate": 7.885012660256696e-06, "loss": 0.8286, "step": 9016 }, { "epoch": 4.08563661078387, "grad_norm": 0.11660710518557607, "learning_rate": 7.87747080273638e-06, "loss": 0.8443, "step": 9017 }, { "epoch": 4.086089714544631, "grad_norm": 0.12406733350858515, "learning_rate": 7.869932159790687e-06, "loss": 0.8508, "step": 9018 }, { "epoch": 4.086542818305392, "grad_norm": 0.12642161516454187, "learning_rate": 7.862396732174016e-06, "loss": 0.8371, "step": 9019 }, { "epoch": 4.086995922066153, "grad_norm": 0.12332981886781527, "learning_rate": 7.854864520640446e-06, "loss": 0.8263, "step": 9020 }, { "epoch": 4.087449025826914, "grad_norm": 0.12713359958743028, "learning_rate": 7.847335525943763e-06, "loss": 0.8517, "step": 9021 }, { "epoch": 4.087902129587675, "grad_norm": 0.12002245039975712, "learning_rate": 7.839809748837384e-06, "loss": 0.8586, "step": 9022 }, { "epoch": 4.0883552333484365, "grad_norm": 0.12105452734895686, "learning_rate": 7.832287190074455e-06, "loss": 0.8367, "step": 9023 }, { "epoch": 4.088808337109198, "grad_norm": 0.11173007112521044, "learning_rate": 7.824767850407755e-06, "loss": 0.851, "step": 9024 }, { "epoch": 4.089261440869959, "grad_norm": 0.11911777323598689, "learning_rate": 7.817251730589786e-06, "loss": 0.838, "step": 9025 }, { "epoch": 4.089714544630721, "grad_norm": 0.13302499587833258, "learning_rate": 7.809738831372677e-06, "loss": 0.8562, "step": 9026 }, { "epoch": 4.090167648391482, "grad_norm": 0.11597999549132577, "learning_rate": 7.802229153508283e-06, "loss": 0.8448, "step": 9027 }, { "epoch": 4.090620752152243, "grad_norm": 0.11523250273041015, "learning_rate": 7.794722697748107e-06, "loss": 0.8297, "step": 9028 }, { "epoch": 4.091073855913004, "grad_norm": 0.1295458507002296, "learning_rate": 7.787219464843324e-06, "loss": 0.8524, "step": 9029 }, { "epoch": 4.091526959673765, "grad_norm": 0.11553641927535747, "learning_rate": 7.779719455544819e-06, "loss": 0.8445, "step": 9030 }, { "epoch": 4.0919800634345265, "grad_norm": 0.12458188703336245, "learning_rate": 7.772222670603118e-06, "loss": 0.8554, "step": 9031 }, { "epoch": 4.092433167195288, "grad_norm": 0.11825849876910417, "learning_rate": 7.764729110768462e-06, "loss": 0.8642, "step": 9032 }, { "epoch": 4.092886270956049, "grad_norm": 0.1284537897107612, "learning_rate": 7.757238776790722e-06, "loss": 0.8439, "step": 9033 }, { "epoch": 4.09333937471681, "grad_norm": 0.12481724696178118, "learning_rate": 7.7497516694195e-06, "loss": 0.8496, "step": 9034 }, { "epoch": 4.093792478477571, "grad_norm": 0.1268059947531071, "learning_rate": 7.742267789404034e-06, "loss": 0.8514, "step": 9035 }, { "epoch": 4.094245582238333, "grad_norm": 0.11442694817717419, "learning_rate": 7.734787137493245e-06, "loss": 0.8436, "step": 9036 }, { "epoch": 4.094698685999094, "grad_norm": 0.12829363792205764, "learning_rate": 7.727309714435756e-06, "loss": 0.8465, "step": 9037 }, { "epoch": 4.095151789759855, "grad_norm": 0.11071122659084918, "learning_rate": 7.719835520979831e-06, "loss": 0.8506, "step": 9038 }, { "epoch": 4.095604893520616, "grad_norm": 0.13759966661861736, "learning_rate": 7.712364557873454e-06, "loss": 0.8396, "step": 9039 }, { "epoch": 4.0960579972813775, "grad_norm": 0.11310704609662894, "learning_rate": 7.704896825864234e-06, "loss": 0.8399, "step": 9040 }, { "epoch": 4.096511101042139, "grad_norm": 0.12443237402339258, "learning_rate": 7.697432325699509e-06, "loss": 0.8597, "step": 9041 }, { "epoch": 4.0969642048029, "grad_norm": 0.13379804719851554, "learning_rate": 7.689971058126247e-06, "loss": 0.8516, "step": 9042 }, { "epoch": 4.097417308563661, "grad_norm": 0.10470233032553201, "learning_rate": 7.682513023891132e-06, "loss": 0.827, "step": 9043 }, { "epoch": 4.097870412324422, "grad_norm": 0.12344460505735878, "learning_rate": 7.675058223740501e-06, "loss": 0.8519, "step": 9044 }, { "epoch": 4.098323516085183, "grad_norm": 0.11439593683836853, "learning_rate": 7.667606658420359e-06, "loss": 0.846, "step": 9045 }, { "epoch": 4.098776619845944, "grad_norm": 0.11955515438214462, "learning_rate": 7.660158328676424e-06, "loss": 0.8534, "step": 9046 }, { "epoch": 4.099229723606706, "grad_norm": 0.12119326715579694, "learning_rate": 7.65271323525404e-06, "loss": 0.8483, "step": 9047 }, { "epoch": 4.0996828273674675, "grad_norm": 0.14256357417632964, "learning_rate": 7.645271378898287e-06, "loss": 0.8593, "step": 9048 }, { "epoch": 4.100135931128229, "grad_norm": 0.12749806923874415, "learning_rate": 7.637832760353858e-06, "loss": 0.8454, "step": 9049 }, { "epoch": 4.10058903488899, "grad_norm": 0.11671171059415257, "learning_rate": 7.630397380365177e-06, "loss": 0.8576, "step": 9050 }, { "epoch": 4.101042138649751, "grad_norm": 0.1361760173262336, "learning_rate": 7.6229652396762985e-06, "loss": 0.8224, "step": 9051 }, { "epoch": 4.101495242410512, "grad_norm": 0.11815280278701824, "learning_rate": 7.6155363390309955e-06, "loss": 0.8626, "step": 9052 }, { "epoch": 4.101948346171273, "grad_norm": 0.1406711122048231, "learning_rate": 7.608110679172682e-06, "loss": 0.8665, "step": 9053 }, { "epoch": 4.102401449932034, "grad_norm": 0.12703796047694607, "learning_rate": 7.600688260844449e-06, "loss": 0.8588, "step": 9054 }, { "epoch": 4.102854553692795, "grad_norm": 0.12894539110758876, "learning_rate": 7.593269084789101e-06, "loss": 0.8578, "step": 9055 }, { "epoch": 4.1033076574535565, "grad_norm": 0.12116924787673708, "learning_rate": 7.585853151749067e-06, "loss": 0.8402, "step": 9056 }, { "epoch": 4.103760761214318, "grad_norm": 0.11456688199625026, "learning_rate": 7.578440462466496e-06, "loss": 0.8661, "step": 9057 }, { "epoch": 4.10421386497508, "grad_norm": 0.10807000551939576, "learning_rate": 7.571031017683177e-06, "loss": 0.8348, "step": 9058 }, { "epoch": 4.104666968735841, "grad_norm": 0.12364457937808551, "learning_rate": 7.563624818140609e-06, "loss": 0.8583, "step": 9059 }, { "epoch": 4.105120072496602, "grad_norm": 0.11336571409940317, "learning_rate": 7.556221864579933e-06, "loss": 0.8376, "step": 9060 }, { "epoch": 4.105573176257363, "grad_norm": 0.1402159262575629, "learning_rate": 7.548822157741971e-06, "loss": 0.8565, "step": 9061 }, { "epoch": 4.106026280018124, "grad_norm": 0.11496711362562277, "learning_rate": 7.541425698367253e-06, "loss": 0.8486, "step": 9062 }, { "epoch": 4.106479383778885, "grad_norm": 0.11771955314418389, "learning_rate": 7.534032487195935e-06, "loss": 0.8331, "step": 9063 }, { "epoch": 4.1069324875396465, "grad_norm": 0.1886040437771919, "learning_rate": 7.526642524967895e-06, "loss": 0.8409, "step": 9064 }, { "epoch": 4.107385591300408, "grad_norm": 0.1296526759469234, "learning_rate": 7.519255812422641e-06, "loss": 0.8637, "step": 9065 }, { "epoch": 4.107838695061169, "grad_norm": 0.11322580152895949, "learning_rate": 7.511872350299407e-06, "loss": 0.8336, "step": 9066 }, { "epoch": 4.10829179882193, "grad_norm": 0.12994218292380272, "learning_rate": 7.504492139337038e-06, "loss": 0.8453, "step": 9067 }, { "epoch": 4.108744902582692, "grad_norm": 0.12872974022509776, "learning_rate": 7.497115180274113e-06, "loss": 0.8283, "step": 9068 }, { "epoch": 4.109198006343453, "grad_norm": 0.12650897379853995, "learning_rate": 7.4897414738488435e-06, "loss": 0.8428, "step": 9069 }, { "epoch": 4.109651110104214, "grad_norm": 0.13153446792437887, "learning_rate": 7.482371020799157e-06, "loss": 0.8554, "step": 9070 }, { "epoch": 4.110104213864975, "grad_norm": 0.12994071245848005, "learning_rate": 7.4750038218626054e-06, "loss": 0.8405, "step": 9071 }, { "epoch": 4.110557317625736, "grad_norm": 0.13091540792030573, "learning_rate": 7.46763987777646e-06, "loss": 0.8484, "step": 9072 }, { "epoch": 4.1110104213864975, "grad_norm": 0.11691523976034777, "learning_rate": 7.460279189277648e-06, "loss": 0.8544, "step": 9073 }, { "epoch": 4.111463525147259, "grad_norm": 0.14231049201806728, "learning_rate": 7.452921757102749e-06, "loss": 0.8226, "step": 9074 }, { "epoch": 4.11191662890802, "grad_norm": 0.129820796470804, "learning_rate": 7.445567581988063e-06, "loss": 0.849, "step": 9075 }, { "epoch": 4.112369732668781, "grad_norm": 0.13562341936611566, "learning_rate": 7.438216664669516e-06, "loss": 0.8498, "step": 9076 }, { "epoch": 4.112822836429542, "grad_norm": 0.1461449281588022, "learning_rate": 7.4308690058827544e-06, "loss": 0.8604, "step": 9077 }, { "epoch": 4.113275940190303, "grad_norm": 0.15262864941073365, "learning_rate": 7.423524606363051e-06, "loss": 0.8559, "step": 9078 }, { "epoch": 4.113729043951065, "grad_norm": 0.11754660666182298, "learning_rate": 7.416183466845406e-06, "loss": 0.8499, "step": 9079 }, { "epoch": 4.114182147711826, "grad_norm": 0.15723157168892696, "learning_rate": 7.408845588064441e-06, "loss": 0.8571, "step": 9080 }, { "epoch": 4.1146352514725875, "grad_norm": 0.1281471586994, "learning_rate": 7.401510970754473e-06, "loss": 0.8347, "step": 9081 }, { "epoch": 4.115088355233349, "grad_norm": 0.12182280299459551, "learning_rate": 7.394179615649513e-06, "loss": 0.8383, "step": 9082 }, { "epoch": 4.11554145899411, "grad_norm": 0.1475734950111609, "learning_rate": 7.386851523483205e-06, "loss": 0.8507, "step": 9083 }, { "epoch": 4.115994562754871, "grad_norm": 0.11897351470442616, "learning_rate": 7.379526694988906e-06, "loss": 0.8485, "step": 9084 }, { "epoch": 4.116447666515632, "grad_norm": 0.13194892308626524, "learning_rate": 7.37220513089961e-06, "loss": 0.8468, "step": 9085 }, { "epoch": 4.116900770276393, "grad_norm": 0.14821323832839015, "learning_rate": 7.364886831948022e-06, "loss": 0.8493, "step": 9086 }, { "epoch": 4.117353874037154, "grad_norm": 0.12353612182199976, "learning_rate": 7.3575717988664826e-06, "loss": 0.8508, "step": 9087 }, { "epoch": 4.117806977797915, "grad_norm": 0.15214958767132924, "learning_rate": 7.35026003238704e-06, "loss": 0.8318, "step": 9088 }, { "epoch": 4.1182600815586765, "grad_norm": 0.13943036895150035, "learning_rate": 7.342951533241396e-06, "loss": 0.8652, "step": 9089 }, { "epoch": 4.1187131853194385, "grad_norm": 0.14089356276678297, "learning_rate": 7.335646302160913e-06, "loss": 0.8553, "step": 9090 }, { "epoch": 4.1191662890802, "grad_norm": 0.13042465943245668, "learning_rate": 7.328344339876663e-06, "loss": 0.8326, "step": 9091 }, { "epoch": 4.119619392840961, "grad_norm": 0.16206790880052516, "learning_rate": 7.321045647119351e-06, "loss": 0.848, "step": 9092 }, { "epoch": 4.120072496601722, "grad_norm": 0.13042928716940486, "learning_rate": 7.313750224619394e-06, "loss": 0.8425, "step": 9093 }, { "epoch": 4.120525600362483, "grad_norm": 0.14032492905586977, "learning_rate": 7.3064580731068415e-06, "loss": 0.8361, "step": 9094 }, { "epoch": 4.120978704123244, "grad_norm": 0.13921198507738242, "learning_rate": 7.299169193311453e-06, "loss": 0.8409, "step": 9095 }, { "epoch": 4.121431807884005, "grad_norm": 0.10746358240548591, "learning_rate": 7.29188358596264e-06, "loss": 0.8479, "step": 9096 }, { "epoch": 4.1218849116447664, "grad_norm": 0.14637941560102088, "learning_rate": 7.2846012517894695e-06, "loss": 0.8558, "step": 9097 }, { "epoch": 4.122338015405528, "grad_norm": 0.11948209597036609, "learning_rate": 7.277322191520731e-06, "loss": 0.8598, "step": 9098 }, { "epoch": 4.122791119166289, "grad_norm": 0.13583001580390625, "learning_rate": 7.270046405884832e-06, "loss": 0.8577, "step": 9099 }, { "epoch": 4.123244222927051, "grad_norm": 0.12770270587056234, "learning_rate": 7.2627738956098985e-06, "loss": 0.84, "step": 9100 }, { "epoch": 4.123697326687812, "grad_norm": 0.1301464311379594, "learning_rate": 7.255504661423684e-06, "loss": 0.8518, "step": 9101 }, { "epoch": 4.124150430448573, "grad_norm": 0.13694401039355292, "learning_rate": 7.248238704053663e-06, "loss": 0.8225, "step": 9102 }, { "epoch": 4.124603534209334, "grad_norm": 0.13389467955587325, "learning_rate": 7.240976024226931e-06, "loss": 0.8358, "step": 9103 }, { "epoch": 4.125056637970095, "grad_norm": 0.13410325975355505, "learning_rate": 7.233716622670304e-06, "loss": 0.832, "step": 9104 }, { "epoch": 4.125509741730856, "grad_norm": 0.1122602266523896, "learning_rate": 7.226460500110235e-06, "loss": 0.8442, "step": 9105 }, { "epoch": 4.1259628454916175, "grad_norm": 0.12194001846882738, "learning_rate": 7.219207657272851e-06, "loss": 0.8438, "step": 9106 }, { "epoch": 4.126415949252379, "grad_norm": 0.13331715677526335, "learning_rate": 7.211958094883984e-06, "loss": 0.8567, "step": 9107 }, { "epoch": 4.12686905301314, "grad_norm": 0.13922224660505614, "learning_rate": 7.204711813669085e-06, "loss": 0.8375, "step": 9108 }, { "epoch": 4.127322156773901, "grad_norm": 0.11369933751781115, "learning_rate": 7.197468814353339e-06, "loss": 0.8433, "step": 9109 }, { "epoch": 4.127775260534662, "grad_norm": 0.12408217513044877, "learning_rate": 7.190229097661539e-06, "loss": 0.8314, "step": 9110 }, { "epoch": 4.128228364295424, "grad_norm": 0.12223128645991954, "learning_rate": 7.1829926643182025e-06, "loss": 0.8438, "step": 9111 }, { "epoch": 4.128681468056185, "grad_norm": 0.1189303599140154, "learning_rate": 7.17575951504748e-06, "loss": 0.8349, "step": 9112 }, { "epoch": 4.129134571816946, "grad_norm": 0.11144994069977611, "learning_rate": 7.1685296505732234e-06, "loss": 0.8375, "step": 9113 }, { "epoch": 4.1295876755777075, "grad_norm": 0.15245854426637992, "learning_rate": 7.161303071618935e-06, "loss": 0.8597, "step": 9114 }, { "epoch": 4.130040779338469, "grad_norm": 0.12530489860157581, "learning_rate": 7.1540797789077855e-06, "loss": 0.8502, "step": 9115 }, { "epoch": 4.13049388309923, "grad_norm": 0.13682282075205704, "learning_rate": 7.146859773162642e-06, "loss": 0.8617, "step": 9116 }, { "epoch": 4.130946986859991, "grad_norm": 0.1427559310596577, "learning_rate": 7.1396430551060115e-06, "loss": 0.8462, "step": 9117 }, { "epoch": 4.131400090620752, "grad_norm": 0.12082582922729904, "learning_rate": 7.1324296254601065e-06, "loss": 0.845, "step": 9118 }, { "epoch": 4.131853194381513, "grad_norm": 0.15116746585353533, "learning_rate": 7.1252194849467685e-06, "loss": 0.8457, "step": 9119 }, { "epoch": 4.132306298142274, "grad_norm": 0.1226792246157594, "learning_rate": 7.118012634287557e-06, "loss": 0.8555, "step": 9120 }, { "epoch": 4.132759401903035, "grad_norm": 0.14903933305907205, "learning_rate": 7.110809074203664e-06, "loss": 0.8094, "step": 9121 }, { "epoch": 4.133212505663797, "grad_norm": 0.13128410903152238, "learning_rate": 7.103608805415962e-06, "loss": 0.855, "step": 9122 }, { "epoch": 4.1336656094245585, "grad_norm": 0.1201384426594708, "learning_rate": 7.096411828645009e-06, "loss": 0.8337, "step": 9123 }, { "epoch": 4.13411871318532, "grad_norm": 0.17226086877431904, "learning_rate": 7.089218144611023e-06, "loss": 0.8532, "step": 9124 }, { "epoch": 4.134571816946081, "grad_norm": 0.12921785219886184, "learning_rate": 7.082027754033882e-06, "loss": 0.847, "step": 9125 }, { "epoch": 4.135024920706842, "grad_norm": 0.16725005352858063, "learning_rate": 7.0748406576331444e-06, "loss": 0.8474, "step": 9126 }, { "epoch": 4.135478024467603, "grad_norm": 0.12931447342823996, "learning_rate": 7.067656856128056e-06, "loss": 0.85, "step": 9127 }, { "epoch": 4.135931128228364, "grad_norm": 0.13565722837786756, "learning_rate": 7.060476350237491e-06, "loss": 0.8523, "step": 9128 }, { "epoch": 4.136384231989125, "grad_norm": 0.13603295146688818, "learning_rate": 7.0532991406800435e-06, "loss": 0.8411, "step": 9129 }, { "epoch": 4.136837335749886, "grad_norm": 0.10834508046124176, "learning_rate": 7.046125228173935e-06, "loss": 0.8372, "step": 9130 }, { "epoch": 4.137290439510648, "grad_norm": 0.12945610340717867, "learning_rate": 7.038954613437092e-06, "loss": 0.8494, "step": 9131 }, { "epoch": 4.137743543271409, "grad_norm": 0.11517696874945275, "learning_rate": 7.031787297187076e-06, "loss": 0.8431, "step": 9132 }, { "epoch": 4.138196647032171, "grad_norm": 0.11041585915829806, "learning_rate": 7.0246232801411566e-06, "loss": 0.8475, "step": 9133 }, { "epoch": 4.138649750792932, "grad_norm": 0.11912328612437509, "learning_rate": 7.017462563016244e-06, "loss": 0.8506, "step": 9134 }, { "epoch": 4.139102854553693, "grad_norm": 0.12061129740494271, "learning_rate": 7.010305146528913e-06, "loss": 0.8574, "step": 9135 }, { "epoch": 4.139555958314454, "grad_norm": 0.13049152850217677, "learning_rate": 7.0031510313954476e-06, "loss": 0.8381, "step": 9136 }, { "epoch": 4.140009062075215, "grad_norm": 0.1489159793021914, "learning_rate": 6.996000218331751e-06, "loss": 0.8438, "step": 9137 }, { "epoch": 4.140462165835976, "grad_norm": 0.13281644632344716, "learning_rate": 6.98885270805345e-06, "loss": 0.8846, "step": 9138 }, { "epoch": 4.1409152695967375, "grad_norm": 0.12003842795808896, "learning_rate": 6.981708501275784e-06, "loss": 0.8442, "step": 9139 }, { "epoch": 4.141368373357499, "grad_norm": 0.1347160444176712, "learning_rate": 6.9745675987137156e-06, "loss": 0.8491, "step": 9140 }, { "epoch": 4.14182147711826, "grad_norm": 0.13059409485443213, "learning_rate": 6.967430001081834e-06, "loss": 0.8463, "step": 9141 }, { "epoch": 4.142274580879021, "grad_norm": 0.13144007707431532, "learning_rate": 6.960295709094413e-06, "loss": 0.8385, "step": 9142 }, { "epoch": 4.142727684639783, "grad_norm": 0.14495867249291172, "learning_rate": 6.953164723465416e-06, "loss": 0.855, "step": 9143 }, { "epoch": 4.143180788400544, "grad_norm": 0.1254721112723603, "learning_rate": 6.946037044908434e-06, "loss": 0.8422, "step": 9144 }, { "epoch": 4.143633892161305, "grad_norm": 0.13708023572844755, "learning_rate": 6.9389126741367685e-06, "loss": 0.8317, "step": 9145 }, { "epoch": 4.144086995922066, "grad_norm": 0.12625053027373795, "learning_rate": 6.931791611863352e-06, "loss": 0.8494, "step": 9146 }, { "epoch": 4.144540099682827, "grad_norm": 0.1286182277143583, "learning_rate": 6.924673858800832e-06, "loss": 0.8361, "step": 9147 }, { "epoch": 4.144993203443589, "grad_norm": 0.1329591809846905, "learning_rate": 6.91755941566147e-06, "loss": 0.8404, "step": 9148 }, { "epoch": 4.14544630720435, "grad_norm": 0.11868812620433784, "learning_rate": 6.9104482831572515e-06, "loss": 0.8566, "step": 9149 }, { "epoch": 4.145899410965111, "grad_norm": 0.11684552033366775, "learning_rate": 6.90334046199979e-06, "loss": 0.84, "step": 9150 }, { "epoch": 4.146352514725872, "grad_norm": 0.13393488929750516, "learning_rate": 6.89623595290037e-06, "loss": 0.8319, "step": 9151 }, { "epoch": 4.146805618486633, "grad_norm": 0.13337418939786044, "learning_rate": 6.889134756569977e-06, "loss": 0.8627, "step": 9152 }, { "epoch": 4.147258722247394, "grad_norm": 0.1279506226596989, "learning_rate": 6.8820368737192264e-06, "loss": 0.8534, "step": 9153 }, { "epoch": 4.147711826008156, "grad_norm": 0.14038520886252162, "learning_rate": 6.874942305058434e-06, "loss": 0.8689, "step": 9154 }, { "epoch": 4.148164929768917, "grad_norm": 0.13195335711440875, "learning_rate": 6.867851051297556e-06, "loss": 0.843, "step": 9155 }, { "epoch": 4.1486180335296785, "grad_norm": 0.15480061951060045, "learning_rate": 6.860763113146247e-06, "loss": 0.8738, "step": 9156 }, { "epoch": 4.14907113729044, "grad_norm": 0.13792102266604236, "learning_rate": 6.853678491313806e-06, "loss": 0.826, "step": 9157 }, { "epoch": 4.149524241051201, "grad_norm": 0.11846601345286187, "learning_rate": 6.846597186509192e-06, "loss": 0.8416, "step": 9158 }, { "epoch": 4.149977344811962, "grad_norm": 0.13550037501699352, "learning_rate": 6.839519199441067e-06, "loss": 0.8454, "step": 9159 }, { "epoch": 4.150430448572723, "grad_norm": 0.14935546814528386, "learning_rate": 6.832444530817728e-06, "loss": 0.8673, "step": 9160 }, { "epoch": 4.150883552333484, "grad_norm": 0.14536851122110114, "learning_rate": 6.825373181347168e-06, "loss": 0.8275, "step": 9161 }, { "epoch": 4.151336656094245, "grad_norm": 0.1376603086801455, "learning_rate": 6.818305151737017e-06, "loss": 0.8244, "step": 9162 }, { "epoch": 4.151789759855006, "grad_norm": 0.12706417230073108, "learning_rate": 6.811240442694602e-06, "loss": 0.8514, "step": 9163 }, { "epoch": 4.152242863615768, "grad_norm": 0.15981329176985082, "learning_rate": 6.804179054926892e-06, "loss": 0.8473, "step": 9164 }, { "epoch": 4.15269596737653, "grad_norm": 0.15193243730239664, "learning_rate": 6.797120989140551e-06, "loss": 0.8491, "step": 9165 }, { "epoch": 4.153149071137291, "grad_norm": 0.12815986833436402, "learning_rate": 6.790066246041891e-06, "loss": 0.861, "step": 9166 }, { "epoch": 4.153602174898052, "grad_norm": 0.178131603467983, "learning_rate": 6.78301482633688e-06, "loss": 0.8333, "step": 9167 }, { "epoch": 4.154055278658813, "grad_norm": 0.13116442569598694, "learning_rate": 6.775966730731194e-06, "loss": 0.8375, "step": 9168 }, { "epoch": 4.154508382419574, "grad_norm": 0.1417945824408333, "learning_rate": 6.768921959930126e-06, "loss": 0.8573, "step": 9169 }, { "epoch": 4.154961486180335, "grad_norm": 0.1577265048377004, "learning_rate": 6.76188051463869e-06, "loss": 0.8591, "step": 9170 }, { "epoch": 4.155414589941096, "grad_norm": 0.11326851931764902, "learning_rate": 6.754842395561518e-06, "loss": 0.8601, "step": 9171 }, { "epoch": 4.1558676937018575, "grad_norm": 0.13202943731229058, "learning_rate": 6.747807603402945e-06, "loss": 0.8387, "step": 9172 }, { "epoch": 4.156320797462619, "grad_norm": 0.14025684436562602, "learning_rate": 6.7407761388669445e-06, "loss": 0.8458, "step": 9173 }, { "epoch": 4.15677390122338, "grad_norm": 0.1102534488936095, "learning_rate": 6.73374800265719e-06, "loss": 0.8446, "step": 9174 }, { "epoch": 4.157227004984142, "grad_norm": 0.131160777809434, "learning_rate": 6.7267231954769895e-06, "loss": 0.8368, "step": 9175 }, { "epoch": 4.157680108744903, "grad_norm": 0.10801560285150741, "learning_rate": 6.719701718029323e-06, "loss": 0.8504, "step": 9176 }, { "epoch": 4.158133212505664, "grad_norm": 0.10343876658514715, "learning_rate": 6.7126835710168694e-06, "loss": 0.831, "step": 9177 }, { "epoch": 4.158586316266425, "grad_norm": 0.10945279325587357, "learning_rate": 6.7056687551419274e-06, "loss": 0.8425, "step": 9178 }, { "epoch": 4.159039420027186, "grad_norm": 0.11869747914930508, "learning_rate": 6.6986572711065055e-06, "loss": 0.8501, "step": 9179 }, { "epoch": 4.159492523787947, "grad_norm": 0.10926608929288399, "learning_rate": 6.691649119612243e-06, "loss": 0.8425, "step": 9180 }, { "epoch": 4.159945627548709, "grad_norm": 0.1753757326052441, "learning_rate": 6.684644301360484e-06, "loss": 0.845, "step": 9181 }, { "epoch": 4.16039873130947, "grad_norm": 0.11577975403456285, "learning_rate": 6.677642817052183e-06, "loss": 0.836, "step": 9182 }, { "epoch": 4.160851835070231, "grad_norm": 0.11761586291515738, "learning_rate": 6.670644667388022e-06, "loss": 0.8504, "step": 9183 }, { "epoch": 4.161304938830992, "grad_norm": 0.11954115688409411, "learning_rate": 6.663649853068302e-06, "loss": 0.8286, "step": 9184 }, { "epoch": 4.161758042591753, "grad_norm": 0.11543995506844111, "learning_rate": 6.6566583747930305e-06, "loss": 0.8531, "step": 9185 }, { "epoch": 4.162211146352515, "grad_norm": 0.11076084755670097, "learning_rate": 6.649670233261845e-06, "loss": 0.8583, "step": 9186 }, { "epoch": 4.162664250113276, "grad_norm": 0.11237931302211701, "learning_rate": 6.642685429174065e-06, "loss": 0.8457, "step": 9187 }, { "epoch": 4.163117353874037, "grad_norm": 0.11262138677872024, "learning_rate": 6.635703963228692e-06, "loss": 0.8529, "step": 9188 }, { "epoch": 4.1635704576347985, "grad_norm": 0.11391763767658401, "learning_rate": 6.62872583612435e-06, "loss": 0.8496, "step": 9189 }, { "epoch": 4.16402356139556, "grad_norm": 0.10664819581603623, "learning_rate": 6.621751048559386e-06, "loss": 0.8513, "step": 9190 }, { "epoch": 4.164476665156321, "grad_norm": 0.12841013255750477, "learning_rate": 6.614779601231758e-06, "loss": 0.8429, "step": 9191 }, { "epoch": 4.164929768917082, "grad_norm": 0.11114170410724991, "learning_rate": 6.607811494839138e-06, "loss": 0.8554, "step": 9192 }, { "epoch": 4.165382872677843, "grad_norm": 0.1439156579534054, "learning_rate": 6.600846730078813e-06, "loss": 0.8625, "step": 9193 }, { "epoch": 4.165835976438604, "grad_norm": 0.12153527870151153, "learning_rate": 6.593885307647791e-06, "loss": 0.8542, "step": 9194 }, { "epoch": 4.166289080199365, "grad_norm": 0.1251090176189254, "learning_rate": 6.586927228242701e-06, "loss": 0.8397, "step": 9195 }, { "epoch": 4.166742183960127, "grad_norm": 0.1379473340904233, "learning_rate": 6.5799724925598475e-06, "loss": 0.8431, "step": 9196 }, { "epoch": 4.167195287720888, "grad_norm": 0.12530345073848556, "learning_rate": 6.57302110129523e-06, "loss": 0.8389, "step": 9197 }, { "epoch": 4.16764839148165, "grad_norm": 0.131711606206146, "learning_rate": 6.56607305514446e-06, "loss": 0.8406, "step": 9198 }, { "epoch": 4.168101495242411, "grad_norm": 0.12618942257664453, "learning_rate": 6.559128354802871e-06, "loss": 0.8371, "step": 9199 }, { "epoch": 4.168554599003172, "grad_norm": 0.12212009693088784, "learning_rate": 6.552187000965418e-06, "loss": 0.8463, "step": 9200 }, { "epoch": 4.169007702763933, "grad_norm": 0.13745604588463373, "learning_rate": 6.545248994326754e-06, "loss": 0.8529, "step": 9201 }, { "epoch": 4.169460806524694, "grad_norm": 0.12791184422926444, "learning_rate": 6.5383143355811685e-06, "loss": 0.8471, "step": 9202 }, { "epoch": 4.169913910285455, "grad_norm": 0.11253740287406824, "learning_rate": 6.5313830254226216e-06, "loss": 0.8487, "step": 9203 }, { "epoch": 4.170367014046216, "grad_norm": 0.1510368722443346, "learning_rate": 6.524455064544768e-06, "loss": 0.8241, "step": 9204 }, { "epoch": 4.1708201178069775, "grad_norm": 0.1127504502708648, "learning_rate": 6.517530453640879e-06, "loss": 0.8302, "step": 9205 }, { "epoch": 4.171273221567739, "grad_norm": 0.12160083938896735, "learning_rate": 6.510609193403939e-06, "loss": 0.8456, "step": 9206 }, { "epoch": 4.171726325328501, "grad_norm": 0.11776621956403696, "learning_rate": 6.5036912845265566e-06, "loss": 0.8374, "step": 9207 }, { "epoch": 4.172179429089262, "grad_norm": 0.11667149616875966, "learning_rate": 6.496776727701038e-06, "loss": 0.8636, "step": 9208 }, { "epoch": 4.172632532850023, "grad_norm": 0.13575970722773437, "learning_rate": 6.489865523619321e-06, "loss": 0.8442, "step": 9209 }, { "epoch": 4.173085636610784, "grad_norm": 0.1056380702467168, "learning_rate": 6.4829576729730446e-06, "loss": 0.848, "step": 9210 }, { "epoch": 4.173538740371545, "grad_norm": 0.12828019005548413, "learning_rate": 6.476053176453487e-06, "loss": 0.8429, "step": 9211 }, { "epoch": 4.173991844132306, "grad_norm": 0.12023476037207725, "learning_rate": 6.46915203475158e-06, "loss": 0.8318, "step": 9212 }, { "epoch": 4.174444947893067, "grad_norm": 0.12073753403283534, "learning_rate": 6.462254248557962e-06, "loss": 0.8478, "step": 9213 }, { "epoch": 4.1748980516538285, "grad_norm": 0.10799083545059839, "learning_rate": 6.45535981856289e-06, "loss": 0.8107, "step": 9214 }, { "epoch": 4.17535115541459, "grad_norm": 0.11764306063480835, "learning_rate": 6.448468745456322e-06, "loss": 0.8278, "step": 9215 }, { "epoch": 4.175804259175351, "grad_norm": 0.11387720975616464, "learning_rate": 6.44158102992785e-06, "loss": 0.8545, "step": 9216 }, { "epoch": 4.176257362936113, "grad_norm": 0.10521136128887634, "learning_rate": 6.434696672666754e-06, "loss": 0.8376, "step": 9217 }, { "epoch": 4.176710466696874, "grad_norm": 0.10796411531244629, "learning_rate": 6.427815674361961e-06, "loss": 0.8377, "step": 9218 }, { "epoch": 4.177163570457635, "grad_norm": 0.11444083280722177, "learning_rate": 6.42093803570206e-06, "loss": 0.8422, "step": 9219 }, { "epoch": 4.177616674218396, "grad_norm": 0.11642370754687122, "learning_rate": 6.41406375737533e-06, "loss": 0.8371, "step": 9220 }, { "epoch": 4.178069777979157, "grad_norm": 0.1062661018133082, "learning_rate": 6.40719284006968e-06, "loss": 0.8411, "step": 9221 }, { "epoch": 4.1785228817399185, "grad_norm": 0.11241036406795896, "learning_rate": 6.400325284472715e-06, "loss": 0.863, "step": 9222 }, { "epoch": 4.17897598550068, "grad_norm": 0.12500979540103824, "learning_rate": 6.393461091271667e-06, "loss": 0.8525, "step": 9223 }, { "epoch": 4.179429089261441, "grad_norm": 0.11260061936288733, "learning_rate": 6.386600261153466e-06, "loss": 0.8639, "step": 9224 }, { "epoch": 4.179882193022202, "grad_norm": 0.11197827965952648, "learning_rate": 6.37974279480468e-06, "loss": 0.8508, "step": 9225 }, { "epoch": 4.180335296782963, "grad_norm": 0.13554617162560112, "learning_rate": 6.372888692911567e-06, "loss": 0.8387, "step": 9226 }, { "epoch": 4.180788400543724, "grad_norm": 0.11962333461388677, "learning_rate": 6.366037956160025e-06, "loss": 0.8549, "step": 9227 }, { "epoch": 4.181241504304486, "grad_norm": 0.1300382181885915, "learning_rate": 6.359190585235611e-06, "loss": 0.8324, "step": 9228 }, { "epoch": 4.181694608065247, "grad_norm": 0.13095581608478976, "learning_rate": 6.352346580823572e-06, "loss": 0.8482, "step": 9229 }, { "epoch": 4.182147711826008, "grad_norm": 0.13142632223714495, "learning_rate": 6.345505943608792e-06, "loss": 0.8427, "step": 9230 }, { "epoch": 4.1826008155867695, "grad_norm": 0.16155089671352252, "learning_rate": 6.338668674275848e-06, "loss": 0.868, "step": 9231 }, { "epoch": 4.183053919347531, "grad_norm": 0.11155242617991314, "learning_rate": 6.331834773508934e-06, "loss": 0.8295, "step": 9232 }, { "epoch": 4.183507023108292, "grad_norm": 0.13827497835275604, "learning_rate": 6.325004241991957e-06, "loss": 0.8441, "step": 9233 }, { "epoch": 4.183960126869053, "grad_norm": 0.13870012261580678, "learning_rate": 6.31817708040845e-06, "loss": 0.8546, "step": 9234 }, { "epoch": 4.184413230629814, "grad_norm": 0.12174695912734597, "learning_rate": 6.311353289441635e-06, "loss": 0.8673, "step": 9235 }, { "epoch": 4.184866334390575, "grad_norm": 0.15727400326717852, "learning_rate": 6.304532869774375e-06, "loss": 0.8513, "step": 9236 }, { "epoch": 4.185319438151336, "grad_norm": 0.12202809218923467, "learning_rate": 6.297715822089201e-06, "loss": 0.8643, "step": 9237 }, { "epoch": 4.1857725419120975, "grad_norm": 0.1435249323864566, "learning_rate": 6.290902147068334e-06, "loss": 0.8328, "step": 9238 }, { "epoch": 4.1862256456728595, "grad_norm": 0.1483445730983601, "learning_rate": 6.284091845393594e-06, "loss": 0.8476, "step": 9239 }, { "epoch": 4.186678749433621, "grad_norm": 0.13155835859854104, "learning_rate": 6.277284917746538e-06, "loss": 0.8491, "step": 9240 }, { "epoch": 4.187131853194382, "grad_norm": 0.15135714522575938, "learning_rate": 6.270481364808332e-06, "loss": 0.8364, "step": 9241 }, { "epoch": 4.187584956955143, "grad_norm": 0.13481517925859532, "learning_rate": 6.263681187259836e-06, "loss": 0.8613, "step": 9242 }, { "epoch": 4.188038060715904, "grad_norm": 0.13479955640747812, "learning_rate": 6.256884385781541e-06, "loss": 0.8291, "step": 9243 }, { "epoch": 4.188491164476665, "grad_norm": 0.12614456032604476, "learning_rate": 6.250090961053645e-06, "loss": 0.8219, "step": 9244 }, { "epoch": 4.188944268237426, "grad_norm": 0.11945160414504596, "learning_rate": 6.243300913755952e-06, "loss": 0.8613, "step": 9245 }, { "epoch": 4.189397371998187, "grad_norm": 0.12144776180963879, "learning_rate": 6.236514244567984e-06, "loss": 0.858, "step": 9246 }, { "epoch": 4.1898504757589485, "grad_norm": 0.12772637664302194, "learning_rate": 6.229730954168887e-06, "loss": 0.8397, "step": 9247 }, { "epoch": 4.19030357951971, "grad_norm": 0.10759951614949051, "learning_rate": 6.222951043237468e-06, "loss": 0.8343, "step": 9248 }, { "epoch": 4.190756683280471, "grad_norm": 0.11926729750889065, "learning_rate": 6.216174512452236e-06, "loss": 0.8299, "step": 9249 }, { "epoch": 4.191209787041233, "grad_norm": 0.12168343351868396, "learning_rate": 6.209401362491303e-06, "loss": 0.8516, "step": 9250 }, { "epoch": 4.191662890801994, "grad_norm": 0.12586332388592497, "learning_rate": 6.202631594032498e-06, "loss": 0.8288, "step": 9251 }, { "epoch": 4.192115994562755, "grad_norm": 0.12457840281538358, "learning_rate": 6.195865207753269e-06, "loss": 0.8537, "step": 9252 }, { "epoch": 4.192569098323516, "grad_norm": 0.12398791838208024, "learning_rate": 6.189102204330759e-06, "loss": 0.8639, "step": 9253 }, { "epoch": 4.193022202084277, "grad_norm": 0.12744207844938293, "learning_rate": 6.182342584441752e-06, "loss": 0.8458, "step": 9254 }, { "epoch": 4.1934753058450385, "grad_norm": 0.12624482531680917, "learning_rate": 6.175586348762688e-06, "loss": 0.8567, "step": 9255 }, { "epoch": 4.1939284096058, "grad_norm": 0.11933623113751857, "learning_rate": 6.168833497969698e-06, "loss": 0.8445, "step": 9256 }, { "epoch": 4.194381513366561, "grad_norm": 0.12476176774429407, "learning_rate": 6.1620840327385335e-06, "loss": 0.8794, "step": 9257 }, { "epoch": 4.194834617127322, "grad_norm": 0.11643289134383912, "learning_rate": 6.155337953744651e-06, "loss": 0.8564, "step": 9258 }, { "epoch": 4.195287720888083, "grad_norm": 0.11898018215397267, "learning_rate": 6.148595261663128e-06, "loss": 0.8404, "step": 9259 }, { "epoch": 4.195740824648845, "grad_norm": 0.12459515169917391, "learning_rate": 6.141855957168736e-06, "loss": 0.8481, "step": 9260 }, { "epoch": 4.196193928409606, "grad_norm": 0.12231082959444312, "learning_rate": 6.135120040935874e-06, "loss": 0.8374, "step": 9261 }, { "epoch": 4.196647032170367, "grad_norm": 0.1427004140508521, "learning_rate": 6.12838751363864e-06, "loss": 0.834, "step": 9262 }, { "epoch": 4.197100135931128, "grad_norm": 0.15302059269086157, "learning_rate": 6.1216583759507696e-06, "loss": 0.8227, "step": 9263 }, { "epoch": 4.1975532396918895, "grad_norm": 0.12093940646527243, "learning_rate": 6.114932628545647e-06, "loss": 0.8559, "step": 9264 }, { "epoch": 4.198006343452651, "grad_norm": 0.12948249618346774, "learning_rate": 6.108210272096351e-06, "loss": 0.8551, "step": 9265 }, { "epoch": 4.198459447213412, "grad_norm": 0.10901146230854065, "learning_rate": 6.1014913072755846e-06, "loss": 0.8504, "step": 9266 }, { "epoch": 4.198912550974173, "grad_norm": 0.11539773766647735, "learning_rate": 6.094775734755756e-06, "loss": 0.858, "step": 9267 }, { "epoch": 4.199365654734934, "grad_norm": 0.12624607628723652, "learning_rate": 6.0880635552088786e-06, "loss": 0.8553, "step": 9268 }, { "epoch": 4.199818758495695, "grad_norm": 0.11883310581296044, "learning_rate": 6.0813547693066865e-06, "loss": 0.8607, "step": 9269 }, { "epoch": 4.200271862256456, "grad_norm": 0.12591566122865122, "learning_rate": 6.074649377720509e-06, "loss": 0.85, "step": 9270 }, { "epoch": 4.200724966017218, "grad_norm": 0.14124557446971467, "learning_rate": 6.067947381121406e-06, "loss": 0.8513, "step": 9271 }, { "epoch": 4.2011780697779795, "grad_norm": 0.12238699313358781, "learning_rate": 6.061248780180036e-06, "loss": 0.8413, "step": 9272 }, { "epoch": 4.201631173538741, "grad_norm": 0.11969214488376352, "learning_rate": 6.054553575566746e-06, "loss": 0.8366, "step": 9273 }, { "epoch": 4.202084277299502, "grad_norm": 0.11524308966784419, "learning_rate": 6.047861767951553e-06, "loss": 0.833, "step": 9274 }, { "epoch": 4.202537381060263, "grad_norm": 0.11558257999987225, "learning_rate": 6.041173358004102e-06, "loss": 0.8275, "step": 9275 }, { "epoch": 4.202990484821024, "grad_norm": 0.13275957415852038, "learning_rate": 6.0344883463937385e-06, "loss": 0.8442, "step": 9276 }, { "epoch": 4.203443588581785, "grad_norm": 0.11986924895813009, "learning_rate": 6.0278067337894295e-06, "loss": 0.8396, "step": 9277 }, { "epoch": 4.203896692342546, "grad_norm": 0.13979739817787162, "learning_rate": 6.021128520859836e-06, "loss": 0.8481, "step": 9278 }, { "epoch": 4.204349796103307, "grad_norm": 0.11369149431842987, "learning_rate": 6.014453708273249e-06, "loss": 0.8374, "step": 9279 }, { "epoch": 4.2048028998640685, "grad_norm": 0.12258320776269338, "learning_rate": 6.007782296697629e-06, "loss": 0.8592, "step": 9280 }, { "epoch": 4.2052560036248305, "grad_norm": 0.13492116798571022, "learning_rate": 6.001114286800618e-06, "loss": 0.8792, "step": 9281 }, { "epoch": 4.205709107385592, "grad_norm": 0.11436385960699504, "learning_rate": 5.994449679249474e-06, "loss": 0.8464, "step": 9282 }, { "epoch": 4.206162211146353, "grad_norm": 0.12447186504905008, "learning_rate": 5.987788474711163e-06, "loss": 0.8528, "step": 9283 }, { "epoch": 4.206615314907114, "grad_norm": 0.12033931074919227, "learning_rate": 5.981130673852269e-06, "loss": 0.8452, "step": 9284 }, { "epoch": 4.207068418667875, "grad_norm": 0.12088122672293881, "learning_rate": 5.974476277339069e-06, "loss": 0.84, "step": 9285 }, { "epoch": 4.207521522428636, "grad_norm": 0.13062204299263505, "learning_rate": 5.967825285837467e-06, "loss": 0.8508, "step": 9286 }, { "epoch": 4.207974626189397, "grad_norm": 0.1344209220615256, "learning_rate": 5.9611777000130585e-06, "loss": 0.8391, "step": 9287 }, { "epoch": 4.2084277299501585, "grad_norm": 0.13137633609009772, "learning_rate": 5.954533520531076e-06, "loss": 0.8361, "step": 9288 }, { "epoch": 4.20888083371092, "grad_norm": 0.11831388632753734, "learning_rate": 5.94789274805641e-06, "loss": 0.8436, "step": 9289 }, { "epoch": 4.209333937471681, "grad_norm": 0.14913869424001905, "learning_rate": 5.94125538325363e-06, "loss": 0.842, "step": 9290 }, { "epoch": 4.209787041232442, "grad_norm": 0.13392930839657624, "learning_rate": 5.93462142678694e-06, "loss": 0.8255, "step": 9291 }, { "epoch": 4.210240144993204, "grad_norm": 0.12103042088501514, "learning_rate": 5.927990879320229e-06, "loss": 0.8576, "step": 9292 }, { "epoch": 4.210693248753965, "grad_norm": 0.1306569933196131, "learning_rate": 5.9213637415170165e-06, "loss": 0.8395, "step": 9293 }, { "epoch": 4.211146352514726, "grad_norm": 0.1461057631348178, "learning_rate": 5.914740014040514e-06, "loss": 0.8409, "step": 9294 }, { "epoch": 4.211599456275487, "grad_norm": 0.12350226804127483, "learning_rate": 5.908119697553561e-06, "loss": 0.8209, "step": 9295 }, { "epoch": 4.212052560036248, "grad_norm": 0.14649756414820467, "learning_rate": 5.901502792718665e-06, "loss": 0.847, "step": 9296 }, { "epoch": 4.2125056637970095, "grad_norm": 0.11813419304539689, "learning_rate": 5.894889300197992e-06, "loss": 0.8459, "step": 9297 }, { "epoch": 4.212958767557771, "grad_norm": 0.13827378479379707, "learning_rate": 5.888279220653381e-06, "loss": 0.8535, "step": 9298 }, { "epoch": 4.213411871318532, "grad_norm": 0.12797877862041976, "learning_rate": 5.881672554746316e-06, "loss": 0.8426, "step": 9299 }, { "epoch": 4.213864975079293, "grad_norm": 0.11993982290416873, "learning_rate": 5.875069303137926e-06, "loss": 0.8273, "step": 9300 }, { "epoch": 4.214318078840054, "grad_norm": 0.15410623049191563, "learning_rate": 5.8684694664890326e-06, "loss": 0.8408, "step": 9301 }, { "epoch": 4.214771182600815, "grad_norm": 0.12625412587556425, "learning_rate": 5.861873045460082e-06, "loss": 0.8518, "step": 9302 }, { "epoch": 4.215224286361577, "grad_norm": 0.11893531572492078, "learning_rate": 5.855280040711209e-06, "loss": 0.8325, "step": 9303 }, { "epoch": 4.215677390122338, "grad_norm": 0.13279443516215605, "learning_rate": 5.848690452902168e-06, "loss": 0.8395, "step": 9304 }, { "epoch": 4.2161304938830995, "grad_norm": 0.12128746337391767, "learning_rate": 5.84210428269242e-06, "loss": 0.8514, "step": 9305 }, { "epoch": 4.216583597643861, "grad_norm": 0.12143799899575228, "learning_rate": 5.835521530741033e-06, "loss": 0.8708, "step": 9306 }, { "epoch": 4.217036701404622, "grad_norm": 0.13094424423259346, "learning_rate": 5.828942197706782e-06, "loss": 0.8374, "step": 9307 }, { "epoch": 4.217489805165383, "grad_norm": 0.11120983573170817, "learning_rate": 5.822366284248064e-06, "loss": 0.8324, "step": 9308 }, { "epoch": 4.217942908926144, "grad_norm": 0.14212498706237353, "learning_rate": 5.815793791022937e-06, "loss": 0.8619, "step": 9309 }, { "epoch": 4.218396012686905, "grad_norm": 0.11197425101578828, "learning_rate": 5.8092247186891395e-06, "loss": 0.8618, "step": 9310 }, { "epoch": 4.218849116447666, "grad_norm": 0.14011431347307035, "learning_rate": 5.802659067904039e-06, "loss": 0.8346, "step": 9311 }, { "epoch": 4.219302220208427, "grad_norm": 0.2125457908241975, "learning_rate": 5.796096839324699e-06, "loss": 0.8753, "step": 9312 }, { "epoch": 4.2197553239691885, "grad_norm": 0.11018485565252124, "learning_rate": 5.789538033607786e-06, "loss": 0.8418, "step": 9313 }, { "epoch": 4.2202084277299505, "grad_norm": 0.12095151811719618, "learning_rate": 5.7829826514096855e-06, "loss": 0.8456, "step": 9314 }, { "epoch": 4.220661531490712, "grad_norm": 0.13345282246100285, "learning_rate": 5.77643069338639e-06, "loss": 0.849, "step": 9315 }, { "epoch": 4.221114635251473, "grad_norm": 0.12000642444911604, "learning_rate": 5.769882160193563e-06, "loss": 0.8456, "step": 9316 }, { "epoch": 4.221567739012234, "grad_norm": 0.11839027820870165, "learning_rate": 5.763337052486555e-06, "loss": 0.8249, "step": 9317 }, { "epoch": 4.222020842772995, "grad_norm": 0.13870923806595964, "learning_rate": 5.756795370920323e-06, "loss": 0.8425, "step": 9318 }, { "epoch": 4.222473946533756, "grad_norm": 0.11080386118919763, "learning_rate": 5.750257116149533e-06, "loss": 0.8381, "step": 9319 }, { "epoch": 4.222927050294517, "grad_norm": 0.13068484578061398, "learning_rate": 5.7437222888284636e-06, "loss": 0.8347, "step": 9320 }, { "epoch": 4.2233801540552784, "grad_norm": 0.11990859218058585, "learning_rate": 5.737190889611084e-06, "loss": 0.8404, "step": 9321 }, { "epoch": 4.22383325781604, "grad_norm": 0.11123428314866407, "learning_rate": 5.730662919150991e-06, "loss": 0.8526, "step": 9322 }, { "epoch": 4.224286361576801, "grad_norm": 0.12893667467480063, "learning_rate": 5.724138378101471e-06, "loss": 0.8427, "step": 9323 }, { "epoch": 4.224739465337563, "grad_norm": 0.12513587643837448, "learning_rate": 5.717617267115438e-06, "loss": 0.8384, "step": 9324 }, { "epoch": 4.225192569098324, "grad_norm": 0.11220021757013757, "learning_rate": 5.711099586845472e-06, "loss": 0.8349, "step": 9325 }, { "epoch": 4.225645672859085, "grad_norm": 0.14865814182476947, "learning_rate": 5.7045853379438285e-06, "loss": 0.8542, "step": 9326 }, { "epoch": 4.226098776619846, "grad_norm": 0.10226954322424482, "learning_rate": 5.698074521062378e-06, "loss": 0.8394, "step": 9327 }, { "epoch": 4.226551880380607, "grad_norm": 0.11786607893388927, "learning_rate": 5.691567136852696e-06, "loss": 0.8462, "step": 9328 }, { "epoch": 4.227004984141368, "grad_norm": 0.10758889388024266, "learning_rate": 5.685063185965977e-06, "loss": 0.8467, "step": 9329 }, { "epoch": 4.2274580879021295, "grad_norm": 0.10944925852606967, "learning_rate": 5.678562669053098e-06, "loss": 0.8349, "step": 9330 }, { "epoch": 4.227911191662891, "grad_norm": 0.09241988213219998, "learning_rate": 5.672065586764564e-06, "loss": 0.8355, "step": 9331 }, { "epoch": 4.228364295423652, "grad_norm": 0.10919384332062408, "learning_rate": 5.665571939750569e-06, "loss": 0.8586, "step": 9332 }, { "epoch": 4.228817399184413, "grad_norm": 0.10168140184912902, "learning_rate": 5.659081728660947e-06, "loss": 0.8473, "step": 9333 }, { "epoch": 4.229270502945174, "grad_norm": 0.11373577506808125, "learning_rate": 5.652594954145167e-06, "loss": 0.8502, "step": 9334 }, { "epoch": 4.229723606705936, "grad_norm": 0.10428562589856713, "learning_rate": 5.646111616852401e-06, "loss": 0.8282, "step": 9335 }, { "epoch": 4.230176710466697, "grad_norm": 0.1103216735634133, "learning_rate": 5.639631717431435e-06, "loss": 0.8475, "step": 9336 }, { "epoch": 4.230629814227458, "grad_norm": 0.10790128566133107, "learning_rate": 5.633155256530738e-06, "loss": 0.828, "step": 9337 }, { "epoch": 4.2310829179882194, "grad_norm": 0.10417064739428328, "learning_rate": 5.626682234798409e-06, "loss": 0.8327, "step": 9338 }, { "epoch": 4.231536021748981, "grad_norm": 0.09513023075148291, "learning_rate": 5.62021265288224e-06, "loss": 0.8563, "step": 9339 }, { "epoch": 4.231989125509742, "grad_norm": 0.10329901937864554, "learning_rate": 5.613746511429643e-06, "loss": 0.8383, "step": 9340 }, { "epoch": 4.232442229270503, "grad_norm": 0.10140412571202234, "learning_rate": 5.607283811087691e-06, "loss": 0.8363, "step": 9341 }, { "epoch": 4.232895333031264, "grad_norm": 0.10569192853343844, "learning_rate": 5.600824552503139e-06, "loss": 0.8406, "step": 9342 }, { "epoch": 4.233348436792025, "grad_norm": 0.10923181555573146, "learning_rate": 5.594368736322367e-06, "loss": 0.848, "step": 9343 }, { "epoch": 4.233801540552786, "grad_norm": 0.10105267206151329, "learning_rate": 5.587916363191435e-06, "loss": 0.8497, "step": 9344 }, { "epoch": 4.234254644313548, "grad_norm": 0.10862126364540738, "learning_rate": 5.581467433756032e-06, "loss": 0.8383, "step": 9345 }, { "epoch": 4.234707748074309, "grad_norm": 0.11704884548010053, "learning_rate": 5.575021948661534e-06, "loss": 0.848, "step": 9346 }, { "epoch": 4.2351608518350705, "grad_norm": 0.10978743480924008, "learning_rate": 5.568579908552934e-06, "loss": 0.8416, "step": 9347 }, { "epoch": 4.235613955595832, "grad_norm": 0.11167619092217068, "learning_rate": 5.562141314074927e-06, "loss": 0.8466, "step": 9348 }, { "epoch": 4.236067059356593, "grad_norm": 0.12412213186238354, "learning_rate": 5.555706165871822e-06, "loss": 0.8651, "step": 9349 }, { "epoch": 4.236520163117354, "grad_norm": 0.10940150213865794, "learning_rate": 5.549274464587595e-06, "loss": 0.8473, "step": 9350 }, { "epoch": 4.236973266878115, "grad_norm": 0.11418952827450396, "learning_rate": 5.542846210865893e-06, "loss": 0.8608, "step": 9351 }, { "epoch": 4.237426370638876, "grad_norm": 0.11550451884479214, "learning_rate": 5.5364214053500055e-06, "loss": 0.8519, "step": 9352 }, { "epoch": 4.237879474399637, "grad_norm": 0.10957540903708597, "learning_rate": 5.530000048682867e-06, "loss": 0.8419, "step": 9353 }, { "epoch": 4.238332578160398, "grad_norm": 0.1042945447023007, "learning_rate": 5.523582141507078e-06, "loss": 0.8439, "step": 9354 }, { "epoch": 4.23878568192116, "grad_norm": 0.1133391689198451, "learning_rate": 5.517167684464904e-06, "loss": 0.8328, "step": 9355 }, { "epoch": 4.239238785681922, "grad_norm": 0.10769796893317639, "learning_rate": 5.510756678198243e-06, "loss": 0.8465, "step": 9356 }, { "epoch": 4.239691889442683, "grad_norm": 0.44266744411857656, "learning_rate": 5.504349123348669e-06, "loss": 0.8832, "step": 9357 }, { "epoch": 4.240144993203444, "grad_norm": 0.11482029309324004, "learning_rate": 5.4979450205573915e-06, "loss": 0.8569, "step": 9358 }, { "epoch": 4.240598096964205, "grad_norm": 0.13490041613400247, "learning_rate": 5.4915443704652935e-06, "loss": 0.8588, "step": 9359 }, { "epoch": 4.241051200724966, "grad_norm": 0.1305390079695546, "learning_rate": 5.485147173712894e-06, "loss": 0.8262, "step": 9360 }, { "epoch": 4.241504304485727, "grad_norm": 0.11658732724768066, "learning_rate": 5.478753430940376e-06, "loss": 0.8415, "step": 9361 }, { "epoch": 4.241957408246488, "grad_norm": 0.14793237767894615, "learning_rate": 5.47236314278758e-06, "loss": 0.8578, "step": 9362 }, { "epoch": 4.2424105120072495, "grad_norm": 0.09879739334825251, "learning_rate": 5.465976309893992e-06, "loss": 0.8384, "step": 9363 }, { "epoch": 4.242863615768011, "grad_norm": 0.15488701161410226, "learning_rate": 5.459592932898768e-06, "loss": 0.8309, "step": 9364 }, { "epoch": 4.243316719528772, "grad_norm": 0.1024258929471322, "learning_rate": 5.453213012440683e-06, "loss": 0.8359, "step": 9365 }, { "epoch": 4.243769823289533, "grad_norm": 0.13111222455907223, "learning_rate": 5.446836549158221e-06, "loss": 0.8439, "step": 9366 }, { "epoch": 4.244222927050295, "grad_norm": 0.13310384185028942, "learning_rate": 5.440463543689465e-06, "loss": 0.832, "step": 9367 }, { "epoch": 4.244676030811056, "grad_norm": 0.11485174636494634, "learning_rate": 5.434093996672189e-06, "loss": 0.8323, "step": 9368 }, { "epoch": 4.245129134571817, "grad_norm": 0.12046906783958891, "learning_rate": 5.42772790874381e-06, "loss": 0.8465, "step": 9369 }, { "epoch": 4.245582238332578, "grad_norm": 0.13226064615891095, "learning_rate": 5.42136528054138e-06, "loss": 0.8415, "step": 9370 }, { "epoch": 4.246035342093339, "grad_norm": 0.12850455441382985, "learning_rate": 5.41500611270164e-06, "loss": 0.8689, "step": 9371 }, { "epoch": 4.246488445854101, "grad_norm": 0.1251399970364291, "learning_rate": 5.408650405860951e-06, "loss": 0.8424, "step": 9372 }, { "epoch": 4.246941549614862, "grad_norm": 0.11080370774445161, "learning_rate": 5.402298160655361e-06, "loss": 0.8351, "step": 9373 }, { "epoch": 4.247394653375623, "grad_norm": 0.1172605396644413, "learning_rate": 5.3959493777205355e-06, "loss": 0.8415, "step": 9374 }, { "epoch": 4.247847757136384, "grad_norm": 0.12238086164171631, "learning_rate": 5.389604057691827e-06, "loss": 0.8483, "step": 9375 }, { "epoch": 4.248300860897145, "grad_norm": 0.11316091510409193, "learning_rate": 5.38326220120422e-06, "loss": 0.8234, "step": 9376 }, { "epoch": 4.248753964657906, "grad_norm": 0.1306563719950506, "learning_rate": 5.376923808892352e-06, "loss": 0.8367, "step": 9377 }, { "epoch": 4.249207068418668, "grad_norm": 0.0955020915015359, "learning_rate": 5.37058888139053e-06, "loss": 0.8431, "step": 9378 }, { "epoch": 4.249660172179429, "grad_norm": 0.11586666104030675, "learning_rate": 5.3642574193326945e-06, "loss": 0.8333, "step": 9379 }, { "epoch": 4.2501132759401905, "grad_norm": 0.11472853581384967, "learning_rate": 5.357929423352466e-06, "loss": 0.8459, "step": 9380 }, { "epoch": 4.250566379700952, "grad_norm": 0.11672070132218292, "learning_rate": 5.35160489408308e-06, "loss": 0.8638, "step": 9381 }, { "epoch": 4.251019483461713, "grad_norm": 0.12204827885435718, "learning_rate": 5.345283832157466e-06, "loss": 0.8447, "step": 9382 }, { "epoch": 4.251472587222474, "grad_norm": 0.12912185924193148, "learning_rate": 5.338966238208167e-06, "loss": 0.8458, "step": 9383 }, { "epoch": 4.251925690983235, "grad_norm": 0.12569730323985454, "learning_rate": 5.332652112867425e-06, "loss": 0.8579, "step": 9384 }, { "epoch": 4.252378794743996, "grad_norm": 0.10897924765400618, "learning_rate": 5.326341456767092e-06, "loss": 0.8549, "step": 9385 }, { "epoch": 4.252831898504757, "grad_norm": 0.12163982094683538, "learning_rate": 5.320034270538683e-06, "loss": 0.8513, "step": 9386 }, { "epoch": 4.253285002265518, "grad_norm": 0.1396794130271068, "learning_rate": 5.31373055481339e-06, "loss": 0.8308, "step": 9387 }, { "epoch": 4.2537381060262796, "grad_norm": 0.11754728943518349, "learning_rate": 5.307430310222024e-06, "loss": 0.8354, "step": 9388 }, { "epoch": 4.254191209787042, "grad_norm": 0.1188380628178035, "learning_rate": 5.30113353739508e-06, "loss": 0.8518, "step": 9389 }, { "epoch": 4.254644313547803, "grad_norm": 0.11018465888033521, "learning_rate": 5.294840236962673e-06, "loss": 0.8373, "step": 9390 }, { "epoch": 4.255097417308564, "grad_norm": 0.12495732824504194, "learning_rate": 5.288550409554609e-06, "loss": 0.8505, "step": 9391 }, { "epoch": 4.255550521069325, "grad_norm": 0.12299068087296423, "learning_rate": 5.2822640558003055e-06, "loss": 0.8467, "step": 9392 }, { "epoch": 4.256003624830086, "grad_norm": 0.1264492272133038, "learning_rate": 5.2759811763288685e-06, "loss": 0.8339, "step": 9393 }, { "epoch": 4.256456728590847, "grad_norm": 0.10755670080431787, "learning_rate": 5.269701771769033e-06, "loss": 0.8569, "step": 9394 }, { "epoch": 4.256909832351608, "grad_norm": 0.10251757090447358, "learning_rate": 5.263425842749188e-06, "loss": 0.8289, "step": 9395 }, { "epoch": 4.2573629361123695, "grad_norm": 0.13012516656544412, "learning_rate": 5.2571533898973895e-06, "loss": 0.8441, "step": 9396 }, { "epoch": 4.257816039873131, "grad_norm": 0.12539682076933192, "learning_rate": 5.2508844138413215e-06, "loss": 0.8337, "step": 9397 }, { "epoch": 4.258269143633892, "grad_norm": 0.10661660115909276, "learning_rate": 5.244618915208355e-06, "loss": 0.8426, "step": 9398 }, { "epoch": 4.258722247394654, "grad_norm": 0.1518082210532112, "learning_rate": 5.238356894625476e-06, "loss": 0.868, "step": 9399 }, { "epoch": 4.259175351155415, "grad_norm": 0.12622725553034483, "learning_rate": 5.232098352719352e-06, "loss": 0.8625, "step": 9400 }, { "epoch": 4.259628454916176, "grad_norm": 0.11592134552291061, "learning_rate": 5.225843290116283e-06, "loss": 0.8278, "step": 9401 }, { "epoch": 4.260081558676937, "grad_norm": 0.1372744252467552, "learning_rate": 5.219591707442218e-06, "loss": 0.8335, "step": 9402 }, { "epoch": 4.260534662437698, "grad_norm": 0.11961693074569449, "learning_rate": 5.213343605322787e-06, "loss": 0.8619, "step": 9403 }, { "epoch": 4.260987766198459, "grad_norm": 0.1441955449684881, "learning_rate": 5.20709898438323e-06, "loss": 0.8461, "step": 9404 }, { "epoch": 4.261440869959221, "grad_norm": 0.11751795209422805, "learning_rate": 5.200857845248481e-06, "loss": 0.8636, "step": 9405 }, { "epoch": 4.261893973719982, "grad_norm": 0.1195941249349553, "learning_rate": 5.1946201885430914e-06, "loss": 0.8572, "step": 9406 }, { "epoch": 4.262347077480743, "grad_norm": 0.11713198070068914, "learning_rate": 5.188386014891285e-06, "loss": 0.8456, "step": 9407 }, { "epoch": 4.262800181241504, "grad_norm": 0.1000294094214751, "learning_rate": 5.182155324916922e-06, "loss": 0.8358, "step": 9408 }, { "epoch": 4.263253285002266, "grad_norm": 0.1268065125996606, "learning_rate": 5.175928119243545e-06, "loss": 0.8308, "step": 9409 }, { "epoch": 4.263706388763027, "grad_norm": 0.10206923701512176, "learning_rate": 5.169704398494286e-06, "loss": 0.8392, "step": 9410 }, { "epoch": 4.264159492523788, "grad_norm": 0.10515736878846393, "learning_rate": 5.163484163291994e-06, "loss": 0.8604, "step": 9411 }, { "epoch": 4.264612596284549, "grad_norm": 0.12689075234496366, "learning_rate": 5.1572674142591306e-06, "loss": 0.872, "step": 9412 }, { "epoch": 4.2650657000453105, "grad_norm": 0.10778204906160448, "learning_rate": 5.1510541520178294e-06, "loss": 0.8592, "step": 9413 }, { "epoch": 4.265518803806072, "grad_norm": 0.11490313740531445, "learning_rate": 5.144844377189864e-06, "loss": 0.8489, "step": 9414 }, { "epoch": 4.265971907566833, "grad_norm": 0.11460178578697108, "learning_rate": 5.138638090396648e-06, "loss": 0.8269, "step": 9415 }, { "epoch": 4.266425011327594, "grad_norm": 0.11129221710223917, "learning_rate": 5.132435292259277e-06, "loss": 0.8348, "step": 9416 }, { "epoch": 4.266878115088355, "grad_norm": 0.1096333086484845, "learning_rate": 5.126235983398462e-06, "loss": 0.8472, "step": 9417 }, { "epoch": 4.267331218849116, "grad_norm": 0.11007499318077495, "learning_rate": 5.120040164434596e-06, "loss": 0.858, "step": 9418 }, { "epoch": 4.267784322609877, "grad_norm": 0.11309928759152735, "learning_rate": 5.1138478359877e-06, "loss": 0.8528, "step": 9419 }, { "epoch": 4.268237426370639, "grad_norm": 0.12349165834240665, "learning_rate": 5.107658998677463e-06, "loss": 0.8463, "step": 9420 }, { "epoch": 4.2686905301314, "grad_norm": 0.1188156413454598, "learning_rate": 5.101473653123212e-06, "loss": 0.8322, "step": 9421 }, { "epoch": 4.269143633892162, "grad_norm": 0.11201973273163869, "learning_rate": 5.0952917999439155e-06, "loss": 0.8528, "step": 9422 }, { "epoch": 4.269596737652923, "grad_norm": 0.10829868187985073, "learning_rate": 5.089113439758229e-06, "loss": 0.8405, "step": 9423 }, { "epoch": 4.270049841413684, "grad_norm": 0.11124905040551235, "learning_rate": 5.082938573184413e-06, "loss": 0.8667, "step": 9424 }, { "epoch": 4.270502945174445, "grad_norm": 0.11989148903461329, "learning_rate": 5.0767672008404175e-06, "loss": 0.8485, "step": 9425 }, { "epoch": 4.270956048935206, "grad_norm": 0.10272646389523789, "learning_rate": 5.070599323343812e-06, "loss": 0.8378, "step": 9426 }, { "epoch": 4.271409152695967, "grad_norm": 0.10421335602146505, "learning_rate": 5.064434941311844e-06, "loss": 0.8585, "step": 9427 }, { "epoch": 4.271862256456728, "grad_norm": 0.19066312708704333, "learning_rate": 5.058274055361385e-06, "loss": 0.866, "step": 9428 }, { "epoch": 4.2723153602174895, "grad_norm": 0.099109832138865, "learning_rate": 5.052116666108981e-06, "loss": 0.8382, "step": 9429 }, { "epoch": 4.272768463978251, "grad_norm": 0.10361352381567909, "learning_rate": 5.045962774170811e-06, "loss": 0.8551, "step": 9430 }, { "epoch": 4.273221567739013, "grad_norm": 0.12859188393081272, "learning_rate": 5.039812380162694e-06, "loss": 0.8474, "step": 9431 }, { "epoch": 4.273674671499774, "grad_norm": 0.10039045010929668, "learning_rate": 5.033665484700141e-06, "loss": 0.8407, "step": 9432 }, { "epoch": 4.274127775260535, "grad_norm": 0.14602123914415252, "learning_rate": 5.02752208839826e-06, "loss": 0.8727, "step": 9433 }, { "epoch": 4.274580879021296, "grad_norm": 0.1194309127894506, "learning_rate": 5.021382191871857e-06, "loss": 0.858, "step": 9434 }, { "epoch": 4.275033982782057, "grad_norm": 0.13708536297963303, "learning_rate": 5.015245795735344e-06, "loss": 0.8516, "step": 9435 }, { "epoch": 4.275487086542818, "grad_norm": 0.13941134995980717, "learning_rate": 5.009112900602824e-06, "loss": 0.8583, "step": 9436 }, { "epoch": 4.275940190303579, "grad_norm": 0.12646804093474925, "learning_rate": 5.002983507088024e-06, "loss": 0.8606, "step": 9437 }, { "epoch": 4.2763932940643405, "grad_norm": 0.13065883586002178, "learning_rate": 4.996857615804312e-06, "loss": 0.8403, "step": 9438 }, { "epoch": 4.276846397825102, "grad_norm": 0.14985548803960794, "learning_rate": 4.990735227364738e-06, "loss": 0.8435, "step": 9439 }, { "epoch": 4.277299501585863, "grad_norm": 0.12129323572420282, "learning_rate": 4.984616342381969e-06, "loss": 0.855, "step": 9440 }, { "epoch": 4.277752605346624, "grad_norm": 0.130212488021489, "learning_rate": 4.978500961468355e-06, "loss": 0.8745, "step": 9441 }, { "epoch": 4.278205709107386, "grad_norm": 0.10789646555785891, "learning_rate": 4.972389085235851e-06, "loss": 0.8335, "step": 9442 }, { "epoch": 4.278658812868147, "grad_norm": 0.1345728037224661, "learning_rate": 4.96628071429611e-06, "loss": 0.8499, "step": 9443 }, { "epoch": 4.279111916628908, "grad_norm": 0.12455261438793845, "learning_rate": 4.9601758492603935e-06, "loss": 0.8552, "step": 9444 }, { "epoch": 4.279565020389669, "grad_norm": 0.11227194256577616, "learning_rate": 4.9540744907396395e-06, "loss": 0.8318, "step": 9445 }, { "epoch": 4.2800181241504305, "grad_norm": 0.12187516644710573, "learning_rate": 4.947976639344428e-06, "loss": 0.854, "step": 9446 }, { "epoch": 4.280471227911192, "grad_norm": 0.11940208685542751, "learning_rate": 4.941882295684966e-06, "loss": 0.8406, "step": 9447 }, { "epoch": 4.280924331671953, "grad_norm": 0.1063377269915506, "learning_rate": 4.935791460371149e-06, "loss": 0.8362, "step": 9448 }, { "epoch": 4.281377435432714, "grad_norm": 0.13515208651492197, "learning_rate": 4.929704134012485e-06, "loss": 0.8458, "step": 9449 }, { "epoch": 4.281830539193475, "grad_norm": 0.10245200175318277, "learning_rate": 4.923620317218158e-06, "loss": 0.8548, "step": 9450 }, { "epoch": 4.282283642954236, "grad_norm": 0.11002855877540017, "learning_rate": 4.9175400105969795e-06, "loss": 0.8409, "step": 9451 }, { "epoch": 4.282736746714997, "grad_norm": 0.11738477818484341, "learning_rate": 4.911463214757435e-06, "loss": 0.862, "step": 9452 }, { "epoch": 4.283189850475759, "grad_norm": 0.09656315901757101, "learning_rate": 4.905389930307625e-06, "loss": 0.8436, "step": 9453 }, { "epoch": 4.28364295423652, "grad_norm": 0.10700091250724424, "learning_rate": 4.899320157855334e-06, "loss": 0.8486, "step": 9454 }, { "epoch": 4.2840960579972815, "grad_norm": 0.11454635220864262, "learning_rate": 4.89325389800797e-06, "loss": 0.8571, "step": 9455 }, { "epoch": 4.284549161758043, "grad_norm": 0.11071077048109614, "learning_rate": 4.887191151372585e-06, "loss": 0.8435, "step": 9456 }, { "epoch": 4.285002265518804, "grad_norm": 0.11783476451873813, "learning_rate": 4.881131918555917e-06, "loss": 0.8542, "step": 9457 }, { "epoch": 4.285455369279565, "grad_norm": 0.12579093267003702, "learning_rate": 4.875076200164302e-06, "loss": 0.847, "step": 9458 }, { "epoch": 4.285908473040326, "grad_norm": 0.11546830619622951, "learning_rate": 4.8690239968037704e-06, "loss": 0.8337, "step": 9459 }, { "epoch": 4.286361576801087, "grad_norm": 0.11116081976369416, "learning_rate": 4.862975309079967e-06, "loss": 0.8495, "step": 9460 }, { "epoch": 4.286814680561848, "grad_norm": 0.12195380416788877, "learning_rate": 4.856930137598204e-06, "loss": 0.8524, "step": 9461 }, { "epoch": 4.2872677843226095, "grad_norm": 0.10490697544749643, "learning_rate": 4.850888482963436e-06, "loss": 0.8444, "step": 9462 }, { "epoch": 4.2877208880833715, "grad_norm": 0.11706519165483385, "learning_rate": 4.844850345780257e-06, "loss": 0.8476, "step": 9463 }, { "epoch": 4.288173991844133, "grad_norm": 0.25331828343276463, "learning_rate": 4.838815726652923e-06, "loss": 0.8907, "step": 9464 }, { "epoch": 4.288627095604894, "grad_norm": 0.11019642609272745, "learning_rate": 4.832784626185332e-06, "loss": 0.8347, "step": 9465 }, { "epoch": 4.289080199365655, "grad_norm": 0.10986844126716817, "learning_rate": 4.8267570449810434e-06, "loss": 0.8765, "step": 9466 }, { "epoch": 4.289533303126416, "grad_norm": 0.10255653950630841, "learning_rate": 4.820732983643219e-06, "loss": 0.844, "step": 9467 }, { "epoch": 4.289986406887177, "grad_norm": 0.11774540450403082, "learning_rate": 4.814712442774725e-06, "loss": 0.8414, "step": 9468 }, { "epoch": 4.290439510647938, "grad_norm": 0.11417747264395643, "learning_rate": 4.808695422978038e-06, "loss": 0.8557, "step": 9469 }, { "epoch": 4.290892614408699, "grad_norm": 0.09922208075473621, "learning_rate": 4.8026819248553125e-06, "loss": 0.8375, "step": 9470 }, { "epoch": 4.2913457181694605, "grad_norm": 0.11378841215673119, "learning_rate": 4.796671949008307e-06, "loss": 0.845, "step": 9471 }, { "epoch": 4.291798821930222, "grad_norm": 0.10696003672477516, "learning_rate": 4.790665496038482e-06, "loss": 0.8497, "step": 9472 }, { "epoch": 4.292251925690984, "grad_norm": 0.10384867154544673, "learning_rate": 4.7846625665469004e-06, "loss": 0.8386, "step": 9473 }, { "epoch": 4.292705029451745, "grad_norm": 0.12907002351002705, "learning_rate": 4.77866316113428e-06, "loss": 0.8369, "step": 9474 }, { "epoch": 4.293158133212506, "grad_norm": 0.10236037189443348, "learning_rate": 4.7726672804010175e-06, "loss": 0.8653, "step": 9475 }, { "epoch": 4.293611236973267, "grad_norm": 0.10345894893351153, "learning_rate": 4.766674924947116e-06, "loss": 0.8318, "step": 9476 }, { "epoch": 4.294064340734028, "grad_norm": 0.11315956849784456, "learning_rate": 4.7606860953722575e-06, "loss": 0.8539, "step": 9477 }, { "epoch": 4.294517444494789, "grad_norm": 0.10005048236788809, "learning_rate": 4.754700792275744e-06, "loss": 0.8295, "step": 9478 }, { "epoch": 4.2949705482555505, "grad_norm": 0.1139744005701545, "learning_rate": 4.748719016256558e-06, "loss": 0.827, "step": 9479 }, { "epoch": 4.295423652016312, "grad_norm": 0.16274209118106464, "learning_rate": 4.742740767913283e-06, "loss": 0.8612, "step": 9480 }, { "epoch": 4.295876755777073, "grad_norm": 0.09760888274614345, "learning_rate": 4.736766047844206e-06, "loss": 0.8429, "step": 9481 }, { "epoch": 4.296329859537834, "grad_norm": 0.11127728473160614, "learning_rate": 4.730794856647207e-06, "loss": 0.8439, "step": 9482 }, { "epoch": 4.296782963298595, "grad_norm": 0.10101227019047349, "learning_rate": 4.724827194919841e-06, "loss": 0.8537, "step": 9483 }, { "epoch": 4.297236067059357, "grad_norm": 0.10198987684504801, "learning_rate": 4.718863063259318e-06, "loss": 0.8305, "step": 9484 }, { "epoch": 4.297689170820118, "grad_norm": 0.11785250277364966, "learning_rate": 4.712902462262468e-06, "loss": 0.8284, "step": 9485 }, { "epoch": 4.298142274580879, "grad_norm": 0.11307947818038419, "learning_rate": 4.70694539252579e-06, "loss": 0.8426, "step": 9486 }, { "epoch": 4.29859537834164, "grad_norm": 0.1128264350798306, "learning_rate": 4.7009918546454135e-06, "loss": 0.8605, "step": 9487 }, { "epoch": 4.2990484821024015, "grad_norm": 0.11125676717130333, "learning_rate": 4.6950418492171365e-06, "loss": 0.852, "step": 9488 }, { "epoch": 4.299501585863163, "grad_norm": 0.1164381773545565, "learning_rate": 4.689095376836376e-06, "loss": 0.8426, "step": 9489 }, { "epoch": 4.299954689623924, "grad_norm": 0.10221850442403699, "learning_rate": 4.683152438098218e-06, "loss": 0.8507, "step": 9490 }, { "epoch": 4.300407793384685, "grad_norm": 0.1269520068676985, "learning_rate": 4.677213033597383e-06, "loss": 0.8401, "step": 9491 }, { "epoch": 4.300860897145446, "grad_norm": 0.11669247691492719, "learning_rate": 4.671277163928234e-06, "loss": 0.8634, "step": 9492 }, { "epoch": 4.301314000906207, "grad_norm": 0.10531721417198733, "learning_rate": 4.6653448296848015e-06, "loss": 0.8468, "step": 9493 }, { "epoch": 4.301767104666968, "grad_norm": 0.1119244269467983, "learning_rate": 4.659416031460726e-06, "loss": 0.861, "step": 9494 }, { "epoch": 4.30222020842773, "grad_norm": 0.11637678032947509, "learning_rate": 4.653490769849342e-06, "loss": 0.8514, "step": 9495 }, { "epoch": 4.3026733121884915, "grad_norm": 0.10992387043586838, "learning_rate": 4.647569045443581e-06, "loss": 0.8231, "step": 9496 }, { "epoch": 4.303126415949253, "grad_norm": 0.11704240370782461, "learning_rate": 4.641650858836064e-06, "loss": 0.856, "step": 9497 }, { "epoch": 4.303579519710014, "grad_norm": 0.11401638534028809, "learning_rate": 4.635736210619022e-06, "loss": 0.8619, "step": 9498 }, { "epoch": 4.304032623470775, "grad_norm": 0.10477555214128331, "learning_rate": 4.629825101384348e-06, "loss": 0.8558, "step": 9499 }, { "epoch": 4.304485727231536, "grad_norm": 0.12890928029251, "learning_rate": 4.62391753172359e-06, "loss": 0.8577, "step": 9500 }, { "epoch": 4.304938830992297, "grad_norm": 0.11166894446146036, "learning_rate": 4.61801350222792e-06, "loss": 0.8419, "step": 9501 }, { "epoch": 4.305391934753058, "grad_norm": 0.09986483450383785, "learning_rate": 4.61211301348818e-06, "loss": 0.8216, "step": 9502 }, { "epoch": 4.305845038513819, "grad_norm": 0.11720733971843368, "learning_rate": 4.606216066094833e-06, "loss": 0.8348, "step": 9503 }, { "epoch": 4.3062981422745805, "grad_norm": 0.12652955031456334, "learning_rate": 4.600322660638013e-06, "loss": 0.8317, "step": 9504 }, { "epoch": 4.306751246035342, "grad_norm": 0.11457552077990382, "learning_rate": 4.594432797707469e-06, "loss": 0.8455, "step": 9505 }, { "epoch": 4.307204349796104, "grad_norm": 0.11523609524340345, "learning_rate": 4.588546477892632e-06, "loss": 0.8241, "step": 9506 }, { "epoch": 4.307657453556865, "grad_norm": 0.13150380432881228, "learning_rate": 4.582663701782557e-06, "loss": 0.8562, "step": 9507 }, { "epoch": 4.308110557317626, "grad_norm": 0.11900593286572031, "learning_rate": 4.576784469965927e-06, "loss": 0.8488, "step": 9508 }, { "epoch": 4.308563661078387, "grad_norm": 0.11030779070885288, "learning_rate": 4.570908783031116e-06, "loss": 0.8387, "step": 9509 }, { "epoch": 4.309016764839148, "grad_norm": 0.13640567950253163, "learning_rate": 4.565036641566094e-06, "loss": 0.8448, "step": 9510 }, { "epoch": 4.309469868599909, "grad_norm": 0.1022518010120055, "learning_rate": 4.5591680461585195e-06, "loss": 0.8458, "step": 9511 }, { "epoch": 4.3099229723606705, "grad_norm": 0.1106447028954754, "learning_rate": 4.553302997395661e-06, "loss": 0.8568, "step": 9512 }, { "epoch": 4.310376076121432, "grad_norm": 0.1355871896124539, "learning_rate": 4.547441495864462e-06, "loss": 0.8506, "step": 9513 }, { "epoch": 4.310829179882193, "grad_norm": 0.11251582481009371, "learning_rate": 4.541583542151484e-06, "loss": 0.8483, "step": 9514 }, { "epoch": 4.311282283642954, "grad_norm": 0.1335231888653641, "learning_rate": 4.535729136842944e-06, "loss": 0.8442, "step": 9515 }, { "epoch": 4.311735387403715, "grad_norm": 0.12872039977838753, "learning_rate": 4.529878280524722e-06, "loss": 0.8446, "step": 9516 }, { "epoch": 4.312188491164477, "grad_norm": 0.11500390247935997, "learning_rate": 4.524030973782308e-06, "loss": 0.8538, "step": 9517 }, { "epoch": 4.312641594925238, "grad_norm": 0.14561001739213253, "learning_rate": 4.518187217200867e-06, "loss": 0.8492, "step": 9518 }, { "epoch": 4.313094698685999, "grad_norm": 0.10017379438638217, "learning_rate": 4.51234701136519e-06, "loss": 0.8398, "step": 9519 }, { "epoch": 4.31354780244676, "grad_norm": 0.09565141880470776, "learning_rate": 4.50651035685973e-06, "loss": 0.8405, "step": 9520 }, { "epoch": 4.3140009062075215, "grad_norm": 0.10279629395092993, "learning_rate": 4.5006772542685575e-06, "loss": 0.8555, "step": 9521 }, { "epoch": 4.314454009968283, "grad_norm": 0.10049574461399298, "learning_rate": 4.494847704175422e-06, "loss": 0.8699, "step": 9522 }, { "epoch": 4.314907113729044, "grad_norm": 0.11141888755195938, "learning_rate": 4.489021707163694e-06, "loss": 0.8445, "step": 9523 }, { "epoch": 4.315360217489805, "grad_norm": 0.09969558899079205, "learning_rate": 4.483199263816392e-06, "loss": 0.8453, "step": 9524 }, { "epoch": 4.315813321250566, "grad_norm": 0.1004899383147178, "learning_rate": 4.477380374716171e-06, "loss": 0.8573, "step": 9525 }, { "epoch": 4.316266425011327, "grad_norm": 0.10738427447567722, "learning_rate": 4.471565040445365e-06, "loss": 0.8685, "step": 9526 }, { "epoch": 4.316719528772089, "grad_norm": 0.0900495341010155, "learning_rate": 4.46575326158591e-06, "loss": 0.8347, "step": 9527 }, { "epoch": 4.31717263253285, "grad_norm": 0.10552168205563152, "learning_rate": 4.4599450387194e-06, "loss": 0.8475, "step": 9528 }, { "epoch": 4.3176257362936115, "grad_norm": 0.10940865501082285, "learning_rate": 4.454140372427089e-06, "loss": 0.8523, "step": 9529 }, { "epoch": 4.318078840054373, "grad_norm": 0.10418988607435138, "learning_rate": 4.448339263289856e-06, "loss": 0.8289, "step": 9530 }, { "epoch": 4.318531943815134, "grad_norm": 0.11416638603233244, "learning_rate": 4.44254171188824e-06, "loss": 0.8596, "step": 9531 }, { "epoch": 4.318985047575895, "grad_norm": 0.09972435487476985, "learning_rate": 4.4367477188024035e-06, "loss": 0.8743, "step": 9532 }, { "epoch": 4.319438151336656, "grad_norm": 0.10817203238754125, "learning_rate": 4.430957284612176e-06, "loss": 0.8297, "step": 9533 }, { "epoch": 4.319891255097417, "grad_norm": 0.09744599389877778, "learning_rate": 4.425170409897015e-06, "loss": 0.8304, "step": 9534 }, { "epoch": 4.320344358858178, "grad_norm": 0.10162031772587546, "learning_rate": 4.419387095236016e-06, "loss": 0.8428, "step": 9535 }, { "epoch": 4.320797462618939, "grad_norm": 0.10876501216855641, "learning_rate": 4.413607341207944e-06, "loss": 0.8553, "step": 9536 }, { "epoch": 4.321250566379701, "grad_norm": 0.09842492717154364, "learning_rate": 4.40783114839118e-06, "loss": 0.8363, "step": 9537 }, { "epoch": 4.3217036701404625, "grad_norm": 0.11048032374041722, "learning_rate": 4.402058517363772e-06, "loss": 0.8408, "step": 9538 }, { "epoch": 4.322156773901224, "grad_norm": 0.10646169154879728, "learning_rate": 4.3962894487033884e-06, "loss": 0.846, "step": 9539 }, { "epoch": 4.322609877661985, "grad_norm": 0.10734303063319181, "learning_rate": 4.390523942987371e-06, "loss": 0.8341, "step": 9540 }, { "epoch": 4.323062981422746, "grad_norm": 0.11487695540284411, "learning_rate": 4.384762000792662e-06, "loss": 0.8373, "step": 9541 }, { "epoch": 4.323516085183507, "grad_norm": 0.11220524528157118, "learning_rate": 4.379003622695894e-06, "loss": 0.8431, "step": 9542 }, { "epoch": 4.323969188944268, "grad_norm": 0.1309056137196574, "learning_rate": 4.373248809273314e-06, "loss": 0.8333, "step": 9543 }, { "epoch": 4.324422292705029, "grad_norm": 0.11106359943219302, "learning_rate": 4.3674975611008105e-06, "loss": 0.8377, "step": 9544 }, { "epoch": 4.3248753964657904, "grad_norm": 0.13602564445094883, "learning_rate": 4.361749878753938e-06, "loss": 0.8439, "step": 9545 }, { "epoch": 4.325328500226552, "grad_norm": 0.1086240474569009, "learning_rate": 4.356005762807867e-06, "loss": 0.832, "step": 9546 }, { "epoch": 4.325781603987313, "grad_norm": 0.11057389917536596, "learning_rate": 4.350265213837435e-06, "loss": 0.8579, "step": 9547 }, { "epoch": 4.326234707748075, "grad_norm": 0.12569419966647272, "learning_rate": 4.344528232417102e-06, "loss": 0.8521, "step": 9548 }, { "epoch": 4.326687811508836, "grad_norm": 0.12130589094043234, "learning_rate": 4.338794819120998e-06, "loss": 0.8291, "step": 9549 }, { "epoch": 4.327140915269597, "grad_norm": 0.10725354917522162, "learning_rate": 4.333064974522852e-06, "loss": 0.8409, "step": 9550 }, { "epoch": 4.327594019030358, "grad_norm": 0.11518671675575264, "learning_rate": 4.327338699196091e-06, "loss": 0.8574, "step": 9551 }, { "epoch": 4.328047122791119, "grad_norm": 0.11311583461849985, "learning_rate": 4.321615993713737e-06, "loss": 0.8602, "step": 9552 }, { "epoch": 4.32850022655188, "grad_norm": 0.09751795240713786, "learning_rate": 4.315896858648478e-06, "loss": 0.8305, "step": 9553 }, { "epoch": 4.3289533303126415, "grad_norm": 0.11242809749285555, "learning_rate": 4.3101812945726484e-06, "loss": 0.8482, "step": 9554 }, { "epoch": 4.329406434073403, "grad_norm": 0.09988904411920507, "learning_rate": 4.304469302058203e-06, "loss": 0.8446, "step": 9555 }, { "epoch": 4.329859537834164, "grad_norm": 0.11434241624034762, "learning_rate": 4.2987608816767735e-06, "loss": 0.8178, "step": 9556 }, { "epoch": 4.330312641594925, "grad_norm": 0.11086261125150987, "learning_rate": 4.293056033999596e-06, "loss": 0.8368, "step": 9557 }, { "epoch": 4.330765745355686, "grad_norm": 0.10171435633966032, "learning_rate": 4.287354759597584e-06, "loss": 0.854, "step": 9558 }, { "epoch": 4.331218849116448, "grad_norm": 0.11032702611408919, "learning_rate": 4.281657059041267e-06, "loss": 0.825, "step": 9559 }, { "epoch": 4.331671952877209, "grad_norm": 0.10908676490619423, "learning_rate": 4.275962932900824e-06, "loss": 0.8712, "step": 9560 }, { "epoch": 4.33212505663797, "grad_norm": 0.11113391943922847, "learning_rate": 4.270272381746088e-06, "loss": 0.8218, "step": 9561 }, { "epoch": 4.3325781603987314, "grad_norm": 0.10745809032609488, "learning_rate": 4.264585406146515e-06, "loss": 0.8488, "step": 9562 }, { "epoch": 4.333031264159493, "grad_norm": 0.10073854416851782, "learning_rate": 4.258902006671229e-06, "loss": 0.8304, "step": 9563 }, { "epoch": 4.333484367920254, "grad_norm": 0.12471689761190383, "learning_rate": 4.2532221838889635e-06, "loss": 0.8329, "step": 9564 }, { "epoch": 4.333937471681015, "grad_norm": 0.10180188992671421, "learning_rate": 4.247545938368123e-06, "loss": 0.8379, "step": 9565 }, { "epoch": 4.334390575441776, "grad_norm": 0.10799708708940926, "learning_rate": 4.241873270676738e-06, "loss": 0.8275, "step": 9566 }, { "epoch": 4.334843679202537, "grad_norm": 0.10604766397743991, "learning_rate": 4.236204181382486e-06, "loss": 0.8527, "step": 9567 }, { "epoch": 4.335296782963298, "grad_norm": 0.10055868399769277, "learning_rate": 4.230538671052693e-06, "loss": 0.8459, "step": 9568 }, { "epoch": 4.335749886724059, "grad_norm": 0.10189530350826434, "learning_rate": 4.2248767402543e-06, "loss": 0.8488, "step": 9569 }, { "epoch": 4.336202990484821, "grad_norm": 0.10552965345475523, "learning_rate": 4.21921838955393e-06, "loss": 0.8544, "step": 9570 }, { "epoch": 4.3366560942455825, "grad_norm": 0.11197724124031609, "learning_rate": 4.213563619517809e-06, "loss": 0.8479, "step": 9571 }, { "epoch": 4.337109198006344, "grad_norm": 0.09722204909383285, "learning_rate": 4.20791243071184e-06, "loss": 0.8557, "step": 9572 }, { "epoch": 4.337562301767105, "grad_norm": 0.09943453587000733, "learning_rate": 4.202264823701536e-06, "loss": 0.843, "step": 9573 }, { "epoch": 4.338015405527866, "grad_norm": 0.11174257997117243, "learning_rate": 4.196620799052084e-06, "loss": 0.8378, "step": 9574 }, { "epoch": 4.338468509288627, "grad_norm": 0.09498930542073845, "learning_rate": 4.190980357328278e-06, "loss": 0.8364, "step": 9575 }, { "epoch": 4.338921613049388, "grad_norm": 0.0972914776273821, "learning_rate": 4.185343499094568e-06, "loss": 0.8379, "step": 9576 }, { "epoch": 4.339374716810149, "grad_norm": 0.1071873265339491, "learning_rate": 4.179710224915061e-06, "loss": 0.8531, "step": 9577 }, { "epoch": 4.33982782057091, "grad_norm": 0.10112341993400072, "learning_rate": 4.174080535353478e-06, "loss": 0.8648, "step": 9578 }, { "epoch": 4.340280924331672, "grad_norm": 0.10846183908141395, "learning_rate": 4.1684544309732104e-06, "loss": 0.8379, "step": 9579 }, { "epoch": 4.340734028092434, "grad_norm": 0.10244813916118767, "learning_rate": 4.162831912337262e-06, "loss": 0.8302, "step": 9580 }, { "epoch": 4.341187131853195, "grad_norm": 0.09929014919258207, "learning_rate": 4.157212980008298e-06, "loss": 0.8394, "step": 9581 }, { "epoch": 4.341640235613956, "grad_norm": 0.11365585369454662, "learning_rate": 4.151597634548608e-06, "loss": 0.8387, "step": 9582 }, { "epoch": 4.342093339374717, "grad_norm": 0.12252493600357711, "learning_rate": 4.145985876520144e-06, "loss": 0.8419, "step": 9583 }, { "epoch": 4.342546443135478, "grad_norm": 0.1012373422849142, "learning_rate": 4.140377706484478e-06, "loss": 0.8619, "step": 9584 }, { "epoch": 4.342999546896239, "grad_norm": 0.13239461857430673, "learning_rate": 4.134773125002842e-06, "loss": 0.8186, "step": 9585 }, { "epoch": 4.343452650657, "grad_norm": 0.10927305913867104, "learning_rate": 4.129172132636088e-06, "loss": 0.8589, "step": 9586 }, { "epoch": 4.3439057544177615, "grad_norm": 0.10458128437148474, "learning_rate": 4.123574729944735e-06, "loss": 0.8479, "step": 9587 }, { "epoch": 4.344358858178523, "grad_norm": 0.11057265151571934, "learning_rate": 4.117980917488917e-06, "loss": 0.8384, "step": 9588 }, { "epoch": 4.344811961939284, "grad_norm": 0.09451751338359049, "learning_rate": 4.112390695828414e-06, "loss": 0.8435, "step": 9589 }, { "epoch": 4.345265065700046, "grad_norm": 0.11996908302065526, "learning_rate": 4.106804065522672e-06, "loss": 0.8548, "step": 9590 }, { "epoch": 4.345718169460807, "grad_norm": 0.10352116916427777, "learning_rate": 4.101221027130731e-06, "loss": 0.8313, "step": 9591 }, { "epoch": 4.346171273221568, "grad_norm": 0.09685827993148592, "learning_rate": 4.095641581211327e-06, "loss": 0.855, "step": 9592 }, { "epoch": 4.346624376982329, "grad_norm": 0.11814248080432226, "learning_rate": 4.090065728322783e-06, "loss": 0.8634, "step": 9593 }, { "epoch": 4.34707748074309, "grad_norm": 0.11616446166995302, "learning_rate": 4.084493469023105e-06, "loss": 0.8531, "step": 9594 }, { "epoch": 4.347530584503851, "grad_norm": 0.09562746212589283, "learning_rate": 4.078924803869919e-06, "loss": 0.8561, "step": 9595 }, { "epoch": 4.347983688264613, "grad_norm": 0.11283460529541157, "learning_rate": 4.073359733420481e-06, "loss": 0.8612, "step": 9596 }, { "epoch": 4.348436792025374, "grad_norm": 0.11072845101958866, "learning_rate": 4.067798258231714e-06, "loss": 0.8367, "step": 9597 }, { "epoch": 4.348889895786135, "grad_norm": 0.10315255483160674, "learning_rate": 4.0622403788601564e-06, "loss": 0.855, "step": 9598 }, { "epoch": 4.349342999546896, "grad_norm": 0.1007989405511028, "learning_rate": 4.056686095862015e-06, "loss": 0.8436, "step": 9599 }, { "epoch": 4.349796103307657, "grad_norm": 0.10625768749637109, "learning_rate": 4.051135409793099e-06, "loss": 0.8505, "step": 9600 }, { "epoch": 4.350249207068419, "grad_norm": 0.11336062947807046, "learning_rate": 4.045588321208893e-06, "loss": 0.8328, "step": 9601 }, { "epoch": 4.35070231082918, "grad_norm": 0.10559642508001049, "learning_rate": 4.040044830664495e-06, "loss": 0.8572, "step": 9602 }, { "epoch": 4.351155414589941, "grad_norm": 0.09355666751052386, "learning_rate": 4.034504938714671e-06, "loss": 0.862, "step": 9603 }, { "epoch": 4.3516085183507025, "grad_norm": 0.11099925869003963, "learning_rate": 4.0289686459138e-06, "loss": 0.8353, "step": 9604 }, { "epoch": 4.352061622111464, "grad_norm": 0.10884063207758402, "learning_rate": 4.023435952815904e-06, "loss": 0.8509, "step": 9605 }, { "epoch": 4.352514725872225, "grad_norm": 0.10100827128527895, "learning_rate": 4.017906859974665e-06, "loss": 0.8379, "step": 9606 }, { "epoch": 4.352967829632986, "grad_norm": 0.10212170160117451, "learning_rate": 4.0123813679433835e-06, "loss": 0.8261, "step": 9607 }, { "epoch": 4.353420933393747, "grad_norm": 0.10946656027246551, "learning_rate": 4.0068594772750206e-06, "loss": 0.8668, "step": 9608 }, { "epoch": 4.353874037154508, "grad_norm": 0.10430949413523, "learning_rate": 4.0013411885221476e-06, "loss": 0.8526, "step": 9609 }, { "epoch": 4.354327140915269, "grad_norm": 0.8833127294340051, "learning_rate": 3.995826502237008e-06, "loss": 0.8472, "step": 9610 }, { "epoch": 4.35478024467603, "grad_norm": 0.10756425506408417, "learning_rate": 3.990315418971453e-06, "loss": 0.8301, "step": 9611 }, { "epoch": 4.355233348436792, "grad_norm": 0.10911593326602421, "learning_rate": 3.9848079392770065e-06, "loss": 0.8499, "step": 9612 }, { "epoch": 4.355686452197554, "grad_norm": 0.1195514428242282, "learning_rate": 3.979304063704805e-06, "loss": 0.849, "step": 9613 }, { "epoch": 4.356139555958315, "grad_norm": 0.11106720214829642, "learning_rate": 3.973803792805626e-06, "loss": 0.8783, "step": 9614 }, { "epoch": 4.356592659719076, "grad_norm": 0.11242015096981227, "learning_rate": 3.9683071271299136e-06, "loss": 0.8598, "step": 9615 }, { "epoch": 4.357045763479837, "grad_norm": 0.11732359533199571, "learning_rate": 3.962814067227707e-06, "loss": 0.8294, "step": 9616 }, { "epoch": 4.357498867240598, "grad_norm": 0.11200188483234126, "learning_rate": 3.957324613648736e-06, "loss": 0.8614, "step": 9617 }, { "epoch": 4.357951971001359, "grad_norm": 0.10613804966788344, "learning_rate": 3.951838766942322e-06, "loss": 0.8466, "step": 9618 }, { "epoch": 4.35840507476212, "grad_norm": 0.1193534620510738, "learning_rate": 3.94635652765746e-06, "loss": 0.8414, "step": 9619 }, { "epoch": 4.3588581785228815, "grad_norm": 0.10706981490364205, "learning_rate": 3.9408778963427655e-06, "loss": 0.8534, "step": 9620 }, { "epoch": 4.359311282283643, "grad_norm": 0.10335346770405064, "learning_rate": 3.935402873546488e-06, "loss": 0.8417, "step": 9621 }, { "epoch": 4.359764386044404, "grad_norm": 0.09798223372676369, "learning_rate": 3.9299314598165365e-06, "loss": 0.8546, "step": 9622 }, { "epoch": 4.360217489805166, "grad_norm": 0.1079367320599906, "learning_rate": 3.9244636557004415e-06, "loss": 0.8361, "step": 9623 }, { "epoch": 4.360670593565927, "grad_norm": 0.10811985853728037, "learning_rate": 3.918999461745392e-06, "loss": 0.8441, "step": 9624 }, { "epoch": 4.361123697326688, "grad_norm": 0.1060622618663791, "learning_rate": 3.913538878498182e-06, "loss": 0.8562, "step": 9625 }, { "epoch": 4.361576801087449, "grad_norm": 0.10273088784010095, "learning_rate": 3.908081906505281e-06, "loss": 0.8668, "step": 9626 }, { "epoch": 4.36202990484821, "grad_norm": 0.1129137526013996, "learning_rate": 3.9026285463127725e-06, "loss": 0.8387, "step": 9627 }, { "epoch": 4.362483008608971, "grad_norm": 0.11670259488407762, "learning_rate": 3.8971787984663925e-06, "loss": 0.8566, "step": 9628 }, { "epoch": 4.3629361123697326, "grad_norm": 0.09629353893077455, "learning_rate": 3.8917326635115045e-06, "loss": 0.826, "step": 9629 }, { "epoch": 4.363389216130494, "grad_norm": 0.10906734624377148, "learning_rate": 3.88629014199311e-06, "loss": 0.8619, "step": 9630 }, { "epoch": 4.363842319891255, "grad_norm": 0.1263150562374139, "learning_rate": 3.880851234455869e-06, "loss": 0.8589, "step": 9631 }, { "epoch": 4.364295423652016, "grad_norm": 0.10693635027438622, "learning_rate": 3.875415941444054e-06, "loss": 0.8507, "step": 9632 }, { "epoch": 4.364748527412777, "grad_norm": 0.10126865500763206, "learning_rate": 3.869984263501594e-06, "loss": 0.877, "step": 9633 }, { "epoch": 4.365201631173539, "grad_norm": 0.10845287475385158, "learning_rate": 3.864556201172041e-06, "loss": 0.8483, "step": 9634 }, { "epoch": 4.3656547349343, "grad_norm": 0.10857330550853403, "learning_rate": 3.8591317549986085e-06, "loss": 0.8329, "step": 9635 }, { "epoch": 4.366107838695061, "grad_norm": 0.0991433153232861, "learning_rate": 3.853710925524121e-06, "loss": 0.8615, "step": 9636 }, { "epoch": 4.3665609424558225, "grad_norm": 0.10496166681035912, "learning_rate": 3.848293713291056e-06, "loss": 0.8201, "step": 9637 }, { "epoch": 4.367014046216584, "grad_norm": 0.10698867245810988, "learning_rate": 3.842880118841516e-06, "loss": 0.8386, "step": 9638 }, { "epoch": 4.367467149977345, "grad_norm": 0.10244228574249013, "learning_rate": 3.837470142717275e-06, "loss": 0.8428, "step": 9639 }, { "epoch": 4.367920253738106, "grad_norm": 0.09978461891763826, "learning_rate": 3.8320637854597055e-06, "loss": 0.8567, "step": 9640 }, { "epoch": 4.368373357498867, "grad_norm": 0.10292084384032686, "learning_rate": 3.82666104760983e-06, "loss": 0.849, "step": 9641 }, { "epoch": 4.368826461259628, "grad_norm": 0.10720748353692898, "learning_rate": 3.821261929708327e-06, "loss": 0.8716, "step": 9642 }, { "epoch": 4.369279565020389, "grad_norm": 0.09848084877696146, "learning_rate": 3.8158664322954785e-06, "loss": 0.8403, "step": 9643 }, { "epoch": 4.369732668781151, "grad_norm": 0.10699879277689822, "learning_rate": 3.81047455591125e-06, "loss": 0.8563, "step": 9644 }, { "epoch": 4.370185772541912, "grad_norm": 0.10104485992835646, "learning_rate": 3.8050863010952e-06, "loss": 0.8387, "step": 9645 }, { "epoch": 4.370638876302674, "grad_norm": 0.09296708027244795, "learning_rate": 3.7997016683865506e-06, "loss": 0.8505, "step": 9646 }, { "epoch": 4.371091980063435, "grad_norm": 0.10779488865370711, "learning_rate": 3.794320658324151e-06, "loss": 0.8263, "step": 9647 }, { "epoch": 4.371545083824196, "grad_norm": 0.10789984363173617, "learning_rate": 3.7889432714464992e-06, "loss": 0.8297, "step": 9648 }, { "epoch": 4.371998187584957, "grad_norm": 0.11368984369484213, "learning_rate": 3.7835695082917156e-06, "loss": 0.8709, "step": 9649 }, { "epoch": 4.372451291345718, "grad_norm": 0.11447262333135748, "learning_rate": 3.7781993693975573e-06, "loss": 0.856, "step": 9650 }, { "epoch": 4.372904395106479, "grad_norm": 0.138459242212683, "learning_rate": 3.7728328553014425e-06, "loss": 0.8496, "step": 9651 }, { "epoch": 4.37335749886724, "grad_norm": 0.10094947078099116, "learning_rate": 3.7674699665403956e-06, "loss": 0.8376, "step": 9652 }, { "epoch": 4.3738106026280015, "grad_norm": 0.11525487410411347, "learning_rate": 3.7621107036511074e-06, "loss": 0.8419, "step": 9653 }, { "epoch": 4.3742637063887635, "grad_norm": 0.1271793095984026, "learning_rate": 3.756755067169877e-06, "loss": 0.8452, "step": 9654 }, { "epoch": 4.374716810149525, "grad_norm": 0.0965468335473844, "learning_rate": 3.7514030576326677e-06, "loss": 0.8395, "step": 9655 }, { "epoch": 4.375169913910286, "grad_norm": 0.08933731072157967, "learning_rate": 3.746054675575064e-06, "loss": 0.8215, "step": 9656 }, { "epoch": 4.375623017671047, "grad_norm": 0.11775507234442201, "learning_rate": 3.7407099215322818e-06, "loss": 0.8415, "step": 9657 }, { "epoch": 4.376076121431808, "grad_norm": 0.11529978189609383, "learning_rate": 3.735368796039196e-06, "loss": 0.8385, "step": 9658 }, { "epoch": 4.376529225192569, "grad_norm": 0.10618417725592243, "learning_rate": 3.73003129963029e-06, "loss": 0.8268, "step": 9659 }, { "epoch": 4.37698232895333, "grad_norm": 0.11139657935040992, "learning_rate": 3.724697432839719e-06, "loss": 0.8581, "step": 9660 }, { "epoch": 4.377435432714091, "grad_norm": 0.10629754865765205, "learning_rate": 3.7193671962012334e-06, "loss": 0.8394, "step": 9661 }, { "epoch": 4.3778885364748525, "grad_norm": 0.09829070007259802, "learning_rate": 3.7140405902482603e-06, "loss": 0.8561, "step": 9662 }, { "epoch": 4.378341640235614, "grad_norm": 0.09389908306548615, "learning_rate": 3.708717615513835e-06, "loss": 0.8261, "step": 9663 }, { "epoch": 4.378794743996375, "grad_norm": 0.09711175335975906, "learning_rate": 3.703398272530647e-06, "loss": 0.8583, "step": 9664 }, { "epoch": 4.379247847757137, "grad_norm": 0.10083580850124887, "learning_rate": 3.698082561831009e-06, "loss": 0.8447, "step": 9665 }, { "epoch": 4.379700951517898, "grad_norm": 0.11018152276296898, "learning_rate": 3.6927704839468725e-06, "loss": 0.8385, "step": 9666 }, { "epoch": 4.380154055278659, "grad_norm": 0.10375954732592745, "learning_rate": 3.6874620394098436e-06, "loss": 0.8475, "step": 9667 }, { "epoch": 4.38060715903942, "grad_norm": 0.10418179812058201, "learning_rate": 3.682157228751129e-06, "loss": 0.8495, "step": 9668 }, { "epoch": 4.381060262800181, "grad_norm": 0.10225604351467966, "learning_rate": 3.6768560525016186e-06, "loss": 0.837, "step": 9669 }, { "epoch": 4.3815133665609425, "grad_norm": 0.09286570629638241, "learning_rate": 3.671558511191791e-06, "loss": 0.8619, "step": 9670 }, { "epoch": 4.381966470321704, "grad_norm": 0.106769314875008, "learning_rate": 3.6662646053517992e-06, "loss": 0.8435, "step": 9671 }, { "epoch": 4.382419574082465, "grad_norm": 0.10661035084314607, "learning_rate": 3.660974335511407e-06, "loss": 0.8486, "step": 9672 }, { "epoch": 4.382872677843226, "grad_norm": 0.11044930059271144, "learning_rate": 3.6556877022000214e-06, "loss": 0.8421, "step": 9673 }, { "epoch": 4.383325781603987, "grad_norm": 0.11130717055525416, "learning_rate": 3.650404705946704e-06, "loss": 0.8376, "step": 9674 }, { "epoch": 4.383778885364748, "grad_norm": 0.10485305403311124, "learning_rate": 3.645125347280112e-06, "loss": 0.8389, "step": 9675 }, { "epoch": 4.38423198912551, "grad_norm": 0.09974371894658637, "learning_rate": 3.6398496267285865e-06, "loss": 0.8457, "step": 9676 }, { "epoch": 4.384685092886271, "grad_norm": 0.11628637119454381, "learning_rate": 3.6345775448200615e-06, "loss": 0.8659, "step": 9677 }, { "epoch": 4.385138196647032, "grad_norm": 0.10530913698547287, "learning_rate": 3.6293091020821413e-06, "loss": 0.8534, "step": 9678 }, { "epoch": 4.3855913004077935, "grad_norm": 0.11462311437951768, "learning_rate": 3.6240442990420397e-06, "loss": 0.8411, "step": 9679 }, { "epoch": 4.386044404168555, "grad_norm": 0.09173199126059907, "learning_rate": 3.618783136226629e-06, "loss": 0.8248, "step": 9680 }, { "epoch": 4.386497507929316, "grad_norm": 0.0979878403468998, "learning_rate": 3.613525614162399e-06, "loss": 0.8457, "step": 9681 }, { "epoch": 4.386950611690077, "grad_norm": 0.1052000300323892, "learning_rate": 3.6082717333754746e-06, "loss": 0.8372, "step": 9682 }, { "epoch": 4.387403715450838, "grad_norm": 0.10183062449761227, "learning_rate": 3.603021494391641e-06, "loss": 0.8224, "step": 9683 }, { "epoch": 4.387856819211599, "grad_norm": 0.10200333859110242, "learning_rate": 3.5977748977362813e-06, "loss": 0.8301, "step": 9684 }, { "epoch": 4.38830992297236, "grad_norm": 0.09495564368515391, "learning_rate": 3.592531943934452e-06, "loss": 0.8444, "step": 9685 }, { "epoch": 4.3887630267331215, "grad_norm": 0.09776448882139965, "learning_rate": 3.587292633510817e-06, "loss": 0.8523, "step": 9686 }, { "epoch": 4.3892161304938835, "grad_norm": 0.09249551092453447, "learning_rate": 3.582056966989691e-06, "loss": 0.8462, "step": 9687 }, { "epoch": 4.389669234254645, "grad_norm": 0.09876205530703953, "learning_rate": 3.5768249448950145e-06, "loss": 0.8362, "step": 9688 }, { "epoch": 4.390122338015406, "grad_norm": 0.1099751315082595, "learning_rate": 3.5715965677503773e-06, "loss": 0.8592, "step": 9689 }, { "epoch": 4.390575441776167, "grad_norm": 0.10618465587443257, "learning_rate": 3.5663718360789877e-06, "loss": 0.8491, "step": 9690 }, { "epoch": 4.391028545536928, "grad_norm": 0.11742898780327564, "learning_rate": 3.561150750403686e-06, "loss": 0.8543, "step": 9691 }, { "epoch": 4.391481649297689, "grad_norm": 0.11100273610051152, "learning_rate": 3.5559333112469816e-06, "loss": 0.8534, "step": 9692 }, { "epoch": 4.39193475305845, "grad_norm": 0.10670796937027528, "learning_rate": 3.550719519130974e-06, "loss": 0.8444, "step": 9693 }, { "epoch": 4.392387856819211, "grad_norm": 0.1194973312315556, "learning_rate": 3.545509374577445e-06, "loss": 0.8627, "step": 9694 }, { "epoch": 4.3928409605799725, "grad_norm": 0.10939315653556778, "learning_rate": 3.540302878107751e-06, "loss": 0.8518, "step": 9695 }, { "epoch": 4.393294064340734, "grad_norm": 0.10048792068553929, "learning_rate": 3.535100030242946e-06, "loss": 0.8446, "step": 9696 }, { "epoch": 4.393747168101495, "grad_norm": 0.1252661131422611, "learning_rate": 3.529900831503672e-06, "loss": 0.8473, "step": 9697 }, { "epoch": 4.394200271862257, "grad_norm": 0.11914606840227952, "learning_rate": 3.524705282410241e-06, "loss": 0.8449, "step": 9698 }, { "epoch": 4.394653375623018, "grad_norm": 0.107209311142689, "learning_rate": 3.519513383482567e-06, "loss": 0.8376, "step": 9699 }, { "epoch": 4.395106479383779, "grad_norm": 0.11237348868426347, "learning_rate": 3.514325135240233e-06, "loss": 0.8239, "step": 9700 }, { "epoch": 4.39555958314454, "grad_norm": 0.11465982493736336, "learning_rate": 3.5091405382024335e-06, "loss": 0.8633, "step": 9701 }, { "epoch": 4.396012686905301, "grad_norm": 0.10211689719577478, "learning_rate": 3.503959592887989e-06, "loss": 0.8352, "step": 9702 }, { "epoch": 4.3964657906660625, "grad_norm": 0.10984508955418579, "learning_rate": 3.498782299815382e-06, "loss": 0.8226, "step": 9703 }, { "epoch": 4.396918894426824, "grad_norm": 0.10931132952286213, "learning_rate": 3.4936086595027095e-06, "loss": 0.8575, "step": 9704 }, { "epoch": 4.397371998187585, "grad_norm": 0.08752746779578455, "learning_rate": 3.4884386724677223e-06, "loss": 0.8444, "step": 9705 }, { "epoch": 4.397825101948346, "grad_norm": 0.10087389774185437, "learning_rate": 3.48327233922777e-06, "loss": 0.8482, "step": 9706 }, { "epoch": 4.398278205709107, "grad_norm": 0.10085426151605363, "learning_rate": 3.4781096602998845e-06, "loss": 0.863, "step": 9707 }, { "epoch": 4.398731309469869, "grad_norm": 0.09079830063324935, "learning_rate": 3.472950636200687e-06, "loss": 0.8462, "step": 9708 }, { "epoch": 4.39918441323063, "grad_norm": 0.1018365333283631, "learning_rate": 3.4677952674464676e-06, "loss": 0.8495, "step": 9709 }, { "epoch": 4.399637516991391, "grad_norm": 0.10127204242676587, "learning_rate": 3.462643554553129e-06, "loss": 0.8557, "step": 9710 }, { "epoch": 4.400090620752152, "grad_norm": 0.10246459823678561, "learning_rate": 3.457495498036205e-06, "loss": 0.8328, "step": 9711 }, { "epoch": 4.4005437245129135, "grad_norm": 0.09531860586490642, "learning_rate": 3.452351098410893e-06, "loss": 0.8529, "step": 9712 }, { "epoch": 4.400996828273675, "grad_norm": 0.09569161114478439, "learning_rate": 3.4472103561919856e-06, "loss": 0.8561, "step": 9713 }, { "epoch": 4.401449932034436, "grad_norm": 0.10284777717845749, "learning_rate": 3.4420732718939465e-06, "loss": 0.8116, "step": 9714 }, { "epoch": 4.401903035795197, "grad_norm": 0.1070984639381634, "learning_rate": 3.43693984603084e-06, "loss": 0.836, "step": 9715 }, { "epoch": 4.402356139555958, "grad_norm": 0.09630323555969329, "learning_rate": 3.4318100791163935e-06, "loss": 0.8448, "step": 9716 }, { "epoch": 4.402809243316719, "grad_norm": 0.10452036221118358, "learning_rate": 3.4266839716639465e-06, "loss": 0.8382, "step": 9717 }, { "epoch": 4.403262347077481, "grad_norm": 0.11305831806963547, "learning_rate": 3.421561524186472e-06, "loss": 0.8405, "step": 9718 }, { "epoch": 4.403715450838242, "grad_norm": 0.1027006764757777, "learning_rate": 3.4164427371966035e-06, "loss": 0.8524, "step": 9719 }, { "epoch": 4.4041685545990035, "grad_norm": 0.10713551151227431, "learning_rate": 3.411327611206572e-06, "loss": 0.8422, "step": 9720 }, { "epoch": 4.404621658359765, "grad_norm": 0.09163104998269914, "learning_rate": 3.4062161467282785e-06, "loss": 0.8455, "step": 9721 }, { "epoch": 4.405074762120526, "grad_norm": 0.09337652796060755, "learning_rate": 3.401108344273216e-06, "loss": 0.8398, "step": 9722 }, { "epoch": 4.405527865881287, "grad_norm": 0.11017341425681879, "learning_rate": 3.3960042043525586e-06, "loss": 0.8531, "step": 9723 }, { "epoch": 4.405980969642048, "grad_norm": 0.10128736142755076, "learning_rate": 3.3909037274770664e-06, "loss": 0.8618, "step": 9724 }, { "epoch": 4.406434073402809, "grad_norm": 0.09483196532457974, "learning_rate": 3.3858069141571747e-06, "loss": 0.8287, "step": 9725 }, { "epoch": 4.40688717716357, "grad_norm": 0.09643106208117465, "learning_rate": 3.3807137649029253e-06, "loss": 0.8472, "step": 9726 }, { "epoch": 4.407340280924331, "grad_norm": 0.1033622700831172, "learning_rate": 3.375624280223995e-06, "loss": 0.8299, "step": 9727 }, { "epoch": 4.4077933846850925, "grad_norm": 0.09504887258276883, "learning_rate": 3.3705384606297088e-06, "loss": 0.8614, "step": 9728 }, { "epoch": 4.4082464884458545, "grad_norm": 0.08942067126388557, "learning_rate": 3.365456306629007e-06, "loss": 0.8531, "step": 9729 }, { "epoch": 4.408699592206616, "grad_norm": 0.11270898904263846, "learning_rate": 3.360377818730487e-06, "loss": 0.849, "step": 9730 }, { "epoch": 4.409152695967377, "grad_norm": 0.10568035564680102, "learning_rate": 3.355302997442351e-06, "loss": 0.8333, "step": 9731 }, { "epoch": 4.409605799728138, "grad_norm": 0.09170328234420638, "learning_rate": 3.3502318432724555e-06, "loss": 0.85, "step": 9732 }, { "epoch": 4.410058903488899, "grad_norm": 0.10029762343709038, "learning_rate": 3.3451643567282832e-06, "loss": 0.8394, "step": 9733 }, { "epoch": 4.41051200724966, "grad_norm": 0.10444324730698126, "learning_rate": 3.3401005383169395e-06, "loss": 0.8294, "step": 9734 }, { "epoch": 4.410965111010421, "grad_norm": 0.11218821210587553, "learning_rate": 3.3350403885451832e-06, "loss": 0.868, "step": 9735 }, { "epoch": 4.4114182147711825, "grad_norm": 0.09842698283788018, "learning_rate": 3.3299839079193874e-06, "loss": 0.8489, "step": 9736 }, { "epoch": 4.411871318531944, "grad_norm": 0.0957330426317515, "learning_rate": 3.324931096945574e-06, "loss": 0.8382, "step": 9737 }, { "epoch": 4.412324422292705, "grad_norm": 0.12325839892949486, "learning_rate": 3.3198819561293736e-06, "loss": 0.8418, "step": 9738 }, { "epoch": 4.412777526053466, "grad_norm": 0.08953109419102143, "learning_rate": 3.314836485976085e-06, "loss": 0.8339, "step": 9739 }, { "epoch": 4.413230629814228, "grad_norm": 0.11209670697081592, "learning_rate": 3.309794686990606e-06, "loss": 0.8446, "step": 9740 }, { "epoch": 4.413683733574989, "grad_norm": 0.1011225719302445, "learning_rate": 3.304756559677489e-06, "loss": 0.8586, "step": 9741 }, { "epoch": 4.41413683733575, "grad_norm": 0.09270134501503641, "learning_rate": 3.2997221045409077e-06, "loss": 0.8469, "step": 9742 }, { "epoch": 4.414589941096511, "grad_norm": 0.10355134497210984, "learning_rate": 3.2946913220846644e-06, "loss": 0.8654, "step": 9743 }, { "epoch": 4.415043044857272, "grad_norm": 0.10137982497392291, "learning_rate": 3.2896642128122138e-06, "loss": 0.8522, "step": 9744 }, { "epoch": 4.4154961486180335, "grad_norm": 0.11350610834662671, "learning_rate": 3.2846407772266197e-06, "loss": 0.8585, "step": 9745 }, { "epoch": 4.415949252378795, "grad_norm": 0.12244848752046943, "learning_rate": 3.2796210158305963e-06, "loss": 0.8614, "step": 9746 }, { "epoch": 4.416402356139556, "grad_norm": 0.1147082498335983, "learning_rate": 3.2746049291264747e-06, "loss": 0.8462, "step": 9747 }, { "epoch": 4.416855459900317, "grad_norm": 0.11852239533351376, "learning_rate": 3.269592517616236e-06, "loss": 0.8546, "step": 9748 }, { "epoch": 4.417308563661078, "grad_norm": 0.09711544167479824, "learning_rate": 3.2645837818014736e-06, "loss": 0.8416, "step": 9749 }, { "epoch": 4.417761667421839, "grad_norm": 0.10434733338910449, "learning_rate": 3.259578722183432e-06, "loss": 0.8667, "step": 9750 }, { "epoch": 4.418214771182601, "grad_norm": 0.10543934187265426, "learning_rate": 3.2545773392629764e-06, "loss": 0.8428, "step": 9751 }, { "epoch": 4.418667874943362, "grad_norm": 0.1036633074676956, "learning_rate": 3.2495796335406047e-06, "loss": 0.8448, "step": 9752 }, { "epoch": 4.4191209787041235, "grad_norm": 0.10934107156759428, "learning_rate": 3.2445856055164504e-06, "loss": 0.8579, "step": 9753 }, { "epoch": 4.419574082464885, "grad_norm": 0.11042244904689116, "learning_rate": 3.2395952556902642e-06, "loss": 0.8664, "step": 9754 }, { "epoch": 4.420027186225646, "grad_norm": 0.10742574304301783, "learning_rate": 3.234608584561465e-06, "loss": 0.8401, "step": 9755 }, { "epoch": 4.420480289986407, "grad_norm": 0.09160116955493469, "learning_rate": 3.2296255926290575e-06, "loss": 0.8706, "step": 9756 }, { "epoch": 4.420933393747168, "grad_norm": 0.1041465447717723, "learning_rate": 3.2246462803917233e-06, "loss": 0.8536, "step": 9757 }, { "epoch": 4.421386497507929, "grad_norm": 0.1042678296978463, "learning_rate": 3.2196706483477304e-06, "loss": 0.8352, "step": 9758 }, { "epoch": 4.42183960126869, "grad_norm": 0.09734048091643796, "learning_rate": 3.214698696995027e-06, "loss": 0.8686, "step": 9759 }, { "epoch": 4.422292705029451, "grad_norm": 0.10061294125486318, "learning_rate": 3.2097304268311392e-06, "loss": 0.8354, "step": 9760 }, { "epoch": 4.4227458087902125, "grad_norm": 0.10320010284966226, "learning_rate": 3.2047658383532832e-06, "loss": 0.8651, "step": 9761 }, { "epoch": 4.4231989125509745, "grad_norm": 0.09410830437849403, "learning_rate": 3.1998049320582568e-06, "loss": 0.8533, "step": 9762 }, { "epoch": 4.423652016311736, "grad_norm": 0.11141237784617304, "learning_rate": 3.194847708442508e-06, "loss": 0.8211, "step": 9763 }, { "epoch": 4.424105120072497, "grad_norm": 0.08918900332083633, "learning_rate": 3.189894168002128e-06, "loss": 0.8494, "step": 9764 }, { "epoch": 4.424558223833258, "grad_norm": 0.0981079154715198, "learning_rate": 3.184944311232818e-06, "loss": 0.8441, "step": 9765 }, { "epoch": 4.425011327594019, "grad_norm": 0.0911362599174987, "learning_rate": 3.1799981386299383e-06, "loss": 0.8407, "step": 9766 }, { "epoch": 4.42546443135478, "grad_norm": 0.10048253983883448, "learning_rate": 3.1750556506884436e-06, "loss": 0.8331, "step": 9767 }, { "epoch": 4.425917535115541, "grad_norm": 0.9831510118824707, "learning_rate": 3.1701168479029555e-06, "loss": 0.8262, "step": 9768 }, { "epoch": 4.426370638876302, "grad_norm": 0.09749074349331126, "learning_rate": 3.1651817307677013e-06, "loss": 0.8493, "step": 9769 }, { "epoch": 4.426823742637064, "grad_norm": 0.09282095743499953, "learning_rate": 3.160250299776557e-06, "loss": 0.8334, "step": 9770 }, { "epoch": 4.427276846397825, "grad_norm": 0.0982832650925157, "learning_rate": 3.155322555423017e-06, "loss": 0.8423, "step": 9771 }, { "epoch": 4.427729950158587, "grad_norm": 0.09634806516482916, "learning_rate": 3.1503984982002066e-06, "loss": 0.8616, "step": 9772 }, { "epoch": 4.428183053919348, "grad_norm": 0.09361369162542321, "learning_rate": 3.145478128600905e-06, "loss": 0.8724, "step": 9773 }, { "epoch": 4.428636157680109, "grad_norm": 0.0906717432226084, "learning_rate": 3.1405614471174827e-06, "loss": 0.8357, "step": 9774 }, { "epoch": 4.42908926144087, "grad_norm": 0.09372393521677387, "learning_rate": 3.135648454241986e-06, "loss": 0.8568, "step": 9775 }, { "epoch": 4.429542365201631, "grad_norm": 0.09003673223411691, "learning_rate": 3.1307391504660446e-06, "loss": 0.8479, "step": 9776 }, { "epoch": 4.429995468962392, "grad_norm": 0.09147977328000928, "learning_rate": 3.1258335362809666e-06, "loss": 0.831, "step": 9777 }, { "epoch": 4.4304485727231535, "grad_norm": 0.10030141554580495, "learning_rate": 3.120931612177658e-06, "loss": 0.8492, "step": 9778 }, { "epoch": 4.430901676483915, "grad_norm": 0.0958447251134262, "learning_rate": 3.116033378646659e-06, "loss": 0.8617, "step": 9779 }, { "epoch": 4.431354780244676, "grad_norm": 0.0882387951453932, "learning_rate": 3.1111388361781603e-06, "loss": 0.8426, "step": 9780 }, { "epoch": 4.431807884005437, "grad_norm": 0.09262009123982881, "learning_rate": 3.1062479852619563e-06, "loss": 0.848, "step": 9781 }, { "epoch": 4.432260987766199, "grad_norm": 0.09548618121756776, "learning_rate": 3.1013608263875007e-06, "loss": 0.8606, "step": 9782 }, { "epoch": 4.43271409152696, "grad_norm": 0.0931403425280639, "learning_rate": 3.0964773600438458e-06, "loss": 0.866, "step": 9783 }, { "epoch": 4.433167195287721, "grad_norm": 0.09720251597586996, "learning_rate": 3.0915975867197032e-06, "loss": 0.8566, "step": 9784 }, { "epoch": 4.433620299048482, "grad_norm": 0.0991632962196483, "learning_rate": 3.0867215069033987e-06, "loss": 0.8489, "step": 9785 }, { "epoch": 4.4340734028092434, "grad_norm": 0.09538403690639068, "learning_rate": 3.0818491210828962e-06, "loss": 0.8598, "step": 9786 }, { "epoch": 4.434526506570005, "grad_norm": 0.08381382808780044, "learning_rate": 3.076980429745784e-06, "loss": 0.853, "step": 9787 }, { "epoch": 4.434979610330766, "grad_norm": 0.09171186239935729, "learning_rate": 3.072115433379277e-06, "loss": 0.851, "step": 9788 }, { "epoch": 4.435432714091527, "grad_norm": 0.09400120623051468, "learning_rate": 3.0672541324702387e-06, "loss": 0.8605, "step": 9789 }, { "epoch": 4.435885817852288, "grad_norm": 0.08720124050265288, "learning_rate": 3.062396527505138e-06, "loss": 0.8509, "step": 9790 }, { "epoch": 4.436338921613049, "grad_norm": 0.09076040833349311, "learning_rate": 3.057542618970102e-06, "loss": 0.8713, "step": 9791 }, { "epoch": 4.43679202537381, "grad_norm": 0.09319867481788968, "learning_rate": 3.052692407350852e-06, "loss": 0.838, "step": 9792 }, { "epoch": 4.437245129134572, "grad_norm": 0.09394835266328949, "learning_rate": 3.0478458931327794e-06, "loss": 0.8441, "step": 9793 }, { "epoch": 4.437698232895333, "grad_norm": 0.09715542755300667, "learning_rate": 3.0430030768008765e-06, "loss": 0.859, "step": 9794 }, { "epoch": 4.4381513366560945, "grad_norm": 0.09423438662515367, "learning_rate": 3.038163958839766e-06, "loss": 0.8414, "step": 9795 }, { "epoch": 4.438604440416856, "grad_norm": 0.10086083454007863, "learning_rate": 3.033328539733731e-06, "loss": 0.8409, "step": 9796 }, { "epoch": 4.439057544177617, "grad_norm": 0.10017697221897731, "learning_rate": 3.028496819966642e-06, "loss": 0.8268, "step": 9797 }, { "epoch": 4.439510647938378, "grad_norm": 0.09904608912365627, "learning_rate": 3.023668800022037e-06, "loss": 0.8362, "step": 9798 }, { "epoch": 4.439963751699139, "grad_norm": 0.1025763345904649, "learning_rate": 3.0188444803830497e-06, "loss": 0.8276, "step": 9799 }, { "epoch": 4.4404168554599, "grad_norm": 0.09226647639768232, "learning_rate": 3.01402386153248e-06, "loss": 0.8324, "step": 9800 }, { "epoch": 4.440869959220661, "grad_norm": 0.1009359393473806, "learning_rate": 3.0092069439527204e-06, "loss": 0.8565, "step": 9801 }, { "epoch": 4.441323062981422, "grad_norm": 0.10527494006910727, "learning_rate": 3.004393728125825e-06, "loss": 0.8502, "step": 9802 }, { "epoch": 4.441776166742184, "grad_norm": 0.107351174831136, "learning_rate": 2.9995842145334573e-06, "loss": 0.8562, "step": 9803 }, { "epoch": 4.442229270502946, "grad_norm": 0.09950901362241374, "learning_rate": 2.994778403656908e-06, "loss": 0.8446, "step": 9804 }, { "epoch": 4.442682374263707, "grad_norm": 0.10618683222660823, "learning_rate": 2.989976295977122e-06, "loss": 0.8441, "step": 9805 }, { "epoch": 4.443135478024468, "grad_norm": 0.10090624200508559, "learning_rate": 2.985177891974642e-06, "loss": 0.855, "step": 9806 }, { "epoch": 4.443588581785229, "grad_norm": 0.09707285219688855, "learning_rate": 2.980383192129668e-06, "loss": 0.8356, "step": 9807 }, { "epoch": 4.44404168554599, "grad_norm": 0.09784762415324746, "learning_rate": 2.97559219692201e-06, "loss": 0.8491, "step": 9808 }, { "epoch": 4.444494789306751, "grad_norm": 0.09995622811558968, "learning_rate": 2.9708049068311174e-06, "loss": 0.8529, "step": 9809 }, { "epoch": 4.444947893067512, "grad_norm": 0.09146127891285206, "learning_rate": 2.9660213223360545e-06, "loss": 0.8518, "step": 9810 }, { "epoch": 4.4454009968282735, "grad_norm": 0.11186299812217197, "learning_rate": 2.961241443915537e-06, "loss": 0.8377, "step": 9811 }, { "epoch": 4.445854100589035, "grad_norm": 0.10796805734102954, "learning_rate": 2.956465272047888e-06, "loss": 0.8444, "step": 9812 }, { "epoch": 4.446307204349796, "grad_norm": 0.09259293910143204, "learning_rate": 2.951692807211086e-06, "loss": 0.8387, "step": 9813 }, { "epoch": 4.446760308110557, "grad_norm": 0.10342316590880887, "learning_rate": 2.946924049882709e-06, "loss": 0.8365, "step": 9814 }, { "epoch": 4.447213411871319, "grad_norm": 0.099075464467083, "learning_rate": 2.942159000539975e-06, "loss": 0.8533, "step": 9815 }, { "epoch": 4.44766651563208, "grad_norm": 0.10135385936670162, "learning_rate": 2.9373976596597465e-06, "loss": 0.8484, "step": 9816 }, { "epoch": 4.448119619392841, "grad_norm": 0.09135311871007568, "learning_rate": 2.9326400277184874e-06, "loss": 0.8591, "step": 9817 }, { "epoch": 4.448572723153602, "grad_norm": 0.08211414249901716, "learning_rate": 2.927886105192319e-06, "loss": 0.8282, "step": 9818 }, { "epoch": 4.449025826914363, "grad_norm": 0.09209551860717062, "learning_rate": 2.923135892556963e-06, "loss": 0.8544, "step": 9819 }, { "epoch": 4.449478930675125, "grad_norm": 0.09646337734339251, "learning_rate": 2.918389390287799e-06, "loss": 0.8423, "step": 9820 }, { "epoch": 4.449932034435886, "grad_norm": 0.08766810223331713, "learning_rate": 2.913646598859803e-06, "loss": 0.8449, "step": 9821 }, { "epoch": 4.450385138196647, "grad_norm": 0.09431627231022074, "learning_rate": 2.908907518747617e-06, "loss": 0.8541, "step": 9822 }, { "epoch": 4.450838241957408, "grad_norm": 0.09139540323960327, "learning_rate": 2.9041721504254795e-06, "loss": 0.8608, "step": 9823 }, { "epoch": 4.451291345718169, "grad_norm": 0.08428077073108195, "learning_rate": 2.8994404943672606e-06, "loss": 0.8173, "step": 9824 }, { "epoch": 4.45174444947893, "grad_norm": 0.1023742716711407, "learning_rate": 2.894712551046488e-06, "loss": 0.8598, "step": 9825 }, { "epoch": 4.452197553239692, "grad_norm": 0.09997619031331917, "learning_rate": 2.8899883209362813e-06, "loss": 0.847, "step": 9826 }, { "epoch": 4.452650657000453, "grad_norm": 0.08884912354404798, "learning_rate": 2.8852678045094173e-06, "loss": 0.8586, "step": 9827 }, { "epoch": 4.4531037607612145, "grad_norm": 0.10234504832753687, "learning_rate": 2.8805510022382788e-06, "loss": 0.8446, "step": 9828 }, { "epoch": 4.453556864521976, "grad_norm": 0.10528922291846714, "learning_rate": 2.875837914594892e-06, "loss": 0.8424, "step": 9829 }, { "epoch": 4.454009968282737, "grad_norm": 0.1045705195924595, "learning_rate": 2.8711285420509028e-06, "loss": 0.826, "step": 9830 }, { "epoch": 4.454463072043498, "grad_norm": 0.09649136148230036, "learning_rate": 2.8664228850775998e-06, "loss": 0.8365, "step": 9831 }, { "epoch": 4.454916175804259, "grad_norm": 0.10957950891545606, "learning_rate": 2.8617209441458784e-06, "loss": 0.8746, "step": 9832 }, { "epoch": 4.45536927956502, "grad_norm": 0.09712440781934753, "learning_rate": 2.857022719726268e-06, "loss": 0.85, "step": 9833 }, { "epoch": 4.455822383325781, "grad_norm": 0.10339751860111243, "learning_rate": 2.852328212288944e-06, "loss": 0.8439, "step": 9834 }, { "epoch": 4.456275487086542, "grad_norm": 0.10007989534412895, "learning_rate": 2.8476374223036864e-06, "loss": 0.8503, "step": 9835 }, { "epoch": 4.456728590847304, "grad_norm": 0.10376695226714198, "learning_rate": 2.8429503502399194e-06, "loss": 0.8518, "step": 9836 }, { "epoch": 4.457181694608066, "grad_norm": 0.09266004263104985, "learning_rate": 2.838266996566681e-06, "loss": 0.8348, "step": 9837 }, { "epoch": 4.457634798368827, "grad_norm": 0.0912089882820244, "learning_rate": 2.8335873617526545e-06, "loss": 0.8321, "step": 9838 }, { "epoch": 4.458087902129588, "grad_norm": 0.10034612152067886, "learning_rate": 2.8289114462661407e-06, "loss": 0.8385, "step": 9839 }, { "epoch": 4.458541005890349, "grad_norm": 0.10490523680060079, "learning_rate": 2.824239250575058e-06, "loss": 0.8525, "step": 9840 }, { "epoch": 4.45899410965111, "grad_norm": 0.0955072920650967, "learning_rate": 2.819570775146976e-06, "loss": 0.8343, "step": 9841 }, { "epoch": 4.459447213411871, "grad_norm": 0.10869895176685558, "learning_rate": 2.8149060204490708e-06, "loss": 0.8585, "step": 9842 }, { "epoch": 4.459900317172632, "grad_norm": 0.10098755157469981, "learning_rate": 2.810244986948165e-06, "loss": 0.8433, "step": 9843 }, { "epoch": 4.4603534209333935, "grad_norm": 0.09409658251772003, "learning_rate": 2.8055876751106858e-06, "loss": 0.8429, "step": 9844 }, { "epoch": 4.460806524694155, "grad_norm": 0.09375689209967374, "learning_rate": 2.800934085402718e-06, "loss": 0.8541, "step": 9845 }, { "epoch": 4.461259628454917, "grad_norm": 0.10191213737384953, "learning_rate": 2.7962842182899376e-06, "loss": 0.8546, "step": 9846 }, { "epoch": 4.461712732215678, "grad_norm": 0.11373811328178629, "learning_rate": 2.7916380742376837e-06, "loss": 0.8671, "step": 9847 }, { "epoch": 4.462165835976439, "grad_norm": 0.10163457005285491, "learning_rate": 2.7869956537109e-06, "loss": 0.8305, "step": 9848 }, { "epoch": 4.4626189397372, "grad_norm": 0.09709538334981739, "learning_rate": 2.782356957174157e-06, "loss": 0.838, "step": 9849 }, { "epoch": 4.463072043497961, "grad_norm": 0.10475672075723226, "learning_rate": 2.777721985091675e-06, "loss": 0.8643, "step": 9850 }, { "epoch": 4.463525147258722, "grad_norm": 0.10391525190299072, "learning_rate": 2.7730907379272685e-06, "loss": 0.8369, "step": 9851 }, { "epoch": 4.463978251019483, "grad_norm": 0.10260384433665669, "learning_rate": 2.768463216144417e-06, "loss": 0.846, "step": 9852 }, { "epoch": 4.4644313547802446, "grad_norm": 0.09893788351100465, "learning_rate": 2.7638394202061847e-06, "loss": 0.8464, "step": 9853 }, { "epoch": 4.464884458541006, "grad_norm": 0.12144189866417393, "learning_rate": 2.7592193505753085e-06, "loss": 0.8441, "step": 9854 }, { "epoch": 4.465337562301767, "grad_norm": 0.12568484751077466, "learning_rate": 2.7546030077141117e-06, "loss": 0.843, "step": 9855 }, { "epoch": 4.465790666062528, "grad_norm": 0.09826952114020263, "learning_rate": 2.749990392084567e-06, "loss": 0.8665, "step": 9856 }, { "epoch": 4.46624376982329, "grad_norm": 0.10124198404962335, "learning_rate": 2.7453815041482746e-06, "loss": 0.8514, "step": 9857 }, { "epoch": 4.466696873584051, "grad_norm": 0.11109622670577293, "learning_rate": 2.740776344366447e-06, "loss": 0.8305, "step": 9858 }, { "epoch": 4.467149977344812, "grad_norm": 0.10501234640304942, "learning_rate": 2.736174913199947e-06, "loss": 0.8371, "step": 9859 }, { "epoch": 4.467603081105573, "grad_norm": 0.09518337283519923, "learning_rate": 2.731577211109233e-06, "loss": 0.8419, "step": 9860 }, { "epoch": 4.4680561848663345, "grad_norm": 0.09389315910128104, "learning_rate": 2.7269832385544215e-06, "loss": 0.8462, "step": 9861 }, { "epoch": 4.468509288627096, "grad_norm": 0.0906283171313862, "learning_rate": 2.7223929959952333e-06, "loss": 0.8514, "step": 9862 }, { "epoch": 4.468962392387857, "grad_norm": 0.08948633836637883, "learning_rate": 2.71780648389103e-06, "loss": 0.8331, "step": 9863 }, { "epoch": 4.469415496148618, "grad_norm": 0.08972585097785697, "learning_rate": 2.7132237027007956e-06, "loss": 0.8576, "step": 9864 }, { "epoch": 4.469868599909379, "grad_norm": 0.08694019842011966, "learning_rate": 2.708644652883132e-06, "loss": 0.8363, "step": 9865 }, { "epoch": 4.47032170367014, "grad_norm": 0.10599887070666408, "learning_rate": 2.7040693348962778e-06, "loss": 0.8476, "step": 9866 }, { "epoch": 4.470774807430901, "grad_norm": 0.10126132890050496, "learning_rate": 2.699497749198101e-06, "loss": 0.8497, "step": 9867 }, { "epoch": 4.471227911191663, "grad_norm": 0.0885735004600968, "learning_rate": 2.6949298962460858e-06, "loss": 0.8426, "step": 9868 }, { "epoch": 4.471681014952424, "grad_norm": 0.1120695030982866, "learning_rate": 2.6903657764973414e-06, "loss": 0.8428, "step": 9869 }, { "epoch": 4.472134118713186, "grad_norm": 0.09966034362119745, "learning_rate": 2.6858053904086223e-06, "loss": 0.8471, "step": 9870 }, { "epoch": 4.472587222473947, "grad_norm": 0.08791700224049334, "learning_rate": 2.681248738436284e-06, "loss": 0.8284, "step": 9871 }, { "epoch": 4.473040326234708, "grad_norm": 0.10595120479678576, "learning_rate": 2.676695821036335e-06, "loss": 0.847, "step": 9872 }, { "epoch": 4.473493429995469, "grad_norm": 0.08995456224535492, "learning_rate": 2.6721466386643836e-06, "loss": 0.8259, "step": 9873 }, { "epoch": 4.47394653375623, "grad_norm": 0.09777023819452703, "learning_rate": 2.6676011917756885e-06, "loss": 0.8395, "step": 9874 }, { "epoch": 4.474399637516991, "grad_norm": 0.08729765400807514, "learning_rate": 2.6630594808251163e-06, "loss": 0.838, "step": 9875 }, { "epoch": 4.474852741277752, "grad_norm": 0.09804825066257816, "learning_rate": 2.6585215062671575e-06, "loss": 0.8354, "step": 9876 }, { "epoch": 4.4753058450385135, "grad_norm": 0.0979849810850959, "learning_rate": 2.6539872685559555e-06, "loss": 0.8434, "step": 9877 }, { "epoch": 4.475758948799275, "grad_norm": 0.09106828322468724, "learning_rate": 2.649456768145249e-06, "loss": 0.8623, "step": 9878 }, { "epoch": 4.476212052560037, "grad_norm": 0.09498296886716251, "learning_rate": 2.644930005488422e-06, "loss": 0.8395, "step": 9879 }, { "epoch": 4.476665156320798, "grad_norm": 0.09798122083503667, "learning_rate": 2.640406981038468e-06, "loss": 0.8527, "step": 9880 }, { "epoch": 4.477118260081559, "grad_norm": 0.08341875143818565, "learning_rate": 2.635887695248034e-06, "loss": 0.8358, "step": 9881 }, { "epoch": 4.47757136384232, "grad_norm": 0.08726427685431751, "learning_rate": 2.631372148569358e-06, "loss": 0.842, "step": 9882 }, { "epoch": 4.478024467603081, "grad_norm": 0.09601180635081377, "learning_rate": 2.6268603414543313e-06, "loss": 0.865, "step": 9883 }, { "epoch": 4.478477571363842, "grad_norm": 0.08703947223335927, "learning_rate": 2.6223522743544604e-06, "loss": 0.8559, "step": 9884 }, { "epoch": 4.478930675124603, "grad_norm": 0.09093875237763634, "learning_rate": 2.617847947720864e-06, "loss": 0.8486, "step": 9885 }, { "epoch": 4.4793837788853645, "grad_norm": 0.09261814788907911, "learning_rate": 2.6133473620043238e-06, "loss": 0.8357, "step": 9886 }, { "epoch": 4.479836882646126, "grad_norm": 0.1031353199981003, "learning_rate": 2.608850517655199e-06, "loss": 0.8492, "step": 9887 }, { "epoch": 4.480289986406887, "grad_norm": 0.10958874756995758, "learning_rate": 2.6043574151235218e-06, "loss": 0.8438, "step": 9888 }, { "epoch": 4.480743090167648, "grad_norm": 0.08723126801110971, "learning_rate": 2.5998680548589094e-06, "loss": 0.8575, "step": 9889 }, { "epoch": 4.48119619392841, "grad_norm": 0.09475775437093373, "learning_rate": 2.595382437310634e-06, "loss": 0.8409, "step": 9890 }, { "epoch": 4.481649297689171, "grad_norm": 0.09358437280819466, "learning_rate": 2.5909005629275762e-06, "loss": 0.8529, "step": 9891 }, { "epoch": 4.482102401449932, "grad_norm": 0.10320755316848797, "learning_rate": 2.5864224321582444e-06, "loss": 0.845, "step": 9892 }, { "epoch": 4.482555505210693, "grad_norm": 0.09366831780883901, "learning_rate": 2.581948045450786e-06, "loss": 0.8583, "step": 9893 }, { "epoch": 4.4830086089714545, "grad_norm": 0.10904279291789962, "learning_rate": 2.577477403252946e-06, "loss": 0.824, "step": 9894 }, { "epoch": 4.483461712732216, "grad_norm": 0.09961668174188099, "learning_rate": 2.5730105060121303e-06, "loss": 0.8386, "step": 9895 }, { "epoch": 4.483914816492977, "grad_norm": 0.09171577380844474, "learning_rate": 2.5685473541753368e-06, "loss": 0.8529, "step": 9896 }, { "epoch": 4.484367920253738, "grad_norm": 0.09730902683060648, "learning_rate": 2.5640879481892132e-06, "loss": 0.8494, "step": 9897 }, { "epoch": 4.484821024014499, "grad_norm": 0.08892501791039423, "learning_rate": 2.5596322885000157e-06, "loss": 0.8529, "step": 9898 }, { "epoch": 4.48527412777526, "grad_norm": 0.09764189729628786, "learning_rate": 2.555180375553641e-06, "loss": 0.8297, "step": 9899 }, { "epoch": 4.485727231536022, "grad_norm": 0.08482042222407628, "learning_rate": 2.5507322097955946e-06, "loss": 0.8511, "step": 9900 }, { "epoch": 4.486180335296783, "grad_norm": 0.08958102703194426, "learning_rate": 2.546287791671014e-06, "loss": 0.8629, "step": 9901 }, { "epoch": 4.486633439057544, "grad_norm": 0.11038169143811258, "learning_rate": 2.5418471216246674e-06, "loss": 0.8503, "step": 9902 }, { "epoch": 4.4870865428183055, "grad_norm": 0.10993667696986234, "learning_rate": 2.537410200100934e-06, "loss": 0.8306, "step": 9903 }, { "epoch": 4.487539646579067, "grad_norm": 0.09247940042423064, "learning_rate": 2.532977027543839e-06, "loss": 0.8608, "step": 9904 }, { "epoch": 4.487992750339828, "grad_norm": 0.08843537863188072, "learning_rate": 2.528547604397007e-06, "loss": 0.8415, "step": 9905 }, { "epoch": 4.488445854100589, "grad_norm": 0.1025841647914945, "learning_rate": 2.5241219311037134e-06, "loss": 0.8568, "step": 9906 }, { "epoch": 4.48889895786135, "grad_norm": 0.10703598632031652, "learning_rate": 2.5197000081068356e-06, "loss": 0.8374, "step": 9907 }, { "epoch": 4.489352061622111, "grad_norm": 0.08827943730395155, "learning_rate": 2.515281835848895e-06, "loss": 0.852, "step": 9908 }, { "epoch": 4.489805165382872, "grad_norm": 0.09657837325083969, "learning_rate": 2.510867414772018e-06, "loss": 0.8401, "step": 9909 }, { "epoch": 4.490258269143634, "grad_norm": 0.10388127206964509, "learning_rate": 2.5064567453179666e-06, "loss": 0.8373, "step": 9910 }, { "epoch": 4.4907113729043955, "grad_norm": 0.0925186315769958, "learning_rate": 2.5020498279281393e-06, "loss": 0.8423, "step": 9911 }, { "epoch": 4.491164476665157, "grad_norm": 0.08366403132813192, "learning_rate": 2.4976466630435248e-06, "loss": 0.8391, "step": 9912 }, { "epoch": 4.491617580425918, "grad_norm": 0.0907059044263618, "learning_rate": 2.49324725110478e-06, "loss": 0.8246, "step": 9913 }, { "epoch": 4.492070684186679, "grad_norm": 0.09909748063768631, "learning_rate": 2.4888515925521483e-06, "loss": 0.8466, "step": 9914 }, { "epoch": 4.49252378794744, "grad_norm": 0.08608171966837717, "learning_rate": 2.484459687825522e-06, "loss": 0.8484, "step": 9915 }, { "epoch": 4.492976891708201, "grad_norm": 0.09625947379899699, "learning_rate": 2.480071537364408e-06, "loss": 0.8475, "step": 9916 }, { "epoch": 4.493429995468962, "grad_norm": 0.094517579709189, "learning_rate": 2.4756871416079296e-06, "loss": 0.8354, "step": 9917 }, { "epoch": 4.493883099229723, "grad_norm": 0.09634713273392943, "learning_rate": 2.4713065009948565e-06, "loss": 0.8416, "step": 9918 }, { "epoch": 4.4943362029904845, "grad_norm": 0.08603639382038401, "learning_rate": 2.466929615963558e-06, "loss": 0.8346, "step": 9919 }, { "epoch": 4.494789306751246, "grad_norm": 0.08658757914629425, "learning_rate": 2.462556486952048e-06, "loss": 0.8669, "step": 9920 }, { "epoch": 4.495242410512008, "grad_norm": 0.09333830260549514, "learning_rate": 2.4581871143979497e-06, "loss": 0.8548, "step": 9921 }, { "epoch": 4.495695514272769, "grad_norm": 0.09200403267150291, "learning_rate": 2.453821498738518e-06, "loss": 0.8669, "step": 9922 }, { "epoch": 4.49614861803353, "grad_norm": 0.09220765116332659, "learning_rate": 2.449459640410634e-06, "loss": 0.8473, "step": 9923 }, { "epoch": 4.496601721794291, "grad_norm": 0.08963007151342914, "learning_rate": 2.4451015398507936e-06, "loss": 0.8374, "step": 9924 }, { "epoch": 4.497054825555052, "grad_norm": 0.09067936954553893, "learning_rate": 2.4407471974951193e-06, "loss": 0.8358, "step": 9925 }, { "epoch": 4.497507929315813, "grad_norm": 0.08879366947831464, "learning_rate": 2.436396613779368e-06, "loss": 0.8318, "step": 9926 }, { "epoch": 4.4979610330765745, "grad_norm": 0.08979608832072362, "learning_rate": 2.4320497891388993e-06, "loss": 0.844, "step": 9927 }, { "epoch": 4.498414136837336, "grad_norm": 0.10808482718909064, "learning_rate": 2.4277067240087294e-06, "loss": 0.8483, "step": 9928 }, { "epoch": 4.498867240598097, "grad_norm": 0.08832619883818088, "learning_rate": 2.4233674188234657e-06, "loss": 0.8468, "step": 9929 }, { "epoch": 4.499320344358858, "grad_norm": 0.08580423865864864, "learning_rate": 2.419031874017348e-06, "loss": 0.8401, "step": 9930 }, { "epoch": 4.499773448119619, "grad_norm": 0.08675235056788491, "learning_rate": 2.41470009002426e-06, "loss": 0.8408, "step": 9931 }, { "epoch": 4.500226551880381, "grad_norm": 0.09190342070044168, "learning_rate": 2.4103720672776733e-06, "loss": 0.8394, "step": 9932 }, { "epoch": 4.500679655641142, "grad_norm": 0.09781168054690657, "learning_rate": 2.406047806210725e-06, "loss": 0.854, "step": 9933 }, { "epoch": 4.501132759401903, "grad_norm": 0.09246489291756912, "learning_rate": 2.401727307256132e-06, "loss": 0.8439, "step": 9934 }, { "epoch": 4.501585863162664, "grad_norm": 0.08838292615672476, "learning_rate": 2.397410570846277e-06, "loss": 0.8294, "step": 9935 }, { "epoch": 4.5020389669234255, "grad_norm": 0.0890512382100774, "learning_rate": 2.3930975974131345e-06, "loss": 0.8367, "step": 9936 }, { "epoch": 4.502492070684187, "grad_norm": 0.10114227345701965, "learning_rate": 2.3887883873883055e-06, "loss": 0.8501, "step": 9937 }, { "epoch": 4.502945174444948, "grad_norm": 0.09194494080349702, "learning_rate": 2.384482941203041e-06, "loss": 0.8353, "step": 9938 }, { "epoch": 4.503398278205709, "grad_norm": 0.08292626600233007, "learning_rate": 2.380181259288179e-06, "loss": 0.8277, "step": 9939 }, { "epoch": 4.50385138196647, "grad_norm": 0.08767601716108059, "learning_rate": 2.3758833420742144e-06, "loss": 0.8435, "step": 9940 }, { "epoch": 4.504304485727231, "grad_norm": 0.10052521700181656, "learning_rate": 2.3715891899912347e-06, "loss": 0.8604, "step": 9941 }, { "epoch": 4.504757589487992, "grad_norm": 0.09661656122780671, "learning_rate": 2.3672988034689803e-06, "loss": 0.8627, "step": 9942 }, { "epoch": 4.505210693248754, "grad_norm": 0.08334379761952733, "learning_rate": 2.363012182936788e-06, "loss": 0.8431, "step": 9943 }, { "epoch": 4.5056637970095155, "grad_norm": 0.09060746239393978, "learning_rate": 2.3587293288236392e-06, "loss": 0.8526, "step": 9944 }, { "epoch": 4.506116900770277, "grad_norm": 0.09586680654583335, "learning_rate": 2.354450241558124e-06, "loss": 0.8491, "step": 9945 }, { "epoch": 4.506570004531038, "grad_norm": 0.0922330525680685, "learning_rate": 2.3501749215684556e-06, "loss": 0.8559, "step": 9946 }, { "epoch": 4.507023108291799, "grad_norm": 0.09728760992224572, "learning_rate": 2.345903369282483e-06, "loss": 0.8611, "step": 9947 }, { "epoch": 4.50747621205256, "grad_norm": 0.0938749492001255, "learning_rate": 2.3416355851276597e-06, "loss": 0.838, "step": 9948 }, { "epoch": 4.507929315813321, "grad_norm": 0.09075292330180608, "learning_rate": 2.3373715695310883e-06, "loss": 0.8298, "step": 9949 }, { "epoch": 4.508382419574082, "grad_norm": 0.08650848754991602, "learning_rate": 2.3331113229194635e-06, "loss": 0.8295, "step": 9950 }, { "epoch": 4.508835523334843, "grad_norm": 0.10514495374456341, "learning_rate": 2.328854845719133e-06, "loss": 0.8568, "step": 9951 }, { "epoch": 4.5092886270956045, "grad_norm": 0.08323542715888903, "learning_rate": 2.32460213835604e-06, "loss": 0.8333, "step": 9952 }, { "epoch": 4.509741730856366, "grad_norm": 0.09099965559106528, "learning_rate": 2.3203532012557606e-06, "loss": 0.8571, "step": 9953 }, { "epoch": 4.510194834617128, "grad_norm": 0.08621620551190837, "learning_rate": 2.3161080348435096e-06, "loss": 0.834, "step": 9954 }, { "epoch": 4.510647938377889, "grad_norm": 0.08414226157567845, "learning_rate": 2.3118666395440938e-06, "loss": 0.8432, "step": 9955 }, { "epoch": 4.51110104213865, "grad_norm": 0.09159239763684562, "learning_rate": 2.3076290157819737e-06, "loss": 0.8557, "step": 9956 }, { "epoch": 4.511554145899411, "grad_norm": 0.08398597334576796, "learning_rate": 2.3033951639812057e-06, "loss": 0.8459, "step": 9957 }, { "epoch": 4.512007249660172, "grad_norm": 0.08305097704613161, "learning_rate": 2.2991650845654956e-06, "loss": 0.8259, "step": 9958 }, { "epoch": 4.512460353420933, "grad_norm": 0.08826584284009154, "learning_rate": 2.29493877795814e-06, "loss": 0.8441, "step": 9959 }, { "epoch": 4.5129134571816945, "grad_norm": 0.08595884810717835, "learning_rate": 2.2907162445820895e-06, "loss": 0.8598, "step": 9960 }, { "epoch": 4.513366560942456, "grad_norm": 0.08400414151903669, "learning_rate": 2.2864974848598996e-06, "loss": 0.8564, "step": 9961 }, { "epoch": 4.513819664703217, "grad_norm": 0.08592327943232993, "learning_rate": 2.2822824992137393e-06, "loss": 0.8266, "step": 9962 }, { "epoch": 4.514272768463979, "grad_norm": 0.09114174316790927, "learning_rate": 2.2780712880654264e-06, "loss": 0.8523, "step": 9963 }, { "epoch": 4.514725872224739, "grad_norm": 0.08968380238646673, "learning_rate": 2.2738638518363754e-06, "loss": 0.8576, "step": 9964 }, { "epoch": 4.515178975985501, "grad_norm": 0.09265393873889834, "learning_rate": 2.2696601909476446e-06, "loss": 0.8308, "step": 9965 }, { "epoch": 4.515632079746262, "grad_norm": 0.09479773828107209, "learning_rate": 2.2654603058198932e-06, "loss": 0.82, "step": 9966 }, { "epoch": 4.516085183507023, "grad_norm": 0.08721059403977063, "learning_rate": 2.2612641968734294e-06, "loss": 0.8642, "step": 9967 }, { "epoch": 4.516538287267784, "grad_norm": 0.08806937894015977, "learning_rate": 2.257071864528144e-06, "loss": 0.8517, "step": 9968 }, { "epoch": 4.5169913910285455, "grad_norm": 0.09694019130188358, "learning_rate": 2.252883309203595e-06, "loss": 0.8572, "step": 9969 }, { "epoch": 4.517444494789307, "grad_norm": 0.08772898184656863, "learning_rate": 2.2486985313189312e-06, "loss": 0.8462, "step": 9970 }, { "epoch": 4.517897598550068, "grad_norm": 0.09233645442474317, "learning_rate": 2.244517531292929e-06, "loss": 0.8354, "step": 9971 }, { "epoch": 4.518350702310829, "grad_norm": 0.08931797659989946, "learning_rate": 2.240340309544e-06, "loss": 0.8449, "step": 9972 }, { "epoch": 4.51880380607159, "grad_norm": 0.09698643407265181, "learning_rate": 2.236166866490157e-06, "loss": 0.8647, "step": 9973 }, { "epoch": 4.519256909832352, "grad_norm": 0.09122596573190785, "learning_rate": 2.2319972025490567e-06, "loss": 0.8454, "step": 9974 }, { "epoch": 4.519710013593113, "grad_norm": 0.10142020547379893, "learning_rate": 2.2278313181379562e-06, "loss": 0.8491, "step": 9975 }, { "epoch": 4.520163117353874, "grad_norm": 0.08335958290295906, "learning_rate": 2.223669213673758e-06, "loss": 0.862, "step": 9976 }, { "epoch": 4.5206162211146355, "grad_norm": 0.09543520253783509, "learning_rate": 2.2195108895729645e-06, "loss": 0.8597, "step": 9977 }, { "epoch": 4.521069324875397, "grad_norm": 0.09709143049665717, "learning_rate": 2.2153563462517092e-06, "loss": 0.8676, "step": 9978 }, { "epoch": 4.521522428636158, "grad_norm": 0.0993366315133414, "learning_rate": 2.211205584125753e-06, "loss": 0.8362, "step": 9979 }, { "epoch": 4.521975532396919, "grad_norm": 0.08686271891103822, "learning_rate": 2.207058603610466e-06, "loss": 0.8401, "step": 9980 }, { "epoch": 4.52242863615768, "grad_norm": 0.08391481571762645, "learning_rate": 2.2029154051208445e-06, "loss": 0.8316, "step": 9981 }, { "epoch": 4.522881739918441, "grad_norm": 0.09017593563027225, "learning_rate": 2.1987759890715087e-06, "loss": 0.8403, "step": 9982 }, { "epoch": 4.523334843679202, "grad_norm": 0.09310464911198013, "learning_rate": 2.194640355876705e-06, "loss": 0.8395, "step": 9983 }, { "epoch": 4.523787947439963, "grad_norm": 0.09044750437854447, "learning_rate": 2.1905085059502882e-06, "loss": 0.831, "step": 9984 }, { "epoch": 4.524241051200725, "grad_norm": 0.09101088511692859, "learning_rate": 2.1863804397057464e-06, "loss": 0.8572, "step": 9985 }, { "epoch": 4.5246941549614865, "grad_norm": 0.09534281628993405, "learning_rate": 2.1822561575561842e-06, "loss": 0.8403, "step": 9986 }, { "epoch": 4.525147258722248, "grad_norm": 0.08752702388782278, "learning_rate": 2.1781356599143287e-06, "loss": 0.8317, "step": 9987 }, { "epoch": 4.525600362483009, "grad_norm": 0.11352424101932695, "learning_rate": 2.174018947192522e-06, "loss": 0.8627, "step": 9988 }, { "epoch": 4.52605346624377, "grad_norm": 0.10270193705472265, "learning_rate": 2.1699060198027455e-06, "loss": 0.8533, "step": 9989 }, { "epoch": 4.526506570004531, "grad_norm": 0.09565016807204718, "learning_rate": 2.1657968781565766e-06, "loss": 0.8402, "step": 9990 }, { "epoch": 4.526959673765292, "grad_norm": 0.10050367783292453, "learning_rate": 2.161691522665228e-06, "loss": 0.8636, "step": 9991 }, { "epoch": 4.527412777526053, "grad_norm": 0.09865130832893958, "learning_rate": 2.157589953739536e-06, "loss": 0.8366, "step": 9992 }, { "epoch": 4.527865881286814, "grad_norm": 0.0903636563560827, "learning_rate": 2.1534921717899504e-06, "loss": 0.8191, "step": 9993 }, { "epoch": 4.528318985047576, "grad_norm": 0.09332788074362541, "learning_rate": 2.149398177226556e-06, "loss": 0.8239, "step": 9994 }, { "epoch": 4.528772088808337, "grad_norm": 0.0990729029913446, "learning_rate": 2.1453079704590295e-06, "loss": 0.8369, "step": 9995 }, { "epoch": 4.529225192569099, "grad_norm": 0.09462205713339424, "learning_rate": 2.141221551896706e-06, "loss": 0.849, "step": 9996 }, { "epoch": 4.52967829632986, "grad_norm": 0.09407764525534908, "learning_rate": 2.1371389219485163e-06, "loss": 0.8604, "step": 9997 }, { "epoch": 4.530131400090621, "grad_norm": 0.10925153599266992, "learning_rate": 2.1330600810230083e-06, "loss": 0.8532, "step": 9998 }, { "epoch": 4.530584503851382, "grad_norm": 0.09848528087857235, "learning_rate": 2.128985029528372e-06, "loss": 0.826, "step": 9999 }, { "epoch": 4.531037607612143, "grad_norm": 0.09205220697309319, "learning_rate": 2.1249137678724006e-06, "loss": 0.8507, "step": 10000 }, { "epoch": 4.531490711372904, "grad_norm": 0.09479455552363486, "learning_rate": 2.1208462964625244e-06, "loss": 0.85, "step": 10001 }, { "epoch": 4.5319438151336655, "grad_norm": 0.09725051730944774, "learning_rate": 2.1167826157057724e-06, "loss": 0.856, "step": 10002 }, { "epoch": 4.532396918894427, "grad_norm": 0.08914476900571623, "learning_rate": 2.11272272600882e-06, "loss": 0.8484, "step": 10003 }, { "epoch": 4.532850022655188, "grad_norm": 0.08618425861487192, "learning_rate": 2.1086666277779333e-06, "loss": 0.8643, "step": 10004 }, { "epoch": 4.533303126415949, "grad_norm": 0.10226278573744423, "learning_rate": 2.1046143214190318e-06, "loss": 0.8519, "step": 10005 }, { "epoch": 4.53375623017671, "grad_norm": 0.10902951145249155, "learning_rate": 2.100565807337631e-06, "loss": 0.8388, "step": 10006 }, { "epoch": 4.534209333937472, "grad_norm": 0.09658221894389632, "learning_rate": 2.096521085938865e-06, "loss": 0.8265, "step": 10007 }, { "epoch": 4.534662437698233, "grad_norm": 0.09114829924630233, "learning_rate": 2.0924801576275165e-06, "loss": 0.8549, "step": 10008 }, { "epoch": 4.535115541458994, "grad_norm": 0.09148508158835707, "learning_rate": 2.08844302280796e-06, "loss": 0.8572, "step": 10009 }, { "epoch": 4.5355686452197554, "grad_norm": 0.10464876375954171, "learning_rate": 2.084409681884205e-06, "loss": 0.8369, "step": 10010 }, { "epoch": 4.536021748980517, "grad_norm": 0.10026096589855873, "learning_rate": 2.0803801352598672e-06, "loss": 0.841, "step": 10011 }, { "epoch": 4.536474852741278, "grad_norm": 0.09043792598395653, "learning_rate": 2.0763543833382105e-06, "loss": 0.8382, "step": 10012 }, { "epoch": 4.536927956502039, "grad_norm": 0.08857537817203492, "learning_rate": 2.07233242652209e-06, "loss": 0.8611, "step": 10013 }, { "epoch": 4.5373810602628, "grad_norm": 0.09475019394671264, "learning_rate": 2.0683142652139844e-06, "loss": 0.8522, "step": 10014 }, { "epoch": 4.537834164023561, "grad_norm": 0.10800176729241033, "learning_rate": 2.064299899816016e-06, "loss": 0.8526, "step": 10015 }, { "epoch": 4.538287267784322, "grad_norm": 0.09445958252234106, "learning_rate": 2.060289330729899e-06, "loss": 0.8416, "step": 10016 }, { "epoch": 4.538740371545083, "grad_norm": 0.091426978433748, "learning_rate": 2.0562825583569924e-06, "loss": 0.8562, "step": 10017 }, { "epoch": 4.539193475305845, "grad_norm": 0.09580662447493232, "learning_rate": 2.052279583098247e-06, "loss": 0.8545, "step": 10018 }, { "epoch": 4.5396465790666065, "grad_norm": 0.09587256538900377, "learning_rate": 2.048280405354266e-06, "loss": 0.8688, "step": 10019 }, { "epoch": 4.540099682827368, "grad_norm": 0.09307673610448002, "learning_rate": 2.0442850255252454e-06, "loss": 0.841, "step": 10020 }, { "epoch": 4.540552786588129, "grad_norm": 0.0875193888321988, "learning_rate": 2.0402934440110166e-06, "loss": 0.8446, "step": 10021 }, { "epoch": 4.54100589034889, "grad_norm": 0.08638108855690028, "learning_rate": 2.036305661211029e-06, "loss": 0.836, "step": 10022 }, { "epoch": 4.541458994109651, "grad_norm": 0.08289244824397236, "learning_rate": 2.0323216775243403e-06, "loss": 0.8493, "step": 10023 }, { "epoch": 4.541912097870412, "grad_norm": 0.08520379786900269, "learning_rate": 2.0283414933496414e-06, "loss": 0.8605, "step": 10024 }, { "epoch": 4.542365201631173, "grad_norm": 0.08797863134440409, "learning_rate": 2.02436510908524e-06, "loss": 0.8591, "step": 10025 }, { "epoch": 4.542818305391934, "grad_norm": 0.08901088615468562, "learning_rate": 2.0203925251290625e-06, "loss": 0.8457, "step": 10026 }, { "epoch": 4.5432714091526964, "grad_norm": 0.09041740813408923, "learning_rate": 2.016423741878648e-06, "loss": 0.8433, "step": 10027 }, { "epoch": 4.543724512913457, "grad_norm": 0.10047573315217093, "learning_rate": 2.012458759731173e-06, "loss": 0.8595, "step": 10028 }, { "epoch": 4.544177616674219, "grad_norm": 0.08412814770982448, "learning_rate": 2.0084975790834083e-06, "loss": 0.8446, "step": 10029 }, { "epoch": 4.54463072043498, "grad_norm": 0.09318376364348514, "learning_rate": 2.0045402003317703e-06, "loss": 0.8297, "step": 10030 }, { "epoch": 4.545083824195741, "grad_norm": 0.08942585819225593, "learning_rate": 2.0005866238722803e-06, "loss": 0.8378, "step": 10031 }, { "epoch": 4.545536927956502, "grad_norm": 0.09996392991016138, "learning_rate": 1.996636850100573e-06, "loss": 0.8421, "step": 10032 }, { "epoch": 4.545990031717263, "grad_norm": 0.09418639425615197, "learning_rate": 1.9926908794119228e-06, "loss": 0.8557, "step": 10033 }, { "epoch": 4.546443135478024, "grad_norm": 0.09212032469330456, "learning_rate": 1.988748712201205e-06, "loss": 0.8444, "step": 10034 }, { "epoch": 4.5468962392387855, "grad_norm": 0.0949649751582218, "learning_rate": 1.9848103488629223e-06, "loss": 0.8446, "step": 10035 }, { "epoch": 4.547349342999547, "grad_norm": 0.1045481734844443, "learning_rate": 1.9808757897911945e-06, "loss": 0.8417, "step": 10036 }, { "epoch": 4.547802446760308, "grad_norm": 0.09055450517120134, "learning_rate": 1.9769450353797783e-06, "loss": 0.8563, "step": 10037 }, { "epoch": 4.54825555052107, "grad_norm": 0.09286555656316552, "learning_rate": 1.9730180860220027e-06, "loss": 0.8497, "step": 10038 }, { "epoch": 4.548708654281831, "grad_norm": 0.09713230504779684, "learning_rate": 1.9690949421108694e-06, "loss": 0.8601, "step": 10039 }, { "epoch": 4.549161758042592, "grad_norm": 0.0964575838005756, "learning_rate": 1.9651756040389625e-06, "loss": 0.8356, "step": 10040 }, { "epoch": 4.549614861803353, "grad_norm": 0.09042735816391088, "learning_rate": 1.9612600721985142e-06, "loss": 0.8478, "step": 10041 }, { "epoch": 4.550067965564114, "grad_norm": 0.09375474992358648, "learning_rate": 1.957348346981354e-06, "loss": 0.8657, "step": 10042 }, { "epoch": 4.550521069324875, "grad_norm": 0.08940278213108381, "learning_rate": 1.9534404287789276e-06, "loss": 0.8427, "step": 10043 }, { "epoch": 4.550974173085637, "grad_norm": 0.10126944639300269, "learning_rate": 1.9495363179823236e-06, "loss": 0.8487, "step": 10044 }, { "epoch": 4.551427276846398, "grad_norm": 0.09376985540614019, "learning_rate": 1.945636014982224e-06, "loss": 0.8613, "step": 10045 }, { "epoch": 4.551880380607159, "grad_norm": 0.09479901659270268, "learning_rate": 1.9417395201689528e-06, "loss": 0.8339, "step": 10046 }, { "epoch": 4.55233348436792, "grad_norm": 0.0844959636472176, "learning_rate": 1.9378468339324285e-06, "loss": 0.8384, "step": 10047 }, { "epoch": 4.552786588128681, "grad_norm": 0.08503658311092531, "learning_rate": 1.933957956662211e-06, "loss": 0.844, "step": 10048 }, { "epoch": 4.553239691889443, "grad_norm": 0.0978593439412502, "learning_rate": 1.9300728887474607e-06, "loss": 0.8595, "step": 10049 }, { "epoch": 4.553692795650204, "grad_norm": 0.1078305352839146, "learning_rate": 1.9261916305769724e-06, "loss": 0.8575, "step": 10050 }, { "epoch": 4.554145899410965, "grad_norm": 0.08211117368854674, "learning_rate": 1.922314182539151e-06, "loss": 0.8383, "step": 10051 }, { "epoch": 4.5545990031717265, "grad_norm": 0.08927532083232038, "learning_rate": 1.918440545022016e-06, "loss": 0.8575, "step": 10052 }, { "epoch": 4.555052106932488, "grad_norm": 0.09635565078493484, "learning_rate": 1.91457071841322e-06, "loss": 0.8365, "step": 10053 }, { "epoch": 4.555505210693249, "grad_norm": 0.09533193126495675, "learning_rate": 1.9107047031000147e-06, "loss": 0.8545, "step": 10054 }, { "epoch": 4.55595831445401, "grad_norm": 0.09871634235208128, "learning_rate": 1.9068424994692946e-06, "loss": 0.8438, "step": 10055 }, { "epoch": 4.556411418214771, "grad_norm": 0.0956604074068091, "learning_rate": 1.9029841079075461e-06, "loss": 0.8375, "step": 10056 }, { "epoch": 4.556864521975532, "grad_norm": 0.09732157069880712, "learning_rate": 1.899129528800896e-06, "loss": 0.8243, "step": 10057 }, { "epoch": 4.557317625736293, "grad_norm": 0.10386026946647998, "learning_rate": 1.89527876253508e-06, "loss": 0.8348, "step": 10058 }, { "epoch": 4.557770729497054, "grad_norm": 0.08978864683149669, "learning_rate": 1.8914318094954432e-06, "loss": 0.848, "step": 10059 }, { "epoch": 4.558223833257816, "grad_norm": 0.10550048117850774, "learning_rate": 1.887588670066971e-06, "loss": 0.8476, "step": 10060 }, { "epoch": 4.558676937018578, "grad_norm": 0.09547808129208166, "learning_rate": 1.8837493446342448e-06, "loss": 0.8309, "step": 10061 }, { "epoch": 4.559130040779339, "grad_norm": 0.08622570795254633, "learning_rate": 1.8799138335814859e-06, "loss": 0.8596, "step": 10062 }, { "epoch": 4.5595831445401, "grad_norm": 0.10183038700064352, "learning_rate": 1.8760821372925165e-06, "loss": 0.8511, "step": 10063 }, { "epoch": 4.560036248300861, "grad_norm": 0.10500417465222761, "learning_rate": 1.8722542561507851e-06, "loss": 0.8625, "step": 10064 }, { "epoch": 4.560489352061622, "grad_norm": 0.09664568121769272, "learning_rate": 1.8684301905393497e-06, "loss": 0.8342, "step": 10065 }, { "epoch": 4.560942455822383, "grad_norm": 0.09125196864219375, "learning_rate": 1.8646099408409046e-06, "loss": 0.8461, "step": 10066 }, { "epoch": 4.561395559583144, "grad_norm": 0.08944407681309954, "learning_rate": 1.8607935074377438e-06, "loss": 0.8456, "step": 10067 }, { "epoch": 4.5618486633439055, "grad_norm": 0.10847952939842573, "learning_rate": 1.8569808907117793e-06, "loss": 0.8437, "step": 10068 }, { "epoch": 4.562301767104667, "grad_norm": 0.09694775100365498, "learning_rate": 1.8531720910445683e-06, "loss": 0.8243, "step": 10069 }, { "epoch": 4.562754870865428, "grad_norm": 0.0918211656334645, "learning_rate": 1.8493671088172461e-06, "loss": 0.8524, "step": 10070 }, { "epoch": 4.56320797462619, "grad_norm": 0.10464255219249403, "learning_rate": 1.8455659444105968e-06, "loss": 0.851, "step": 10071 }, { "epoch": 4.563661078386951, "grad_norm": 0.10148895121274935, "learning_rate": 1.8417685982050094e-06, "loss": 0.8523, "step": 10072 }, { "epoch": 4.564114182147712, "grad_norm": 0.09261872735690307, "learning_rate": 1.8379750705804956e-06, "loss": 0.8252, "step": 10073 }, { "epoch": 4.564567285908473, "grad_norm": 0.10966806122738638, "learning_rate": 1.834185361916676e-06, "loss": 0.8285, "step": 10074 }, { "epoch": 4.565020389669234, "grad_norm": 0.10526421053076379, "learning_rate": 1.8303994725927942e-06, "loss": 0.8326, "step": 10075 }, { "epoch": 4.565473493429995, "grad_norm": 0.09221769967038786, "learning_rate": 1.8266174029877247e-06, "loss": 0.834, "step": 10076 }, { "epoch": 4.5659265971907566, "grad_norm": 0.08964753269388158, "learning_rate": 1.8228391534799384e-06, "loss": 0.836, "step": 10077 }, { "epoch": 4.566379700951518, "grad_norm": 0.09862712351469809, "learning_rate": 1.8190647244475324e-06, "loss": 0.8412, "step": 10078 }, { "epoch": 4.566832804712279, "grad_norm": 0.082346970770469, "learning_rate": 1.815294116268227e-06, "loss": 0.824, "step": 10079 }, { "epoch": 4.56728590847304, "grad_norm": 0.08701639450369623, "learning_rate": 1.8115273293193558e-06, "loss": 0.8394, "step": 10080 }, { "epoch": 4.567739012233801, "grad_norm": 0.09399881067646844, "learning_rate": 1.807764363977862e-06, "loss": 0.8256, "step": 10081 }, { "epoch": 4.568192115994563, "grad_norm": 0.08335258730961395, "learning_rate": 1.8040052206203285e-06, "loss": 0.8368, "step": 10082 }, { "epoch": 4.568645219755324, "grad_norm": 0.08251354682701136, "learning_rate": 1.80024989962293e-06, "loss": 0.8359, "step": 10083 }, { "epoch": 4.569098323516085, "grad_norm": 0.09053577513669328, "learning_rate": 1.7964984013614684e-06, "loss": 0.8442, "step": 10084 }, { "epoch": 4.5695514272768465, "grad_norm": 0.09153548233925879, "learning_rate": 1.792750726211372e-06, "loss": 0.8507, "step": 10085 }, { "epoch": 4.570004531037608, "grad_norm": 0.08972692822091036, "learning_rate": 1.7890068745476696e-06, "loss": 0.8572, "step": 10086 }, { "epoch": 4.570457634798369, "grad_norm": 0.09085316142188755, "learning_rate": 1.7852668467450352e-06, "loss": 0.8402, "step": 10087 }, { "epoch": 4.57091073855913, "grad_norm": 0.09261654481784634, "learning_rate": 1.7815306431777201e-06, "loss": 0.8233, "step": 10088 }, { "epoch": 4.571363842319891, "grad_norm": 0.09873843697583506, "learning_rate": 1.7777982642196301e-06, "loss": 0.8514, "step": 10089 }, { "epoch": 4.571816946080652, "grad_norm": 0.07848575683907844, "learning_rate": 1.7740697102442662e-06, "loss": 0.8334, "step": 10090 }, { "epoch": 4.572270049841414, "grad_norm": 0.08446086919059226, "learning_rate": 1.7703449816247565e-06, "loss": 0.8467, "step": 10091 }, { "epoch": 4.572723153602174, "grad_norm": 0.09068243994039435, "learning_rate": 1.766624078733843e-06, "loss": 0.826, "step": 10092 }, { "epoch": 4.573176257362936, "grad_norm": 0.09042132114294107, "learning_rate": 1.7629070019438765e-06, "loss": 0.8475, "step": 10093 }, { "epoch": 4.573629361123698, "grad_norm": 0.08464889723846124, "learning_rate": 1.7591937516268488e-06, "loss": 0.849, "step": 10094 }, { "epoch": 4.574082464884459, "grad_norm": 0.08678354915901038, "learning_rate": 1.7554843281543377e-06, "loss": 0.8443, "step": 10095 }, { "epoch": 4.57453556864522, "grad_norm": 0.09323224759029287, "learning_rate": 1.751778731897562e-06, "loss": 0.8595, "step": 10096 }, { "epoch": 4.574988672405981, "grad_norm": 0.09761563750513139, "learning_rate": 1.7480769632273454e-06, "loss": 0.8446, "step": 10097 }, { "epoch": 4.575441776166742, "grad_norm": 0.09603553459489256, "learning_rate": 1.7443790225141377e-06, "loss": 0.8435, "step": 10098 }, { "epoch": 4.575894879927503, "grad_norm": 0.08928432566573323, "learning_rate": 1.7406849101279943e-06, "loss": 0.8391, "step": 10099 }, { "epoch": 4.576347983688264, "grad_norm": 0.08554498623010122, "learning_rate": 1.7369946264385973e-06, "loss": 0.8265, "step": 10100 }, { "epoch": 4.5768010874490255, "grad_norm": 0.09510985012165142, "learning_rate": 1.7333081718152377e-06, "loss": 0.844, "step": 10101 }, { "epoch": 4.5772541912097875, "grad_norm": 0.0947448500621973, "learning_rate": 1.7296255466268385e-06, "loss": 0.8338, "step": 10102 }, { "epoch": 4.577707294970549, "grad_norm": 0.08575289707432636, "learning_rate": 1.725946751241918e-06, "loss": 0.844, "step": 10103 }, { "epoch": 4.57816039873131, "grad_norm": 0.09381441972796635, "learning_rate": 1.7222717860286176e-06, "loss": 0.8494, "step": 10104 }, { "epoch": 4.578613502492071, "grad_norm": 0.10379860383260146, "learning_rate": 1.7186006513547093e-06, "loss": 0.8386, "step": 10105 }, { "epoch": 4.579066606252832, "grad_norm": 0.08838284155973777, "learning_rate": 1.7149333475875663e-06, "loss": 0.8509, "step": 10106 }, { "epoch": 4.579519710013593, "grad_norm": 0.09104566660986015, "learning_rate": 1.7112698750941925e-06, "loss": 0.8518, "step": 10107 }, { "epoch": 4.579972813774354, "grad_norm": 0.08819014876373835, "learning_rate": 1.7076102342411838e-06, "loss": 0.8378, "step": 10108 }, { "epoch": 4.580425917535115, "grad_norm": 0.09104508649425781, "learning_rate": 1.7039544253947848e-06, "loss": 0.8328, "step": 10109 }, { "epoch": 4.5808790212958765, "grad_norm": 0.09112251757227724, "learning_rate": 1.7003024489208364e-06, "loss": 0.8384, "step": 10110 }, { "epoch": 4.581332125056638, "grad_norm": 0.09203754679471088, "learning_rate": 1.6966543051847929e-06, "loss": 0.8296, "step": 10111 }, { "epoch": 4.581785228817399, "grad_norm": 0.09074753073553336, "learning_rate": 1.693009994551744e-06, "loss": 0.8475, "step": 10112 }, { "epoch": 4.582238332578161, "grad_norm": 0.08902408996379203, "learning_rate": 1.6893695173863679e-06, "loss": 0.8619, "step": 10113 }, { "epoch": 4.582691436338922, "grad_norm": 0.08704144693521776, "learning_rate": 1.6857328740529944e-06, "loss": 0.845, "step": 10114 }, { "epoch": 4.583144540099683, "grad_norm": 0.09316605485616822, "learning_rate": 1.6821000649155327e-06, "loss": 0.8654, "step": 10115 }, { "epoch": 4.583597643860444, "grad_norm": 0.0865355895612808, "learning_rate": 1.6784710903375457e-06, "loss": 0.8576, "step": 10116 }, { "epoch": 4.584050747621205, "grad_norm": 0.08657913710363982, "learning_rate": 1.6748459506821735e-06, "loss": 0.8468, "step": 10117 }, { "epoch": 4.5845038513819665, "grad_norm": 0.09086518826036441, "learning_rate": 1.6712246463122062e-06, "loss": 0.8634, "step": 10118 }, { "epoch": 4.584956955142728, "grad_norm": 0.08516853417041133, "learning_rate": 1.6676071775900294e-06, "loss": 0.8492, "step": 10119 }, { "epoch": 4.585410058903489, "grad_norm": 0.08721722727577248, "learning_rate": 1.6639935448776512e-06, "loss": 0.8415, "step": 10120 }, { "epoch": 4.58586316266425, "grad_norm": 0.09486072087644946, "learning_rate": 1.6603837485367025e-06, "loss": 0.8512, "step": 10121 }, { "epoch": 4.586316266425011, "grad_norm": 0.08869153211218089, "learning_rate": 1.6567777889284142e-06, "loss": 0.8146, "step": 10122 }, { "epoch": 4.586769370185772, "grad_norm": 0.08927408754230133, "learning_rate": 1.653175666413649e-06, "loss": 0.8293, "step": 10123 }, { "epoch": 4.587222473946534, "grad_norm": 0.09237906994378293, "learning_rate": 1.6495773813528782e-06, "loss": 0.8266, "step": 10124 }, { "epoch": 4.587675577707295, "grad_norm": 0.09828422895947984, "learning_rate": 1.6459829341061916e-06, "loss": 0.8648, "step": 10125 }, { "epoch": 4.588128681468056, "grad_norm": 0.09258955531918067, "learning_rate": 1.6423923250332885e-06, "loss": 0.8249, "step": 10126 }, { "epoch": 4.5885817852288175, "grad_norm": 0.08986989604953809, "learning_rate": 1.6388055544934989e-06, "loss": 0.855, "step": 10127 }, { "epoch": 4.589034888989579, "grad_norm": 0.0887022999843824, "learning_rate": 1.6352226228457534e-06, "loss": 0.8471, "step": 10128 }, { "epoch": 4.58948799275034, "grad_norm": 0.0929313396914355, "learning_rate": 1.631643530448601e-06, "loss": 0.8399, "step": 10129 }, { "epoch": 4.589941096511101, "grad_norm": 0.08890443051218623, "learning_rate": 1.6280682776602174e-06, "loss": 0.8457, "step": 10130 }, { "epoch": 4.590394200271862, "grad_norm": 0.09880628034294012, "learning_rate": 1.624496864838374e-06, "loss": 0.8627, "step": 10131 }, { "epoch": 4.590847304032623, "grad_norm": 0.08444311175707367, "learning_rate": 1.6209292923404873e-06, "loss": 0.8467, "step": 10132 }, { "epoch": 4.591300407793384, "grad_norm": 0.08356038924980612, "learning_rate": 1.6173655605235561e-06, "loss": 0.8442, "step": 10133 }, { "epoch": 4.5917535115541455, "grad_norm": 0.10014504801551065, "learning_rate": 1.613805669744224e-06, "loss": 0.8517, "step": 10134 }, { "epoch": 4.5922066153149075, "grad_norm": 0.09186261914576747, "learning_rate": 1.6102496203587304e-06, "loss": 0.8488, "step": 10135 }, { "epoch": 4.592659719075669, "grad_norm": 0.08902972757506224, "learning_rate": 1.6066974127229328e-06, "loss": 0.8421, "step": 10136 }, { "epoch": 4.59311282283643, "grad_norm": 0.07909309765944918, "learning_rate": 1.6031490471923206e-06, "loss": 0.8457, "step": 10137 }, { "epoch": 4.593565926597191, "grad_norm": 0.0847711007283463, "learning_rate": 1.5996045241219738e-06, "loss": 0.8436, "step": 10138 }, { "epoch": 4.594019030357952, "grad_norm": 0.09073323131196302, "learning_rate": 1.5960638438666133e-06, "loss": 0.8541, "step": 10139 }, { "epoch": 4.594472134118713, "grad_norm": 0.08651041152494286, "learning_rate": 1.5925270067805554e-06, "loss": 0.8222, "step": 10140 }, { "epoch": 4.594925237879474, "grad_norm": 0.08590403532026429, "learning_rate": 1.588994013217744e-06, "loss": 0.8499, "step": 10141 }, { "epoch": 4.595378341640235, "grad_norm": 0.0928409296917738, "learning_rate": 1.585464863531727e-06, "loss": 0.8437, "step": 10142 }, { "epoch": 4.5958314454009965, "grad_norm": 0.079341400865888, "learning_rate": 1.5819395580756803e-06, "loss": 0.8509, "step": 10143 }, { "epoch": 4.596284549161758, "grad_norm": 0.08659948524076264, "learning_rate": 1.578418097202392e-06, "loss": 0.842, "step": 10144 }, { "epoch": 4.596737652922519, "grad_norm": 0.10041918680223896, "learning_rate": 1.5749004812642521e-06, "loss": 0.836, "step": 10145 }, { "epoch": 4.597190756683281, "grad_norm": 0.09474382922275165, "learning_rate": 1.5713867106132853e-06, "loss": 0.8357, "step": 10146 }, { "epoch": 4.597643860444042, "grad_norm": 0.08756537632157962, "learning_rate": 1.5678767856011169e-06, "loss": 0.841, "step": 10147 }, { "epoch": 4.598096964204803, "grad_norm": 0.08934098366203005, "learning_rate": 1.5643707065789992e-06, "loss": 0.8322, "step": 10148 }, { "epoch": 4.598550067965564, "grad_norm": 0.09167996750019526, "learning_rate": 1.5608684738977897e-06, "loss": 0.8487, "step": 10149 }, { "epoch": 4.599003171726325, "grad_norm": 0.08659998473360299, "learning_rate": 1.5573700879079678e-06, "loss": 0.8443, "step": 10150 }, { "epoch": 4.5994562754870865, "grad_norm": 0.09262107105089765, "learning_rate": 1.5538755489596224e-06, "loss": 0.8587, "step": 10151 }, { "epoch": 4.599909379247848, "grad_norm": 0.08469798017428971, "learning_rate": 1.5503848574024605e-06, "loss": 0.8366, "step": 10152 }, { "epoch": 4.600362483008609, "grad_norm": 0.08449900613238084, "learning_rate": 1.546898013585798e-06, "loss": 0.8587, "step": 10153 }, { "epoch": 4.60081558676937, "grad_norm": 0.07860329751378413, "learning_rate": 1.5434150178585827e-06, "loss": 0.8425, "step": 10154 }, { "epoch": 4.601268690530132, "grad_norm": 0.08503796070463514, "learning_rate": 1.539935870569358e-06, "loss": 0.8352, "step": 10155 }, { "epoch": 4.601721794290892, "grad_norm": 0.08994418604831465, "learning_rate": 1.5364605720662896e-06, "loss": 0.8513, "step": 10156 }, { "epoch": 4.602174898051654, "grad_norm": 0.08676715241433464, "learning_rate": 1.5329891226971616e-06, "loss": 0.8263, "step": 10157 }, { "epoch": 4.602628001812415, "grad_norm": 0.08352491027599587, "learning_rate": 1.5295215228093674e-06, "loss": 0.8507, "step": 10158 }, { "epoch": 4.603081105573176, "grad_norm": 0.08654336662857173, "learning_rate": 1.5260577727499225e-06, "loss": 0.8535, "step": 10159 }, { "epoch": 4.6035342093339375, "grad_norm": 0.08739690798709218, "learning_rate": 1.5225978728654478e-06, "loss": 0.8465, "step": 10160 }, { "epoch": 4.603987313094699, "grad_norm": 0.09003070262580162, "learning_rate": 1.519141823502186e-06, "loss": 0.8307, "step": 10161 }, { "epoch": 4.60444041685546, "grad_norm": 0.08849448283837097, "learning_rate": 1.5156896250059893e-06, "loss": 0.849, "step": 10162 }, { "epoch": 4.604893520616221, "grad_norm": 0.0779556107179329, "learning_rate": 1.5122412777223283e-06, "loss": 0.8578, "step": 10163 }, { "epoch": 4.605346624376982, "grad_norm": 0.0861637800908185, "learning_rate": 1.5087967819962913e-06, "loss": 0.8419, "step": 10164 }, { "epoch": 4.605799728137743, "grad_norm": 0.0828378915547317, "learning_rate": 1.5053561381725624e-06, "loss": 0.8381, "step": 10165 }, { "epoch": 4.606252831898505, "grad_norm": 0.09528790424313334, "learning_rate": 1.5019193465954706e-06, "loss": 0.846, "step": 10166 }, { "epoch": 4.606705935659266, "grad_norm": 0.09472407368896009, "learning_rate": 1.4984864076089367e-06, "loss": 0.8319, "step": 10167 }, { "epoch": 4.6071590394200275, "grad_norm": 0.08610005616866483, "learning_rate": 1.4950573215565034e-06, "loss": 0.8353, "step": 10168 }, { "epoch": 4.607612143180789, "grad_norm": 0.0877800218978183, "learning_rate": 1.4916320887813273e-06, "loss": 0.8507, "step": 10169 }, { "epoch": 4.60806524694155, "grad_norm": 0.09949467419271803, "learning_rate": 1.4882107096261833e-06, "loss": 0.8592, "step": 10170 }, { "epoch": 4.608518350702311, "grad_norm": 0.09085272923386309, "learning_rate": 1.4847931844334507e-06, "loss": 0.8398, "step": 10171 }, { "epoch": 4.608971454463072, "grad_norm": 0.0823261544858658, "learning_rate": 1.4813795135451225e-06, "loss": 0.842, "step": 10172 }, { "epoch": 4.609424558223833, "grad_norm": 0.08094960104772667, "learning_rate": 1.4779696973028324e-06, "loss": 0.8554, "step": 10173 }, { "epoch": 4.609877661984594, "grad_norm": 0.08912128702329995, "learning_rate": 1.474563736047787e-06, "loss": 0.8662, "step": 10174 }, { "epoch": 4.610330765745355, "grad_norm": 0.0850822513727664, "learning_rate": 1.4711616301208476e-06, "loss": 0.8433, "step": 10175 }, { "epoch": 4.6107838695061165, "grad_norm": 0.08112116943054545, "learning_rate": 1.4677633798624525e-06, "loss": 0.8636, "step": 10176 }, { "epoch": 4.6112369732668785, "grad_norm": 0.08207288804784658, "learning_rate": 1.4643689856126898e-06, "loss": 0.8458, "step": 10177 }, { "epoch": 4.61169007702764, "grad_norm": 0.08370571182861276, "learning_rate": 1.4609784477112298e-06, "loss": 0.8351, "step": 10178 }, { "epoch": 4.612143180788401, "grad_norm": 0.0870837722233374, "learning_rate": 1.4575917664973792e-06, "loss": 0.8323, "step": 10179 }, { "epoch": 4.612596284549162, "grad_norm": 0.08664015288597086, "learning_rate": 1.4542089423100537e-06, "loss": 0.8362, "step": 10180 }, { "epoch": 4.613049388309923, "grad_norm": 0.08216963542096177, "learning_rate": 1.4508299754877685e-06, "loss": 0.8313, "step": 10181 }, { "epoch": 4.613502492070684, "grad_norm": 0.08220738556215952, "learning_rate": 1.447454866368676e-06, "loss": 0.8382, "step": 10182 }, { "epoch": 4.613955595831445, "grad_norm": 0.08668157347644101, "learning_rate": 1.4440836152905234e-06, "loss": 0.8477, "step": 10183 }, { "epoch": 4.6144086995922065, "grad_norm": 0.08523362005290994, "learning_rate": 1.4407162225906856e-06, "loss": 0.8385, "step": 10184 }, { "epoch": 4.614861803352968, "grad_norm": 0.08249250501652981, "learning_rate": 1.4373526886061418e-06, "loss": 0.8372, "step": 10185 }, { "epoch": 4.615314907113729, "grad_norm": 0.08568729203687722, "learning_rate": 1.4339930136734891e-06, "loss": 0.8566, "step": 10186 }, { "epoch": 4.61576801087449, "grad_norm": 0.08684764948043357, "learning_rate": 1.4306371981289347e-06, "loss": 0.839, "step": 10187 }, { "epoch": 4.616221114635252, "grad_norm": 0.08226155247443502, "learning_rate": 1.4272852423083116e-06, "loss": 0.8751, "step": 10188 }, { "epoch": 4.616674218396013, "grad_norm": 0.07915843559932985, "learning_rate": 1.4239371465470498e-06, "loss": 0.8586, "step": 10189 }, { "epoch": 4.617127322156774, "grad_norm": 0.09154861269763061, "learning_rate": 1.4205929111802008e-06, "loss": 0.8462, "step": 10190 }, { "epoch": 4.617580425917535, "grad_norm": 0.08854594219889735, "learning_rate": 1.417252536542435e-06, "loss": 0.841, "step": 10191 }, { "epoch": 4.618033529678296, "grad_norm": 0.08685499181344597, "learning_rate": 1.4139160229680272e-06, "loss": 0.8281, "step": 10192 }, { "epoch": 4.6184866334390575, "grad_norm": 0.08563476827107377, "learning_rate": 1.4105833707908701e-06, "loss": 0.8719, "step": 10193 }, { "epoch": 4.618939737199819, "grad_norm": 0.09171348651551234, "learning_rate": 1.4072545803444703e-06, "loss": 0.8382, "step": 10194 }, { "epoch": 4.61939284096058, "grad_norm": 0.09052707795833465, "learning_rate": 1.4039296519619483e-06, "loss": 0.8552, "step": 10195 }, { "epoch": 4.619845944721341, "grad_norm": 0.08532115562012844, "learning_rate": 1.400608585976042e-06, "loss": 0.8148, "step": 10196 }, { "epoch": 4.620299048482102, "grad_norm": 0.08811318349788214, "learning_rate": 1.3972913827190859e-06, "loss": 0.8501, "step": 10197 }, { "epoch": 4.620752152242863, "grad_norm": 0.08660680878001695, "learning_rate": 1.3939780425230497e-06, "loss": 0.8443, "step": 10198 }, { "epoch": 4.621205256003625, "grad_norm": 0.08729472215854911, "learning_rate": 1.3906685657194995e-06, "loss": 0.8608, "step": 10199 }, { "epoch": 4.621658359764386, "grad_norm": 0.2530695656791465, "learning_rate": 1.3873629526396326e-06, "loss": 0.8652, "step": 10200 }, { "epoch": 4.6221114635251475, "grad_norm": 0.09909733801970995, "learning_rate": 1.384061203614242e-06, "loss": 0.8645, "step": 10201 }, { "epoch": 4.622564567285909, "grad_norm": 0.08215329231369659, "learning_rate": 1.3807633189737435e-06, "loss": 0.843, "step": 10202 }, { "epoch": 4.62301767104667, "grad_norm": 0.08590557925090037, "learning_rate": 1.3774692990481575e-06, "loss": 0.8599, "step": 10203 }, { "epoch": 4.623470774807431, "grad_norm": 0.09335759705256967, "learning_rate": 1.374179144167136e-06, "loss": 0.822, "step": 10204 }, { "epoch": 4.623923878568192, "grad_norm": 0.10529282589683445, "learning_rate": 1.3708928546599265e-06, "loss": 0.8477, "step": 10205 }, { "epoch": 4.624376982328953, "grad_norm": 0.08229900078663875, "learning_rate": 1.3676104308553906e-06, "loss": 0.824, "step": 10206 }, { "epoch": 4.624830086089714, "grad_norm": 0.0813627899807839, "learning_rate": 1.3643318730820166e-06, "loss": 0.8244, "step": 10207 }, { "epoch": 4.625283189850475, "grad_norm": 0.08759881049099004, "learning_rate": 1.361057181667893e-06, "loss": 0.8454, "step": 10208 }, { "epoch": 4.6257362936112365, "grad_norm": 0.09390169107766026, "learning_rate": 1.3577863569407269e-06, "loss": 0.8515, "step": 10209 }, { "epoch": 4.6261893973719985, "grad_norm": 0.09338814882249696, "learning_rate": 1.3545193992278293e-06, "loss": 0.8457, "step": 10210 }, { "epoch": 4.62664250113276, "grad_norm": 0.09011496922636687, "learning_rate": 1.3512563088561436e-06, "loss": 0.8471, "step": 10211 }, { "epoch": 4.627095604893521, "grad_norm": 0.09220492417927525, "learning_rate": 1.347997086152204e-06, "loss": 0.8387, "step": 10212 }, { "epoch": 4.627548708654282, "grad_norm": 0.08456809018185853, "learning_rate": 1.3447417314421807e-06, "loss": 0.8423, "step": 10213 }, { "epoch": 4.628001812415043, "grad_norm": 0.0805544257298872, "learning_rate": 1.3414902450518353e-06, "loss": 0.8413, "step": 10214 }, { "epoch": 4.628454916175804, "grad_norm": 0.08735214084356911, "learning_rate": 1.338242627306552e-06, "loss": 0.839, "step": 10215 }, { "epoch": 4.628908019936565, "grad_norm": 0.08701413718962071, "learning_rate": 1.3349988785313328e-06, "loss": 0.8381, "step": 10216 }, { "epoch": 4.629361123697326, "grad_norm": 0.08560676701147255, "learning_rate": 1.3317589990507806e-06, "loss": 0.8505, "step": 10217 }, { "epoch": 4.629814227458088, "grad_norm": 0.0782170972761906, "learning_rate": 1.328522989189125e-06, "loss": 0.8503, "step": 10218 }, { "epoch": 4.63026733121885, "grad_norm": 0.08410237325325386, "learning_rate": 1.3252908492701866e-06, "loss": 0.8467, "step": 10219 }, { "epoch": 4.63072043497961, "grad_norm": 0.0820292263353073, "learning_rate": 1.3220625796174314e-06, "loss": 0.8513, "step": 10220 }, { "epoch": 4.631173538740372, "grad_norm": 0.09080880784412608, "learning_rate": 1.3188381805539074e-06, "loss": 0.8349, "step": 10221 }, { "epoch": 4.631626642501133, "grad_norm": 0.08729049295861696, "learning_rate": 1.3156176524022945e-06, "loss": 0.8613, "step": 10222 }, { "epoch": 4.632079746261894, "grad_norm": 0.12544587873285115, "learning_rate": 1.3124009954848682e-06, "loss": 0.8461, "step": 10223 }, { "epoch": 4.632532850022655, "grad_norm": 0.08912367840003622, "learning_rate": 1.3091882101235399e-06, "loss": 0.8417, "step": 10224 }, { "epoch": 4.632985953783416, "grad_norm": 0.09274047503559973, "learning_rate": 1.3059792966398122e-06, "loss": 0.8559, "step": 10225 }, { "epoch": 4.6334390575441775, "grad_norm": 0.08380551986195722, "learning_rate": 1.3027742553548062e-06, "loss": 0.8312, "step": 10226 }, { "epoch": 4.633892161304939, "grad_norm": 0.10203691357375252, "learning_rate": 1.2995730865892654e-06, "loss": 0.8345, "step": 10227 }, { "epoch": 4.6343452650657, "grad_norm": 0.0918005381694916, "learning_rate": 1.2963757906635333e-06, "loss": 0.8524, "step": 10228 }, { "epoch": 4.634798368826461, "grad_norm": 0.110703224733389, "learning_rate": 1.2931823678975719e-06, "loss": 0.8564, "step": 10229 }, { "epoch": 4.635251472587223, "grad_norm": 0.08509361259866878, "learning_rate": 1.2899928186109523e-06, "loss": 0.8485, "step": 10230 }, { "epoch": 4.635704576347984, "grad_norm": 0.08157822335103838, "learning_rate": 1.2868071431228635e-06, "loss": 0.8336, "step": 10231 }, { "epoch": 4.636157680108745, "grad_norm": 0.09073999323919243, "learning_rate": 1.2836253417520994e-06, "loss": 0.852, "step": 10232 }, { "epoch": 4.636610783869506, "grad_norm": 0.08739978340598423, "learning_rate": 1.280447414817072e-06, "loss": 0.8508, "step": 10233 }, { "epoch": 4.6370638876302674, "grad_norm": 0.08289697502037159, "learning_rate": 1.2772733626358068e-06, "loss": 0.849, "step": 10234 }, { "epoch": 4.637516991391029, "grad_norm": 0.0823114564237036, "learning_rate": 1.27410318552593e-06, "loss": 0.8488, "step": 10235 }, { "epoch": 4.63797009515179, "grad_norm": 0.07988407200382301, "learning_rate": 1.2709368838046943e-06, "loss": 0.8611, "step": 10236 }, { "epoch": 4.638423198912551, "grad_norm": 0.07834213635420234, "learning_rate": 1.2677744577889573e-06, "loss": 0.849, "step": 10237 }, { "epoch": 4.638876302673312, "grad_norm": 0.08424571110948852, "learning_rate": 1.2646159077951991e-06, "loss": 0.8481, "step": 10238 }, { "epoch": 4.639329406434073, "grad_norm": 0.08530107230771068, "learning_rate": 1.2614612341394872e-06, "loss": 0.8642, "step": 10239 }, { "epoch": 4.639782510194834, "grad_norm": 0.07939831726554046, "learning_rate": 1.2583104371375244e-06, "loss": 0.8267, "step": 10240 }, { "epoch": 4.640235613955596, "grad_norm": 0.08237914475375172, "learning_rate": 1.2551635171046227e-06, "loss": 0.8586, "step": 10241 }, { "epoch": 4.640688717716357, "grad_norm": 0.07973148875920141, "learning_rate": 1.2520204743556952e-06, "loss": 0.8355, "step": 10242 }, { "epoch": 4.6411418214771185, "grad_norm": 0.08785478551855735, "learning_rate": 1.2488813092052764e-06, "loss": 0.841, "step": 10243 }, { "epoch": 4.64159492523788, "grad_norm": 0.08036288604074594, "learning_rate": 1.2457460219675066e-06, "loss": 0.8484, "step": 10244 }, { "epoch": 4.642048028998641, "grad_norm": 0.07991547704526593, "learning_rate": 1.2426146129561433e-06, "loss": 0.8442, "step": 10245 }, { "epoch": 4.642501132759402, "grad_norm": 0.08229817088673336, "learning_rate": 1.239487082484554e-06, "loss": 0.8474, "step": 10246 }, { "epoch": 4.642954236520163, "grad_norm": 0.0861856997489365, "learning_rate": 1.2363634308657192e-06, "loss": 0.8471, "step": 10247 }, { "epoch": 4.643407340280924, "grad_norm": 0.07835146124564925, "learning_rate": 1.2332436584122243e-06, "loss": 0.844, "step": 10248 }, { "epoch": 4.643860444041685, "grad_norm": 0.08945201446144327, "learning_rate": 1.230127765436282e-06, "loss": 0.8242, "step": 10249 }, { "epoch": 4.644313547802446, "grad_norm": 0.08270449962097016, "learning_rate": 1.2270157522496962e-06, "loss": 0.8231, "step": 10250 }, { "epoch": 4.644766651563208, "grad_norm": 0.08007954461012173, "learning_rate": 1.2239076191638977e-06, "loss": 0.8471, "step": 10251 }, { "epoch": 4.64521975532397, "grad_norm": 0.08426636124910195, "learning_rate": 1.2208033664899266e-06, "loss": 0.8391, "step": 10252 }, { "epoch": 4.645672859084731, "grad_norm": 0.07935825494546284, "learning_rate": 1.2177029945384233e-06, "loss": 0.8536, "step": 10253 }, { "epoch": 4.646125962845492, "grad_norm": 0.08208965085174509, "learning_rate": 1.214606503619664e-06, "loss": 0.8315, "step": 10254 }, { "epoch": 4.646579066606253, "grad_norm": 0.0794430112965293, "learning_rate": 1.2115138940435078e-06, "loss": 0.8383, "step": 10255 }, { "epoch": 4.647032170367014, "grad_norm": 0.08187512338828803, "learning_rate": 1.208425166119449e-06, "loss": 0.8349, "step": 10256 }, { "epoch": 4.647485274127775, "grad_norm": 0.07950408273967192, "learning_rate": 1.2053403201565783e-06, "loss": 0.8685, "step": 10257 }, { "epoch": 4.647938377888536, "grad_norm": 0.08064132458464625, "learning_rate": 1.2022593564636043e-06, "loss": 0.8616, "step": 10258 }, { "epoch": 4.6483914816492975, "grad_norm": 0.08380923018998762, "learning_rate": 1.1991822753488447e-06, "loss": 0.8612, "step": 10259 }, { "epoch": 4.648844585410059, "grad_norm": 0.07890424654210228, "learning_rate": 1.1961090771202311e-06, "loss": 0.8536, "step": 10260 }, { "epoch": 4.64929768917082, "grad_norm": 0.08099779141192273, "learning_rate": 1.1930397620853128e-06, "loss": 0.8509, "step": 10261 }, { "epoch": 4.649750792931581, "grad_norm": 0.08329857529412016, "learning_rate": 1.1899743305512267e-06, "loss": 0.861, "step": 10262 }, { "epoch": 4.650203896692343, "grad_norm": 0.08219450885576465, "learning_rate": 1.1869127828247539e-06, "loss": 0.8462, "step": 10263 }, { "epoch": 4.650657000453104, "grad_norm": 0.07867040105219547, "learning_rate": 1.1838551192122583e-06, "loss": 0.8325, "step": 10264 }, { "epoch": 4.651110104213865, "grad_norm": 0.0872727924451238, "learning_rate": 1.1808013400197437e-06, "loss": 0.8389, "step": 10265 }, { "epoch": 4.651563207974626, "grad_norm": 0.08627905806671829, "learning_rate": 1.1777514455527882e-06, "loss": 0.8494, "step": 10266 }, { "epoch": 4.652016311735387, "grad_norm": 0.0859862564026225, "learning_rate": 1.1747054361166144e-06, "loss": 0.8481, "step": 10267 }, { "epoch": 4.652469415496149, "grad_norm": 0.09091964823770575, "learning_rate": 1.1716633120160359e-06, "loss": 0.8342, "step": 10268 }, { "epoch": 4.65292251925691, "grad_norm": 0.08718849930408253, "learning_rate": 1.1686250735554939e-06, "loss": 0.8378, "step": 10269 }, { "epoch": 4.653375623017671, "grad_norm": 0.08891997088945376, "learning_rate": 1.165590721039025e-06, "loss": 0.8616, "step": 10270 }, { "epoch": 4.653828726778432, "grad_norm": 0.0854437548474897, "learning_rate": 1.162560254770284e-06, "loss": 0.8743, "step": 10271 }, { "epoch": 4.654281830539193, "grad_norm": 0.10091512010517018, "learning_rate": 1.15953367505254e-06, "loss": 0.8605, "step": 10272 }, { "epoch": 4.654734934299954, "grad_norm": 0.08378362278904807, "learning_rate": 1.15651098218867e-06, "loss": 0.8628, "step": 10273 }, { "epoch": 4.655188038060716, "grad_norm": 0.08488430829686872, "learning_rate": 1.153492176481157e-06, "loss": 0.836, "step": 10274 }, { "epoch": 4.655641141821477, "grad_norm": 0.087285439415351, "learning_rate": 1.1504772582321011e-06, "loss": 0.831, "step": 10275 }, { "epoch": 4.6560942455822385, "grad_norm": 0.09687653398451629, "learning_rate": 1.1474662277432215e-06, "loss": 0.8627, "step": 10276 }, { "epoch": 4.656547349343, "grad_norm": 0.0871472477316641, "learning_rate": 1.1444590853158278e-06, "loss": 0.8423, "step": 10277 }, { "epoch": 4.657000453103761, "grad_norm": 0.08782433977967322, "learning_rate": 1.1414558312508484e-06, "loss": 0.8678, "step": 10278 }, { "epoch": 4.657453556864522, "grad_norm": 0.08551361327017466, "learning_rate": 1.1384564658488383e-06, "loss": 0.8582, "step": 10279 }, { "epoch": 4.657906660625283, "grad_norm": 0.10809193685008783, "learning_rate": 1.13546098940994e-06, "loss": 0.8637, "step": 10280 }, { "epoch": 4.658359764386044, "grad_norm": 0.08885897664239317, "learning_rate": 1.1324694022339267e-06, "loss": 0.8517, "step": 10281 }, { "epoch": 4.658812868146805, "grad_norm": 0.08873976198799526, "learning_rate": 1.1294817046201677e-06, "loss": 0.8555, "step": 10282 }, { "epoch": 4.659265971907567, "grad_norm": 0.0864844503185706, "learning_rate": 1.126497896867651e-06, "loss": 0.8503, "step": 10283 }, { "epoch": 4.6597190756683275, "grad_norm": 0.10243383643876404, "learning_rate": 1.1235179792749685e-06, "loss": 0.8627, "step": 10284 }, { "epoch": 4.66017217942909, "grad_norm": 0.09571206788948666, "learning_rate": 1.120541952140335e-06, "loss": 0.8519, "step": 10285 }, { "epoch": 4.660625283189851, "grad_norm": 0.10455316161223253, "learning_rate": 1.1175698157615656e-06, "loss": 0.84, "step": 10286 }, { "epoch": 4.661078386950612, "grad_norm": 0.08199921130813768, "learning_rate": 1.1146015704360802e-06, "loss": 0.8454, "step": 10287 }, { "epoch": 4.661531490711373, "grad_norm": 0.08309767579128488, "learning_rate": 1.1116372164609301e-06, "loss": 0.8282, "step": 10288 }, { "epoch": 4.661984594472134, "grad_norm": 0.08331330297724464, "learning_rate": 1.108676754132758e-06, "loss": 0.832, "step": 10289 }, { "epoch": 4.662437698232895, "grad_norm": 0.11098628681768845, "learning_rate": 1.105720183747825e-06, "loss": 0.8586, "step": 10290 }, { "epoch": 4.662890801993656, "grad_norm": 0.08956106272409012, "learning_rate": 1.1027675056020048e-06, "loss": 0.8223, "step": 10291 }, { "epoch": 4.6633439057544175, "grad_norm": 0.08482260025221104, "learning_rate": 1.0998187199907773e-06, "loss": 0.8386, "step": 10292 }, { "epoch": 4.663797009515179, "grad_norm": 0.0837569173047949, "learning_rate": 1.0968738272092305e-06, "loss": 0.8369, "step": 10293 }, { "epoch": 4.664250113275941, "grad_norm": 0.08398504070565298, "learning_rate": 1.093932827552071e-06, "loss": 0.8546, "step": 10294 }, { "epoch": 4.664703217036702, "grad_norm": 0.09466537859437356, "learning_rate": 1.0909957213136102e-06, "loss": 0.8358, "step": 10295 }, { "epoch": 4.665156320797463, "grad_norm": 0.1076208487520524, "learning_rate": 1.0880625087877683e-06, "loss": 0.8174, "step": 10296 }, { "epoch": 4.665609424558224, "grad_norm": 0.08437711087891984, "learning_rate": 1.085133190268084e-06, "loss": 0.8472, "step": 10297 }, { "epoch": 4.666062528318985, "grad_norm": 0.0899574465964186, "learning_rate": 1.082207766047696e-06, "loss": 0.8613, "step": 10298 }, { "epoch": 4.666515632079746, "grad_norm": 0.08233206630142556, "learning_rate": 1.0792862364193612e-06, "loss": 0.829, "step": 10299 }, { "epoch": 4.666968735840507, "grad_norm": 0.09314784208148495, "learning_rate": 1.076368601675437e-06, "loss": 0.8533, "step": 10300 }, { "epoch": 4.6674218396012686, "grad_norm": 0.08865540994989388, "learning_rate": 1.0734548621079121e-06, "loss": 0.851, "step": 10301 }, { "epoch": 4.66787494336203, "grad_norm": 0.09404583729576045, "learning_rate": 1.0705450180083577e-06, "loss": 0.8424, "step": 10302 }, { "epoch": 4.668328047122791, "grad_norm": 0.08045436855361174, "learning_rate": 1.0676390696679718e-06, "loss": 0.8424, "step": 10303 }, { "epoch": 4.668781150883552, "grad_norm": 0.08314104036191002, "learning_rate": 1.0647370173775618e-06, "loss": 0.8498, "step": 10304 }, { "epoch": 4.669234254644314, "grad_norm": 0.08434133537010223, "learning_rate": 1.0618388614275399e-06, "loss": 0.8338, "step": 10305 }, { "epoch": 4.669687358405075, "grad_norm": 0.09151456518825248, "learning_rate": 1.0589446021079363e-06, "loss": 0.8704, "step": 10306 }, { "epoch": 4.670140462165836, "grad_norm": 0.08941085048870305, "learning_rate": 1.0560542397083773e-06, "loss": 0.8435, "step": 10307 }, { "epoch": 4.670593565926597, "grad_norm": 0.08024447438384671, "learning_rate": 1.0531677745181158e-06, "loss": 0.8547, "step": 10308 }, { "epoch": 4.6710466696873585, "grad_norm": 0.07829338352161021, "learning_rate": 1.050285206826005e-06, "loss": 0.857, "step": 10309 }, { "epoch": 4.67149977344812, "grad_norm": 0.08801882497251821, "learning_rate": 1.0474065369205077e-06, "loss": 0.8495, "step": 10310 }, { "epoch": 4.671952877208881, "grad_norm": 0.09522345846140584, "learning_rate": 1.0445317650897002e-06, "loss": 0.845, "step": 10311 }, { "epoch": 4.672405980969642, "grad_norm": 0.1005114668133518, "learning_rate": 1.0416608916212678e-06, "loss": 0.8529, "step": 10312 }, { "epoch": 4.672859084730403, "grad_norm": 0.08404920723127426, "learning_rate": 1.0387939168025096e-06, "loss": 0.8362, "step": 10313 }, { "epoch": 4.673312188491164, "grad_norm": 0.08096869989209365, "learning_rate": 1.035930840920316e-06, "loss": 0.8631, "step": 10314 }, { "epoch": 4.673765292251925, "grad_norm": 0.09526381536069449, "learning_rate": 1.0330716642612227e-06, "loss": 0.8567, "step": 10315 }, { "epoch": 4.674218396012687, "grad_norm": 0.08029598086205536, "learning_rate": 1.0302163871113335e-06, "loss": 0.8681, "step": 10316 }, { "epoch": 4.674671499773448, "grad_norm": 0.07988009403073178, "learning_rate": 1.027365009756398e-06, "loss": 0.8442, "step": 10317 }, { "epoch": 4.6751246035342096, "grad_norm": 0.08240341019813852, "learning_rate": 1.024517532481757e-06, "loss": 0.8659, "step": 10318 }, { "epoch": 4.675577707294971, "grad_norm": 0.08154592315612644, "learning_rate": 1.021673955572351e-06, "loss": 0.8619, "step": 10319 }, { "epoch": 4.676030811055732, "grad_norm": 0.08770957605343772, "learning_rate": 1.0188342793127614e-06, "loss": 0.864, "step": 10320 }, { "epoch": 4.676483914816493, "grad_norm": 0.08425701872550878, "learning_rate": 1.015998503987148e-06, "loss": 0.8511, "step": 10321 }, { "epoch": 4.676937018577254, "grad_norm": 0.08012510272431973, "learning_rate": 1.0131666298793052e-06, "loss": 0.8355, "step": 10322 }, { "epoch": 4.677390122338015, "grad_norm": 0.08132957135897398, "learning_rate": 1.010338657272607e-06, "loss": 0.8582, "step": 10323 }, { "epoch": 4.677843226098776, "grad_norm": 0.10912398450767494, "learning_rate": 1.0075145864500757e-06, "loss": 0.8506, "step": 10324 }, { "epoch": 4.6782963298595375, "grad_norm": 0.08591024428911993, "learning_rate": 1.004694417694303e-06, "loss": 0.8446, "step": 10325 }, { "epoch": 4.678749433620299, "grad_norm": 0.09012212105901143, "learning_rate": 1.0018781512875297e-06, "loss": 0.851, "step": 10326 }, { "epoch": 4.679202537381061, "grad_norm": 0.0797626231190478, "learning_rate": 9.990657875115662e-07, "loss": 0.84, "step": 10327 }, { "epoch": 4.679655641141822, "grad_norm": 0.08298924820366672, "learning_rate": 9.96257326647867e-07, "loss": 0.8305, "step": 10328 }, { "epoch": 4.680108744902583, "grad_norm": 0.0824885262840542, "learning_rate": 9.934527689774786e-07, "loss": 0.8382, "step": 10329 }, { "epoch": 4.680561848663344, "grad_norm": 0.08813656464256023, "learning_rate": 9.90652114781052e-07, "loss": 0.8334, "step": 10330 }, { "epoch": 4.681014952424105, "grad_norm": 0.07733515866568044, "learning_rate": 9.87855364338861e-07, "loss": 0.84, "step": 10331 }, { "epoch": 4.681468056184866, "grad_norm": 0.08237640871629837, "learning_rate": 9.850625179307793e-07, "loss": 0.8491, "step": 10332 }, { "epoch": 4.681921159945627, "grad_norm": 0.08601914930057038, "learning_rate": 9.82273575836299e-07, "loss": 0.8375, "step": 10333 }, { "epoch": 4.6823742637063885, "grad_norm": 0.08403754080426803, "learning_rate": 9.794885383345075e-07, "loss": 0.8564, "step": 10334 }, { "epoch": 4.68282736746715, "grad_norm": 0.08695696997674969, "learning_rate": 9.767074057041205e-07, "loss": 0.8389, "step": 10335 }, { "epoch": 4.683280471227912, "grad_norm": 0.08000678141980462, "learning_rate": 9.739301782234434e-07, "loss": 0.8589, "step": 10336 }, { "epoch": 4.683733574988672, "grad_norm": 0.09304681014009239, "learning_rate": 9.711568561704053e-07, "loss": 0.8353, "step": 10337 }, { "epoch": 4.684186678749434, "grad_norm": 0.0913878394362702, "learning_rate": 9.683874398225357e-07, "loss": 0.841, "step": 10338 }, { "epoch": 4.684639782510195, "grad_norm": 0.08385052071044707, "learning_rate": 9.656219294569723e-07, "loss": 0.8384, "step": 10339 }, { "epoch": 4.685092886270956, "grad_norm": 0.07744261530842471, "learning_rate": 9.628603253504765e-07, "loss": 0.8367, "step": 10340 }, { "epoch": 4.685545990031717, "grad_norm": 0.08544886365777528, "learning_rate": 9.601026277794001e-07, "loss": 0.8388, "step": 10341 }, { "epoch": 4.6859990937924785, "grad_norm": 0.08364959658821909, "learning_rate": 9.573488370197182e-07, "loss": 0.8379, "step": 10342 }, { "epoch": 4.68645219755324, "grad_norm": 0.07705990230709177, "learning_rate": 9.545989533469968e-07, "loss": 0.853, "step": 10343 }, { "epoch": 4.686905301314001, "grad_norm": 0.09151446316674781, "learning_rate": 9.518529770364427e-07, "loss": 0.8404, "step": 10344 }, { "epoch": 4.687358405074762, "grad_norm": 0.08303998433365394, "learning_rate": 9.491109083628313e-07, "loss": 0.8247, "step": 10345 }, { "epoch": 4.687811508835523, "grad_norm": 0.0834301968403448, "learning_rate": 9.463727476005835e-07, "loss": 0.8562, "step": 10346 }, { "epoch": 4.688264612596285, "grad_norm": 0.08773124449731003, "learning_rate": 9.436384950237065e-07, "loss": 0.8499, "step": 10347 }, { "epoch": 4.688717716357046, "grad_norm": 0.08008240822248397, "learning_rate": 9.409081509058216e-07, "loss": 0.849, "step": 10348 }, { "epoch": 4.689170820117807, "grad_norm": 0.07910363136407111, "learning_rate": 9.381817155201634e-07, "loss": 0.8334, "step": 10349 }, { "epoch": 4.689623923878568, "grad_norm": 0.08358524952077054, "learning_rate": 9.354591891395715e-07, "loss": 0.8308, "step": 10350 }, { "epoch": 4.6900770276393295, "grad_norm": 0.07950457522999849, "learning_rate": 9.327405720364991e-07, "loss": 0.8232, "step": 10351 }, { "epoch": 4.690530131400091, "grad_norm": 0.08350508424156196, "learning_rate": 9.300258644829951e-07, "loss": 0.8564, "step": 10352 }, { "epoch": 4.690983235160852, "grad_norm": 0.09264167357574664, "learning_rate": 9.273150667507358e-07, "loss": 0.8521, "step": 10353 }, { "epoch": 4.691436338921613, "grad_norm": 0.08432603182084905, "learning_rate": 9.246081791109929e-07, "loss": 0.8615, "step": 10354 }, { "epoch": 4.691889442682374, "grad_norm": 0.08247240102766831, "learning_rate": 9.219052018346519e-07, "loss": 0.8598, "step": 10355 }, { "epoch": 4.692342546443135, "grad_norm": 0.08356498418776225, "learning_rate": 9.192061351922033e-07, "loss": 0.8492, "step": 10356 }, { "epoch": 4.692795650203896, "grad_norm": 0.08997720477334845, "learning_rate": 9.165109794537507e-07, "loss": 0.8353, "step": 10357 }, { "epoch": 4.693248753964658, "grad_norm": 0.08373675468644634, "learning_rate": 9.138197348890076e-07, "loss": 0.8452, "step": 10358 }, { "epoch": 4.6937018577254195, "grad_norm": 0.09167255451433534, "learning_rate": 9.111324017672874e-07, "loss": 0.8574, "step": 10359 }, { "epoch": 4.694154961486181, "grad_norm": 0.12032513849778641, "learning_rate": 9.084489803575213e-07, "loss": 0.8314, "step": 10360 }, { "epoch": 4.694608065246942, "grad_norm": 0.0884475963158489, "learning_rate": 9.057694709282417e-07, "loss": 0.855, "step": 10361 }, { "epoch": 4.695061169007703, "grad_norm": 0.07944738216321236, "learning_rate": 9.03093873747598e-07, "loss": 0.8467, "step": 10362 }, { "epoch": 4.695514272768464, "grad_norm": 0.0800278406799963, "learning_rate": 9.004221890833453e-07, "loss": 0.8275, "step": 10363 }, { "epoch": 4.695967376529225, "grad_norm": 0.0869056675400083, "learning_rate": 8.97754417202834e-07, "loss": 0.8325, "step": 10364 }, { "epoch": 4.696420480289986, "grad_norm": 0.08257395192381753, "learning_rate": 8.950905583730463e-07, "loss": 0.8534, "step": 10365 }, { "epoch": 4.696873584050747, "grad_norm": 0.08516031397698394, "learning_rate": 8.924306128605509e-07, "loss": 0.8702, "step": 10366 }, { "epoch": 4.6973266878115085, "grad_norm": 0.08792169571005075, "learning_rate": 8.897745809315395e-07, "loss": 0.8341, "step": 10367 }, { "epoch": 4.69777979157227, "grad_norm": 0.07486905250303384, "learning_rate": 8.871224628518082e-07, "loss": 0.86, "step": 10368 }, { "epoch": 4.698232895333032, "grad_norm": 0.07833337153771479, "learning_rate": 8.844742588867628e-07, "loss": 0.8306, "step": 10369 }, { "epoch": 4.698685999093793, "grad_norm": 0.07486760672623058, "learning_rate": 8.818299693014088e-07, "loss": 0.855, "step": 10370 }, { "epoch": 4.699139102854554, "grad_norm": 0.08502787604617497, "learning_rate": 8.791895943603657e-07, "loss": 0.8406, "step": 10371 }, { "epoch": 4.699592206615315, "grad_norm": 0.07901214548212655, "learning_rate": 8.765531343278666e-07, "loss": 0.8237, "step": 10372 }, { "epoch": 4.700045310376076, "grad_norm": 0.0895076827913647, "learning_rate": 8.73920589467745e-07, "loss": 0.8579, "step": 10373 }, { "epoch": 4.700498414136837, "grad_norm": 0.07997611087295145, "learning_rate": 8.712919600434521e-07, "loss": 0.8433, "step": 10374 }, { "epoch": 4.7009515178975985, "grad_norm": 0.08042253554942742, "learning_rate": 8.686672463180312e-07, "loss": 0.8375, "step": 10375 }, { "epoch": 4.70140462165836, "grad_norm": 0.08752293151173084, "learning_rate": 8.66046448554152e-07, "loss": 0.8369, "step": 10376 }, { "epoch": 4.701857725419121, "grad_norm": 0.08468914591462466, "learning_rate": 8.634295670140757e-07, "loss": 0.8486, "step": 10377 }, { "epoch": 4.702310829179882, "grad_norm": 0.07605715262847253, "learning_rate": 8.608166019596864e-07, "loss": 0.8432, "step": 10378 }, { "epoch": 4.702763932940643, "grad_norm": 0.07572762456966625, "learning_rate": 8.582075536524681e-07, "loss": 0.8366, "step": 10379 }, { "epoch": 4.703217036701405, "grad_norm": 0.0772453175157251, "learning_rate": 8.556024223535097e-07, "loss": 0.8511, "step": 10380 }, { "epoch": 4.703670140462166, "grad_norm": 0.07424646165074497, "learning_rate": 8.530012083235139e-07, "loss": 0.8376, "step": 10381 }, { "epoch": 4.704123244222927, "grad_norm": 0.0839938129136572, "learning_rate": 8.504039118227969e-07, "loss": 0.8364, "step": 10382 }, { "epoch": 4.704576347983688, "grad_norm": 0.08897512990034351, "learning_rate": 8.478105331112707e-07, "loss": 0.8426, "step": 10383 }, { "epoch": 4.7050294517444495, "grad_norm": 0.07888944496699614, "learning_rate": 8.452210724484567e-07, "loss": 0.85, "step": 10384 }, { "epoch": 4.705482555505211, "grad_norm": 0.08300947279242732, "learning_rate": 8.426355300934985e-07, "loss": 0.854, "step": 10385 }, { "epoch": 4.705935659265972, "grad_norm": 0.08954798792620976, "learning_rate": 8.400539063051228e-07, "loss": 0.8625, "step": 10386 }, { "epoch": 4.706388763026733, "grad_norm": 0.08578597027887765, "learning_rate": 8.374762013416959e-07, "loss": 0.8429, "step": 10387 }, { "epoch": 4.706841866787494, "grad_norm": 0.07620567791602563, "learning_rate": 8.349024154611584e-07, "loss": 0.8381, "step": 10388 }, { "epoch": 4.707294970548255, "grad_norm": 0.08000572640412118, "learning_rate": 8.323325489210865e-07, "loss": 0.8515, "step": 10389 }, { "epoch": 4.707748074309016, "grad_norm": 0.09931608336979317, "learning_rate": 8.297666019786476e-07, "loss": 0.8315, "step": 10390 }, { "epoch": 4.708201178069778, "grad_norm": 0.08243190445265516, "learning_rate": 8.272045748906188e-07, "loss": 0.8495, "step": 10391 }, { "epoch": 4.7086542818305395, "grad_norm": 0.07824693037109846, "learning_rate": 8.246464679133992e-07, "loss": 0.839, "step": 10392 }, { "epoch": 4.709107385591301, "grad_norm": 0.07847894141827856, "learning_rate": 8.220922813029708e-07, "loss": 0.8248, "step": 10393 }, { "epoch": 4.709560489352062, "grad_norm": 0.08427732934298537, "learning_rate": 8.195420153149425e-07, "loss": 0.8512, "step": 10394 }, { "epoch": 4.710013593112823, "grad_norm": 0.0760484481876022, "learning_rate": 8.169956702045279e-07, "loss": 0.8398, "step": 10395 }, { "epoch": 4.710466696873584, "grad_norm": 0.08803543640182714, "learning_rate": 8.144532462265409e-07, "loss": 0.8549, "step": 10396 }, { "epoch": 4.710919800634345, "grad_norm": 0.08120102746155325, "learning_rate": 8.119147436354135e-07, "loss": 0.8355, "step": 10397 }, { "epoch": 4.711372904395106, "grad_norm": 0.07884062960994993, "learning_rate": 8.093801626851738e-07, "loss": 0.8459, "step": 10398 }, { "epoch": 4.711826008155867, "grad_norm": 0.08319707394330955, "learning_rate": 8.068495036294677e-07, "loss": 0.8629, "step": 10399 }, { "epoch": 4.712279111916629, "grad_norm": 0.0766580984022047, "learning_rate": 8.043227667215369e-07, "loss": 0.8611, "step": 10400 }, { "epoch": 4.71273221567739, "grad_norm": 0.08203476509388134, "learning_rate": 8.017999522142461e-07, "loss": 0.8361, "step": 10401 }, { "epoch": 4.713185319438152, "grad_norm": 0.08903693576593556, "learning_rate": 7.992810603600554e-07, "loss": 0.848, "step": 10402 }, { "epoch": 4.713638423198913, "grad_norm": 0.08027121946481505, "learning_rate": 7.967660914110342e-07, "loss": 0.8408, "step": 10403 }, { "epoch": 4.714091526959674, "grad_norm": 0.08109071515805119, "learning_rate": 7.942550456188614e-07, "loss": 0.8339, "step": 10404 }, { "epoch": 4.714544630720435, "grad_norm": 0.0846705876822488, "learning_rate": 7.91747923234829e-07, "loss": 0.8374, "step": 10405 }, { "epoch": 4.714997734481196, "grad_norm": 0.07724077599317102, "learning_rate": 7.892447245098211e-07, "loss": 0.8385, "step": 10406 }, { "epoch": 4.715450838241957, "grad_norm": 0.07647964908178179, "learning_rate": 7.867454496943483e-07, "loss": 0.8336, "step": 10407 }, { "epoch": 4.7159039420027185, "grad_norm": 0.07570327096835101, "learning_rate": 7.842500990385171e-07, "loss": 0.8453, "step": 10408 }, { "epoch": 4.71635704576348, "grad_norm": 0.08934939889242967, "learning_rate": 7.817586727920301e-07, "loss": 0.8527, "step": 10409 }, { "epoch": 4.716810149524241, "grad_norm": 0.07913271043035643, "learning_rate": 7.792711712042256e-07, "loss": 0.8426, "step": 10410 }, { "epoch": 4.717263253285003, "grad_norm": 0.0891144857899787, "learning_rate": 7.767875945240244e-07, "loss": 0.8442, "step": 10411 }, { "epoch": 4.717716357045764, "grad_norm": 0.07931286297554159, "learning_rate": 7.7430794299997e-07, "loss": 0.8766, "step": 10412 }, { "epoch": 4.718169460806525, "grad_norm": 0.0808014805885503, "learning_rate": 7.718322168802017e-07, "loss": 0.8465, "step": 10413 }, { "epoch": 4.718622564567286, "grad_norm": 0.08828593759623163, "learning_rate": 7.693604164124769e-07, "loss": 0.8629, "step": 10414 }, { "epoch": 4.719075668328047, "grad_norm": 0.08718204575789827, "learning_rate": 7.668925418441442e-07, "loss": 0.8625, "step": 10415 }, { "epoch": 4.719528772088808, "grad_norm": 0.07955208772767176, "learning_rate": 7.644285934221795e-07, "loss": 0.8498, "step": 10416 }, { "epoch": 4.7199818758495695, "grad_norm": 0.08217715058221996, "learning_rate": 7.619685713931501e-07, "loss": 0.8586, "step": 10417 }, { "epoch": 4.720434979610331, "grad_norm": 0.07927911638903669, "learning_rate": 7.595124760032369e-07, "loss": 0.8456, "step": 10418 }, { "epoch": 4.720888083371092, "grad_norm": 0.07697332519782855, "learning_rate": 7.570603074982297e-07, "loss": 0.839, "step": 10419 }, { "epoch": 4.721341187131853, "grad_norm": 0.08071056562185257, "learning_rate": 7.54612066123519e-07, "loss": 0.8513, "step": 10420 }, { "epoch": 4.721794290892614, "grad_norm": 0.07287248946475781, "learning_rate": 7.521677521241088e-07, "loss": 0.839, "step": 10421 }, { "epoch": 4.722247394653376, "grad_norm": 0.07857937110232106, "learning_rate": 7.497273657446036e-07, "loss": 0.8709, "step": 10422 }, { "epoch": 4.722700498414137, "grad_norm": 0.07595588101483901, "learning_rate": 7.472909072292256e-07, "loss": 0.8387, "step": 10423 }, { "epoch": 4.723153602174898, "grad_norm": 0.08401699860434209, "learning_rate": 7.448583768217932e-07, "loss": 0.8553, "step": 10424 }, { "epoch": 4.7236067059356595, "grad_norm": 0.08348647717219973, "learning_rate": 7.42429774765725e-07, "loss": 0.839, "step": 10425 }, { "epoch": 4.724059809696421, "grad_norm": 0.07788638795904756, "learning_rate": 7.400051013040709e-07, "loss": 0.8372, "step": 10426 }, { "epoch": 4.724512913457182, "grad_norm": 0.07272345378696465, "learning_rate": 7.375843566794683e-07, "loss": 0.8513, "step": 10427 }, { "epoch": 4.724966017217943, "grad_norm": 0.09325259309986862, "learning_rate": 7.351675411341674e-07, "loss": 0.8566, "step": 10428 }, { "epoch": 4.725419120978704, "grad_norm": 0.09548053649869202, "learning_rate": 7.327546549100195e-07, "loss": 0.8505, "step": 10429 }, { "epoch": 4.725872224739465, "grad_norm": 0.08198563187749901, "learning_rate": 7.30345698248498e-07, "loss": 0.8243, "step": 10430 }, { "epoch": 4.726325328500226, "grad_norm": 0.07857556250184372, "learning_rate": 7.279406713906633e-07, "loss": 0.8454, "step": 10431 }, { "epoch": 4.726778432260987, "grad_norm": 0.07990700213850893, "learning_rate": 7.255395745771943e-07, "loss": 0.8321, "step": 10432 }, { "epoch": 4.727231536021749, "grad_norm": 0.07302506368486696, "learning_rate": 7.23142408048374e-07, "loss": 0.8399, "step": 10433 }, { "epoch": 4.7276846397825105, "grad_norm": 0.07797131659583019, "learning_rate": 7.207491720440952e-07, "loss": 0.8435, "step": 10434 }, { "epoch": 4.728137743543272, "grad_norm": 0.08226861862824572, "learning_rate": 7.183598668038505e-07, "loss": 0.843, "step": 10435 }, { "epoch": 4.728590847304033, "grad_norm": 0.0797831632869817, "learning_rate": 7.159744925667511e-07, "loss": 0.8538, "step": 10436 }, { "epoch": 4.729043951064794, "grad_norm": 0.07490512545517336, "learning_rate": 7.135930495714948e-07, "loss": 0.8376, "step": 10437 }, { "epoch": 4.729497054825555, "grad_norm": 0.07542467989734497, "learning_rate": 7.112155380564068e-07, "loss": 0.8402, "step": 10438 }, { "epoch": 4.729950158586316, "grad_norm": 0.07601360745051403, "learning_rate": 7.088419582594075e-07, "loss": 0.8401, "step": 10439 }, { "epoch": 4.730403262347077, "grad_norm": 0.07520171873521829, "learning_rate": 7.064723104180227e-07, "loss": 0.8404, "step": 10440 }, { "epoch": 4.730856366107838, "grad_norm": 0.08385503249212202, "learning_rate": 7.041065947693959e-07, "loss": 0.8693, "step": 10441 }, { "epoch": 4.7313094698686, "grad_norm": 0.08169760994991662, "learning_rate": 7.017448115502623e-07, "loss": 0.8634, "step": 10442 }, { "epoch": 4.731762573629361, "grad_norm": 0.07510619763030026, "learning_rate": 6.99386960996975e-07, "loss": 0.8589, "step": 10443 }, { "epoch": 4.732215677390123, "grad_norm": 0.07690807825523933, "learning_rate": 6.970330433454919e-07, "loss": 0.8472, "step": 10444 }, { "epoch": 4.732668781150884, "grad_norm": 0.07606540883924783, "learning_rate": 6.946830588313714e-07, "loss": 0.8519, "step": 10445 }, { "epoch": 4.733121884911645, "grad_norm": 0.07653725842943727, "learning_rate": 6.923370076897807e-07, "loss": 0.8345, "step": 10446 }, { "epoch": 4.733574988672406, "grad_norm": 0.07933607187333905, "learning_rate": 6.899948901554965e-07, "loss": 0.8388, "step": 10447 }, { "epoch": 4.734028092433167, "grad_norm": 0.08594203154837686, "learning_rate": 6.876567064629003e-07, "loss": 0.8652, "step": 10448 }, { "epoch": 4.734481196193928, "grad_norm": 0.08785662329030419, "learning_rate": 6.853224568459738e-07, "loss": 0.8598, "step": 10449 }, { "epoch": 4.7349342999546895, "grad_norm": 0.0834103980770277, "learning_rate": 6.829921415383211e-07, "loss": 0.873, "step": 10450 }, { "epoch": 4.735387403715451, "grad_norm": 0.07683855163219543, "learning_rate": 6.806657607731382e-07, "loss": 0.8521, "step": 10451 }, { "epoch": 4.735840507476212, "grad_norm": 0.08318954680077281, "learning_rate": 6.783433147832252e-07, "loss": 0.8334, "step": 10452 }, { "epoch": 4.736293611236973, "grad_norm": 0.07903134265759783, "learning_rate": 6.760248038010009e-07, "loss": 0.8494, "step": 10453 }, { "epoch": 4.736746714997734, "grad_norm": 0.080500838814744, "learning_rate": 6.737102280584795e-07, "loss": 0.8493, "step": 10454 }, { "epoch": 4.737199818758496, "grad_norm": 0.07813557499867056, "learning_rate": 6.713995877872892e-07, "loss": 0.8271, "step": 10455 }, { "epoch": 4.737652922519257, "grad_norm": 0.07896649056874981, "learning_rate": 6.690928832186627e-07, "loss": 0.8265, "step": 10456 }, { "epoch": 4.738106026280018, "grad_norm": 0.07650157592069144, "learning_rate": 6.667901145834333e-07, "loss": 0.8451, "step": 10457 }, { "epoch": 4.738559130040779, "grad_norm": 0.08400145356657315, "learning_rate": 6.644912821120431e-07, "loss": 0.8301, "step": 10458 }, { "epoch": 4.739012233801541, "grad_norm": 0.08226354173923832, "learning_rate": 6.621963860345482e-07, "loss": 0.853, "step": 10459 }, { "epoch": 4.739465337562302, "grad_norm": 0.073795145037968, "learning_rate": 6.599054265806004e-07, "loss": 0.8395, "step": 10460 }, { "epoch": 4.739918441323063, "grad_norm": 0.07894946260296604, "learning_rate": 6.576184039794564e-07, "loss": 0.8513, "step": 10461 }, { "epoch": 4.740371545083824, "grad_norm": 0.07432760732221684, "learning_rate": 6.553353184599909e-07, "loss": 0.8505, "step": 10462 }, { "epoch": 4.740824648844585, "grad_norm": 0.07229856166314663, "learning_rate": 6.530561702506744e-07, "loss": 0.834, "step": 10463 }, { "epoch": 4.741277752605347, "grad_norm": 0.08152076797101422, "learning_rate": 6.507809595795867e-07, "loss": 0.8326, "step": 10464 }, { "epoch": 4.741730856366107, "grad_norm": 0.0828824217707067, "learning_rate": 6.485096866744123e-07, "loss": 0.8469, "step": 10465 }, { "epoch": 4.742183960126869, "grad_norm": 0.07048443600231595, "learning_rate": 6.462423517624495e-07, "loss": 0.8524, "step": 10466 }, { "epoch": 4.7426370638876305, "grad_norm": 0.08001924391420823, "learning_rate": 6.439789550705833e-07, "loss": 0.8533, "step": 10467 }, { "epoch": 4.743090167648392, "grad_norm": 0.08452450908991214, "learning_rate": 6.41719496825326e-07, "loss": 0.8369, "step": 10468 }, { "epoch": 4.743543271409153, "grad_norm": 0.07590802335428239, "learning_rate": 6.394639772527855e-07, "loss": 0.8416, "step": 10469 }, { "epoch": 4.743996375169914, "grad_norm": 0.08062702995334212, "learning_rate": 6.372123965786703e-07, "loss": 0.8494, "step": 10470 }, { "epoch": 4.744449478930675, "grad_norm": 0.07922715996601916, "learning_rate": 6.349647550283111e-07, "loss": 0.8386, "step": 10471 }, { "epoch": 4.744902582691436, "grad_norm": 0.09070485746875058, "learning_rate": 6.327210528266303e-07, "loss": 0.8496, "step": 10472 }, { "epoch": 4.745355686452197, "grad_norm": 0.07441284223524255, "learning_rate": 6.304812901981594e-07, "loss": 0.8326, "step": 10473 }, { "epoch": 4.745808790212958, "grad_norm": 0.0858879275091019, "learning_rate": 6.282454673670347e-07, "loss": 0.8655, "step": 10474 }, { "epoch": 4.7462618939737204, "grad_norm": 0.08524385786561871, "learning_rate": 6.260135845570059e-07, "loss": 0.8645, "step": 10475 }, { "epoch": 4.746714997734482, "grad_norm": 0.07713638717322192, "learning_rate": 6.23785641991419e-07, "loss": 0.8523, "step": 10476 }, { "epoch": 4.747168101495243, "grad_norm": 0.08358358780540119, "learning_rate": 6.215616398932245e-07, "loss": 0.8413, "step": 10477 }, { "epoch": 4.747621205256004, "grad_norm": 0.0785720822389651, "learning_rate": 6.193415784849909e-07, "loss": 0.8707, "step": 10478 }, { "epoch": 4.748074309016765, "grad_norm": 0.08449292641121099, "learning_rate": 6.17125457988883e-07, "loss": 0.852, "step": 10479 }, { "epoch": 4.748527412777526, "grad_norm": 0.08168237205332185, "learning_rate": 6.149132786266743e-07, "loss": 0.8599, "step": 10480 }, { "epoch": 4.748980516538287, "grad_norm": 0.08378952098154827, "learning_rate": 6.127050406197343e-07, "loss": 0.843, "step": 10481 }, { "epoch": 4.749433620299048, "grad_norm": 0.07891882149916668, "learning_rate": 6.105007441890598e-07, "loss": 0.8416, "step": 10482 }, { "epoch": 4.7498867240598095, "grad_norm": 0.08320786693824561, "learning_rate": 6.083003895552253e-07, "loss": 0.8315, "step": 10483 }, { "epoch": 4.750339827820571, "grad_norm": 0.08333225328980894, "learning_rate": 6.061039769384369e-07, "loss": 0.8452, "step": 10484 }, { "epoch": 4.750792931581332, "grad_norm": 0.076251404841909, "learning_rate": 6.039115065584922e-07, "loss": 0.847, "step": 10485 }, { "epoch": 4.751246035342094, "grad_norm": 0.08536793515736779, "learning_rate": 6.017229786347889e-07, "loss": 0.8705, "step": 10486 }, { "epoch": 4.751699139102855, "grad_norm": 0.08185257526104454, "learning_rate": 5.995383933863475e-07, "loss": 0.8485, "step": 10487 }, { "epoch": 4.752152242863616, "grad_norm": 0.0816876658209833, "learning_rate": 5.973577510317797e-07, "loss": 0.8535, "step": 10488 }, { "epoch": 4.752605346624377, "grad_norm": 0.07799573808525904, "learning_rate": 5.951810517893108e-07, "loss": 0.8297, "step": 10489 }, { "epoch": 4.753058450385138, "grad_norm": 0.07804863895524485, "learning_rate": 5.930082958767624e-07, "loss": 0.8461, "step": 10490 }, { "epoch": 4.753511554145899, "grad_norm": 0.0763365864721077, "learning_rate": 5.908394835115738e-07, "loss": 0.8542, "step": 10491 }, { "epoch": 4.753964657906661, "grad_norm": 0.07630970700305502, "learning_rate": 5.886746149107758e-07, "loss": 0.8303, "step": 10492 }, { "epoch": 4.754417761667422, "grad_norm": 0.08842140970252264, "learning_rate": 5.865136902910218e-07, "loss": 0.8439, "step": 10493 }, { "epoch": 4.754870865428183, "grad_norm": 0.0786892040815694, "learning_rate": 5.843567098685476e-07, "loss": 0.8398, "step": 10494 }, { "epoch": 4.755323969188944, "grad_norm": 0.08221103248677482, "learning_rate": 5.822036738592162e-07, "loss": 0.8708, "step": 10495 }, { "epoch": 4.755777072949705, "grad_norm": 0.0755108819565888, "learning_rate": 5.80054582478482e-07, "loss": 0.8479, "step": 10496 }, { "epoch": 4.756230176710467, "grad_norm": 0.07367932715067133, "learning_rate": 5.779094359414128e-07, "loss": 0.8672, "step": 10497 }, { "epoch": 4.756683280471228, "grad_norm": 0.08973981704946954, "learning_rate": 5.757682344626814e-07, "loss": 0.8231, "step": 10498 }, { "epoch": 4.757136384231989, "grad_norm": 0.08348723145178102, "learning_rate": 5.736309782565519e-07, "loss": 0.8668, "step": 10499 }, { "epoch": 4.7575894879927505, "grad_norm": 0.07739491185271086, "learning_rate": 5.714976675369155e-07, "loss": 0.8323, "step": 10500 }, { "epoch": 4.758042591753512, "grad_norm": 0.08158551042042915, "learning_rate": 5.693683025172503e-07, "loss": 0.8418, "step": 10501 }, { "epoch": 4.758495695514273, "grad_norm": 0.08125995060396399, "learning_rate": 5.672428834106524e-07, "loss": 0.8469, "step": 10502 }, { "epoch": 4.758948799275034, "grad_norm": 0.08147778075646099, "learning_rate": 5.651214104298098e-07, "loss": 0.8507, "step": 10503 }, { "epoch": 4.759401903035795, "grad_norm": 0.07279263017872577, "learning_rate": 5.630038837870322e-07, "loss": 0.8502, "step": 10504 }, { "epoch": 4.759855006796556, "grad_norm": 0.08080109808524583, "learning_rate": 5.608903036942215e-07, "loss": 0.8485, "step": 10505 }, { "epoch": 4.760308110557317, "grad_norm": 0.07651154927334164, "learning_rate": 5.58780670362884e-07, "loss": 0.8614, "step": 10506 }, { "epoch": 4.760761214318078, "grad_norm": 0.07578553428130415, "learning_rate": 5.566749840041397e-07, "loss": 0.8605, "step": 10507 }, { "epoch": 4.76121431807884, "grad_norm": 0.07432105702688264, "learning_rate": 5.545732448287089e-07, "loss": 0.8416, "step": 10508 }, { "epoch": 4.761667421839602, "grad_norm": 0.0897254169782533, "learning_rate": 5.524754530469212e-07, "loss": 0.8761, "step": 10509 }, { "epoch": 4.762120525600363, "grad_norm": 0.07410857752921829, "learning_rate": 5.503816088686976e-07, "loss": 0.8601, "step": 10510 }, { "epoch": 4.762573629361124, "grad_norm": 0.08381726890904419, "learning_rate": 5.482917125035858e-07, "loss": 0.8343, "step": 10511 }, { "epoch": 4.763026733121885, "grad_norm": 0.07276706969841941, "learning_rate": 5.462057641607166e-07, "loss": 0.8439, "step": 10512 }, { "epoch": 4.763479836882646, "grad_norm": 0.07752068154105386, "learning_rate": 5.441237640488428e-07, "loss": 0.8353, "step": 10513 }, { "epoch": 4.763932940643407, "grad_norm": 0.08106093588454341, "learning_rate": 5.42045712376309e-07, "loss": 0.8428, "step": 10514 }, { "epoch": 4.764386044404168, "grad_norm": 0.08720972804872673, "learning_rate": 5.399716093510732e-07, "loss": 0.8414, "step": 10515 }, { "epoch": 4.7648391481649295, "grad_norm": 0.08307207278766633, "learning_rate": 5.379014551806982e-07, "loss": 0.8274, "step": 10516 }, { "epoch": 4.765292251925691, "grad_norm": 0.08305009181015958, "learning_rate": 5.35835250072343e-07, "loss": 0.8515, "step": 10517 }, { "epoch": 4.765745355686452, "grad_norm": 0.07508144051884298, "learning_rate": 5.337729942327841e-07, "loss": 0.8285, "step": 10518 }, { "epoch": 4.766198459447214, "grad_norm": 0.0721816787441974, "learning_rate": 5.317146878683899e-07, "loss": 0.84, "step": 10519 }, { "epoch": 4.766651563207975, "grad_norm": 0.08178353253268948, "learning_rate": 5.296603311851467e-07, "loss": 0.8575, "step": 10520 }, { "epoch": 4.767104666968736, "grad_norm": 0.07558948331445876, "learning_rate": 5.276099243886368e-07, "loss": 0.8454, "step": 10521 }, { "epoch": 4.767557770729497, "grad_norm": 0.07648660072473813, "learning_rate": 5.255634676840427e-07, "loss": 0.8542, "step": 10522 }, { "epoch": 4.768010874490258, "grad_norm": 0.07500714960749723, "learning_rate": 5.235209612761649e-07, "loss": 0.8271, "step": 10523 }, { "epoch": 4.768463978251019, "grad_norm": 0.07472487231072429, "learning_rate": 5.214824053693956e-07, "loss": 0.8418, "step": 10524 }, { "epoch": 4.7689170820117806, "grad_norm": 0.08404402393911106, "learning_rate": 5.194478001677449e-07, "loss": 0.8414, "step": 10525 }, { "epoch": 4.769370185772542, "grad_norm": 0.07192301414353534, "learning_rate": 5.174171458748101e-07, "loss": 0.8343, "step": 10526 }, { "epoch": 4.769823289533303, "grad_norm": 0.07600053801916426, "learning_rate": 5.153904426938194e-07, "loss": 0.8403, "step": 10527 }, { "epoch": 4.770276393294065, "grad_norm": 0.07718438796455311, "learning_rate": 5.133676908275754e-07, "loss": 0.8363, "step": 10528 }, { "epoch": 4.770729497054825, "grad_norm": 0.07551747110957815, "learning_rate": 5.113488904785069e-07, "loss": 0.8447, "step": 10529 }, { "epoch": 4.771182600815587, "grad_norm": 0.07724678521096365, "learning_rate": 5.093340418486348e-07, "loss": 0.8643, "step": 10530 }, { "epoch": 4.771635704576348, "grad_norm": 0.595486640725092, "learning_rate": 5.073231451395932e-07, "loss": 0.8587, "step": 10531 }, { "epoch": 4.772088808337109, "grad_norm": 0.0777791690973502, "learning_rate": 5.053162005526168e-07, "loss": 0.8289, "step": 10532 }, { "epoch": 4.7725419120978705, "grad_norm": 0.07781055330585415, "learning_rate": 5.033132082885406e-07, "loss": 0.8611, "step": 10533 }, { "epoch": 4.772995015858632, "grad_norm": 0.07973876286435172, "learning_rate": 5.013141685478173e-07, "loss": 0.8519, "step": 10534 }, { "epoch": 4.773448119619393, "grad_norm": 0.08325194237271877, "learning_rate": 4.993190815304871e-07, "loss": 0.8527, "step": 10535 }, { "epoch": 4.773901223380154, "grad_norm": 0.07316722492112633, "learning_rate": 4.97327947436208e-07, "loss": 0.8575, "step": 10536 }, { "epoch": 4.774354327140915, "grad_norm": 0.0723643909333989, "learning_rate": 4.95340766464234e-07, "loss": 0.847, "step": 10537 }, { "epoch": 4.774807430901676, "grad_norm": 0.07431331264713585, "learning_rate": 4.933575388134282e-07, "loss": 0.8419, "step": 10538 }, { "epoch": 4.775260534662438, "grad_norm": 0.0746904764100976, "learning_rate": 4.913782646822585e-07, "loss": 0.8425, "step": 10539 }, { "epoch": 4.775713638423199, "grad_norm": 0.0761709148986116, "learning_rate": 4.894029442687931e-07, "loss": 0.8363, "step": 10540 }, { "epoch": 4.77616674218396, "grad_norm": 0.08375279466877274, "learning_rate": 4.874315777707095e-07, "loss": 0.8366, "step": 10541 }, { "epoch": 4.7766198459447216, "grad_norm": 0.07568666242871018, "learning_rate": 4.85464165385281e-07, "loss": 0.855, "step": 10542 }, { "epoch": 4.777072949705483, "grad_norm": 0.07094676884998123, "learning_rate": 4.835007073093945e-07, "loss": 0.8255, "step": 10543 }, { "epoch": 4.777526053466244, "grad_norm": 0.07295949457669028, "learning_rate": 4.815412037395373e-07, "loss": 0.8474, "step": 10544 }, { "epoch": 4.777979157227005, "grad_norm": 0.08183949560847861, "learning_rate": 4.795856548718059e-07, "loss": 0.8543, "step": 10545 }, { "epoch": 4.778432260987766, "grad_norm": 0.07649393985729736, "learning_rate": 4.776340609018926e-07, "loss": 0.8551, "step": 10546 }, { "epoch": 4.778885364748527, "grad_norm": 0.08028836374545474, "learning_rate": 4.7568642202509895e-07, "loss": 0.8683, "step": 10547 }, { "epoch": 4.779338468509288, "grad_norm": 0.0776167460357392, "learning_rate": 4.737427384363269e-07, "loss": 0.854, "step": 10548 }, { "epoch": 4.7797915722700495, "grad_norm": 0.07696031718821081, "learning_rate": 4.71803010330083e-07, "loss": 0.8611, "step": 10549 }, { "epoch": 4.7802446760308115, "grad_norm": 0.08488155323791192, "learning_rate": 4.6986723790049205e-07, "loss": 0.8457, "step": 10550 }, { "epoch": 4.780697779791573, "grad_norm": 0.0767409023384723, "learning_rate": 4.679354213412568e-07, "loss": 0.8487, "step": 10551 }, { "epoch": 4.781150883552334, "grad_norm": 0.07591983640850963, "learning_rate": 4.6600756084571154e-07, "loss": 0.8533, "step": 10552 }, { "epoch": 4.781603987313095, "grad_norm": 0.07473144428569066, "learning_rate": 4.640836566067686e-07, "loss": 0.8429, "step": 10553 }, { "epoch": 4.782057091073856, "grad_norm": 0.06976327830500569, "learning_rate": 4.6216370881696726e-07, "loss": 0.8511, "step": 10554 }, { "epoch": 4.782510194834617, "grad_norm": 0.07367950473833436, "learning_rate": 4.6024771766843837e-07, "loss": 0.8397, "step": 10555 }, { "epoch": 4.782963298595378, "grad_norm": 0.07255158689632364, "learning_rate": 4.5833568335291736e-07, "loss": 0.8506, "step": 10556 }, { "epoch": 4.783416402356139, "grad_norm": 0.07798525455263582, "learning_rate": 4.56427606061749e-07, "loss": 0.8256, "step": 10557 }, { "epoch": 4.7838695061169005, "grad_norm": 0.07957527357358132, "learning_rate": 4.5452348598587823e-07, "loss": 0.8479, "step": 10558 }, { "epoch": 4.784322609877662, "grad_norm": 0.08040831390332452, "learning_rate": 4.5262332331585056e-07, "loss": 0.8199, "step": 10559 }, { "epoch": 4.784775713638423, "grad_norm": 0.0757611227918307, "learning_rate": 4.5072711824182493e-07, "loss": 0.816, "step": 10560 }, { "epoch": 4.785228817399185, "grad_norm": 0.07095883857157224, "learning_rate": 4.488348709535606e-07, "loss": 0.8587, "step": 10561 }, { "epoch": 4.785681921159946, "grad_norm": 0.07880481195496972, "learning_rate": 4.4694658164040837e-07, "loss": 0.8409, "step": 10562 }, { "epoch": 4.786135024920707, "grad_norm": 0.07230570490773311, "learning_rate": 4.4506225049134597e-07, "loss": 0.8338, "step": 10563 }, { "epoch": 4.786588128681468, "grad_norm": 0.07856366290370909, "learning_rate": 4.431818776949337e-07, "loss": 0.8496, "step": 10564 }, { "epoch": 4.787041232442229, "grad_norm": 0.07987195496683405, "learning_rate": 4.4130546343935433e-07, "loss": 0.855, "step": 10565 }, { "epoch": 4.7874943362029905, "grad_norm": 0.07457733480857044, "learning_rate": 4.394330079123732e-07, "loss": 0.8206, "step": 10566 }, { "epoch": 4.787947439963752, "grad_norm": 0.08969660614168999, "learning_rate": 4.375645113013782e-07, "loss": 0.8471, "step": 10567 }, { "epoch": 4.788400543724513, "grad_norm": 0.07609638809037945, "learning_rate": 4.356999737933576e-07, "loss": 0.8418, "step": 10568 }, { "epoch": 4.788853647485274, "grad_norm": 0.07157685801209832, "learning_rate": 4.338393955748865e-07, "loss": 0.8429, "step": 10569 }, { "epoch": 4.789306751246035, "grad_norm": 0.07230462950359456, "learning_rate": 4.319827768321716e-07, "loss": 0.843, "step": 10570 }, { "epoch": 4.789759855006796, "grad_norm": 0.07477046444477423, "learning_rate": 4.30130117751002e-07, "loss": 0.8318, "step": 10571 }, { "epoch": 4.790212958767558, "grad_norm": 0.07781294086178636, "learning_rate": 4.282814185167805e-07, "loss": 0.8654, "step": 10572 }, { "epoch": 4.790666062528319, "grad_norm": 0.077520545021483, "learning_rate": 4.264366793145103e-07, "loss": 0.8411, "step": 10573 }, { "epoch": 4.79111916628908, "grad_norm": 0.08055572687779304, "learning_rate": 4.245959003287947e-07, "loss": 0.85, "step": 10574 }, { "epoch": 4.7915722700498415, "grad_norm": 0.07962244356279113, "learning_rate": 4.227590817438465e-07, "loss": 0.8634, "step": 10575 }, { "epoch": 4.792025373810603, "grad_norm": 0.06804642860904744, "learning_rate": 4.2092622374347857e-07, "loss": 0.839, "step": 10576 }, { "epoch": 4.792478477571364, "grad_norm": 0.08113596341467265, "learning_rate": 4.1909732651111755e-07, "loss": 0.8532, "step": 10577 }, { "epoch": 4.792931581332125, "grad_norm": 0.07881451227390832, "learning_rate": 4.172723902297771e-07, "loss": 0.8428, "step": 10578 }, { "epoch": 4.793384685092886, "grad_norm": 0.07438319589080154, "learning_rate": 4.154514150820843e-07, "loss": 0.8297, "step": 10579 }, { "epoch": 4.793837788853647, "grad_norm": 0.07858036868478797, "learning_rate": 4.136344012502669e-07, "loss": 0.8427, "step": 10580 }, { "epoch": 4.794290892614408, "grad_norm": 0.08165883898895374, "learning_rate": 4.11821348916166e-07, "loss": 0.8544, "step": 10581 }, { "epoch": 4.7947439963751695, "grad_norm": 0.08635106005814312, "learning_rate": 4.100122582612098e-07, "loss": 0.8615, "step": 10582 }, { "epoch": 4.7951971001359315, "grad_norm": 0.08405247786324305, "learning_rate": 4.0820712946643583e-07, "loss": 0.8545, "step": 10583 }, { "epoch": 4.795650203896693, "grad_norm": 0.07364734743494042, "learning_rate": 4.06405962712495e-07, "loss": 0.8316, "step": 10584 }, { "epoch": 4.796103307657454, "grad_norm": 0.07499299707341804, "learning_rate": 4.046087581796254e-07, "loss": 0.8657, "step": 10585 }, { "epoch": 4.796556411418215, "grad_norm": 0.07814359552614678, "learning_rate": 4.028155160476921e-07, "loss": 0.8394, "step": 10586 }, { "epoch": 4.797009515178976, "grad_norm": 0.07977832853421438, "learning_rate": 4.0102623649612926e-07, "loss": 0.8382, "step": 10587 }, { "epoch": 4.797462618939737, "grad_norm": 0.07643122173712183, "learning_rate": 3.9924091970401145e-07, "loss": 0.8365, "step": 10588 }, { "epoch": 4.797915722700498, "grad_norm": 0.07246445476699251, "learning_rate": 3.9745956584999134e-07, "loss": 0.847, "step": 10589 }, { "epoch": 4.798368826461259, "grad_norm": 0.07056952491348496, "learning_rate": 3.956821751123352e-07, "loss": 0.8299, "step": 10590 }, { "epoch": 4.7988219302220205, "grad_norm": 0.0731466230012473, "learning_rate": 3.9390874766890517e-07, "loss": 0.8509, "step": 10591 }, { "epoch": 4.7992750339827825, "grad_norm": 0.07135630599350926, "learning_rate": 3.921392836971771e-07, "loss": 0.8363, "step": 10592 }, { "epoch": 4.799728137743543, "grad_norm": 0.07203505182673664, "learning_rate": 3.903737833742316e-07, "loss": 0.8635, "step": 10593 }, { "epoch": 4.800181241504305, "grad_norm": 0.07228605182872906, "learning_rate": 3.886122468767317e-07, "loss": 0.8711, "step": 10594 }, { "epoch": 4.800634345265066, "grad_norm": 0.0783530777058216, "learning_rate": 3.868546743809676e-07, "loss": 0.8259, "step": 10595 }, { "epoch": 4.801087449025827, "grad_norm": 0.07706272101573707, "learning_rate": 3.851010660628207e-07, "loss": 0.8315, "step": 10596 }, { "epoch": 4.801540552786588, "grad_norm": 0.07153494138301589, "learning_rate": 3.833514220977863e-07, "loss": 0.8417, "step": 10597 }, { "epoch": 4.801993656547349, "grad_norm": 0.07824405126369022, "learning_rate": 3.8160574266094653e-07, "loss": 0.8426, "step": 10598 }, { "epoch": 4.8024467603081105, "grad_norm": 0.06885309338500875, "learning_rate": 3.798640279269927e-07, "loss": 0.8451, "step": 10599 }, { "epoch": 4.802899864068872, "grad_norm": 0.0776314622294646, "learning_rate": 3.7812627807022995e-07, "loss": 0.8467, "step": 10600 }, { "epoch": 4.803352967829633, "grad_norm": 0.07337911828799454, "learning_rate": 3.763924932645591e-07, "loss": 0.8503, "step": 10601 }, { "epoch": 4.803806071590394, "grad_norm": 0.07222710015703776, "learning_rate": 3.7466267368347687e-07, "loss": 0.8563, "step": 10602 }, { "epoch": 4.804259175351156, "grad_norm": 0.08127887852493428, "learning_rate": 3.7293681950009376e-07, "loss": 0.852, "step": 10603 }, { "epoch": 4.804712279111917, "grad_norm": 0.07396370316721146, "learning_rate": 3.712149308871249e-07, "loss": 0.834, "step": 10604 }, { "epoch": 4.805165382872678, "grad_norm": 0.07426887342603891, "learning_rate": 3.6949700801687694e-07, "loss": 0.8605, "step": 10605 }, { "epoch": 4.805618486633439, "grad_norm": 0.0735397676327576, "learning_rate": 3.677830510612701e-07, "loss": 0.8442, "step": 10606 }, { "epoch": 4.8060715903942, "grad_norm": 0.08075657377039755, "learning_rate": 3.66073060191825e-07, "loss": 0.8687, "step": 10607 }, { "epoch": 4.8065246941549615, "grad_norm": 0.07921056881627792, "learning_rate": 3.64367035579658e-07, "loss": 0.8462, "step": 10608 }, { "epoch": 4.806977797915723, "grad_norm": 0.07460011821329103, "learning_rate": 3.626649773954993e-07, "loss": 0.8261, "step": 10609 }, { "epoch": 4.807430901676484, "grad_norm": 0.07808284469722201, "learning_rate": 3.609668858096749e-07, "loss": 0.8503, "step": 10610 }, { "epoch": 4.807884005437245, "grad_norm": 0.07578323580140625, "learning_rate": 3.5927276099212426e-07, "loss": 0.8417, "step": 10611 }, { "epoch": 4.808337109198006, "grad_norm": 0.07411667556076841, "learning_rate": 3.575826031123697e-07, "loss": 0.8376, "step": 10612 }, { "epoch": 4.808790212958767, "grad_norm": 0.08335465766606236, "learning_rate": 3.558964123395647e-07, "loss": 0.8379, "step": 10613 }, { "epoch": 4.809243316719529, "grad_norm": 0.08243723536312639, "learning_rate": 3.542141888424322e-07, "loss": 0.8333, "step": 10614 }, { "epoch": 4.80969642048029, "grad_norm": 0.07391971927729221, "learning_rate": 3.525359327893352e-07, "loss": 0.8481, "step": 10615 }, { "epoch": 4.8101495242410515, "grad_norm": 0.07362675950365796, "learning_rate": 3.50861644348206e-07, "loss": 0.849, "step": 10616 }, { "epoch": 4.810602628001813, "grad_norm": 0.08634698909759503, "learning_rate": 3.4919132368659956e-07, "loss": 0.8595, "step": 10617 }, { "epoch": 4.811055731762574, "grad_norm": 0.07550509656600156, "learning_rate": 3.475249709716755e-07, "loss": 0.8579, "step": 10618 }, { "epoch": 4.811508835523335, "grad_norm": 0.07695441132415809, "learning_rate": 3.458625863701759e-07, "loss": 0.8588, "step": 10619 }, { "epoch": 4.811961939284096, "grad_norm": 0.0709429246714921, "learning_rate": 3.4420417004846994e-07, "loss": 0.8265, "step": 10620 }, { "epoch": 4.812415043044857, "grad_norm": 0.07198083230930112, "learning_rate": 3.425497221725094e-07, "loss": 0.8429, "step": 10621 }, { "epoch": 4.812868146805618, "grad_norm": 0.07543099471138429, "learning_rate": 3.408992429078728e-07, "loss": 0.8588, "step": 10622 }, { "epoch": 4.813321250566379, "grad_norm": 0.07542852643317383, "learning_rate": 3.392527324197126e-07, "loss": 0.8416, "step": 10623 }, { "epoch": 4.8137743543271405, "grad_norm": 0.07841029080883888, "learning_rate": 3.376101908728124e-07, "loss": 0.8494, "step": 10624 }, { "epoch": 4.8142274580879025, "grad_norm": 0.07252841225946235, "learning_rate": 3.3597161843152983e-07, "loss": 0.8443, "step": 10625 }, { "epoch": 4.814680561848664, "grad_norm": 0.07147119880776141, "learning_rate": 3.3433701525985794e-07, "loss": 0.8661, "step": 10626 }, { "epoch": 4.815133665609425, "grad_norm": 0.07688784976657895, "learning_rate": 3.327063815213594e-07, "loss": 0.8369, "step": 10627 }, { "epoch": 4.815586769370186, "grad_norm": 0.07272057578548272, "learning_rate": 3.3107971737922793e-07, "loss": 0.8411, "step": 10628 }, { "epoch": 4.816039873130947, "grad_norm": 0.0733567861382636, "learning_rate": 3.294570229962357e-07, "loss": 0.8452, "step": 10629 }, { "epoch": 4.816492976891708, "grad_norm": 0.07141493173072772, "learning_rate": 3.278382985347817e-07, "loss": 0.8129, "step": 10630 }, { "epoch": 4.816946080652469, "grad_norm": 0.07456512453903708, "learning_rate": 3.262235441568473e-07, "loss": 0.8541, "step": 10631 }, { "epoch": 4.8173991844132305, "grad_norm": 0.07278026709691598, "learning_rate": 3.2461276002402344e-07, "loss": 0.8427, "step": 10632 }, { "epoch": 4.817852288173992, "grad_norm": 0.07192121900981391, "learning_rate": 3.2300594629750994e-07, "loss": 0.8551, "step": 10633 }, { "epoch": 4.818305391934753, "grad_norm": 0.0773239198722489, "learning_rate": 3.214031031381026e-07, "loss": 0.8433, "step": 10634 }, { "epoch": 4.818758495695514, "grad_norm": 0.0760401747545378, "learning_rate": 3.1980423070620216e-07, "loss": 0.8395, "step": 10635 }, { "epoch": 4.819211599456276, "grad_norm": 0.07543834675160911, "learning_rate": 3.1820932916180933e-07, "loss": 0.843, "step": 10636 }, { "epoch": 4.819664703217037, "grad_norm": 0.0798616908288237, "learning_rate": 3.1661839866453435e-07, "loss": 0.8281, "step": 10637 }, { "epoch": 4.820117806977798, "grad_norm": 0.08375701244472954, "learning_rate": 3.150314393735787e-07, "loss": 0.8352, "step": 10638 }, { "epoch": 4.820570910738559, "grad_norm": 0.07291421930300639, "learning_rate": 3.134484514477576e-07, "loss": 0.8447, "step": 10639 }, { "epoch": 4.82102401449932, "grad_norm": 0.07651122847858895, "learning_rate": 3.118694350454821e-07, "loss": 0.8381, "step": 10640 }, { "epoch": 4.8214771182600815, "grad_norm": 0.07958751268112516, "learning_rate": 3.1029439032476794e-07, "loss": 0.8501, "step": 10641 }, { "epoch": 4.821930222020843, "grad_norm": 0.0787377400160604, "learning_rate": 3.087233174432358e-07, "loss": 0.8227, "step": 10642 }, { "epoch": 4.822383325781604, "grad_norm": 0.07385535701679134, "learning_rate": 3.071562165581021e-07, "loss": 0.8329, "step": 10643 }, { "epoch": 4.822836429542365, "grad_norm": 0.07317153111926786, "learning_rate": 3.05593087826197e-07, "loss": 0.8457, "step": 10644 }, { "epoch": 4.823289533303126, "grad_norm": 0.0784212524699256, "learning_rate": 3.04033931403942e-07, "loss": 0.8381, "step": 10645 }, { "epoch": 4.823742637063887, "grad_norm": 0.07657708036169564, "learning_rate": 3.0247874744735897e-07, "loss": 0.825, "step": 10646 }, { "epoch": 4.824195740824649, "grad_norm": 0.07054772820922209, "learning_rate": 3.0092753611209225e-07, "loss": 0.8375, "step": 10647 }, { "epoch": 4.82464884458541, "grad_norm": 0.07753145040378392, "learning_rate": 2.993802975533644e-07, "loss": 0.8331, "step": 10648 }, { "epoch": 4.8251019483461715, "grad_norm": 0.07719271998717255, "learning_rate": 2.978370319260204e-07, "loss": 0.8602, "step": 10649 }, { "epoch": 4.825555052106933, "grad_norm": 0.07321074238782138, "learning_rate": 2.9629773938449237e-07, "loss": 0.848, "step": 10650 }, { "epoch": 4.826008155867694, "grad_norm": 0.07778541960477063, "learning_rate": 2.9476242008281697e-07, "loss": 0.8337, "step": 10651 }, { "epoch": 4.826461259628455, "grad_norm": 0.07756437017286195, "learning_rate": 2.932310741746447e-07, "loss": 0.8421, "step": 10652 }, { "epoch": 4.826914363389216, "grad_norm": 0.0776438961198303, "learning_rate": 2.9170370181321296e-07, "loss": 0.8502, "step": 10653 }, { "epoch": 4.827367467149977, "grad_norm": 0.07963671360483568, "learning_rate": 2.901803031513817e-07, "loss": 0.8249, "step": 10654 }, { "epoch": 4.827820570910738, "grad_norm": 0.07299918790990505, "learning_rate": 2.886608783415845e-07, "loss": 0.8572, "step": 10655 }, { "epoch": 4.8282736746715, "grad_norm": 0.07246039938843495, "learning_rate": 2.871454275358909e-07, "loss": 0.8706, "step": 10656 }, { "epoch": 4.8287267784322605, "grad_norm": 0.07654879017460331, "learning_rate": 2.8563395088594403e-07, "loss": 0.8404, "step": 10657 }, { "epoch": 4.8291798821930225, "grad_norm": 0.07731904944862628, "learning_rate": 2.8412644854300066e-07, "loss": 0.8358, "step": 10658 }, { "epoch": 4.829632985953784, "grad_norm": 0.07810290143988424, "learning_rate": 2.8262292065792673e-07, "loss": 0.8622, "step": 10659 }, { "epoch": 4.830086089714545, "grad_norm": 0.07551871546158218, "learning_rate": 2.8112336738117976e-07, "loss": 0.8597, "step": 10660 }, { "epoch": 4.830539193475306, "grad_norm": 0.07672617335577363, "learning_rate": 2.7962778886282624e-07, "loss": 0.882, "step": 10661 }, { "epoch": 4.830992297236067, "grad_norm": 0.071753976110385, "learning_rate": 2.781361852525244e-07, "loss": 0.8299, "step": 10662 }, { "epoch": 4.831445400996828, "grad_norm": 0.08405601176185298, "learning_rate": 2.7664855669955025e-07, "loss": 0.8411, "step": 10663 }, { "epoch": 4.831898504757589, "grad_norm": 0.07155125703328684, "learning_rate": 2.7516490335277147e-07, "loss": 0.8474, "step": 10664 }, { "epoch": 4.83235160851835, "grad_norm": 0.07543925684604515, "learning_rate": 2.736852253606648e-07, "loss": 0.8499, "step": 10665 }, { "epoch": 4.832804712279112, "grad_norm": 0.09366564038956608, "learning_rate": 2.7220952287129397e-07, "loss": 0.8401, "step": 10666 }, { "epoch": 4.833257816039874, "grad_norm": 0.07425011373781415, "learning_rate": 2.707377960323454e-07, "loss": 0.8297, "step": 10667 }, { "epoch": 4.833710919800635, "grad_norm": 0.07564414468311251, "learning_rate": 2.6927004499109676e-07, "loss": 0.8749, "step": 10668 }, { "epoch": 4.834164023561396, "grad_norm": 0.07292189299624915, "learning_rate": 2.678062698944306e-07, "loss": 0.8266, "step": 10669 }, { "epoch": 4.834617127322157, "grad_norm": 0.07258696291002373, "learning_rate": 2.6634647088882524e-07, "loss": 0.8571, "step": 10670 }, { "epoch": 4.835070231082918, "grad_norm": 0.08032032556339364, "learning_rate": 2.648906481203639e-07, "loss": 0.8442, "step": 10671 }, { "epoch": 4.835523334843679, "grad_norm": 0.08136539567676745, "learning_rate": 2.634388017347478e-07, "loss": 0.8629, "step": 10672 }, { "epoch": 4.83597643860444, "grad_norm": 0.08034664515305566, "learning_rate": 2.619909318772473e-07, "loss": 0.8574, "step": 10673 }, { "epoch": 4.8364295423652015, "grad_norm": 0.07300671343852777, "learning_rate": 2.605470386927733e-07, "loss": 0.8324, "step": 10674 }, { "epoch": 4.836882646125963, "grad_norm": 0.07548743162409492, "learning_rate": 2.591071223258057e-07, "loss": 0.8549, "step": 10675 }, { "epoch": 4.837335749886724, "grad_norm": 0.07382601556652822, "learning_rate": 2.57671182920447e-07, "loss": 0.8414, "step": 10676 }, { "epoch": 4.837788853647485, "grad_norm": 0.07237052612593776, "learning_rate": 2.562392206203912e-07, "loss": 0.852, "step": 10677 }, { "epoch": 4.838241957408247, "grad_norm": 0.07418440262804513, "learning_rate": 2.5481123556894136e-07, "loss": 0.8313, "step": 10678 }, { "epoch": 4.838695061169008, "grad_norm": 0.0748017482099704, "learning_rate": 2.533872279089966e-07, "loss": 0.8372, "step": 10679 }, { "epoch": 4.839148164929769, "grad_norm": 0.073043823028191, "learning_rate": 2.519671977830607e-07, "loss": 0.8398, "step": 10680 }, { "epoch": 4.83960126869053, "grad_norm": 0.07416236126959544, "learning_rate": 2.505511453332421e-07, "loss": 0.8428, "step": 10681 }, { "epoch": 4.840054372451291, "grad_norm": 0.07612532587177438, "learning_rate": 2.491390707012409e-07, "loss": 0.8512, "step": 10682 }, { "epoch": 4.840507476212053, "grad_norm": 0.07269472292923024, "learning_rate": 2.477309740283795e-07, "loss": 0.8395, "step": 10683 }, { "epoch": 4.840960579972814, "grad_norm": 0.07311575011230761, "learning_rate": 2.4632685545555865e-07, "loss": 0.8569, "step": 10684 }, { "epoch": 4.841413683733575, "grad_norm": 0.07511218754298389, "learning_rate": 2.449267151232926e-07, "loss": 0.8488, "step": 10685 }, { "epoch": 4.841866787494336, "grad_norm": 0.07145757244855566, "learning_rate": 2.4353055317170025e-07, "loss": 0.8273, "step": 10686 }, { "epoch": 4.842319891255097, "grad_norm": 0.07843838452641685, "learning_rate": 2.4213836974050107e-07, "loss": 0.8409, "step": 10687 }, { "epoch": 4.842772995015858, "grad_norm": 0.07423593439108579, "learning_rate": 2.407501649690103e-07, "loss": 0.8548, "step": 10688 }, { "epoch": 4.84322609877662, "grad_norm": 0.07537989462057937, "learning_rate": 2.3936593899614334e-07, "loss": 0.8643, "step": 10689 }, { "epoch": 4.843679202537381, "grad_norm": 0.06971876396918476, "learning_rate": 2.3798569196043397e-07, "loss": 0.8428, "step": 10690 }, { "epoch": 4.8441323062981425, "grad_norm": 0.07640022494116623, "learning_rate": 2.366094239999983e-07, "loss": 0.86, "step": 10691 }, { "epoch": 4.844585410058904, "grad_norm": 0.07313658334351647, "learning_rate": 2.3523713525256618e-07, "loss": 0.8465, "step": 10692 }, { "epoch": 4.845038513819665, "grad_norm": 0.07259702828628996, "learning_rate": 2.3386882585546332e-07, "loss": 0.8252, "step": 10693 }, { "epoch": 4.845491617580426, "grad_norm": 0.08331440711129529, "learning_rate": 2.325044959456202e-07, "loss": 0.8439, "step": 10694 }, { "epoch": 4.845944721341187, "grad_norm": 0.07511566999516874, "learning_rate": 2.31144145659572e-07, "loss": 0.8484, "step": 10695 }, { "epoch": 4.846397825101948, "grad_norm": 0.06749146792845653, "learning_rate": 2.2978777513344985e-07, "loss": 0.8266, "step": 10696 }, { "epoch": 4.846850928862709, "grad_norm": 0.07360869099344798, "learning_rate": 2.284353845029852e-07, "loss": 0.8226, "step": 10697 }, { "epoch": 4.84730403262347, "grad_norm": 0.07272079665595153, "learning_rate": 2.2708697390351864e-07, "loss": 0.8488, "step": 10698 }, { "epoch": 4.847757136384232, "grad_norm": 0.07279194539051972, "learning_rate": 2.2574254346999113e-07, "loss": 0.8431, "step": 10699 }, { "epoch": 4.848210240144994, "grad_norm": 0.0734192626874798, "learning_rate": 2.244020933369351e-07, "loss": 0.8296, "step": 10700 }, { "epoch": 4.848663343905755, "grad_norm": 0.0728535582442329, "learning_rate": 2.2306562363850093e-07, "loss": 0.8594, "step": 10701 }, { "epoch": 4.849116447666516, "grad_norm": 0.1156096225089671, "learning_rate": 2.2173313450842615e-07, "loss": 0.8448, "step": 10702 }, { "epoch": 4.849569551427277, "grad_norm": 0.07430740848277612, "learning_rate": 2.2040462608006186e-07, "loss": 0.8475, "step": 10703 }, { "epoch": 4.850022655188038, "grad_norm": 0.07502397085071705, "learning_rate": 2.1908009848634616e-07, "loss": 0.8553, "step": 10704 }, { "epoch": 4.850475758948799, "grad_norm": 0.07434710606632425, "learning_rate": 2.177595518598352e-07, "loss": 0.8418, "step": 10705 }, { "epoch": 4.85092886270956, "grad_norm": 0.08427156495991539, "learning_rate": 2.164429863326767e-07, "loss": 0.8567, "step": 10706 }, { "epoch": 4.8513819664703215, "grad_norm": 0.07220505016538495, "learning_rate": 2.1513040203662295e-07, "loss": 0.8741, "step": 10707 }, { "epoch": 4.851835070231083, "grad_norm": 0.07206284821639158, "learning_rate": 2.138217991030267e-07, "loss": 0.8511, "step": 10708 }, { "epoch": 4.852288173991844, "grad_norm": 0.07543470817255622, "learning_rate": 2.1251717766283653e-07, "loss": 0.8597, "step": 10709 }, { "epoch": 4.852741277752605, "grad_norm": 0.07042194991758129, "learning_rate": 2.1121653784662356e-07, "loss": 0.8762, "step": 10710 }, { "epoch": 4.853194381513367, "grad_norm": 0.07258600948923405, "learning_rate": 2.0991987978453253e-07, "loss": 0.8402, "step": 10711 }, { "epoch": 4.853647485274128, "grad_norm": 0.07491807551004327, "learning_rate": 2.0862720360632638e-07, "loss": 0.8406, "step": 10712 }, { "epoch": 4.854100589034889, "grad_norm": 0.07702272435554282, "learning_rate": 2.073385094413727e-07, "loss": 0.8409, "step": 10713 }, { "epoch": 4.85455369279565, "grad_norm": 0.07475516401566118, "learning_rate": 2.0605379741862608e-07, "loss": 0.8359, "step": 10714 }, { "epoch": 4.855006796556411, "grad_norm": 0.06992721636580575, "learning_rate": 2.0477306766665038e-07, "loss": 0.8352, "step": 10715 }, { "epoch": 4.855459900317173, "grad_norm": 0.0741781188735044, "learning_rate": 2.0349632031361867e-07, "loss": 0.8547, "step": 10716 }, { "epoch": 4.855913004077934, "grad_norm": 0.07287561559216742, "learning_rate": 2.0222355548729088e-07, "loss": 0.863, "step": 10717 }, { "epoch": 4.856366107838695, "grad_norm": 0.07299840746598066, "learning_rate": 2.0095477331503631e-07, "loss": 0.8579, "step": 10718 }, { "epoch": 4.856819211599456, "grad_norm": 0.07721914589004611, "learning_rate": 1.9968997392382894e-07, "loss": 0.8638, "step": 10719 }, { "epoch": 4.857272315360218, "grad_norm": 0.07775308452530676, "learning_rate": 1.984291574402386e-07, "loss": 0.8408, "step": 10720 }, { "epoch": 4.857725419120978, "grad_norm": 0.0720188868045803, "learning_rate": 1.9717232399043551e-07, "loss": 0.8391, "step": 10721 }, { "epoch": 4.85817852288174, "grad_norm": 0.06959289907941184, "learning_rate": 1.959194737001946e-07, "loss": 0.8512, "step": 10722 }, { "epoch": 4.858631626642501, "grad_norm": 0.07887603774365082, "learning_rate": 1.9467060669489557e-07, "loss": 0.863, "step": 10723 }, { "epoch": 4.8590847304032625, "grad_norm": 0.07598850387199528, "learning_rate": 1.9342572309950957e-07, "loss": 0.8426, "step": 10724 }, { "epoch": 4.859537834164024, "grad_norm": 0.07924794678448209, "learning_rate": 1.9218482303861696e-07, "loss": 0.8537, "step": 10725 }, { "epoch": 4.859990937924785, "grad_norm": 0.07567350770335272, "learning_rate": 1.909479066363984e-07, "loss": 0.8525, "step": 10726 }, { "epoch": 4.860444041685546, "grad_norm": 0.06865490220399612, "learning_rate": 1.8971497401663487e-07, "loss": 0.8444, "step": 10727 }, { "epoch": 4.860897145446307, "grad_norm": 0.07097674444722758, "learning_rate": 1.8848602530270765e-07, "loss": 0.8252, "step": 10728 }, { "epoch": 4.861350249207068, "grad_norm": 0.07022699359216138, "learning_rate": 1.8726106061760285e-07, "loss": 0.8477, "step": 10729 }, { "epoch": 4.861803352967829, "grad_norm": 0.07256671895756482, "learning_rate": 1.860400800839024e-07, "loss": 0.8458, "step": 10730 }, { "epoch": 4.862256456728591, "grad_norm": 0.07243525187864445, "learning_rate": 1.8482308382379743e-07, "loss": 0.8313, "step": 10731 }, { "epoch": 4.862709560489352, "grad_norm": 0.07396186965891725, "learning_rate": 1.836100719590661e-07, "loss": 0.8517, "step": 10732 }, { "epoch": 4.863162664250114, "grad_norm": 0.07239738965351157, "learning_rate": 1.8240104461110907e-07, "loss": 0.8539, "step": 10733 }, { "epoch": 4.863615768010875, "grad_norm": 0.077467331533432, "learning_rate": 1.8119600190090515e-07, "loss": 0.844, "step": 10734 }, { "epoch": 4.864068871771636, "grad_norm": 0.08116381691579233, "learning_rate": 1.799949439490556e-07, "loss": 0.8526, "step": 10735 }, { "epoch": 4.864521975532397, "grad_norm": 0.07495820374881819, "learning_rate": 1.7879787087574428e-07, "loss": 0.8393, "step": 10736 }, { "epoch": 4.864975079293158, "grad_norm": 0.07282413589519977, "learning_rate": 1.776047828007732e-07, "loss": 0.8487, "step": 10737 }, { "epoch": 4.865428183053919, "grad_norm": 0.07721697961282277, "learning_rate": 1.7641567984353124e-07, "loss": 0.8575, "step": 10738 }, { "epoch": 4.86588128681468, "grad_norm": 0.07878005441258266, "learning_rate": 1.7523056212302104e-07, "loss": 0.8482, "step": 10739 }, { "epoch": 4.8663343905754415, "grad_norm": 0.06849662529936132, "learning_rate": 1.740494297578321e-07, "loss": 0.8258, "step": 10740 }, { "epoch": 4.866787494336203, "grad_norm": 0.07241983437741649, "learning_rate": 1.7287228286616774e-07, "loss": 0.844, "step": 10741 }, { "epoch": 4.867240598096965, "grad_norm": 0.07547352372030354, "learning_rate": 1.7169912156582703e-07, "loss": 0.8413, "step": 10742 }, { "epoch": 4.867693701857726, "grad_norm": 0.07759940725733465, "learning_rate": 1.7052994597420935e-07, "loss": 0.8508, "step": 10743 }, { "epoch": 4.868146805618487, "grad_norm": 0.0782595837232302, "learning_rate": 1.6936475620831894e-07, "loss": 0.8451, "step": 10744 }, { "epoch": 4.868599909379248, "grad_norm": 0.07324644960940334, "learning_rate": 1.6820355238475583e-07, "loss": 0.8483, "step": 10745 }, { "epoch": 4.869053013140009, "grad_norm": 0.07323577014483909, "learning_rate": 1.6704633461973374e-07, "loss": 0.8419, "step": 10746 }, { "epoch": 4.86950611690077, "grad_norm": 0.07623158107186298, "learning_rate": 1.658931030290445e-07, "loss": 0.8351, "step": 10747 }, { "epoch": 4.869959220661531, "grad_norm": 0.07041136253536737, "learning_rate": 1.6474385772810242e-07, "loss": 0.842, "step": 10748 }, { "epoch": 4.8704123244222925, "grad_norm": 0.07543034868942045, "learning_rate": 1.6359859883191776e-07, "loss": 0.8441, "step": 10749 }, { "epoch": 4.870865428183054, "grad_norm": 0.07433249469889455, "learning_rate": 1.6245732645509217e-07, "loss": 0.8306, "step": 10750 }, { "epoch": 4.871318531943815, "grad_norm": 0.07369316347777276, "learning_rate": 1.613200407118365e-07, "loss": 0.8454, "step": 10751 }, { "epoch": 4.871771635704576, "grad_norm": 0.0746001249110642, "learning_rate": 1.6018674171596637e-07, "loss": 0.8509, "step": 10752 }, { "epoch": 4.872224739465338, "grad_norm": 0.0739501679313314, "learning_rate": 1.5905742958088888e-07, "loss": 0.8492, "step": 10753 }, { "epoch": 4.872677843226099, "grad_norm": 0.07301962665766176, "learning_rate": 1.5793210441962025e-07, "loss": 0.8527, "step": 10754 }, { "epoch": 4.87313094698686, "grad_norm": 0.06990483696359123, "learning_rate": 1.5681076634477265e-07, "loss": 0.8349, "step": 10755 }, { "epoch": 4.873584050747621, "grad_norm": 0.07252997272967748, "learning_rate": 1.5569341546855855e-07, "loss": 0.8392, "step": 10756 }, { "epoch": 4.8740371545083825, "grad_norm": 0.07565358666376952, "learning_rate": 1.5458005190279956e-07, "loss": 0.8548, "step": 10757 }, { "epoch": 4.874490258269144, "grad_norm": 0.07202158815358263, "learning_rate": 1.5347067575890883e-07, "loss": 0.8396, "step": 10758 }, { "epoch": 4.874943362029905, "grad_norm": 0.07104484662735921, "learning_rate": 1.5236528714790422e-07, "loss": 0.8271, "step": 10759 }, { "epoch": 4.875396465790666, "grad_norm": 0.07696180220945209, "learning_rate": 1.512638861804039e-07, "loss": 0.8321, "step": 10760 }, { "epoch": 4.875849569551427, "grad_norm": 0.07707748201796391, "learning_rate": 1.5016647296662634e-07, "loss": 0.8386, "step": 10761 }, { "epoch": 4.876302673312188, "grad_norm": 0.07170991095541426, "learning_rate": 1.4907304761639928e-07, "loss": 0.8667, "step": 10762 }, { "epoch": 4.876755777072949, "grad_norm": 0.06936572024224145, "learning_rate": 1.47983610239133e-07, "loss": 0.8297, "step": 10763 }, { "epoch": 4.877208880833711, "grad_norm": 0.07380803805560886, "learning_rate": 1.4689816094386466e-07, "loss": 0.854, "step": 10764 }, { "epoch": 4.877661984594472, "grad_norm": 0.06992210291697501, "learning_rate": 1.458166998392052e-07, "loss": 0.8491, "step": 10765 }, { "epoch": 4.8781150883552336, "grad_norm": 0.07449296127815021, "learning_rate": 1.4473922703337917e-07, "loss": 0.8559, "step": 10766 }, { "epoch": 4.878568192115995, "grad_norm": 0.07733544429669642, "learning_rate": 1.436657426342203e-07, "loss": 0.8352, "step": 10767 }, { "epoch": 4.879021295876756, "grad_norm": 0.0720357345771361, "learning_rate": 1.425962467491493e-07, "loss": 0.8355, "step": 10768 }, { "epoch": 4.879474399637517, "grad_norm": 0.07529433702050715, "learning_rate": 1.415307394851917e-07, "loss": 0.8519, "step": 10769 }, { "epoch": 4.879927503398278, "grad_norm": 0.07900112924618333, "learning_rate": 1.4046922094897775e-07, "loss": 0.8346, "step": 10770 }, { "epoch": 4.880380607159039, "grad_norm": 0.06975733634398758, "learning_rate": 1.3941169124673805e-07, "loss": 0.8388, "step": 10771 }, { "epoch": 4.8808337109198, "grad_norm": 0.06979000116518086, "learning_rate": 1.3835815048429902e-07, "loss": 0.8429, "step": 10772 }, { "epoch": 4.8812868146805615, "grad_norm": 0.07203578308564139, "learning_rate": 1.3730859876708746e-07, "loss": 0.8465, "step": 10773 }, { "epoch": 4.881739918441323, "grad_norm": 0.07258988272126411, "learning_rate": 1.3626303620013936e-07, "loss": 0.8542, "step": 10774 }, { "epoch": 4.882193022202085, "grad_norm": 0.07581005936853805, "learning_rate": 1.3522146288808657e-07, "loss": 0.8594, "step": 10775 }, { "epoch": 4.882646125962846, "grad_norm": 0.0693153184136431, "learning_rate": 1.3418387893516126e-07, "loss": 0.8414, "step": 10776 }, { "epoch": 4.883099229723607, "grad_norm": 0.06932580226814986, "learning_rate": 1.3315028444519596e-07, "loss": 0.8353, "step": 10777 }, { "epoch": 4.883552333484368, "grad_norm": 0.08250064127133469, "learning_rate": 1.3212067952162345e-07, "loss": 0.8598, "step": 10778 }, { "epoch": 4.884005437245129, "grad_norm": 0.07309997478981221, "learning_rate": 1.310950642674813e-07, "loss": 0.8438, "step": 10779 }, { "epoch": 4.88445854100589, "grad_norm": 0.07022934347043812, "learning_rate": 1.3007343878540302e-07, "loss": 0.8409, "step": 10780 }, { "epoch": 4.884911644766651, "grad_norm": 0.07162090243668154, "learning_rate": 1.290558031776268e-07, "loss": 0.8456, "step": 10781 }, { "epoch": 4.8853647485274125, "grad_norm": 0.07520478710948648, "learning_rate": 1.2804215754599115e-07, "loss": 0.869, "step": 10782 }, { "epoch": 4.885817852288174, "grad_norm": 0.07675790969005386, "learning_rate": 1.2703250199192607e-07, "loss": 0.8242, "step": 10783 }, { "epoch": 4.886270956048936, "grad_norm": 0.07159877042858875, "learning_rate": 1.2602683661647964e-07, "loss": 0.8387, "step": 10784 }, { "epoch": 4.886724059809696, "grad_norm": 0.07126351427413023, "learning_rate": 1.2502516152028686e-07, "loss": 0.8352, "step": 10785 }, { "epoch": 4.887177163570458, "grad_norm": 0.07508775813534325, "learning_rate": 1.2402747680358763e-07, "loss": 0.8394, "step": 10786 }, { "epoch": 4.887630267331219, "grad_norm": 0.07322681919242742, "learning_rate": 1.23033782566222e-07, "loss": 0.86, "step": 10787 }, { "epoch": 4.88808337109198, "grad_norm": 0.06749472103815256, "learning_rate": 1.2204407890763047e-07, "loss": 0.8311, "step": 10788 }, { "epoch": 4.888536474852741, "grad_norm": 0.07555263844843199, "learning_rate": 1.210583659268627e-07, "loss": 0.8577, "step": 10789 }, { "epoch": 4.8889895786135025, "grad_norm": 0.0720726024984554, "learning_rate": 1.2007664372254647e-07, "loss": 0.8406, "step": 10790 }, { "epoch": 4.889442682374264, "grad_norm": 0.0737059936676319, "learning_rate": 1.1909891239294091e-07, "loss": 0.8375, "step": 10791 }, { "epoch": 4.889895786135025, "grad_norm": 0.07095919286563955, "learning_rate": 1.1812517203587892e-07, "loss": 0.8464, "step": 10792 }, { "epoch": 4.890348889895786, "grad_norm": 0.06990090055882554, "learning_rate": 1.1715542274880698e-07, "loss": 0.845, "step": 10793 }, { "epoch": 4.890801993656547, "grad_norm": 0.07545042881457098, "learning_rate": 1.1618966462877635e-07, "loss": 0.8267, "step": 10794 }, { "epoch": 4.891255097417309, "grad_norm": 0.07790862684280499, "learning_rate": 1.1522789777242082e-07, "loss": 0.8507, "step": 10795 }, { "epoch": 4.89170820117807, "grad_norm": 0.07275247896084218, "learning_rate": 1.1427012227600121e-07, "loss": 0.8429, "step": 10796 }, { "epoch": 4.892161304938831, "grad_norm": 0.07667367349082851, "learning_rate": 1.1331633823535193e-07, "loss": 0.8542, "step": 10797 }, { "epoch": 4.892614408699592, "grad_norm": 0.07644110457403974, "learning_rate": 1.1236654574592554e-07, "loss": 0.8399, "step": 10798 }, { "epoch": 4.8930675124603535, "grad_norm": 0.07090913549711128, "learning_rate": 1.1142074490277044e-07, "loss": 0.8368, "step": 10799 }, { "epoch": 4.893520616221115, "grad_norm": 0.07807568782175121, "learning_rate": 1.1047893580053537e-07, "loss": 0.8604, "step": 10800 }, { "epoch": 4.893973719981876, "grad_norm": 0.07614269573898794, "learning_rate": 1.0954111853346494e-07, "loss": 0.8496, "step": 10801 }, { "epoch": 4.894426823742637, "grad_norm": 0.07267995973614569, "learning_rate": 1.0860729319541297e-07, "loss": 0.8485, "step": 10802 }, { "epoch": 4.894879927503398, "grad_norm": 0.06965532214606976, "learning_rate": 1.0767745987982914e-07, "loss": 0.8662, "step": 10803 }, { "epoch": 4.895333031264159, "grad_norm": 0.07259449776887608, "learning_rate": 1.0675161867976347e-07, "loss": 0.8513, "step": 10804 }, { "epoch": 4.89578613502492, "grad_norm": 0.07588783724780687, "learning_rate": 1.058297696878663e-07, "loss": 0.8389, "step": 10805 }, { "epoch": 4.896239238785682, "grad_norm": 0.07143474875724776, "learning_rate": 1.0491191299638381e-07, "loss": 0.8478, "step": 10806 }, { "epoch": 4.8966923425464435, "grad_norm": 0.07019677574246821, "learning_rate": 1.0399804869718033e-07, "loss": 0.8368, "step": 10807 }, { "epoch": 4.897145446307205, "grad_norm": 0.07240783607778091, "learning_rate": 1.030881768817027e-07, "loss": 0.8314, "step": 10808 }, { "epoch": 4.897598550067966, "grad_norm": 0.07320675508786419, "learning_rate": 1.0218229764099807e-07, "loss": 0.8588, "step": 10809 }, { "epoch": 4.898051653828727, "grad_norm": 0.07314721464651788, "learning_rate": 1.0128041106572728e-07, "loss": 0.8181, "step": 10810 }, { "epoch": 4.898504757589488, "grad_norm": 0.07181426308371285, "learning_rate": 1.0038251724614256e-07, "loss": 0.8434, "step": 10811 }, { "epoch": 4.898957861350249, "grad_norm": 0.07472745285490227, "learning_rate": 9.948861627209649e-08, "loss": 0.8368, "step": 10812 }, { "epoch": 4.89941096511101, "grad_norm": 0.07059827134234074, "learning_rate": 9.859870823304196e-08, "loss": 0.8339, "step": 10813 }, { "epoch": 4.899864068871771, "grad_norm": 0.0710743788723303, "learning_rate": 9.771279321803662e-08, "loss": 0.8321, "step": 10814 }, { "epoch": 4.9003171726325325, "grad_norm": 0.07290149196844847, "learning_rate": 9.683087131573842e-08, "loss": 0.8404, "step": 10815 }, { "epoch": 4.900770276393294, "grad_norm": 0.07750296835885738, "learning_rate": 9.595294261440125e-08, "loss": 0.8522, "step": 10816 }, { "epoch": 4.901223380154056, "grad_norm": 0.06994680244618472, "learning_rate": 9.507900720187924e-08, "loss": 0.846, "step": 10817 }, { "epoch": 4.901676483914817, "grad_norm": 0.07626991715066322, "learning_rate": 9.420906516563133e-08, "loss": 0.8483, "step": 10818 }, { "epoch": 4.902129587675578, "grad_norm": 0.0692771336831942, "learning_rate": 9.334311659271677e-08, "loss": 0.8433, "step": 10819 }, { "epoch": 4.902582691436339, "grad_norm": 0.07323862573151227, "learning_rate": 9.248116156978626e-08, "loss": 0.838, "step": 10820 }, { "epoch": 4.9030357951971, "grad_norm": 0.07149388637187284, "learning_rate": 9.162320018310411e-08, "loss": 0.8453, "step": 10821 }, { "epoch": 4.903488898957861, "grad_norm": 0.07711335216235808, "learning_rate": 9.076923251852165e-08, "loss": 0.8376, "step": 10822 }, { "epoch": 4.9039420027186225, "grad_norm": 0.07128303487990699, "learning_rate": 8.99192586615083e-08, "loss": 0.8254, "step": 10823 }, { "epoch": 4.904395106479384, "grad_norm": 0.06947390335667207, "learning_rate": 8.907327869711158e-08, "loss": 0.8423, "step": 10824 }, { "epoch": 4.904848210240145, "grad_norm": 0.07693482006993345, "learning_rate": 8.823129270999709e-08, "loss": 0.8368, "step": 10825 }, { "epoch": 4.905301314000906, "grad_norm": 0.07842127217861287, "learning_rate": 8.739330078442632e-08, "loss": 0.82, "step": 10826 }, { "epoch": 4.905754417761667, "grad_norm": 0.0696765356014471, "learning_rate": 8.65593030042522e-08, "loss": 0.8536, "step": 10827 }, { "epoch": 4.906207521522429, "grad_norm": 0.07384425175837439, "learning_rate": 8.572929945293684e-08, "loss": 0.8723, "step": 10828 }, { "epoch": 4.90666062528319, "grad_norm": 0.07578623916212371, "learning_rate": 8.49032902135427e-08, "loss": 0.8436, "step": 10829 }, { "epoch": 4.907113729043951, "grad_norm": 0.07699926115716839, "learning_rate": 8.408127536872812e-08, "loss": 0.8514, "step": 10830 }, { "epoch": 4.907566832804712, "grad_norm": 0.07301396075757072, "learning_rate": 8.326325500075616e-08, "loss": 0.8316, "step": 10831 }, { "epoch": 4.9080199365654735, "grad_norm": 0.0717894484488198, "learning_rate": 8.24492291914858e-08, "loss": 0.8469, "step": 10832 }, { "epoch": 4.908473040326235, "grad_norm": 0.0694691006965545, "learning_rate": 8.163919802238519e-08, "loss": 0.8272, "step": 10833 }, { "epoch": 4.908926144086996, "grad_norm": 0.07085578738976205, "learning_rate": 8.083316157450505e-08, "loss": 0.8576, "step": 10834 }, { "epoch": 4.909379247847757, "grad_norm": 0.07137326666094672, "learning_rate": 8.00311199285142e-08, "loss": 0.8353, "step": 10835 }, { "epoch": 4.909832351608518, "grad_norm": 0.07339018135494224, "learning_rate": 7.923307316467289e-08, "loss": 0.8589, "step": 10836 }, { "epoch": 4.910285455369279, "grad_norm": 0.08015912452440944, "learning_rate": 7.843902136284609e-08, "loss": 0.851, "step": 10837 }, { "epoch": 4.91073855913004, "grad_norm": 0.07358323746130549, "learning_rate": 7.764896460249027e-08, "loss": 0.8538, "step": 10838 }, { "epoch": 4.911191662890802, "grad_norm": 0.0710719047832769, "learning_rate": 7.686290296267552e-08, "loss": 0.8562, "step": 10839 }, { "epoch": 4.9116447666515635, "grad_norm": 0.0727270400804039, "learning_rate": 7.608083652205889e-08, "loss": 0.8461, "step": 10840 }, { "epoch": 4.912097870412325, "grad_norm": 0.07788732409174083, "learning_rate": 7.530276535890668e-08, "loss": 0.8474, "step": 10841 }, { "epoch": 4.912550974173086, "grad_norm": 0.0671197690174914, "learning_rate": 7.452868955108105e-08, "loss": 0.8529, "step": 10842 }, { "epoch": 4.913004077933847, "grad_norm": 0.07165559906902598, "learning_rate": 7.375860917604893e-08, "loss": 0.866, "step": 10843 }, { "epoch": 4.913457181694608, "grad_norm": 0.0768761663453882, "learning_rate": 7.299252431086868e-08, "loss": 0.8409, "step": 10844 }, { "epoch": 4.913910285455369, "grad_norm": 0.07009249431739016, "learning_rate": 7.223043503220784e-08, "loss": 0.8421, "step": 10845 }, { "epoch": 4.91436338921613, "grad_norm": 0.07029273466651297, "learning_rate": 7.147234141632986e-08, "loss": 0.8371, "step": 10846 }, { "epoch": 4.914816492976891, "grad_norm": 0.07272163805228557, "learning_rate": 7.07182435390985e-08, "loss": 0.8673, "step": 10847 }, { "epoch": 4.915269596737653, "grad_norm": 0.07193349536183158, "learning_rate": 6.996814147597786e-08, "loss": 0.8527, "step": 10848 }, { "epoch": 4.915722700498414, "grad_norm": 0.07232467302622402, "learning_rate": 6.922203530202787e-08, "loss": 0.8585, "step": 10849 }, { "epoch": 4.916175804259176, "grad_norm": 0.07253173301897235, "learning_rate": 6.847992509192214e-08, "loss": 0.8464, "step": 10850 }, { "epoch": 4.916628908019937, "grad_norm": 0.07175066563311613, "learning_rate": 6.774181091992127e-08, "loss": 0.8493, "step": 10851 }, { "epoch": 4.917082011780698, "grad_norm": 0.07058124901779257, "learning_rate": 6.700769285988618e-08, "loss": 0.8325, "step": 10852 }, { "epoch": 4.917535115541459, "grad_norm": 0.07441096062720279, "learning_rate": 6.627757098528697e-08, "loss": 0.845, "step": 10853 }, { "epoch": 4.91798821930222, "grad_norm": 0.06843919243280305, "learning_rate": 6.555144536918523e-08, "loss": 0.8395, "step": 10854 }, { "epoch": 4.918441323062981, "grad_norm": 0.07648933454873903, "learning_rate": 6.482931608425169e-08, "loss": 0.8651, "step": 10855 }, { "epoch": 4.9188944268237424, "grad_norm": 0.07252924834614186, "learning_rate": 6.411118320274413e-08, "loss": 0.848, "step": 10856 }, { "epoch": 4.919347530584504, "grad_norm": 0.07301176513606382, "learning_rate": 6.339704679652947e-08, "loss": 0.8449, "step": 10857 }, { "epoch": 4.919800634345265, "grad_norm": 0.06789770373162599, "learning_rate": 6.268690693707503e-08, "loss": 0.8485, "step": 10858 }, { "epoch": 4.920253738106027, "grad_norm": 0.07704079865102462, "learning_rate": 6.198076369544836e-08, "loss": 0.8478, "step": 10859 }, { "epoch": 4.920706841866788, "grad_norm": 0.07264851544110772, "learning_rate": 6.127861714230854e-08, "loss": 0.8383, "step": 10860 }, { "epoch": 4.921159945627549, "grad_norm": 0.07035044292335381, "learning_rate": 6.058046734792822e-08, "loss": 0.8449, "step": 10861 }, { "epoch": 4.92161304938831, "grad_norm": 0.06983288384419602, "learning_rate": 5.988631438216708e-08, "loss": 0.8444, "step": 10862 }, { "epoch": 4.922066153149071, "grad_norm": 0.07570108796423204, "learning_rate": 5.919615831448955e-08, "loss": 0.8578, "step": 10863 }, { "epoch": 4.922519256909832, "grad_norm": 0.07190186965213759, "learning_rate": 5.850999921396483e-08, "loss": 0.8388, "step": 10864 }, { "epoch": 4.9229723606705935, "grad_norm": 0.0705517838864322, "learning_rate": 5.782783714925799e-08, "loss": 0.8344, "step": 10865 }, { "epoch": 4.923425464431355, "grad_norm": 0.06671086978930478, "learning_rate": 5.714967218863887e-08, "loss": 0.8479, "step": 10866 }, { "epoch": 4.923878568192116, "grad_norm": 0.06799393781665901, "learning_rate": 5.6475504399964295e-08, "loss": 0.8456, "step": 10867 }, { "epoch": 4.924331671952877, "grad_norm": 0.0728644350333181, "learning_rate": 5.580533385070475e-08, "loss": 0.8595, "step": 10868 }, { "epoch": 4.924784775713638, "grad_norm": 0.07410249174078835, "learning_rate": 5.5139160607922124e-08, "loss": 0.8504, "step": 10869 }, { "epoch": 4.9252378794744, "grad_norm": 0.07002286785116671, "learning_rate": 5.4476984738287556e-08, "loss": 0.8469, "step": 10870 }, { "epoch": 4.925690983235161, "grad_norm": 0.06974944894531698, "learning_rate": 5.381880630805914e-08, "loss": 0.843, "step": 10871 }, { "epoch": 4.926144086995922, "grad_norm": 0.07726870532038857, "learning_rate": 5.3164625383113064e-08, "loss": 0.8282, "step": 10872 }, { "epoch": 4.9265971907566835, "grad_norm": 0.07098956208144763, "learning_rate": 5.2514442028903654e-08, "loss": 0.8477, "step": 10873 }, { "epoch": 4.927050294517445, "grad_norm": 0.07189175411371534, "learning_rate": 5.1868256310503294e-08, "loss": 0.8313, "step": 10874 }, { "epoch": 4.927503398278206, "grad_norm": 0.07549679438347448, "learning_rate": 5.1226068292571374e-08, "loss": 0.8495, "step": 10875 }, { "epoch": 4.927956502038967, "grad_norm": 0.0738492275958582, "learning_rate": 5.058787803938092e-08, "loss": 0.8276, "step": 10876 }, { "epoch": 4.928409605799728, "grad_norm": 0.07285009090770653, "learning_rate": 4.9953685614787525e-08, "loss": 0.8263, "step": 10877 }, { "epoch": 4.928862709560489, "grad_norm": 0.07186044993059672, "learning_rate": 4.9323491082264864e-08, "loss": 0.8325, "step": 10878 }, { "epoch": 4.92931581332125, "grad_norm": 0.07291702021323336, "learning_rate": 4.8697294504873595e-08, "loss": 0.8311, "step": 10879 }, { "epoch": 4.929768917082011, "grad_norm": 0.07138271400910692, "learning_rate": 4.8075095945279147e-08, "loss": 0.8379, "step": 10880 }, { "epoch": 4.930222020842773, "grad_norm": 0.07066677206470634, "learning_rate": 4.7456895465747276e-08, "loss": 0.8365, "step": 10881 }, { "epoch": 4.9306751246035345, "grad_norm": 0.07408985072067277, "learning_rate": 4.6842693128144044e-08, "loss": 0.8626, "step": 10882 }, { "epoch": 4.931128228364296, "grad_norm": 0.07698443364301817, "learning_rate": 4.62324889939314e-08, "loss": 0.8361, "step": 10883 }, { "epoch": 4.931581332125057, "grad_norm": 0.07185964770653341, "learning_rate": 4.562628312417161e-08, "loss": 0.8221, "step": 10884 }, { "epoch": 4.932034435885818, "grad_norm": 0.07491437288935715, "learning_rate": 4.5024075579536145e-08, "loss": 0.8402, "step": 10885 }, { "epoch": 4.932487539646579, "grad_norm": 0.0695936282754856, "learning_rate": 4.442586642028346e-08, "loss": 0.8401, "step": 10886 }, { "epoch": 4.93294064340734, "grad_norm": 0.07183035137656323, "learning_rate": 4.383165570628123e-08, "loss": 0.8609, "step": 10887 }, { "epoch": 4.933393747168101, "grad_norm": 0.06567659393420737, "learning_rate": 4.324144349698856e-08, "loss": 0.8605, "step": 10888 }, { "epoch": 4.933846850928862, "grad_norm": 0.0717454134797749, "learning_rate": 4.2655229851473746e-08, "loss": 0.8344, "step": 10889 }, { "epoch": 4.934299954689624, "grad_norm": 0.06805489999555013, "learning_rate": 4.207301482839654e-08, "loss": 0.8408, "step": 10890 }, { "epoch": 4.934753058450385, "grad_norm": 0.07359622782597854, "learning_rate": 4.149479848603033e-08, "loss": 0.8379, "step": 10891 }, { "epoch": 4.935206162211147, "grad_norm": 0.07186713574147478, "learning_rate": 4.0920580882226615e-08, "loss": 0.8382, "step": 10892 }, { "epoch": 4.935659265971908, "grad_norm": 0.07250422432456395, "learning_rate": 4.0350362074454974e-08, "loss": 0.871, "step": 10893 }, { "epoch": 4.936112369732669, "grad_norm": 0.07516625052353038, "learning_rate": 3.9784142119776434e-08, "loss": 0.8515, "step": 10894 }, { "epoch": 4.93656547349343, "grad_norm": 0.07275098710744927, "learning_rate": 3.9221921074852345e-08, "loss": 0.8308, "step": 10895 }, { "epoch": 4.937018577254191, "grad_norm": 0.07024302181713943, "learning_rate": 3.8663698995948795e-08, "loss": 0.8509, "step": 10896 }, { "epoch": 4.937471681014952, "grad_norm": 0.07826006833380039, "learning_rate": 3.8109475938932214e-08, "loss": 0.8577, "step": 10897 }, { "epoch": 4.9379247847757135, "grad_norm": 0.07199957627583137, "learning_rate": 3.755925195925603e-08, "loss": 0.8627, "step": 10898 }, { "epoch": 4.938377888536475, "grad_norm": 0.07456855868313798, "learning_rate": 3.7013027111982845e-08, "loss": 0.8424, "step": 10899 }, { "epoch": 4.938830992297236, "grad_norm": 0.07415187667386759, "learning_rate": 3.64708014517845e-08, "loss": 0.8425, "step": 10900 }, { "epoch": 4.939284096057997, "grad_norm": 0.07183645230079458, "learning_rate": 3.5932575032910924e-08, "loss": 0.8462, "step": 10901 }, { "epoch": 4.939737199818758, "grad_norm": 0.0709109328529422, "learning_rate": 3.539834790923458e-08, "loss": 0.8432, "step": 10902 }, { "epoch": 4.94019030357952, "grad_norm": 0.07541286373285899, "learning_rate": 3.486812013420604e-08, "loss": 0.8429, "step": 10903 }, { "epoch": 4.940643407340281, "grad_norm": 0.07238782553071568, "learning_rate": 3.434189176089398e-08, "loss": 0.8613, "step": 10904 }, { "epoch": 4.941096511101042, "grad_norm": 0.07481766264242558, "learning_rate": 3.381966284195848e-08, "loss": 0.8464, "step": 10905 }, { "epoch": 4.941549614861803, "grad_norm": 0.07111473117836185, "learning_rate": 3.330143342965553e-08, "loss": 0.8325, "step": 10906 }, { "epoch": 4.942002718622565, "grad_norm": 0.0746480064653445, "learning_rate": 3.27872035758503e-08, "loss": 0.8451, "step": 10907 }, { "epoch": 4.942455822383326, "grad_norm": 0.07743673548856274, "learning_rate": 3.2276973331999414e-08, "loss": 0.8377, "step": 10908 }, { "epoch": 4.942908926144087, "grad_norm": 0.07056507790175351, "learning_rate": 3.177074274916869e-08, "loss": 0.8605, "step": 10909 }, { "epoch": 4.943362029904848, "grad_norm": 0.07053713220884879, "learning_rate": 3.126851187801094e-08, "loss": 0.8413, "step": 10910 }, { "epoch": 4.943815133665609, "grad_norm": 0.07117170235884734, "learning_rate": 3.0770280768788186e-08, "loss": 0.8529, "step": 10911 }, { "epoch": 4.944268237426371, "grad_norm": 0.07552720162839976, "learning_rate": 3.0276049471358316e-08, "loss": 0.8384, "step": 10912 }, { "epoch": 4.944721341187131, "grad_norm": 0.0737682597811539, "learning_rate": 2.978581803518399e-08, "loss": 0.8534, "step": 10913 }, { "epoch": 4.945174444947893, "grad_norm": 0.07773726743727194, "learning_rate": 2.929958650932374e-08, "loss": 0.8596, "step": 10914 }, { "epoch": 4.9456275487086545, "grad_norm": 0.07587291163982446, "learning_rate": 2.8817354942427544e-08, "loss": 0.8544, "step": 10915 }, { "epoch": 4.946080652469416, "grad_norm": 0.06972683755475392, "learning_rate": 2.8339123382763455e-08, "loss": 0.8238, "step": 10916 }, { "epoch": 4.946533756230177, "grad_norm": 0.0838330777922925, "learning_rate": 2.7864891878186528e-08, "loss": 0.841, "step": 10917 }, { "epoch": 4.946986859990938, "grad_norm": 0.07315819073026983, "learning_rate": 2.7394660476147695e-08, "loss": 0.8437, "step": 10918 }, { "epoch": 4.947439963751699, "grad_norm": 0.07826454578961596, "learning_rate": 2.692842922371153e-08, "loss": 0.8313, "step": 10919 }, { "epoch": 4.94789306751246, "grad_norm": 0.07274421111848085, "learning_rate": 2.646619816753404e-08, "loss": 0.8584, "step": 10920 }, { "epoch": 4.948346171273221, "grad_norm": 0.07113948532604145, "learning_rate": 2.6007967353867125e-08, "loss": 0.8338, "step": 10921 }, { "epoch": 4.948799275033982, "grad_norm": 0.06729170188202362, "learning_rate": 2.5553736828571874e-08, "loss": 0.8454, "step": 10922 }, { "epoch": 4.9492523787947444, "grad_norm": 0.07019657809102224, "learning_rate": 2.5103506637105257e-08, "loss": 0.8245, "step": 10923 }, { "epoch": 4.949705482555506, "grad_norm": 0.0795167344796708, "learning_rate": 2.4657276824515687e-08, "loss": 0.8562, "step": 10924 }, { "epoch": 4.950158586316267, "grad_norm": 0.07154302549253046, "learning_rate": 2.421504743546521e-08, "loss": 0.8573, "step": 10925 }, { "epoch": 4.950611690077028, "grad_norm": 0.0737763676131615, "learning_rate": 2.377681851420288e-08, "loss": 0.8629, "step": 10926 }, { "epoch": 4.951064793837789, "grad_norm": 0.07382693867508465, "learning_rate": 2.3342590104586947e-08, "loss": 0.8414, "step": 10927 }, { "epoch": 4.95151789759855, "grad_norm": 0.07229728477563442, "learning_rate": 2.2912362250075982e-08, "loss": 0.8407, "step": 10928 }, { "epoch": 4.951971001359311, "grad_norm": 0.07510797001076584, "learning_rate": 2.248613499371555e-08, "loss": 0.8581, "step": 10929 }, { "epoch": 4.952424105120072, "grad_norm": 0.07413135432890172, "learning_rate": 2.2063908378160416e-08, "loss": 0.8514, "step": 10930 }, { "epoch": 4.9528772088808335, "grad_norm": 0.0693893693827091, "learning_rate": 2.1645682445670113e-08, "loss": 0.842, "step": 10931 }, { "epoch": 4.953330312641595, "grad_norm": 0.06732325045141394, "learning_rate": 2.1231457238091168e-08, "loss": 0.8361, "step": 10932 }, { "epoch": 4.953783416402356, "grad_norm": 0.0728331085977062, "learning_rate": 2.082123279687931e-08, "loss": 0.8605, "step": 10933 }, { "epoch": 4.954236520163118, "grad_norm": 0.06904782977304443, "learning_rate": 2.0415009163086143e-08, "loss": 0.8366, "step": 10934 }, { "epoch": 4.954689623923879, "grad_norm": 0.07538690497015335, "learning_rate": 2.001278637735915e-08, "loss": 0.8361, "step": 10935 }, { "epoch": 4.95514272768464, "grad_norm": 0.07545152681013258, "learning_rate": 1.9614564479959464e-08, "loss": 0.829, "step": 10936 }, { "epoch": 4.955595831445401, "grad_norm": 0.06896728611808531, "learning_rate": 1.9220343510726324e-08, "loss": 0.8334, "step": 10937 }, { "epoch": 4.956048935206162, "grad_norm": 0.06993565137933305, "learning_rate": 1.883012350911706e-08, "loss": 0.8562, "step": 10938 }, { "epoch": 4.956502038966923, "grad_norm": 0.07128235975660362, "learning_rate": 1.8443904514184874e-08, "loss": 0.8586, "step": 10939 }, { "epoch": 4.956955142727685, "grad_norm": 0.07459892119816697, "learning_rate": 1.8061686564569968e-08, "loss": 0.8422, "step": 10940 }, { "epoch": 4.957408246488446, "grad_norm": 0.07378462697842347, "learning_rate": 1.7683469698530632e-08, "loss": 0.8519, "step": 10941 }, { "epoch": 4.957861350249207, "grad_norm": 0.0717977719087971, "learning_rate": 1.730925395391214e-08, "loss": 0.8452, "step": 10942 }, { "epoch": 4.958314454009968, "grad_norm": 0.07523032027796193, "learning_rate": 1.6939039368164544e-08, "loss": 0.8412, "step": 10943 }, { "epoch": 4.958767557770729, "grad_norm": 0.07229228164582671, "learning_rate": 1.657282597833376e-08, "loss": 0.8214, "step": 10944 }, { "epoch": 4.959220661531491, "grad_norm": 0.07035127168867364, "learning_rate": 1.6210613821070475e-08, "loss": 0.8397, "step": 10945 }, { "epoch": 4.959673765292252, "grad_norm": 0.08370299347919584, "learning_rate": 1.5852402932616807e-08, "loss": 0.8444, "step": 10946 }, { "epoch": 4.960126869053013, "grad_norm": 0.07142118847140318, "learning_rate": 1.5498193348824076e-08, "loss": 0.8243, "step": 10947 }, { "epoch": 4.9605799728137745, "grad_norm": 0.06908917453105073, "learning_rate": 1.5147985105139483e-08, "loss": 0.8582, "step": 10948 }, { "epoch": 4.961033076574536, "grad_norm": 0.07004786476631589, "learning_rate": 1.4801778236610554e-08, "loss": 0.8624, "step": 10949 }, { "epoch": 4.961486180335297, "grad_norm": 0.06966547600548893, "learning_rate": 1.445957277787624e-08, "loss": 0.8523, "step": 10950 }, { "epoch": 4.961939284096058, "grad_norm": 0.07137600792444662, "learning_rate": 1.4121368763189147e-08, "loss": 0.8135, "step": 10951 }, { "epoch": 4.962392387856819, "grad_norm": 0.0686346791308783, "learning_rate": 1.3787166226388871e-08, "loss": 0.8455, "step": 10952 }, { "epoch": 4.96284549161758, "grad_norm": 0.07150054939532091, "learning_rate": 1.345696520091977e-08, "loss": 0.8377, "step": 10953 }, { "epoch": 4.963298595378341, "grad_norm": 0.06784362069439731, "learning_rate": 1.3130765719830962e-08, "loss": 0.8535, "step": 10954 }, { "epoch": 4.963751699139102, "grad_norm": 0.07243525722248399, "learning_rate": 1.2808567815763007e-08, "loss": 0.8663, "step": 10955 }, { "epoch": 4.964204802899864, "grad_norm": 0.07222791484358705, "learning_rate": 1.2490371520956779e-08, "loss": 0.8406, "step": 10956 }, { "epoch": 4.964657906660626, "grad_norm": 0.06873509655659216, "learning_rate": 1.2176176867257916e-08, "loss": 0.8462, "step": 10957 }, { "epoch": 4.965111010421387, "grad_norm": 0.09953952125112224, "learning_rate": 1.1865983886107935e-08, "loss": 0.8358, "step": 10958 }, { "epoch": 4.965564114182148, "grad_norm": 0.07298132073594477, "learning_rate": 1.155979260854867e-08, "loss": 0.8515, "step": 10959 }, { "epoch": 4.966017217942909, "grad_norm": 0.06850620219188551, "learning_rate": 1.1257603065217838e-08, "loss": 0.8241, "step": 10960 }, { "epoch": 4.96647032170367, "grad_norm": 0.07214963403461996, "learning_rate": 1.0959415286362352e-08, "loss": 0.8453, "step": 10961 }, { "epoch": 4.966923425464431, "grad_norm": 0.07158370824953054, "learning_rate": 1.0665229301820568e-08, "loss": 0.851, "step": 10962 }, { "epoch": 4.967376529225192, "grad_norm": 0.07327074272931244, "learning_rate": 1.0375045141026719e-08, "loss": 0.8525, "step": 10963 }, { "epoch": 4.9678296329859535, "grad_norm": 0.07112301668932164, "learning_rate": 1.008886283302868e-08, "loss": 0.8514, "step": 10964 }, { "epoch": 4.968282736746715, "grad_norm": 0.07232761276675351, "learning_rate": 9.80668240646132e-09, "loss": 0.8434, "step": 10965 }, { "epoch": 4.968735840507476, "grad_norm": 0.0799263846204185, "learning_rate": 9.528503889564277e-09, "loss": 0.84, "step": 10966 }, { "epoch": 4.969188944268238, "grad_norm": 0.0742053990169122, "learning_rate": 9.25432731016862e-09, "loss": 0.829, "step": 10967 }, { "epoch": 4.969642048028999, "grad_norm": 0.07264984033612208, "learning_rate": 8.984152695723503e-09, "loss": 0.8459, "step": 10968 }, { "epoch": 4.97009515178976, "grad_norm": 0.07321360210785878, "learning_rate": 8.717980073260635e-09, "loss": 0.8417, "step": 10969 }, { "epoch": 4.970548255550521, "grad_norm": 0.07303535991790239, "learning_rate": 8.455809469412046e-09, "loss": 0.8609, "step": 10970 }, { "epoch": 4.971001359311282, "grad_norm": 0.06980639881995214, "learning_rate": 8.19764091041897e-09, "loss": 0.8397, "step": 10971 }, { "epoch": 4.971454463072043, "grad_norm": 0.07286703837916308, "learning_rate": 7.943474422114073e-09, "loss": 0.8309, "step": 10972 }, { "epoch": 4.9719075668328045, "grad_norm": 0.07419982307095961, "learning_rate": 7.693310029939228e-09, "loss": 0.8443, "step": 10973 }, { "epoch": 4.972360670593566, "grad_norm": 0.07138543298534006, "learning_rate": 7.44714775891886e-09, "loss": 0.8472, "step": 10974 }, { "epoch": 4.972813774354327, "grad_norm": 0.07044878473525279, "learning_rate": 7.204987633691041e-09, "loss": 0.8287, "step": 10975 }, { "epoch": 4.973266878115089, "grad_norm": 0.07367832510007696, "learning_rate": 6.966829678494158e-09, "loss": 0.8374, "step": 10976 }, { "epoch": 4.973719981875849, "grad_norm": 0.07205199573870547, "learning_rate": 6.7326739171535985e-09, "loss": 0.8497, "step": 10977 }, { "epoch": 4.974173085636611, "grad_norm": 0.0737969530993137, "learning_rate": 6.5025203731083894e-09, "loss": 0.835, "step": 10978 }, { "epoch": 4.974626189397372, "grad_norm": 0.06967118765973712, "learning_rate": 6.276369069384558e-09, "loss": 0.8504, "step": 10979 }, { "epoch": 4.975079293158133, "grad_norm": 0.06967120522818539, "learning_rate": 6.0542200286173305e-09, "loss": 0.8262, "step": 10980 }, { "epoch": 4.9755323969188945, "grad_norm": 0.07290807066104632, "learning_rate": 5.8360732730378125e-09, "loss": 0.8476, "step": 10981 }, { "epoch": 4.975985500679656, "grad_norm": 0.06932939881753404, "learning_rate": 5.621928824472988e-09, "loss": 0.861, "step": 10982 }, { "epoch": 4.976438604440417, "grad_norm": 0.0760297051323237, "learning_rate": 5.4117867043590455e-09, "loss": 0.8435, "step": 10983 }, { "epoch": 4.976891708201178, "grad_norm": 0.07127550842127228, "learning_rate": 5.205646933719166e-09, "loss": 0.8556, "step": 10984 }, { "epoch": 4.977344811961939, "grad_norm": 0.06873370729071894, "learning_rate": 5.0035095331857356e-09, "loss": 0.8514, "step": 10985 }, { "epoch": 4.9777979157227, "grad_norm": 0.07195496778268606, "learning_rate": 4.805374522987016e-09, "loss": 0.8634, "step": 10986 }, { "epoch": 4.978251019483462, "grad_norm": 0.06949091621271908, "learning_rate": 4.611241922947152e-09, "loss": 0.8565, "step": 10987 }, { "epoch": 4.978704123244223, "grad_norm": 0.07868013616494378, "learning_rate": 4.421111752495044e-09, "loss": 0.845, "step": 10988 }, { "epoch": 4.979157227004984, "grad_norm": 0.06999291273059569, "learning_rate": 4.234984030664357e-09, "loss": 0.8442, "step": 10989 }, { "epoch": 4.9796103307657456, "grad_norm": 0.06873689792960584, "learning_rate": 4.0528587760713114e-09, "loss": 0.8655, "step": 10990 }, { "epoch": 4.980063434526507, "grad_norm": 0.07523646602378495, "learning_rate": 3.874736006945767e-09, "loss": 0.866, "step": 10991 }, { "epoch": 4.980516538287268, "grad_norm": 0.06933280223557642, "learning_rate": 3.700615741113467e-09, "loss": 0.8542, "step": 10992 }, { "epoch": 4.980969642048029, "grad_norm": 0.0715248072113804, "learning_rate": 3.5304979959960293e-09, "loss": 0.8424, "step": 10993 }, { "epoch": 4.98142274580879, "grad_norm": 0.07785651276659412, "learning_rate": 3.3643827886242764e-09, "loss": 0.8447, "step": 10994 }, { "epoch": 4.981875849569551, "grad_norm": 0.07247311835790814, "learning_rate": 3.2022701356160256e-09, "loss": 0.8457, "step": 10995 }, { "epoch": 4.982328953330312, "grad_norm": 0.0678748805255402, "learning_rate": 3.0441600531938564e-09, "loss": 0.8524, "step": 10996 }, { "epoch": 4.9827820570910735, "grad_norm": 0.06885429897088684, "learning_rate": 2.890052557185108e-09, "loss": 0.8427, "step": 10997 }, { "epoch": 4.9832351608518355, "grad_norm": 0.07266617880691939, "learning_rate": 2.7399476630041167e-09, "loss": 0.8302, "step": 10998 }, { "epoch": 4.983688264612597, "grad_norm": 0.07403672820626671, "learning_rate": 2.5938453856788615e-09, "loss": 0.8508, "step": 10999 }, { "epoch": 4.984141368373358, "grad_norm": 0.0717707292177581, "learning_rate": 2.451745739828759e-09, "loss": 0.8342, "step": 11000 }, { "epoch": 4.984594472134119, "grad_norm": 0.07456525770476599, "learning_rate": 2.3136487396735465e-09, "loss": 0.8502, "step": 11001 }, { "epoch": 4.98504757589488, "grad_norm": 0.06997826304865759, "learning_rate": 2.179554399033279e-09, "loss": 0.855, "step": 11002 }, { "epoch": 4.985500679655641, "grad_norm": 0.07371754470299258, "learning_rate": 2.0494627313238923e-09, "loss": 0.8667, "step": 11003 }, { "epoch": 4.985953783416402, "grad_norm": 0.06617464104806317, "learning_rate": 1.923373749566082e-09, "loss": 0.8563, "step": 11004 }, { "epoch": 4.986406887177163, "grad_norm": 0.06935946453352268, "learning_rate": 1.8012874663808633e-09, "loss": 0.8365, "step": 11005 }, { "epoch": 4.9868599909379245, "grad_norm": 0.07188333161685757, "learning_rate": 1.6832038939806893e-09, "loss": 0.8339, "step": 11006 }, { "epoch": 4.987313094698686, "grad_norm": 0.0697669497529652, "learning_rate": 1.569123044187215e-09, "loss": 0.8543, "step": 11007 }, { "epoch": 4.987766198459447, "grad_norm": 0.07267865331450656, "learning_rate": 1.4590449284090924e-09, "loss": 0.8365, "step": 11008 }, { "epoch": 4.988219302220209, "grad_norm": 0.07276628597629843, "learning_rate": 1.3529695576730562e-09, "loss": 0.8505, "step": 11009 }, { "epoch": 4.98867240598097, "grad_norm": 0.07206373604332171, "learning_rate": 1.2508969425839568e-09, "loss": 0.8588, "step": 11010 }, { "epoch": 4.989125509741731, "grad_norm": 0.07181923554726904, "learning_rate": 1.1528270933647279e-09, "loss": 0.8525, "step": 11011 }, { "epoch": 4.989578613502492, "grad_norm": 0.06776708525660625, "learning_rate": 1.0587600198253001e-09, "loss": 0.8317, "step": 11012 }, { "epoch": 4.990031717263253, "grad_norm": 0.06854730273630649, "learning_rate": 9.686957313803647e-10, "loss": 0.8332, "step": 11013 }, { "epoch": 4.9904848210240145, "grad_norm": 0.07514646098908943, "learning_rate": 8.826342370404916e-10, "loss": 0.8482, "step": 11014 }, { "epoch": 4.990937924784776, "grad_norm": 0.06840233305679691, "learning_rate": 8.005755454210118e-10, "loss": 0.835, "step": 11015 }, { "epoch": 4.991391028545537, "grad_norm": 0.07085020595752846, "learning_rate": 7.225196647286936e-10, "loss": 0.8408, "step": 11016 }, { "epoch": 4.991844132306298, "grad_norm": 0.07395682234800999, "learning_rate": 6.484666027839481e-10, "loss": 0.8532, "step": 11017 }, { "epoch": 4.992297236067059, "grad_norm": 0.071245657346068, "learning_rate": 5.784163669853016e-10, "loss": 0.857, "step": 11018 }, { "epoch": 4.99275033982782, "grad_norm": 0.07611341489036269, "learning_rate": 5.123689643538043e-10, "loss": 0.8453, "step": 11019 }, { "epoch": 4.993203443588582, "grad_norm": 0.06724606995727175, "learning_rate": 4.503244014975039e-10, "loss": 0.8467, "step": 11020 }, { "epoch": 4.993656547349343, "grad_norm": 0.06828452152608225, "learning_rate": 3.922826846158856e-10, "loss": 0.8516, "step": 11021 }, { "epoch": 4.994109651110104, "grad_norm": 0.06982889042121782, "learning_rate": 3.38243819530959e-10, "loss": 0.8413, "step": 11022 }, { "epoch": 4.9945627548708655, "grad_norm": 0.07076145027454751, "learning_rate": 2.8820781163840796e-10, "loss": 0.8341, "step": 11023 }, { "epoch": 4.995015858631627, "grad_norm": 0.07023813842604175, "learning_rate": 2.421746659519997e-10, "loss": 0.8238, "step": 11024 }, { "epoch": 4.995468962392388, "grad_norm": 0.07389347427074382, "learning_rate": 2.0014438707693927e-10, "loss": 0.8276, "step": 11025 }, { "epoch": 4.995922066153149, "grad_norm": 0.06890311746416358, "learning_rate": 1.6211697921875158e-10, "loss": 0.8448, "step": 11026 }, { "epoch": 4.99637516991391, "grad_norm": 0.06962005993920836, "learning_rate": 1.280924461832811e-10, "loss": 0.8345, "step": 11027 }, { "epoch": 4.996828273674671, "grad_norm": 0.07215682487325359, "learning_rate": 9.807079137669207e-11, "loss": 0.8509, "step": 11028 }, { "epoch": 4.997281377435432, "grad_norm": 0.07483164192715988, "learning_rate": 7.205201780102755e-11, "loss": 0.8399, "step": 11029 }, { "epoch": 4.9977344811961935, "grad_norm": 0.06878222628987432, "learning_rate": 5.0036128063091214e-11, "loss": 0.8523, "step": 11030 }, { "epoch": 4.9981875849569555, "grad_norm": 0.07076289114723995, "learning_rate": 3.202312436556554e-11, "loss": 0.8624, "step": 11031 }, { "epoch": 4.998640688717717, "grad_norm": 0.06873152215606901, "learning_rate": 1.801300850701182e-11, "loss": 0.8327, "step": 11032 }, { "epoch": 4.999093792478478, "grad_norm": 0.07162044401841175, "learning_rate": 8.005781890751962e-12, "loss": 0.8559, "step": 11033 }, { "epoch": 4.999546896239239, "grad_norm": 0.07290920079421356, "learning_rate": 2.0014455248684726e-12, "loss": 0.8299, "step": 11034 }, { "epoch": 5.0, "grad_norm": 0.07171933390087797, "learning_rate": 0.0, "loss": 0.8303, "step": 11035 }, { "epoch": 5.0, "step": 11035, "total_flos": 1.8512472981897216e+17, "train_loss": 0.16914351084786516, "train_runtime": 35655.0509, "train_samples_per_second": 158.443, "train_steps_per_second": 0.309 } ], "logging_steps": 1, "max_steps": 11035, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.8512472981897216e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }