{ "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 11040, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0004528985507246377, "grad_norm": 1.5901069643977495, "learning_rate": 7.246376811594204e-08, "loss": 1.1219, "step": 1 }, { "epoch": 0.0009057971014492754, "grad_norm": 1.6117533719753023, "learning_rate": 1.4492753623188408e-07, "loss": 1.1314, "step": 2 }, { "epoch": 0.001358695652173913, "grad_norm": 1.621196694130602, "learning_rate": 2.173913043478261e-07, "loss": 1.1424, "step": 3 }, { "epoch": 0.0018115942028985507, "grad_norm": 1.6139610993399076, "learning_rate": 2.8985507246376816e-07, "loss": 1.1683, "step": 4 }, { "epoch": 0.0022644927536231885, "grad_norm": 1.5999627533459875, "learning_rate": 3.623188405797102e-07, "loss": 1.1658, "step": 5 }, { "epoch": 0.002717391304347826, "grad_norm": 1.6213884696207874, "learning_rate": 4.347826086956522e-07, "loss": 1.1346, "step": 6 }, { "epoch": 0.003170289855072464, "grad_norm": 1.6182784750282349, "learning_rate": 5.072463768115942e-07, "loss": 1.1289, "step": 7 }, { "epoch": 0.0036231884057971015, "grad_norm": 1.5977164608668764, "learning_rate": 5.797101449275363e-07, "loss": 1.1266, "step": 8 }, { "epoch": 0.004076086956521739, "grad_norm": 1.6795891530660272, "learning_rate": 6.521739130434783e-07, "loss": 1.1372, "step": 9 }, { "epoch": 0.004528985507246377, "grad_norm": 1.583548599023431, "learning_rate": 7.246376811594204e-07, "loss": 1.1191, "step": 10 }, { "epoch": 0.004981884057971014, "grad_norm": 1.6425056158637035, "learning_rate": 7.971014492753623e-07, "loss": 1.1518, "step": 11 }, { "epoch": 0.005434782608695652, "grad_norm": 1.590669951349057, "learning_rate": 8.695652173913044e-07, "loss": 1.1305, "step": 12 }, { "epoch": 0.00588768115942029, "grad_norm": 1.6069637252064564, "learning_rate": 9.420289855072465e-07, "loss": 1.1541, "step": 13 }, { "epoch": 0.006340579710144928, "grad_norm": 1.5981612334803836, "learning_rate": 1.0144927536231885e-06, "loss": 1.1213, "step": 14 }, { "epoch": 0.006793478260869565, "grad_norm": 1.5799990914329285, "learning_rate": 1.0869565217391306e-06, "loss": 1.1529, "step": 15 }, { "epoch": 0.007246376811594203, "grad_norm": 1.5633757785306062, "learning_rate": 1.1594202898550726e-06, "loss": 1.1269, "step": 16 }, { "epoch": 0.007699275362318841, "grad_norm": 1.5223074898080975, "learning_rate": 1.2318840579710147e-06, "loss": 1.1191, "step": 17 }, { "epoch": 0.008152173913043478, "grad_norm": 1.5248263458550437, "learning_rate": 1.3043478260869566e-06, "loss": 1.1388, "step": 18 }, { "epoch": 0.008605072463768116, "grad_norm": 1.5385619684480067, "learning_rate": 1.3768115942028987e-06, "loss": 1.132, "step": 19 }, { "epoch": 0.009057971014492754, "grad_norm": 1.5190653927718736, "learning_rate": 1.4492753623188408e-06, "loss": 1.1149, "step": 20 }, { "epoch": 0.009510869565217392, "grad_norm": 1.5156794216876193, "learning_rate": 1.521739130434783e-06, "loss": 1.093, "step": 21 }, { "epoch": 0.009963768115942028, "grad_norm": 1.4740873683856235, "learning_rate": 1.5942028985507246e-06, "loss": 1.0975, "step": 22 }, { "epoch": 0.010416666666666666, "grad_norm": 1.4071843665111285, "learning_rate": 1.6666666666666667e-06, "loss": 1.1308, "step": 23 }, { "epoch": 0.010869565217391304, "grad_norm": 1.3173182907605212, "learning_rate": 1.7391304347826088e-06, "loss": 1.1153, "step": 24 }, { "epoch": 0.011322463768115942, "grad_norm": 1.272939891988781, "learning_rate": 1.8115942028985508e-06, "loss": 1.0779, "step": 25 }, { "epoch": 0.01177536231884058, "grad_norm": 1.2921987901393217, "learning_rate": 1.884057971014493e-06, "loss": 1.1017, "step": 26 }, { "epoch": 0.012228260869565218, "grad_norm": 1.2637801459144555, "learning_rate": 1.956521739130435e-06, "loss": 1.1212, "step": 27 }, { "epoch": 0.012681159420289856, "grad_norm": 1.232918527857806, "learning_rate": 2.028985507246377e-06, "loss": 1.0977, "step": 28 }, { "epoch": 0.013134057971014492, "grad_norm": 1.2274720603976357, "learning_rate": 2.101449275362319e-06, "loss": 1.1202, "step": 29 }, { "epoch": 0.01358695652173913, "grad_norm": 1.17429092722276, "learning_rate": 2.173913043478261e-06, "loss": 1.0843, "step": 30 }, { "epoch": 0.014039855072463768, "grad_norm": 0.8082155192945081, "learning_rate": 2.246376811594203e-06, "loss": 1.0943, "step": 31 }, { "epoch": 0.014492753623188406, "grad_norm": 0.716883720712335, "learning_rate": 2.3188405797101453e-06, "loss": 1.0957, "step": 32 }, { "epoch": 0.014945652173913044, "grad_norm": 0.6911293649297556, "learning_rate": 2.391304347826087e-06, "loss": 1.0807, "step": 33 }, { "epoch": 0.015398550724637682, "grad_norm": 0.6720117901081291, "learning_rate": 2.4637681159420295e-06, "loss": 1.0887, "step": 34 }, { "epoch": 0.01585144927536232, "grad_norm": 0.6643065298510867, "learning_rate": 2.5362318840579714e-06, "loss": 1.0974, "step": 35 }, { "epoch": 0.016304347826086956, "grad_norm": 0.6618263036405281, "learning_rate": 2.6086956521739132e-06, "loss": 1.1162, "step": 36 }, { "epoch": 0.016757246376811596, "grad_norm": 0.6206441292854996, "learning_rate": 2.6811594202898555e-06, "loss": 1.0797, "step": 37 }, { "epoch": 0.017210144927536232, "grad_norm": 0.5738383436510391, "learning_rate": 2.7536231884057974e-06, "loss": 1.0688, "step": 38 }, { "epoch": 0.017663043478260868, "grad_norm": 0.555291068059015, "learning_rate": 2.8260869565217393e-06, "loss": 1.0981, "step": 39 }, { "epoch": 0.018115942028985508, "grad_norm": 0.5415968682349533, "learning_rate": 2.8985507246376816e-06, "loss": 1.1629, "step": 40 }, { "epoch": 0.018568840579710144, "grad_norm": 0.48931440845954055, "learning_rate": 2.9710144927536235e-06, "loss": 1.0907, "step": 41 }, { "epoch": 0.019021739130434784, "grad_norm": 0.47631444198930334, "learning_rate": 3.043478260869566e-06, "loss": 1.0991, "step": 42 }, { "epoch": 0.01947463768115942, "grad_norm": 0.3902599991231055, "learning_rate": 3.1159420289855073e-06, "loss": 1.0843, "step": 43 }, { "epoch": 0.019927536231884056, "grad_norm": 0.36559210044022483, "learning_rate": 3.188405797101449e-06, "loss": 1.079, "step": 44 }, { "epoch": 0.020380434782608696, "grad_norm": 0.4268503657344299, "learning_rate": 3.2608695652173914e-06, "loss": 1.0671, "step": 45 }, { "epoch": 0.020833333333333332, "grad_norm": 0.5419310833091411, "learning_rate": 3.3333333333333333e-06, "loss": 1.0821, "step": 46 }, { "epoch": 0.021286231884057972, "grad_norm": 0.6032555283043047, "learning_rate": 3.4057971014492756e-06, "loss": 1.0772, "step": 47 }, { "epoch": 0.021739130434782608, "grad_norm": 0.6614550052494519, "learning_rate": 3.4782608695652175e-06, "loss": 1.0677, "step": 48 }, { "epoch": 0.022192028985507248, "grad_norm": 0.6784323504761308, "learning_rate": 3.55072463768116e-06, "loss": 1.0687, "step": 49 }, { "epoch": 0.022644927536231884, "grad_norm": 0.6397392771007636, "learning_rate": 3.6231884057971017e-06, "loss": 1.0611, "step": 50 }, { "epoch": 0.02309782608695652, "grad_norm": 0.5882007208533963, "learning_rate": 3.6956521739130436e-06, "loss": 1.0585, "step": 51 }, { "epoch": 0.02355072463768116, "grad_norm": 0.5127677425957103, "learning_rate": 3.768115942028986e-06, "loss": 1.0471, "step": 52 }, { "epoch": 0.024003623188405796, "grad_norm": 0.4321474379450098, "learning_rate": 3.840579710144928e-06, "loss": 1.0548, "step": 53 }, { "epoch": 0.024456521739130436, "grad_norm": 0.3610603324277635, "learning_rate": 3.91304347826087e-06, "loss": 1.0292, "step": 54 }, { "epoch": 0.024909420289855072, "grad_norm": 0.27984449568580977, "learning_rate": 3.9855072463768115e-06, "loss": 1.0459, "step": 55 }, { "epoch": 0.025362318840579712, "grad_norm": 0.20943577633688404, "learning_rate": 4.057971014492754e-06, "loss": 1.0453, "step": 56 }, { "epoch": 0.025815217391304348, "grad_norm": 0.17210507511541204, "learning_rate": 4.130434782608696e-06, "loss": 1.071, "step": 57 }, { "epoch": 0.026268115942028984, "grad_norm": 0.16943455861222656, "learning_rate": 4.202898550724638e-06, "loss": 1.0678, "step": 58 }, { "epoch": 0.026721014492753624, "grad_norm": 0.2049392077688072, "learning_rate": 4.27536231884058e-06, "loss": 1.0428, "step": 59 }, { "epoch": 0.02717391304347826, "grad_norm": 0.24097360463107356, "learning_rate": 4.347826086956522e-06, "loss": 1.0571, "step": 60 }, { "epoch": 0.0276268115942029, "grad_norm": 0.282540568069492, "learning_rate": 4.4202898550724645e-06, "loss": 1.0536, "step": 61 }, { "epoch": 0.028079710144927536, "grad_norm": 0.2936365736634283, "learning_rate": 4.492753623188406e-06, "loss": 1.0385, "step": 62 }, { "epoch": 0.028532608695652172, "grad_norm": 0.29702599516780614, "learning_rate": 4.565217391304348e-06, "loss": 1.0597, "step": 63 }, { "epoch": 0.028985507246376812, "grad_norm": 0.2984475513414003, "learning_rate": 4.637681159420291e-06, "loss": 1.0592, "step": 64 }, { "epoch": 0.029438405797101448, "grad_norm": 0.28425400719814115, "learning_rate": 4.710144927536232e-06, "loss": 1.0372, "step": 65 }, { "epoch": 0.029891304347826088, "grad_norm": 0.2791369568340619, "learning_rate": 4.782608695652174e-06, "loss": 1.0072, "step": 66 }, { "epoch": 0.030344202898550724, "grad_norm": 0.2613763448599239, "learning_rate": 4.855072463768117e-06, "loss": 1.0289, "step": 67 }, { "epoch": 0.030797101449275364, "grad_norm": 0.22090283798981866, "learning_rate": 4.927536231884059e-06, "loss": 1.0259, "step": 68 }, { "epoch": 0.03125, "grad_norm": 0.18893538768132867, "learning_rate": 5e-06, "loss": 1.0655, "step": 69 }, { "epoch": 0.03170289855072464, "grad_norm": 0.14579535180455097, "learning_rate": 5.072463768115943e-06, "loss": 1.0566, "step": 70 }, { "epoch": 0.03215579710144927, "grad_norm": 0.12237525713541561, "learning_rate": 5.144927536231884e-06, "loss": 1.0467, "step": 71 }, { "epoch": 0.03260869565217391, "grad_norm": 0.10653949268026358, "learning_rate": 5.2173913043478265e-06, "loss": 1.0166, "step": 72 }, { "epoch": 0.03306159420289855, "grad_norm": 0.10003863429642575, "learning_rate": 5.289855072463769e-06, "loss": 1.0291, "step": 73 }, { "epoch": 0.03351449275362319, "grad_norm": 0.12964358041313304, "learning_rate": 5.362318840579711e-06, "loss": 1.0493, "step": 74 }, { "epoch": 0.033967391304347824, "grad_norm": 0.15313596428273693, "learning_rate": 5.4347826086956525e-06, "loss": 1.0289, "step": 75 }, { "epoch": 0.034420289855072464, "grad_norm": 0.15582956031416417, "learning_rate": 5.507246376811595e-06, "loss": 1.0387, "step": 76 }, { "epoch": 0.034873188405797104, "grad_norm": 0.1588250117537107, "learning_rate": 5.579710144927537e-06, "loss": 1.0421, "step": 77 }, { "epoch": 0.035326086956521736, "grad_norm": 0.1690315614405794, "learning_rate": 5.652173913043479e-06, "loss": 1.0672, "step": 78 }, { "epoch": 0.035778985507246376, "grad_norm": 0.1481025258836326, "learning_rate": 5.724637681159421e-06, "loss": 1.0116, "step": 79 }, { "epoch": 0.036231884057971016, "grad_norm": 0.14079902239747846, "learning_rate": 5.797101449275363e-06, "loss": 1.0332, "step": 80 }, { "epoch": 0.036684782608695655, "grad_norm": 0.12128333391665107, "learning_rate": 5.8695652173913055e-06, "loss": 1.0412, "step": 81 }, { "epoch": 0.03713768115942029, "grad_norm": 0.10258496673156997, "learning_rate": 5.942028985507247e-06, "loss": 1.0473, "step": 82 }, { "epoch": 0.03759057971014493, "grad_norm": 0.09022335656349463, "learning_rate": 6.014492753623189e-06, "loss": 1.0194, "step": 83 }, { "epoch": 0.03804347826086957, "grad_norm": 0.08392571630901273, "learning_rate": 6.086956521739132e-06, "loss": 1.034, "step": 84 }, { "epoch": 0.0384963768115942, "grad_norm": 0.0734757574798607, "learning_rate": 6.159420289855072e-06, "loss": 1.0443, "step": 85 }, { "epoch": 0.03894927536231884, "grad_norm": 0.08318188978589057, "learning_rate": 6.2318840579710145e-06, "loss": 1.0384, "step": 86 }, { "epoch": 0.03940217391304348, "grad_norm": 0.08506826963590854, "learning_rate": 6.304347826086958e-06, "loss": 1.0367, "step": 87 }, { "epoch": 0.03985507246376811, "grad_norm": 0.07912759321885039, "learning_rate": 6.376811594202898e-06, "loss": 1.0266, "step": 88 }, { "epoch": 0.04030797101449275, "grad_norm": 0.08039695987167923, "learning_rate": 6.449275362318841e-06, "loss": 1.0508, "step": 89 }, { "epoch": 0.04076086956521739, "grad_norm": 0.14815463694575687, "learning_rate": 6.521739130434783e-06, "loss": 1.0452, "step": 90 }, { "epoch": 0.04121376811594203, "grad_norm": 0.09419054591315917, "learning_rate": 6.594202898550725e-06, "loss": 1.0436, "step": 91 }, { "epoch": 0.041666666666666664, "grad_norm": 0.06449517378497384, "learning_rate": 6.666666666666667e-06, "loss": 1.0538, "step": 92 }, { "epoch": 0.042119565217391304, "grad_norm": 0.0954893931680474, "learning_rate": 6.739130434782609e-06, "loss": 1.0547, "step": 93 }, { "epoch": 0.042572463768115944, "grad_norm": 0.06433502355323459, "learning_rate": 6.811594202898551e-06, "loss": 1.0479, "step": 94 }, { "epoch": 0.043025362318840576, "grad_norm": 0.0687966103419091, "learning_rate": 6.884057971014493e-06, "loss": 1.0428, "step": 95 }, { "epoch": 0.043478260869565216, "grad_norm": 0.06250415510937757, "learning_rate": 6.956521739130435e-06, "loss": 1.0307, "step": 96 }, { "epoch": 0.043931159420289856, "grad_norm": 0.06420983350179202, "learning_rate": 7.028985507246377e-06, "loss": 1.0393, "step": 97 }, { "epoch": 0.044384057971014496, "grad_norm": 0.06182112653160365, "learning_rate": 7.10144927536232e-06, "loss": 1.0338, "step": 98 }, { "epoch": 0.04483695652173913, "grad_norm": 0.06201691004280668, "learning_rate": 7.173913043478261e-06, "loss": 1.0337, "step": 99 }, { "epoch": 0.04528985507246377, "grad_norm": 0.06058940617881889, "learning_rate": 7.246376811594203e-06, "loss": 1.03, "step": 100 }, { "epoch": 0.04574275362318841, "grad_norm": 0.060356111687981454, "learning_rate": 7.318840579710146e-06, "loss": 1.0469, "step": 101 }, { "epoch": 0.04619565217391304, "grad_norm": 0.058598457043853414, "learning_rate": 7.391304347826087e-06, "loss": 1.0276, "step": 102 }, { "epoch": 0.04664855072463768, "grad_norm": 0.056288331067057294, "learning_rate": 7.4637681159420295e-06, "loss": 1.0298, "step": 103 }, { "epoch": 0.04710144927536232, "grad_norm": 0.06021573661490323, "learning_rate": 7.536231884057972e-06, "loss": 1.0207, "step": 104 }, { "epoch": 0.04755434782608696, "grad_norm": 0.05857618345243527, "learning_rate": 7.608695652173914e-06, "loss": 1.0222, "step": 105 }, { "epoch": 0.04800724637681159, "grad_norm": 0.056753313513232075, "learning_rate": 7.681159420289856e-06, "loss": 1.0352, "step": 106 }, { "epoch": 0.04846014492753623, "grad_norm": 0.05317980050150279, "learning_rate": 7.753623188405797e-06, "loss": 1.0335, "step": 107 }, { "epoch": 0.04891304347826087, "grad_norm": 0.15788213779262528, "learning_rate": 7.82608695652174e-06, "loss": 1.0488, "step": 108 }, { "epoch": 0.049365942028985504, "grad_norm": 0.055437871815666305, "learning_rate": 7.898550724637682e-06, "loss": 1.0377, "step": 109 }, { "epoch": 0.049818840579710144, "grad_norm": 0.053317068585697916, "learning_rate": 7.971014492753623e-06, "loss": 1.023, "step": 110 }, { "epoch": 0.050271739130434784, "grad_norm": 0.053816794571742085, "learning_rate": 8.043478260869566e-06, "loss": 1.0475, "step": 111 }, { "epoch": 0.050724637681159424, "grad_norm": 0.05068059401274972, "learning_rate": 8.115942028985508e-06, "loss": 1.0366, "step": 112 }, { "epoch": 0.051177536231884056, "grad_norm": 0.05067202515541829, "learning_rate": 8.188405797101449e-06, "loss": 1.0374, "step": 113 }, { "epoch": 0.051630434782608696, "grad_norm": 0.05161060709043937, "learning_rate": 8.260869565217392e-06, "loss": 1.0339, "step": 114 }, { "epoch": 0.052083333333333336, "grad_norm": 0.05238334768844599, "learning_rate": 8.333333333333334e-06, "loss": 1.0386, "step": 115 }, { "epoch": 0.05253623188405797, "grad_norm": 0.05171819831902412, "learning_rate": 8.405797101449275e-06, "loss": 1.0265, "step": 116 }, { "epoch": 0.05298913043478261, "grad_norm": 0.05122795555501833, "learning_rate": 8.478260869565218e-06, "loss": 1.0383, "step": 117 }, { "epoch": 0.05344202898550725, "grad_norm": 0.0486865213385425, "learning_rate": 8.55072463768116e-06, "loss": 1.0261, "step": 118 }, { "epoch": 0.05389492753623189, "grad_norm": 0.0518003292393712, "learning_rate": 8.623188405797103e-06, "loss": 1.0579, "step": 119 }, { "epoch": 0.05434782608695652, "grad_norm": 0.0505960080756749, "learning_rate": 8.695652173913044e-06, "loss": 1.0204, "step": 120 }, { "epoch": 0.05480072463768116, "grad_norm": 0.04867523208517553, "learning_rate": 8.768115942028986e-06, "loss": 0.9972, "step": 121 }, { "epoch": 0.0552536231884058, "grad_norm": 0.04962563656763363, "learning_rate": 8.840579710144929e-06, "loss": 1.0311, "step": 122 }, { "epoch": 0.05570652173913043, "grad_norm": 0.04762335043389719, "learning_rate": 8.91304347826087e-06, "loss": 1.0145, "step": 123 }, { "epoch": 0.05615942028985507, "grad_norm": 0.04900307442489759, "learning_rate": 8.985507246376812e-06, "loss": 1.038, "step": 124 }, { "epoch": 0.05661231884057971, "grad_norm": 0.05024011914534655, "learning_rate": 9.057971014492755e-06, "loss": 1.0365, "step": 125 }, { "epoch": 0.057065217391304345, "grad_norm": 0.0490949576200117, "learning_rate": 9.130434782608697e-06, "loss": 1.0421, "step": 126 }, { "epoch": 0.057518115942028984, "grad_norm": 0.04933884187983614, "learning_rate": 9.202898550724638e-06, "loss": 1.0587, "step": 127 }, { "epoch": 0.057971014492753624, "grad_norm": 0.05812547057697181, "learning_rate": 9.275362318840581e-06, "loss": 1.0425, "step": 128 }, { "epoch": 0.058423913043478264, "grad_norm": 0.050857306264276564, "learning_rate": 9.347826086956523e-06, "loss": 1.0187, "step": 129 }, { "epoch": 0.058876811594202896, "grad_norm": 0.05045319157950115, "learning_rate": 9.420289855072464e-06, "loss": 1.0237, "step": 130 }, { "epoch": 0.059329710144927536, "grad_norm": 0.04899806915709597, "learning_rate": 9.492753623188407e-06, "loss": 1.0214, "step": 131 }, { "epoch": 0.059782608695652176, "grad_norm": 0.04562449190492754, "learning_rate": 9.565217391304349e-06, "loss": 1.0066, "step": 132 }, { "epoch": 0.06023550724637681, "grad_norm": 0.04742626780970863, "learning_rate": 9.63768115942029e-06, "loss": 1.0355, "step": 133 }, { "epoch": 0.06068840579710145, "grad_norm": 0.05012145067372524, "learning_rate": 9.710144927536233e-06, "loss": 1.0265, "step": 134 }, { "epoch": 0.06114130434782609, "grad_norm": 0.07968940507432014, "learning_rate": 9.782608695652175e-06, "loss": 1.0314, "step": 135 }, { "epoch": 0.06159420289855073, "grad_norm": 0.05175137259085276, "learning_rate": 9.855072463768118e-06, "loss": 1.0448, "step": 136 }, { "epoch": 0.06204710144927536, "grad_norm": 0.05014796609168596, "learning_rate": 9.927536231884058e-06, "loss": 1.0236, "step": 137 }, { "epoch": 0.0625, "grad_norm": 0.04735262275994438, "learning_rate": 1e-05, "loss": 1.0186, "step": 138 }, { "epoch": 0.06295289855072464, "grad_norm": 0.04982151870380111, "learning_rate": 1.0072463768115944e-05, "loss": 1.0321, "step": 139 }, { "epoch": 0.06340579710144928, "grad_norm": 0.05036579748838779, "learning_rate": 1.0144927536231885e-05, "loss": 1.0189, "step": 140 }, { "epoch": 0.06385869565217392, "grad_norm": 0.04801986853315039, "learning_rate": 1.0217391304347829e-05, "loss": 1.0165, "step": 141 }, { "epoch": 0.06431159420289854, "grad_norm": 0.04560432053966447, "learning_rate": 1.0289855072463768e-05, "loss": 1.0194, "step": 142 }, { "epoch": 0.06476449275362318, "grad_norm": 0.04827390949804055, "learning_rate": 1.036231884057971e-05, "loss": 1.0411, "step": 143 }, { "epoch": 0.06521739130434782, "grad_norm": 0.04868231432032205, "learning_rate": 1.0434782608695653e-05, "loss": 1.041, "step": 144 }, { "epoch": 0.06567028985507246, "grad_norm": 0.04766717346048875, "learning_rate": 1.0507246376811594e-05, "loss": 1.0205, "step": 145 }, { "epoch": 0.0661231884057971, "grad_norm": 0.049193130057443454, "learning_rate": 1.0579710144927538e-05, "loss": 1.0384, "step": 146 }, { "epoch": 0.06657608695652174, "grad_norm": 0.04389660711213791, "learning_rate": 1.0652173913043479e-05, "loss": 1.0176, "step": 147 }, { "epoch": 0.06702898550724638, "grad_norm": 0.04423667871976201, "learning_rate": 1.0724637681159422e-05, "loss": 1.0407, "step": 148 }, { "epoch": 0.06748188405797101, "grad_norm": 0.048757324326660606, "learning_rate": 1.0797101449275362e-05, "loss": 1.0434, "step": 149 }, { "epoch": 0.06793478260869565, "grad_norm": 0.05331003133739277, "learning_rate": 1.0869565217391305e-05, "loss": 1.0224, "step": 150 }, { "epoch": 0.06838768115942029, "grad_norm": 0.0477160624509604, "learning_rate": 1.0942028985507247e-05, "loss": 0.9997, "step": 151 }, { "epoch": 0.06884057971014493, "grad_norm": 0.04666255213909519, "learning_rate": 1.101449275362319e-05, "loss": 1.0072, "step": 152 }, { "epoch": 0.06929347826086957, "grad_norm": 0.04649986191261088, "learning_rate": 1.1086956521739131e-05, "loss": 1.0246, "step": 153 }, { "epoch": 0.06974637681159421, "grad_norm": 0.05151782175387768, "learning_rate": 1.1159420289855074e-05, "loss": 1.0185, "step": 154 }, { "epoch": 0.07019927536231885, "grad_norm": 0.051837452012233914, "learning_rate": 1.1231884057971016e-05, "loss": 1.0279, "step": 155 }, { "epoch": 0.07065217391304347, "grad_norm": 0.18492309718025485, "learning_rate": 1.1304347826086957e-05, "loss": 1.0406, "step": 156 }, { "epoch": 0.07110507246376811, "grad_norm": 0.0507493983164224, "learning_rate": 1.1376811594202899e-05, "loss": 1.035, "step": 157 }, { "epoch": 0.07155797101449275, "grad_norm": 0.04667486564069905, "learning_rate": 1.1449275362318842e-05, "loss": 1.0378, "step": 158 }, { "epoch": 0.07201086956521739, "grad_norm": 0.050693192466185075, "learning_rate": 1.1521739130434783e-05, "loss": 1.0232, "step": 159 }, { "epoch": 0.07246376811594203, "grad_norm": 0.045370371554142586, "learning_rate": 1.1594202898550726e-05, "loss": 1.0493, "step": 160 }, { "epoch": 0.07291666666666667, "grad_norm": 0.04924367312972138, "learning_rate": 1.1666666666666668e-05, "loss": 1.016, "step": 161 }, { "epoch": 0.07336956521739131, "grad_norm": 0.04617798399389067, "learning_rate": 1.1739130434782611e-05, "loss": 1.01, "step": 162 }, { "epoch": 0.07382246376811594, "grad_norm": 0.04841218501373299, "learning_rate": 1.181159420289855e-05, "loss": 1.0328, "step": 163 }, { "epoch": 0.07427536231884058, "grad_norm": 0.04922716525870731, "learning_rate": 1.1884057971014494e-05, "loss": 1.0247, "step": 164 }, { "epoch": 0.07472826086956522, "grad_norm": 0.05341424859953671, "learning_rate": 1.1956521739130435e-05, "loss": 1.031, "step": 165 }, { "epoch": 0.07518115942028986, "grad_norm": 0.05413079580404668, "learning_rate": 1.2028985507246379e-05, "loss": 0.9864, "step": 166 }, { "epoch": 0.0756340579710145, "grad_norm": 0.047654286680961236, "learning_rate": 1.210144927536232e-05, "loss": 1.0433, "step": 167 }, { "epoch": 0.07608695652173914, "grad_norm": 0.33177190077308655, "learning_rate": 1.2173913043478263e-05, "loss": 1.0349, "step": 168 }, { "epoch": 0.07653985507246377, "grad_norm": 0.0462599677792753, "learning_rate": 1.2246376811594205e-05, "loss": 1.0094, "step": 169 }, { "epoch": 0.0769927536231884, "grad_norm": 0.05122935117285231, "learning_rate": 1.2318840579710144e-05, "loss": 1.0145, "step": 170 }, { "epoch": 0.07744565217391304, "grad_norm": 0.05171596713221746, "learning_rate": 1.2391304347826088e-05, "loss": 1.0531, "step": 171 }, { "epoch": 0.07789855072463768, "grad_norm": 0.05182805246013076, "learning_rate": 1.2463768115942029e-05, "loss": 1.0395, "step": 172 }, { "epoch": 0.07835144927536232, "grad_norm": 0.05155425944396115, "learning_rate": 1.2536231884057972e-05, "loss": 1.0279, "step": 173 }, { "epoch": 0.07880434782608696, "grad_norm": 0.05086377570873808, "learning_rate": 1.2608695652173915e-05, "loss": 1.0513, "step": 174 }, { "epoch": 0.0792572463768116, "grad_norm": 0.050657040993367784, "learning_rate": 1.2681159420289857e-05, "loss": 1.0095, "step": 175 }, { "epoch": 0.07971014492753623, "grad_norm": 0.04951253221240517, "learning_rate": 1.2753623188405797e-05, "loss": 1.0193, "step": 176 }, { "epoch": 0.08016304347826086, "grad_norm": 0.056615822526777465, "learning_rate": 1.282608695652174e-05, "loss": 1.0072, "step": 177 }, { "epoch": 0.0806159420289855, "grad_norm": 0.050788640018191784, "learning_rate": 1.2898550724637681e-05, "loss": 1.0346, "step": 178 }, { "epoch": 0.08106884057971014, "grad_norm": 0.04596724056018303, "learning_rate": 1.2971014492753624e-05, "loss": 1.0036, "step": 179 }, { "epoch": 0.08152173913043478, "grad_norm": 0.0516493444790056, "learning_rate": 1.3043478260869566e-05, "loss": 1.0259, "step": 180 }, { "epoch": 0.08197463768115942, "grad_norm": 0.04927114414799267, "learning_rate": 1.3115942028985509e-05, "loss": 1.0148, "step": 181 }, { "epoch": 0.08242753623188406, "grad_norm": 0.0514350861508676, "learning_rate": 1.318840579710145e-05, "loss": 1.0177, "step": 182 }, { "epoch": 0.08288043478260869, "grad_norm": 0.12602460008664818, "learning_rate": 1.3260869565217392e-05, "loss": 1.0387, "step": 183 }, { "epoch": 0.08333333333333333, "grad_norm": 0.04951426448126516, "learning_rate": 1.3333333333333333e-05, "loss": 1.033, "step": 184 }, { "epoch": 0.08378623188405797, "grad_norm": 0.049090937758837036, "learning_rate": 1.3405797101449276e-05, "loss": 1.0349, "step": 185 }, { "epoch": 0.08423913043478261, "grad_norm": 0.05045410133787918, "learning_rate": 1.3478260869565218e-05, "loss": 1.0082, "step": 186 }, { "epoch": 0.08469202898550725, "grad_norm": 0.04650360604086939, "learning_rate": 1.3550724637681161e-05, "loss": 1.0148, "step": 187 }, { "epoch": 0.08514492753623189, "grad_norm": 0.05405165363784102, "learning_rate": 1.3623188405797103e-05, "loss": 1.0254, "step": 188 }, { "epoch": 0.08559782608695653, "grad_norm": 0.052031068922495836, "learning_rate": 1.3695652173913046e-05, "loss": 1.0417, "step": 189 }, { "epoch": 0.08605072463768115, "grad_norm": 0.04968289314482604, "learning_rate": 1.3768115942028985e-05, "loss": 0.9787, "step": 190 }, { "epoch": 0.08650362318840579, "grad_norm": 0.10499085273933548, "learning_rate": 1.3840579710144929e-05, "loss": 1.0395, "step": 191 }, { "epoch": 0.08695652173913043, "grad_norm": 0.054096027078466606, "learning_rate": 1.391304347826087e-05, "loss": 1.0111, "step": 192 }, { "epoch": 0.08740942028985507, "grad_norm": 0.05069736731803074, "learning_rate": 1.3985507246376813e-05, "loss": 1.0174, "step": 193 }, { "epoch": 0.08786231884057971, "grad_norm": 0.048994360573904897, "learning_rate": 1.4057971014492755e-05, "loss": 1.0597, "step": 194 }, { "epoch": 0.08831521739130435, "grad_norm": 0.048128360250208536, "learning_rate": 1.4130434782608698e-05, "loss": 1.0237, "step": 195 }, { "epoch": 0.08876811594202899, "grad_norm": 0.05176028104586335, "learning_rate": 1.420289855072464e-05, "loss": 1.015, "step": 196 }, { "epoch": 0.08922101449275362, "grad_norm": 0.05957958712615865, "learning_rate": 1.427536231884058e-05, "loss": 1.0325, "step": 197 }, { "epoch": 0.08967391304347826, "grad_norm": 0.04916322188686555, "learning_rate": 1.4347826086956522e-05, "loss": 1.0294, "step": 198 }, { "epoch": 0.0901268115942029, "grad_norm": 0.04988283108022925, "learning_rate": 1.4420289855072465e-05, "loss": 1.0278, "step": 199 }, { "epoch": 0.09057971014492754, "grad_norm": 0.05567674378965187, "learning_rate": 1.4492753623188407e-05, "loss": 1.034, "step": 200 }, { "epoch": 0.09103260869565218, "grad_norm": 0.05138012243945143, "learning_rate": 1.456521739130435e-05, "loss": 1.0104, "step": 201 }, { "epoch": 0.09148550724637682, "grad_norm": 0.05888168725649267, "learning_rate": 1.4637681159420291e-05, "loss": 1.0225, "step": 202 }, { "epoch": 0.09193840579710146, "grad_norm": 0.05124914009155225, "learning_rate": 1.4710144927536235e-05, "loss": 1.0405, "step": 203 }, { "epoch": 0.09239130434782608, "grad_norm": 0.0551345842078635, "learning_rate": 1.4782608695652174e-05, "loss": 1.0154, "step": 204 }, { "epoch": 0.09284420289855072, "grad_norm": 0.05267695330650723, "learning_rate": 1.4855072463768116e-05, "loss": 1.0225, "step": 205 }, { "epoch": 0.09329710144927536, "grad_norm": 0.048321248263807674, "learning_rate": 1.4927536231884059e-05, "loss": 1.0226, "step": 206 }, { "epoch": 0.09375, "grad_norm": 0.049388869524367114, "learning_rate": 1.5000000000000002e-05, "loss": 1.0353, "step": 207 }, { "epoch": 0.09420289855072464, "grad_norm": 0.33192889213110444, "learning_rate": 1.5072463768115944e-05, "loss": 1.017, "step": 208 }, { "epoch": 0.09465579710144928, "grad_norm": 0.051003864811695776, "learning_rate": 1.5144927536231887e-05, "loss": 1.024, "step": 209 }, { "epoch": 0.09510869565217392, "grad_norm": 0.05380295298056812, "learning_rate": 1.5217391304347828e-05, "loss": 1.0308, "step": 210 }, { "epoch": 0.09556159420289854, "grad_norm": 0.04761666105176789, "learning_rate": 1.5289855072463768e-05, "loss": 1.0205, "step": 211 }, { "epoch": 0.09601449275362318, "grad_norm": 0.05105580388909024, "learning_rate": 1.536231884057971e-05, "loss": 1.0257, "step": 212 }, { "epoch": 0.09646739130434782, "grad_norm": 0.052586486006659584, "learning_rate": 1.5434782608695654e-05, "loss": 1.0123, "step": 213 }, { "epoch": 0.09692028985507246, "grad_norm": 0.059317864248778066, "learning_rate": 1.5507246376811594e-05, "loss": 1.0218, "step": 214 }, { "epoch": 0.0973731884057971, "grad_norm": 0.06016350447857732, "learning_rate": 1.5579710144927537e-05, "loss": 1.018, "step": 215 }, { "epoch": 0.09782608695652174, "grad_norm": 0.05299679303579843, "learning_rate": 1.565217391304348e-05, "loss": 1.0167, "step": 216 }, { "epoch": 0.09827898550724638, "grad_norm": 0.05956857866524685, "learning_rate": 1.5724637681159423e-05, "loss": 1.0127, "step": 217 }, { "epoch": 0.09873188405797101, "grad_norm": 0.05841142433863173, "learning_rate": 1.5797101449275363e-05, "loss": 1.0371, "step": 218 }, { "epoch": 0.09918478260869565, "grad_norm": 0.05270413063932214, "learning_rate": 1.5869565217391306e-05, "loss": 1.0176, "step": 219 }, { "epoch": 0.09963768115942029, "grad_norm": 0.06254068903608159, "learning_rate": 1.5942028985507246e-05, "loss": 1.0331, "step": 220 }, { "epoch": 0.10009057971014493, "grad_norm": 0.06171546974850116, "learning_rate": 1.601449275362319e-05, "loss": 1.03, "step": 221 }, { "epoch": 0.10054347826086957, "grad_norm": 0.05036968992046223, "learning_rate": 1.6086956521739132e-05, "loss": 0.9956, "step": 222 }, { "epoch": 0.10099637681159421, "grad_norm": 0.05192457655249771, "learning_rate": 1.6159420289855076e-05, "loss": 1.0348, "step": 223 }, { "epoch": 0.10144927536231885, "grad_norm": 0.056484191168921664, "learning_rate": 1.6231884057971015e-05, "loss": 1.002, "step": 224 }, { "epoch": 0.10190217391304347, "grad_norm": 0.06098636939723506, "learning_rate": 1.630434782608696e-05, "loss": 1.0449, "step": 225 }, { "epoch": 0.10235507246376811, "grad_norm": 0.053395039172723056, "learning_rate": 1.6376811594202898e-05, "loss": 1.0112, "step": 226 }, { "epoch": 0.10280797101449275, "grad_norm": 0.05490153021139311, "learning_rate": 1.644927536231884e-05, "loss": 1.0359, "step": 227 }, { "epoch": 0.10326086956521739, "grad_norm": 0.05930377264748966, "learning_rate": 1.6521739130434785e-05, "loss": 0.995, "step": 228 }, { "epoch": 0.10371376811594203, "grad_norm": 0.0514914375952604, "learning_rate": 1.6594202898550728e-05, "loss": 1.014, "step": 229 }, { "epoch": 0.10416666666666667, "grad_norm": 0.051027774772176474, "learning_rate": 1.6666666666666667e-05, "loss": 1.0206, "step": 230 }, { "epoch": 0.10461956521739131, "grad_norm": 0.048642770912491584, "learning_rate": 1.673913043478261e-05, "loss": 1.0161, "step": 231 }, { "epoch": 0.10507246376811594, "grad_norm": 0.048345117385253204, "learning_rate": 1.681159420289855e-05, "loss": 1.0374, "step": 232 }, { "epoch": 0.10552536231884058, "grad_norm": 0.05574070503751507, "learning_rate": 1.6884057971014494e-05, "loss": 1.0516, "step": 233 }, { "epoch": 0.10597826086956522, "grad_norm": 0.048268155431951454, "learning_rate": 1.6956521739130437e-05, "loss": 1.0037, "step": 234 }, { "epoch": 0.10643115942028986, "grad_norm": 0.08816402396527556, "learning_rate": 1.702898550724638e-05, "loss": 1.0242, "step": 235 }, { "epoch": 0.1068840579710145, "grad_norm": 0.056051495993752184, "learning_rate": 1.710144927536232e-05, "loss": 1.0174, "step": 236 }, { "epoch": 0.10733695652173914, "grad_norm": 0.0520377770156378, "learning_rate": 1.7173913043478263e-05, "loss": 1.0163, "step": 237 }, { "epoch": 0.10778985507246377, "grad_norm": 0.05655404616328451, "learning_rate": 1.7246376811594206e-05, "loss": 1.0538, "step": 238 }, { "epoch": 0.1082427536231884, "grad_norm": 0.05261470845517279, "learning_rate": 1.7318840579710146e-05, "loss": 1.0393, "step": 239 }, { "epoch": 0.10869565217391304, "grad_norm": 0.05097564829999274, "learning_rate": 1.739130434782609e-05, "loss": 1.0519, "step": 240 }, { "epoch": 0.10914855072463768, "grad_norm": 0.2449070596608215, "learning_rate": 1.746376811594203e-05, "loss": 1.0358, "step": 241 }, { "epoch": 0.10960144927536232, "grad_norm": 0.28622421342357945, "learning_rate": 1.7536231884057972e-05, "loss": 1.0335, "step": 242 }, { "epoch": 0.11005434782608696, "grad_norm": 0.059098985626536774, "learning_rate": 1.7608695652173915e-05, "loss": 1.0205, "step": 243 }, { "epoch": 0.1105072463768116, "grad_norm": 0.06638695494707708, "learning_rate": 1.7681159420289858e-05, "loss": 1.0463, "step": 244 }, { "epoch": 0.11096014492753623, "grad_norm": 0.0560259517941731, "learning_rate": 1.7753623188405798e-05, "loss": 1.014, "step": 245 }, { "epoch": 0.11141304347826086, "grad_norm": 0.05623147407366819, "learning_rate": 1.782608695652174e-05, "loss": 1.0114, "step": 246 }, { "epoch": 0.1118659420289855, "grad_norm": 0.05969327174399637, "learning_rate": 1.789855072463768e-05, "loss": 1.0163, "step": 247 }, { "epoch": 0.11231884057971014, "grad_norm": 0.0522960391272492, "learning_rate": 1.7971014492753624e-05, "loss": 1.0194, "step": 248 }, { "epoch": 0.11277173913043478, "grad_norm": 0.05562303798609464, "learning_rate": 1.8043478260869567e-05, "loss": 1.0183, "step": 249 }, { "epoch": 0.11322463768115942, "grad_norm": 0.06118681432266691, "learning_rate": 1.811594202898551e-05, "loss": 1.0404, "step": 250 }, { "epoch": 0.11367753623188406, "grad_norm": 0.051101688767996915, "learning_rate": 1.818840579710145e-05, "loss": 1.0284, "step": 251 }, { "epoch": 0.11413043478260869, "grad_norm": 0.05260262274337756, "learning_rate": 1.8260869565217393e-05, "loss": 1.0208, "step": 252 }, { "epoch": 0.11458333333333333, "grad_norm": 0.049063425210795576, "learning_rate": 1.8333333333333333e-05, "loss": 1.0066, "step": 253 }, { "epoch": 0.11503623188405797, "grad_norm": 0.059723570319032394, "learning_rate": 1.8405797101449276e-05, "loss": 1.0008, "step": 254 }, { "epoch": 0.11548913043478261, "grad_norm": 0.05390911946777677, "learning_rate": 1.847826086956522e-05, "loss": 1.0177, "step": 255 }, { "epoch": 0.11594202898550725, "grad_norm": 0.052445425438059454, "learning_rate": 1.8550724637681162e-05, "loss": 1.0228, "step": 256 }, { "epoch": 0.11639492753623189, "grad_norm": 0.05616876625058797, "learning_rate": 1.8623188405797102e-05, "loss": 1.0226, "step": 257 }, { "epoch": 0.11684782608695653, "grad_norm": 0.06394645417830841, "learning_rate": 1.8695652173913045e-05, "loss": 1.0086, "step": 258 }, { "epoch": 0.11730072463768115, "grad_norm": 0.05431260801684643, "learning_rate": 1.8768115942028985e-05, "loss": 1.02, "step": 259 }, { "epoch": 0.11775362318840579, "grad_norm": 0.05742888713521318, "learning_rate": 1.8840579710144928e-05, "loss": 1.0183, "step": 260 }, { "epoch": 0.11820652173913043, "grad_norm": 0.05637411293976161, "learning_rate": 1.891304347826087e-05, "loss": 1.0403, "step": 261 }, { "epoch": 0.11865942028985507, "grad_norm": 0.057949050292597476, "learning_rate": 1.8985507246376814e-05, "loss": 0.9838, "step": 262 }, { "epoch": 0.11911231884057971, "grad_norm": 0.05428975937782594, "learning_rate": 1.9057971014492754e-05, "loss": 1.0212, "step": 263 }, { "epoch": 0.11956521739130435, "grad_norm": 0.0602875265720204, "learning_rate": 1.9130434782608697e-05, "loss": 1.0096, "step": 264 }, { "epoch": 0.12001811594202899, "grad_norm": 0.052527118609720286, "learning_rate": 1.920289855072464e-05, "loss": 1.0337, "step": 265 }, { "epoch": 0.12047101449275362, "grad_norm": 0.053037150858072715, "learning_rate": 1.927536231884058e-05, "loss": 1.0198, "step": 266 }, { "epoch": 0.12092391304347826, "grad_norm": 0.05169070508609366, "learning_rate": 1.9347826086956523e-05, "loss": 1.0349, "step": 267 }, { "epoch": 0.1213768115942029, "grad_norm": 0.057474554318268686, "learning_rate": 1.9420289855072467e-05, "loss": 0.9947, "step": 268 }, { "epoch": 0.12182971014492754, "grad_norm": 0.05099768544600623, "learning_rate": 1.9492753623188406e-05, "loss": 1.0108, "step": 269 }, { "epoch": 0.12228260869565218, "grad_norm": 0.05355095313709384, "learning_rate": 1.956521739130435e-05, "loss": 1.0215, "step": 270 }, { "epoch": 0.12273550724637682, "grad_norm": 0.059107365652766836, "learning_rate": 1.9637681159420293e-05, "loss": 1.0223, "step": 271 }, { "epoch": 0.12318840579710146, "grad_norm": 0.05554556014250621, "learning_rate": 1.9710144927536236e-05, "loss": 1.0103, "step": 272 }, { "epoch": 0.12364130434782608, "grad_norm": 0.05290612453552974, "learning_rate": 1.9782608695652176e-05, "loss": 1.0068, "step": 273 }, { "epoch": 0.12409420289855072, "grad_norm": 0.05893048020731583, "learning_rate": 1.9855072463768115e-05, "loss": 0.9955, "step": 274 }, { "epoch": 0.12454710144927536, "grad_norm": 0.05856689551970904, "learning_rate": 1.992753623188406e-05, "loss": 1.0048, "step": 275 }, { "epoch": 0.125, "grad_norm": 0.04900083349355953, "learning_rate": 2e-05, "loss": 0.9949, "step": 276 }, { "epoch": 0.12545289855072464, "grad_norm": 0.06168827582663445, "learning_rate": 2.0072463768115945e-05, "loss": 1.0076, "step": 277 }, { "epoch": 0.12590579710144928, "grad_norm": 0.05542476177293149, "learning_rate": 2.0144927536231888e-05, "loss": 0.9967, "step": 278 }, { "epoch": 0.12635869565217392, "grad_norm": 0.05080181381045281, "learning_rate": 2.0217391304347828e-05, "loss": 1.0089, "step": 279 }, { "epoch": 0.12681159420289856, "grad_norm": 0.054762216028630614, "learning_rate": 2.028985507246377e-05, "loss": 1.0145, "step": 280 }, { "epoch": 0.1272644927536232, "grad_norm": 0.05798292741109185, "learning_rate": 2.0362318840579714e-05, "loss": 1.0366, "step": 281 }, { "epoch": 0.12771739130434784, "grad_norm": 0.05044455612753559, "learning_rate": 2.0434782608695657e-05, "loss": 0.9972, "step": 282 }, { "epoch": 0.12817028985507245, "grad_norm": 0.05773642956304356, "learning_rate": 2.0507246376811594e-05, "loss": 0.9915, "step": 283 }, { "epoch": 0.1286231884057971, "grad_norm": 0.05517780190678795, "learning_rate": 2.0579710144927537e-05, "loss": 0.9942, "step": 284 }, { "epoch": 0.12907608695652173, "grad_norm": 0.061477414496427955, "learning_rate": 2.065217391304348e-05, "loss": 1.0339, "step": 285 }, { "epoch": 0.12952898550724637, "grad_norm": 0.06072199593080023, "learning_rate": 2.072463768115942e-05, "loss": 1.034, "step": 286 }, { "epoch": 0.129981884057971, "grad_norm": 0.05752880280984298, "learning_rate": 2.0797101449275363e-05, "loss": 0.997, "step": 287 }, { "epoch": 0.13043478260869565, "grad_norm": 0.05574987801320059, "learning_rate": 2.0869565217391306e-05, "loss": 1.0029, "step": 288 }, { "epoch": 0.1308876811594203, "grad_norm": 0.056949845335220614, "learning_rate": 2.094202898550725e-05, "loss": 1.0021, "step": 289 }, { "epoch": 0.13134057971014493, "grad_norm": 0.0514482783882186, "learning_rate": 2.101449275362319e-05, "loss": 0.9963, "step": 290 }, { "epoch": 0.13179347826086957, "grad_norm": 0.05864900533148441, "learning_rate": 2.1086956521739132e-05, "loss": 1.0248, "step": 291 }, { "epoch": 0.1322463768115942, "grad_norm": 0.05011767390879436, "learning_rate": 2.1159420289855075e-05, "loss": 1.0214, "step": 292 }, { "epoch": 0.13269927536231885, "grad_norm": 0.053724149824252665, "learning_rate": 2.1231884057971018e-05, "loss": 1.0031, "step": 293 }, { "epoch": 0.1331521739130435, "grad_norm": 0.05660854467320067, "learning_rate": 2.1304347826086958e-05, "loss": 1.0041, "step": 294 }, { "epoch": 0.13360507246376813, "grad_norm": 0.05527845007410962, "learning_rate": 2.13768115942029e-05, "loss": 1.012, "step": 295 }, { "epoch": 0.13405797101449277, "grad_norm": 0.06084567425602997, "learning_rate": 2.1449275362318844e-05, "loss": 1.002, "step": 296 }, { "epoch": 0.13451086956521738, "grad_norm": 0.057732702131153925, "learning_rate": 2.152173913043478e-05, "loss": 1.0023, "step": 297 }, { "epoch": 0.13496376811594202, "grad_norm": 0.058394311309446244, "learning_rate": 2.1594202898550724e-05, "loss": 1.0118, "step": 298 }, { "epoch": 0.13541666666666666, "grad_norm": 0.061061635721612044, "learning_rate": 2.1666666666666667e-05, "loss": 1.0306, "step": 299 }, { "epoch": 0.1358695652173913, "grad_norm": 0.05723893287510821, "learning_rate": 2.173913043478261e-05, "loss": 1.0064, "step": 300 }, { "epoch": 0.13632246376811594, "grad_norm": 0.05725115691278677, "learning_rate": 2.1811594202898553e-05, "loss": 1.0129, "step": 301 }, { "epoch": 0.13677536231884058, "grad_norm": 0.05380975223014023, "learning_rate": 2.1884057971014493e-05, "loss": 1.0174, "step": 302 }, { "epoch": 0.13722826086956522, "grad_norm": 0.056225039772875435, "learning_rate": 2.1956521739130436e-05, "loss": 1.0314, "step": 303 }, { "epoch": 0.13768115942028986, "grad_norm": 0.05936930420813653, "learning_rate": 2.202898550724638e-05, "loss": 1.0066, "step": 304 }, { "epoch": 0.1381340579710145, "grad_norm": 0.05622218418876298, "learning_rate": 2.2101449275362323e-05, "loss": 0.9986, "step": 305 }, { "epoch": 0.13858695652173914, "grad_norm": 0.06339828410260841, "learning_rate": 2.2173913043478262e-05, "loss": 1.0013, "step": 306 }, { "epoch": 0.13903985507246377, "grad_norm": 0.06049294001688934, "learning_rate": 2.2246376811594205e-05, "loss": 0.9966, "step": 307 }, { "epoch": 0.13949275362318841, "grad_norm": 0.058838205686178625, "learning_rate": 2.231884057971015e-05, "loss": 1.0246, "step": 308 }, { "epoch": 0.13994565217391305, "grad_norm": 0.05746468908575312, "learning_rate": 2.2391304347826092e-05, "loss": 1.0255, "step": 309 }, { "epoch": 0.1403985507246377, "grad_norm": 0.05909272704461537, "learning_rate": 2.246376811594203e-05, "loss": 1.0339, "step": 310 }, { "epoch": 0.1408514492753623, "grad_norm": 0.05762626362935249, "learning_rate": 2.253623188405797e-05, "loss": 1.0044, "step": 311 }, { "epoch": 0.14130434782608695, "grad_norm": 0.06117212391758945, "learning_rate": 2.2608695652173914e-05, "loss": 0.9911, "step": 312 }, { "epoch": 0.14175724637681159, "grad_norm": 0.07375321478142723, "learning_rate": 2.2681159420289854e-05, "loss": 1.0057, "step": 313 }, { "epoch": 0.14221014492753623, "grad_norm": 0.06070313606656927, "learning_rate": 2.2753623188405797e-05, "loss": 1.0073, "step": 314 }, { "epoch": 0.14266304347826086, "grad_norm": 0.06725624249192266, "learning_rate": 2.282608695652174e-05, "loss": 0.9974, "step": 315 }, { "epoch": 0.1431159420289855, "grad_norm": 0.058438586043938985, "learning_rate": 2.2898550724637684e-05, "loss": 1.0188, "step": 316 }, { "epoch": 0.14356884057971014, "grad_norm": 0.06108113863697551, "learning_rate": 2.2971014492753623e-05, "loss": 0.9972, "step": 317 }, { "epoch": 0.14402173913043478, "grad_norm": 0.06156088375961059, "learning_rate": 2.3043478260869567e-05, "loss": 1.0288, "step": 318 }, { "epoch": 0.14447463768115942, "grad_norm": 0.05356833778622037, "learning_rate": 2.311594202898551e-05, "loss": 1.0074, "step": 319 }, { "epoch": 0.14492753623188406, "grad_norm": 0.06431350895020135, "learning_rate": 2.3188405797101453e-05, "loss": 1.0124, "step": 320 }, { "epoch": 0.1453804347826087, "grad_norm": 0.05450433873868162, "learning_rate": 2.3260869565217396e-05, "loss": 1.0057, "step": 321 }, { "epoch": 0.14583333333333334, "grad_norm": 0.06684214117283244, "learning_rate": 2.3333333333333336e-05, "loss": 1.0109, "step": 322 }, { "epoch": 0.14628623188405798, "grad_norm": 0.05623126444944044, "learning_rate": 2.340579710144928e-05, "loss": 1.0083, "step": 323 }, { "epoch": 0.14673913043478262, "grad_norm": 0.0639211438185882, "learning_rate": 2.3478260869565222e-05, "loss": 1.0214, "step": 324 }, { "epoch": 0.14719202898550723, "grad_norm": 0.05959619075360274, "learning_rate": 2.355072463768116e-05, "loss": 1.0313, "step": 325 }, { "epoch": 0.14764492753623187, "grad_norm": 0.06918182664170716, "learning_rate": 2.36231884057971e-05, "loss": 1.0078, "step": 326 }, { "epoch": 0.1480978260869565, "grad_norm": 0.06505806525100104, "learning_rate": 2.3695652173913045e-05, "loss": 1.0183, "step": 327 }, { "epoch": 0.14855072463768115, "grad_norm": 0.057584888075937556, "learning_rate": 2.3768115942028988e-05, "loss": 1.0047, "step": 328 }, { "epoch": 0.1490036231884058, "grad_norm": 0.07676561421226301, "learning_rate": 2.3840579710144928e-05, "loss": 1.0172, "step": 329 }, { "epoch": 0.14945652173913043, "grad_norm": 0.06647007100969092, "learning_rate": 2.391304347826087e-05, "loss": 1.0183, "step": 330 }, { "epoch": 0.14990942028985507, "grad_norm": 0.06613596238939064, "learning_rate": 2.3985507246376814e-05, "loss": 0.9929, "step": 331 }, { "epoch": 0.1503623188405797, "grad_norm": 0.07004694111469989, "learning_rate": 2.4057971014492757e-05, "loss": 1.0295, "step": 332 }, { "epoch": 0.15081521739130435, "grad_norm": 0.06271800011617518, "learning_rate": 2.4130434782608697e-05, "loss": 0.9886, "step": 333 }, { "epoch": 0.151268115942029, "grad_norm": 0.07046226727267711, "learning_rate": 2.420289855072464e-05, "loss": 1.0017, "step": 334 }, { "epoch": 0.15172101449275363, "grad_norm": 0.05585887751265214, "learning_rate": 2.4275362318840583e-05, "loss": 1.0129, "step": 335 }, { "epoch": 0.15217391304347827, "grad_norm": 0.06522435574598465, "learning_rate": 2.4347826086956526e-05, "loss": 1.0135, "step": 336 }, { "epoch": 0.1526268115942029, "grad_norm": 0.06375274782084969, "learning_rate": 2.4420289855072466e-05, "loss": 1.0086, "step": 337 }, { "epoch": 0.15307971014492755, "grad_norm": 0.06280509102906394, "learning_rate": 2.449275362318841e-05, "loss": 1.0041, "step": 338 }, { "epoch": 0.15353260869565216, "grad_norm": 0.08443642895079492, "learning_rate": 2.456521739130435e-05, "loss": 1.0202, "step": 339 }, { "epoch": 0.1539855072463768, "grad_norm": 0.05876629286428871, "learning_rate": 2.463768115942029e-05, "loss": 1.0109, "step": 340 }, { "epoch": 0.15443840579710144, "grad_norm": 0.07888562603072422, "learning_rate": 2.4710144927536232e-05, "loss": 1.0073, "step": 341 }, { "epoch": 0.15489130434782608, "grad_norm": 0.06301816782977575, "learning_rate": 2.4782608695652175e-05, "loss": 1.0346, "step": 342 }, { "epoch": 0.15534420289855072, "grad_norm": 0.06420826592342997, "learning_rate": 2.4855072463768118e-05, "loss": 1.0086, "step": 343 }, { "epoch": 0.15579710144927536, "grad_norm": 0.06303436289287817, "learning_rate": 2.4927536231884058e-05, "loss": 1.0163, "step": 344 }, { "epoch": 0.15625, "grad_norm": 0.06872990881212297, "learning_rate": 2.5e-05, "loss": 1.0127, "step": 345 }, { "epoch": 0.15670289855072464, "grad_norm": 0.06421781142805305, "learning_rate": 2.5072463768115944e-05, "loss": 0.9943, "step": 346 }, { "epoch": 0.15715579710144928, "grad_norm": 0.06289600789421183, "learning_rate": 2.5144927536231888e-05, "loss": 1.0078, "step": 347 }, { "epoch": 0.15760869565217392, "grad_norm": 0.06647258800206292, "learning_rate": 2.521739130434783e-05, "loss": 1.0037, "step": 348 }, { "epoch": 0.15806159420289856, "grad_norm": 0.0636243423046751, "learning_rate": 2.528985507246377e-05, "loss": 1.0171, "step": 349 }, { "epoch": 0.1585144927536232, "grad_norm": 0.05736339224441743, "learning_rate": 2.5362318840579714e-05, "loss": 0.9974, "step": 350 }, { "epoch": 0.15896739130434784, "grad_norm": 0.05760736302700793, "learning_rate": 2.5434782608695657e-05, "loss": 1.0164, "step": 351 }, { "epoch": 0.15942028985507245, "grad_norm": 0.06678390041613036, "learning_rate": 2.5507246376811593e-05, "loss": 0.989, "step": 352 }, { "epoch": 0.1598731884057971, "grad_norm": 0.05816263153497166, "learning_rate": 2.5579710144927536e-05, "loss": 1.0043, "step": 353 }, { "epoch": 0.16032608695652173, "grad_norm": 0.06866585609082708, "learning_rate": 2.565217391304348e-05, "loss": 1.0196, "step": 354 }, { "epoch": 0.16077898550724637, "grad_norm": 0.06164435732122373, "learning_rate": 2.5724637681159423e-05, "loss": 0.9948, "step": 355 }, { "epoch": 0.161231884057971, "grad_norm": 0.06831818444963494, "learning_rate": 2.5797101449275362e-05, "loss": 1.0069, "step": 356 }, { "epoch": 0.16168478260869565, "grad_norm": 0.0784355882792059, "learning_rate": 2.5869565217391305e-05, "loss": 0.9783, "step": 357 }, { "epoch": 0.1621376811594203, "grad_norm": 0.06365875614133694, "learning_rate": 2.594202898550725e-05, "loss": 1.0064, "step": 358 }, { "epoch": 0.16259057971014493, "grad_norm": 0.07327292422557567, "learning_rate": 2.6014492753623192e-05, "loss": 1.003, "step": 359 }, { "epoch": 0.16304347826086957, "grad_norm": 0.06680041968240953, "learning_rate": 2.608695652173913e-05, "loss": 1.0314, "step": 360 }, { "epoch": 0.1634963768115942, "grad_norm": 0.07092944545556835, "learning_rate": 2.6159420289855075e-05, "loss": 1.0106, "step": 361 }, { "epoch": 0.16394927536231885, "grad_norm": 0.07929259910823967, "learning_rate": 2.6231884057971018e-05, "loss": 1.0203, "step": 362 }, { "epoch": 0.1644021739130435, "grad_norm": 0.06794803342723711, "learning_rate": 2.630434782608696e-05, "loss": 0.9985, "step": 363 }, { "epoch": 0.16485507246376813, "grad_norm": 0.06033480016355719, "learning_rate": 2.63768115942029e-05, "loss": 0.9988, "step": 364 }, { "epoch": 0.16530797101449277, "grad_norm": 0.07636292921715673, "learning_rate": 2.6449275362318844e-05, "loss": 0.9819, "step": 365 }, { "epoch": 0.16576086956521738, "grad_norm": 0.06721022539704834, "learning_rate": 2.6521739130434784e-05, "loss": 1.0076, "step": 366 }, { "epoch": 0.16621376811594202, "grad_norm": 0.08023381427703848, "learning_rate": 2.6594202898550723e-05, "loss": 1.0363, "step": 367 }, { "epoch": 0.16666666666666666, "grad_norm": 0.07434685367453489, "learning_rate": 2.6666666666666667e-05, "loss": 1.02, "step": 368 }, { "epoch": 0.1671195652173913, "grad_norm": 0.06878614646413228, "learning_rate": 2.673913043478261e-05, "loss": 1.0118, "step": 369 }, { "epoch": 0.16757246376811594, "grad_norm": 0.0875407106775544, "learning_rate": 2.6811594202898553e-05, "loss": 1.015, "step": 370 }, { "epoch": 0.16802536231884058, "grad_norm": 0.06968308320035835, "learning_rate": 2.6884057971014496e-05, "loss": 1.0222, "step": 371 }, { "epoch": 0.16847826086956522, "grad_norm": 0.07947065425019344, "learning_rate": 2.6956521739130436e-05, "loss": 1.0135, "step": 372 }, { "epoch": 0.16893115942028986, "grad_norm": 0.07659958161129127, "learning_rate": 2.702898550724638e-05, "loss": 0.9942, "step": 373 }, { "epoch": 0.1693840579710145, "grad_norm": 0.06704543841403225, "learning_rate": 2.7101449275362322e-05, "loss": 1.022, "step": 374 }, { "epoch": 0.16983695652173914, "grad_norm": 0.07505641833782215, "learning_rate": 2.7173913043478265e-05, "loss": 1.0091, "step": 375 }, { "epoch": 0.17028985507246377, "grad_norm": 0.07665831683376437, "learning_rate": 2.7246376811594205e-05, "loss": 1.0321, "step": 376 }, { "epoch": 0.17074275362318841, "grad_norm": 0.0714023383337561, "learning_rate": 2.7318840579710148e-05, "loss": 1.042, "step": 377 }, { "epoch": 0.17119565217391305, "grad_norm": 0.06279619540881025, "learning_rate": 2.739130434782609e-05, "loss": 1.0018, "step": 378 }, { "epoch": 0.1716485507246377, "grad_norm": 0.06525681257151941, "learning_rate": 2.7463768115942034e-05, "loss": 0.9966, "step": 379 }, { "epoch": 0.1721014492753623, "grad_norm": 0.05891569663094091, "learning_rate": 2.753623188405797e-05, "loss": 1.0135, "step": 380 }, { "epoch": 0.17255434782608695, "grad_norm": 0.07268738764975437, "learning_rate": 2.7608695652173914e-05, "loss": 1.004, "step": 381 }, { "epoch": 0.17300724637681159, "grad_norm": 0.0672207977279611, "learning_rate": 2.7681159420289857e-05, "loss": 1.0365, "step": 382 }, { "epoch": 0.17346014492753623, "grad_norm": 0.06252320551503958, "learning_rate": 2.7753623188405797e-05, "loss": 1.0158, "step": 383 }, { "epoch": 0.17391304347826086, "grad_norm": 0.07446182190784309, "learning_rate": 2.782608695652174e-05, "loss": 1.0033, "step": 384 }, { "epoch": 0.1743659420289855, "grad_norm": 0.06586117830167777, "learning_rate": 2.7898550724637683e-05, "loss": 1.0184, "step": 385 }, { "epoch": 0.17481884057971014, "grad_norm": 0.06963838109519649, "learning_rate": 2.7971014492753626e-05, "loss": 1.0282, "step": 386 }, { "epoch": 0.17527173913043478, "grad_norm": 0.06732742138125385, "learning_rate": 2.8043478260869566e-05, "loss": 1.0074, "step": 387 }, { "epoch": 0.17572463768115942, "grad_norm": 0.06656854168679593, "learning_rate": 2.811594202898551e-05, "loss": 1.0149, "step": 388 }, { "epoch": 0.17617753623188406, "grad_norm": 0.06561469473671247, "learning_rate": 2.8188405797101452e-05, "loss": 1.0124, "step": 389 }, { "epoch": 0.1766304347826087, "grad_norm": 0.06468451420454915, "learning_rate": 2.8260869565217396e-05, "loss": 1.017, "step": 390 }, { "epoch": 0.17708333333333334, "grad_norm": 0.06325412693250146, "learning_rate": 2.833333333333334e-05, "loss": 1.0162, "step": 391 }, { "epoch": 0.17753623188405798, "grad_norm": 0.07022989418093471, "learning_rate": 2.840579710144928e-05, "loss": 0.9942, "step": 392 }, { "epoch": 0.17798913043478262, "grad_norm": 0.05928198116351145, "learning_rate": 2.847826086956522e-05, "loss": 1.0057, "step": 393 }, { "epoch": 0.17844202898550723, "grad_norm": 0.10478259120302291, "learning_rate": 2.855072463768116e-05, "loss": 1.0251, "step": 394 }, { "epoch": 0.17889492753623187, "grad_norm": 0.06888608846581931, "learning_rate": 2.86231884057971e-05, "loss": 1.0195, "step": 395 }, { "epoch": 0.1793478260869565, "grad_norm": 0.06601419020499148, "learning_rate": 2.8695652173913044e-05, "loss": 1.019, "step": 396 }, { "epoch": 0.17980072463768115, "grad_norm": 0.07462304300988706, "learning_rate": 2.8768115942028988e-05, "loss": 1.0273, "step": 397 }, { "epoch": 0.1802536231884058, "grad_norm": 0.06425518300997263, "learning_rate": 2.884057971014493e-05, "loss": 0.9937, "step": 398 }, { "epoch": 0.18070652173913043, "grad_norm": 0.07018349702324125, "learning_rate": 2.891304347826087e-05, "loss": 1.0013, "step": 399 }, { "epoch": 0.18115942028985507, "grad_norm": 0.0828807349428865, "learning_rate": 2.8985507246376814e-05, "loss": 1.0058, "step": 400 }, { "epoch": 0.1816123188405797, "grad_norm": 0.07765458082573422, "learning_rate": 2.9057971014492757e-05, "loss": 1.0219, "step": 401 }, { "epoch": 0.18206521739130435, "grad_norm": 0.06901561424967564, "learning_rate": 2.91304347826087e-05, "loss": 1.0019, "step": 402 }, { "epoch": 0.182518115942029, "grad_norm": 0.06406428058016536, "learning_rate": 2.920289855072464e-05, "loss": 0.9943, "step": 403 }, { "epoch": 0.18297101449275363, "grad_norm": 0.08059834858084852, "learning_rate": 2.9275362318840583e-05, "loss": 1.0067, "step": 404 }, { "epoch": 0.18342391304347827, "grad_norm": 0.0678643217889346, "learning_rate": 2.9347826086956526e-05, "loss": 1.0082, "step": 405 }, { "epoch": 0.1838768115942029, "grad_norm": 0.07639507474952519, "learning_rate": 2.942028985507247e-05, "loss": 0.9994, "step": 406 }, { "epoch": 0.18432971014492755, "grad_norm": 0.07720957776384488, "learning_rate": 2.949275362318841e-05, "loss": 1.0094, "step": 407 }, { "epoch": 0.18478260869565216, "grad_norm": 0.06751976043879625, "learning_rate": 2.956521739130435e-05, "loss": 0.9886, "step": 408 }, { "epoch": 0.1852355072463768, "grad_norm": 0.06754338154947619, "learning_rate": 2.9637681159420292e-05, "loss": 1.0197, "step": 409 }, { "epoch": 0.18568840579710144, "grad_norm": 0.07361573155981536, "learning_rate": 2.971014492753623e-05, "loss": 0.9992, "step": 410 }, { "epoch": 0.18614130434782608, "grad_norm": 0.06057363117476252, "learning_rate": 2.9782608695652175e-05, "loss": 1.0296, "step": 411 }, { "epoch": 0.18659420289855072, "grad_norm": 0.07879846310858206, "learning_rate": 2.9855072463768118e-05, "loss": 0.9818, "step": 412 }, { "epoch": 0.18704710144927536, "grad_norm": 0.08019999655342816, "learning_rate": 2.992753623188406e-05, "loss": 1.0046, "step": 413 }, { "epoch": 0.1875, "grad_norm": 0.07239426522032245, "learning_rate": 3.0000000000000004e-05, "loss": 1.0138, "step": 414 }, { "epoch": 0.18795289855072464, "grad_norm": 0.07015674224004155, "learning_rate": 3.0072463768115944e-05, "loss": 0.996, "step": 415 }, { "epoch": 0.18840579710144928, "grad_norm": 0.06463909003839798, "learning_rate": 3.0144927536231887e-05, "loss": 0.9988, "step": 416 }, { "epoch": 0.18885869565217392, "grad_norm": 0.06693482905683373, "learning_rate": 3.021739130434783e-05, "loss": 1.023, "step": 417 }, { "epoch": 0.18931159420289856, "grad_norm": 0.07242183618640823, "learning_rate": 3.0289855072463773e-05, "loss": 1.0004, "step": 418 }, { "epoch": 0.1897644927536232, "grad_norm": 0.07726604412912749, "learning_rate": 3.0362318840579713e-05, "loss": 0.9921, "step": 419 }, { "epoch": 0.19021739130434784, "grad_norm": 0.07695919182242932, "learning_rate": 3.0434782608695656e-05, "loss": 0.9932, "step": 420 }, { "epoch": 0.19067028985507245, "grad_norm": 0.07278909879284003, "learning_rate": 3.0507246376811596e-05, "loss": 1.0043, "step": 421 }, { "epoch": 0.1911231884057971, "grad_norm": 0.07616520417230756, "learning_rate": 3.0579710144927536e-05, "loss": 0.9779, "step": 422 }, { "epoch": 0.19157608695652173, "grad_norm": 0.07262399693226429, "learning_rate": 3.065217391304348e-05, "loss": 1.0025, "step": 423 }, { "epoch": 0.19202898550724637, "grad_norm": 0.07993637188183222, "learning_rate": 3.072463768115942e-05, "loss": 1.0095, "step": 424 }, { "epoch": 0.192481884057971, "grad_norm": 0.0818984474302543, "learning_rate": 3.079710144927536e-05, "loss": 1.0213, "step": 425 }, { "epoch": 0.19293478260869565, "grad_norm": 0.06717711502620274, "learning_rate": 3.086956521739131e-05, "loss": 1.0023, "step": 426 }, { "epoch": 0.1933876811594203, "grad_norm": 0.07793294203877965, "learning_rate": 3.094202898550725e-05, "loss": 1.0249, "step": 427 }, { "epoch": 0.19384057971014493, "grad_norm": 0.07286528582371184, "learning_rate": 3.101449275362319e-05, "loss": 1.0124, "step": 428 }, { "epoch": 0.19429347826086957, "grad_norm": 0.0713850640842334, "learning_rate": 3.1086956521739134e-05, "loss": 1.0056, "step": 429 }, { "epoch": 0.1947463768115942, "grad_norm": 0.07331623651884277, "learning_rate": 3.1159420289855074e-05, "loss": 1.0145, "step": 430 }, { "epoch": 0.19519927536231885, "grad_norm": 0.08372146099888514, "learning_rate": 3.123188405797102e-05, "loss": 1.0172, "step": 431 }, { "epoch": 0.1956521739130435, "grad_norm": 0.07302026921682202, "learning_rate": 3.130434782608696e-05, "loss": 1.0046, "step": 432 }, { "epoch": 0.19610507246376813, "grad_norm": 0.08127362374082193, "learning_rate": 3.13768115942029e-05, "loss": 1.0038, "step": 433 }, { "epoch": 0.19655797101449277, "grad_norm": 0.07154528354585965, "learning_rate": 3.144927536231885e-05, "loss": 0.9985, "step": 434 }, { "epoch": 0.19701086956521738, "grad_norm": 0.08409335167949847, "learning_rate": 3.152173913043478e-05, "loss": 0.9944, "step": 435 }, { "epoch": 0.19746376811594202, "grad_norm": 0.06853804136652666, "learning_rate": 3.1594202898550726e-05, "loss": 0.9753, "step": 436 }, { "epoch": 0.19791666666666666, "grad_norm": 0.08320757904673536, "learning_rate": 3.1666666666666666e-05, "loss": 0.9985, "step": 437 }, { "epoch": 0.1983695652173913, "grad_norm": 0.07180977151603964, "learning_rate": 3.173913043478261e-05, "loss": 0.9964, "step": 438 }, { "epoch": 0.19882246376811594, "grad_norm": 0.08956370135332803, "learning_rate": 3.181159420289855e-05, "loss": 0.9951, "step": 439 }, { "epoch": 0.19927536231884058, "grad_norm": 0.07384013911134317, "learning_rate": 3.188405797101449e-05, "loss": 1.0174, "step": 440 }, { "epoch": 0.19972826086956522, "grad_norm": 0.0812082969594339, "learning_rate": 3.195652173913044e-05, "loss": 1.0189, "step": 441 }, { "epoch": 0.20018115942028986, "grad_norm": 0.08321856746840668, "learning_rate": 3.202898550724638e-05, "loss": 0.9966, "step": 442 }, { "epoch": 0.2006340579710145, "grad_norm": 0.08541099944483777, "learning_rate": 3.2101449275362325e-05, "loss": 1.0085, "step": 443 }, { "epoch": 0.20108695652173914, "grad_norm": 0.08646981171043905, "learning_rate": 3.2173913043478265e-05, "loss": 1.0106, "step": 444 }, { "epoch": 0.20153985507246377, "grad_norm": 0.07582172474664009, "learning_rate": 3.2246376811594205e-05, "loss": 1.0059, "step": 445 }, { "epoch": 0.20199275362318841, "grad_norm": 0.1025616498491228, "learning_rate": 3.231884057971015e-05, "loss": 1.0056, "step": 446 }, { "epoch": 0.20244565217391305, "grad_norm": 0.07176347946707469, "learning_rate": 3.239130434782609e-05, "loss": 1.0309, "step": 447 }, { "epoch": 0.2028985507246377, "grad_norm": 0.11226571780682586, "learning_rate": 3.246376811594203e-05, "loss": 1.0048, "step": 448 }, { "epoch": 0.2033514492753623, "grad_norm": 0.07656229048142874, "learning_rate": 3.253623188405797e-05, "loss": 1.0085, "step": 449 }, { "epoch": 0.20380434782608695, "grad_norm": 0.09402091830452639, "learning_rate": 3.260869565217392e-05, "loss": 1.0209, "step": 450 }, { "epoch": 0.20425724637681159, "grad_norm": 0.07395362732652021, "learning_rate": 3.268115942028986e-05, "loss": 0.9942, "step": 451 }, { "epoch": 0.20471014492753623, "grad_norm": 0.08609467299321379, "learning_rate": 3.2753623188405796e-05, "loss": 1.006, "step": 452 }, { "epoch": 0.20516304347826086, "grad_norm": 0.06974144350115861, "learning_rate": 3.282608695652174e-05, "loss": 1.0046, "step": 453 }, { "epoch": 0.2056159420289855, "grad_norm": 0.07051389820457159, "learning_rate": 3.289855072463768e-05, "loss": 1.0077, "step": 454 }, { "epoch": 0.20606884057971014, "grad_norm": 0.08611430070442272, "learning_rate": 3.297101449275362e-05, "loss": 1.0151, "step": 455 }, { "epoch": 0.20652173913043478, "grad_norm": 0.0732685081463276, "learning_rate": 3.304347826086957e-05, "loss": 0.9891, "step": 456 }, { "epoch": 0.20697463768115942, "grad_norm": 0.07112314572043645, "learning_rate": 3.311594202898551e-05, "loss": 1.0205, "step": 457 }, { "epoch": 0.20742753623188406, "grad_norm": 0.0792020511036897, "learning_rate": 3.3188405797101455e-05, "loss": 0.9953, "step": 458 }, { "epoch": 0.2078804347826087, "grad_norm": 0.07141398751343121, "learning_rate": 3.3260869565217395e-05, "loss": 0.9901, "step": 459 }, { "epoch": 0.20833333333333334, "grad_norm": 0.09733750276804215, "learning_rate": 3.3333333333333335e-05, "loss": 0.9981, "step": 460 }, { "epoch": 0.20878623188405798, "grad_norm": 0.07788785399340323, "learning_rate": 3.340579710144928e-05, "loss": 1.0092, "step": 461 }, { "epoch": 0.20923913043478262, "grad_norm": 0.09209771235477002, "learning_rate": 3.347826086956522e-05, "loss": 0.9929, "step": 462 }, { "epoch": 0.20969202898550723, "grad_norm": 0.11415810920105313, "learning_rate": 3.355072463768116e-05, "loss": 1.0115, "step": 463 }, { "epoch": 0.21014492753623187, "grad_norm": 0.0702177104614932, "learning_rate": 3.36231884057971e-05, "loss": 1.0168, "step": 464 }, { "epoch": 0.2105978260869565, "grad_norm": 0.09701759766119245, "learning_rate": 3.369565217391305e-05, "loss": 0.9833, "step": 465 }, { "epoch": 0.21105072463768115, "grad_norm": 0.0952791499453263, "learning_rate": 3.376811594202899e-05, "loss": 1.0054, "step": 466 }, { "epoch": 0.2115036231884058, "grad_norm": 0.08069280613970328, "learning_rate": 3.384057971014493e-05, "loss": 1.0019, "step": 467 }, { "epoch": 0.21195652173913043, "grad_norm": 0.07822932440220953, "learning_rate": 3.391304347826087e-05, "loss": 1.0042, "step": 468 }, { "epoch": 0.21240942028985507, "grad_norm": 0.08206011267535804, "learning_rate": 3.398550724637681e-05, "loss": 0.9871, "step": 469 }, { "epoch": 0.2128623188405797, "grad_norm": 0.0747779683497238, "learning_rate": 3.405797101449276e-05, "loss": 0.9998, "step": 470 }, { "epoch": 0.21331521739130435, "grad_norm": 0.07614821837085059, "learning_rate": 3.41304347826087e-05, "loss": 0.9764, "step": 471 }, { "epoch": 0.213768115942029, "grad_norm": 0.07764998136642176, "learning_rate": 3.420289855072464e-05, "loss": 1.0215, "step": 472 }, { "epoch": 0.21422101449275363, "grad_norm": 0.08346439855530972, "learning_rate": 3.4275362318840586e-05, "loss": 1.0199, "step": 473 }, { "epoch": 0.21467391304347827, "grad_norm": 0.07803577204134331, "learning_rate": 3.4347826086956526e-05, "loss": 1.0145, "step": 474 }, { "epoch": 0.2151268115942029, "grad_norm": 0.08239366163828747, "learning_rate": 3.4420289855072465e-05, "loss": 1.0194, "step": 475 }, { "epoch": 0.21557971014492755, "grad_norm": 0.08041168681583867, "learning_rate": 3.449275362318841e-05, "loss": 0.9968, "step": 476 }, { "epoch": 0.21603260869565216, "grad_norm": 0.0776238574317334, "learning_rate": 3.456521739130435e-05, "loss": 1.0138, "step": 477 }, { "epoch": 0.2164855072463768, "grad_norm": 0.08637515596356421, "learning_rate": 3.463768115942029e-05, "loss": 0.9634, "step": 478 }, { "epoch": 0.21693840579710144, "grad_norm": 0.085062835934078, "learning_rate": 3.471014492753623e-05, "loss": 1.0126, "step": 479 }, { "epoch": 0.21739130434782608, "grad_norm": 0.07493289373383903, "learning_rate": 3.478260869565218e-05, "loss": 0.9969, "step": 480 }, { "epoch": 0.21784420289855072, "grad_norm": 0.10639460758869673, "learning_rate": 3.485507246376812e-05, "loss": 0.9925, "step": 481 }, { "epoch": 0.21829710144927536, "grad_norm": 0.06836302314172224, "learning_rate": 3.492753623188406e-05, "loss": 0.9983, "step": 482 }, { "epoch": 0.21875, "grad_norm": 0.08791962682007909, "learning_rate": 3.5000000000000004e-05, "loss": 1.007, "step": 483 }, { "epoch": 0.21920289855072464, "grad_norm": 0.08595266652891387, "learning_rate": 3.5072463768115943e-05, "loss": 1.008, "step": 484 }, { "epoch": 0.21965579710144928, "grad_norm": 0.07598468353552972, "learning_rate": 3.514492753623189e-05, "loss": 1.021, "step": 485 }, { "epoch": 0.22010869565217392, "grad_norm": 0.07634039354409475, "learning_rate": 3.521739130434783e-05, "loss": 1.0056, "step": 486 }, { "epoch": 0.22056159420289856, "grad_norm": 0.07536844209239699, "learning_rate": 3.528985507246377e-05, "loss": 1.0219, "step": 487 }, { "epoch": 0.2210144927536232, "grad_norm": 0.06579840490938284, "learning_rate": 3.5362318840579716e-05, "loss": 1.0186, "step": 488 }, { "epoch": 0.22146739130434784, "grad_norm": 0.07821823626850472, "learning_rate": 3.5434782608695656e-05, "loss": 1.012, "step": 489 }, { "epoch": 0.22192028985507245, "grad_norm": 0.07073327902712771, "learning_rate": 3.5507246376811596e-05, "loss": 0.9982, "step": 490 }, { "epoch": 0.2223731884057971, "grad_norm": 0.06904911621260942, "learning_rate": 3.5579710144927535e-05, "loss": 0.9808, "step": 491 }, { "epoch": 0.22282608695652173, "grad_norm": 0.23846761362630314, "learning_rate": 3.565217391304348e-05, "loss": 1.0114, "step": 492 }, { "epoch": 0.22327898550724637, "grad_norm": 0.0886440083403472, "learning_rate": 3.572463768115942e-05, "loss": 0.9817, "step": 493 }, { "epoch": 0.223731884057971, "grad_norm": 0.08991264381765893, "learning_rate": 3.579710144927536e-05, "loss": 1.0152, "step": 494 }, { "epoch": 0.22418478260869565, "grad_norm": 0.07790105334467667, "learning_rate": 3.586956521739131e-05, "loss": 0.9921, "step": 495 }, { "epoch": 0.2246376811594203, "grad_norm": 0.08274626663233109, "learning_rate": 3.594202898550725e-05, "loss": 0.9721, "step": 496 }, { "epoch": 0.22509057971014493, "grad_norm": 0.08177548653672646, "learning_rate": 3.6014492753623194e-05, "loss": 0.9952, "step": 497 }, { "epoch": 0.22554347826086957, "grad_norm": 0.08296061786950348, "learning_rate": 3.6086956521739134e-05, "loss": 1.0072, "step": 498 }, { "epoch": 0.2259963768115942, "grad_norm": 0.08837426810727417, "learning_rate": 3.6159420289855074e-05, "loss": 0.9923, "step": 499 }, { "epoch": 0.22644927536231885, "grad_norm": 0.09255073104886523, "learning_rate": 3.623188405797102e-05, "loss": 1.0062, "step": 500 }, { "epoch": 0.2269021739130435, "grad_norm": 0.10042581333235773, "learning_rate": 3.630434782608696e-05, "loss": 0.9882, "step": 501 }, { "epoch": 0.22735507246376813, "grad_norm": 0.16748469257276183, "learning_rate": 3.63768115942029e-05, "loss": 1.0194, "step": 502 }, { "epoch": 0.22780797101449277, "grad_norm": 0.1096834033167894, "learning_rate": 3.6449275362318846e-05, "loss": 1.0053, "step": 503 }, { "epoch": 0.22826086956521738, "grad_norm": 0.0829311353994745, "learning_rate": 3.6521739130434786e-05, "loss": 1.0038, "step": 504 }, { "epoch": 0.22871376811594202, "grad_norm": 0.09713437071514369, "learning_rate": 3.6594202898550726e-05, "loss": 0.99, "step": 505 }, { "epoch": 0.22916666666666666, "grad_norm": 0.07831798320044123, "learning_rate": 3.6666666666666666e-05, "loss": 0.9989, "step": 506 }, { "epoch": 0.2296195652173913, "grad_norm": 0.10423590049956488, "learning_rate": 3.673913043478261e-05, "loss": 1.0005, "step": 507 }, { "epoch": 0.23007246376811594, "grad_norm": 0.06441915905356396, "learning_rate": 3.681159420289855e-05, "loss": 0.9988, "step": 508 }, { "epoch": 0.23052536231884058, "grad_norm": 0.08802245442052255, "learning_rate": 3.68840579710145e-05, "loss": 0.9942, "step": 509 }, { "epoch": 0.23097826086956522, "grad_norm": 0.08206725899069667, "learning_rate": 3.695652173913044e-05, "loss": 0.9968, "step": 510 }, { "epoch": 0.23143115942028986, "grad_norm": 0.09596102731339985, "learning_rate": 3.702898550724638e-05, "loss": 0.9896, "step": 511 }, { "epoch": 0.2318840579710145, "grad_norm": 0.10528479915797566, "learning_rate": 3.7101449275362325e-05, "loss": 1.0064, "step": 512 }, { "epoch": 0.23233695652173914, "grad_norm": 0.09407125120776315, "learning_rate": 3.7173913043478264e-05, "loss": 1.0129, "step": 513 }, { "epoch": 0.23278985507246377, "grad_norm": 0.07019426171463299, "learning_rate": 3.7246376811594204e-05, "loss": 1.0122, "step": 514 }, { "epoch": 0.23324275362318841, "grad_norm": 0.0931000630597928, "learning_rate": 3.731884057971015e-05, "loss": 1.0075, "step": 515 }, { "epoch": 0.23369565217391305, "grad_norm": 0.0885045704870399, "learning_rate": 3.739130434782609e-05, "loss": 1.0085, "step": 516 }, { "epoch": 0.2341485507246377, "grad_norm": 0.08182372482121965, "learning_rate": 3.746376811594204e-05, "loss": 1.0287, "step": 517 }, { "epoch": 0.2346014492753623, "grad_norm": 0.08016758533774189, "learning_rate": 3.753623188405797e-05, "loss": 1.0223, "step": 518 }, { "epoch": 0.23505434782608695, "grad_norm": 0.0998594884623428, "learning_rate": 3.7608695652173917e-05, "loss": 0.9906, "step": 519 }, { "epoch": 0.23550724637681159, "grad_norm": 0.08835253183628095, "learning_rate": 3.7681159420289856e-05, "loss": 1.04, "step": 520 }, { "epoch": 0.23596014492753623, "grad_norm": 0.0795005349631111, "learning_rate": 3.7753623188405796e-05, "loss": 1.0047, "step": 521 }, { "epoch": 0.23641304347826086, "grad_norm": 0.09411363589225284, "learning_rate": 3.782608695652174e-05, "loss": 1.017, "step": 522 }, { "epoch": 0.2368659420289855, "grad_norm": 0.08116714167557026, "learning_rate": 3.789855072463768e-05, "loss": 1.0131, "step": 523 }, { "epoch": 0.23731884057971014, "grad_norm": 0.08926061048037695, "learning_rate": 3.797101449275363e-05, "loss": 0.9759, "step": 524 }, { "epoch": 0.23777173913043478, "grad_norm": 0.08422795224630779, "learning_rate": 3.804347826086957e-05, "loss": 0.9875, "step": 525 }, { "epoch": 0.23822463768115942, "grad_norm": 0.08125282653583696, "learning_rate": 3.811594202898551e-05, "loss": 0.9794, "step": 526 }, { "epoch": 0.23867753623188406, "grad_norm": 0.09955360924464852, "learning_rate": 3.8188405797101455e-05, "loss": 0.9971, "step": 527 }, { "epoch": 0.2391304347826087, "grad_norm": 0.09238329196106147, "learning_rate": 3.8260869565217395e-05, "loss": 0.9825, "step": 528 }, { "epoch": 0.23958333333333334, "grad_norm": 0.0799474591169727, "learning_rate": 3.833333333333334e-05, "loss": 1.0205, "step": 529 }, { "epoch": 0.24003623188405798, "grad_norm": 0.08085466023430826, "learning_rate": 3.840579710144928e-05, "loss": 1.0013, "step": 530 }, { "epoch": 0.24048913043478262, "grad_norm": 0.0925775446010138, "learning_rate": 3.847826086956522e-05, "loss": 1.0025, "step": 531 }, { "epoch": 0.24094202898550723, "grad_norm": 0.08966048910240286, "learning_rate": 3.855072463768116e-05, "loss": 1.0128, "step": 532 }, { "epoch": 0.24139492753623187, "grad_norm": 0.09261854411186876, "learning_rate": 3.86231884057971e-05, "loss": 0.9831, "step": 533 }, { "epoch": 0.2418478260869565, "grad_norm": 0.10229474132973025, "learning_rate": 3.869565217391305e-05, "loss": 0.9967, "step": 534 }, { "epoch": 0.24230072463768115, "grad_norm": 0.11945208938298275, "learning_rate": 3.876811594202899e-05, "loss": 1.013, "step": 535 }, { "epoch": 0.2427536231884058, "grad_norm": 0.11548834086183449, "learning_rate": 3.884057971014493e-05, "loss": 0.9894, "step": 536 }, { "epoch": 0.24320652173913043, "grad_norm": 0.11231207526449571, "learning_rate": 3.891304347826087e-05, "loss": 1.0072, "step": 537 }, { "epoch": 0.24365942028985507, "grad_norm": 0.10631530136618546, "learning_rate": 3.898550724637681e-05, "loss": 0.9798, "step": 538 }, { "epoch": 0.2441123188405797, "grad_norm": 0.09299018567794949, "learning_rate": 3.905797101449276e-05, "loss": 0.993, "step": 539 }, { "epoch": 0.24456521739130435, "grad_norm": 0.08348495124577522, "learning_rate": 3.91304347826087e-05, "loss": 0.9921, "step": 540 }, { "epoch": 0.245018115942029, "grad_norm": 0.08587300919096898, "learning_rate": 3.920289855072464e-05, "loss": 0.997, "step": 541 }, { "epoch": 0.24547101449275363, "grad_norm": 0.10915479234719509, "learning_rate": 3.9275362318840585e-05, "loss": 1.0078, "step": 542 }, { "epoch": 0.24592391304347827, "grad_norm": 0.10154755919745943, "learning_rate": 3.9347826086956525e-05, "loss": 1.0077, "step": 543 }, { "epoch": 0.2463768115942029, "grad_norm": 0.08735344105543849, "learning_rate": 3.942028985507247e-05, "loss": 1.0168, "step": 544 }, { "epoch": 0.24682971014492755, "grad_norm": 0.08254250125472279, "learning_rate": 3.949275362318841e-05, "loss": 1.0002, "step": 545 }, { "epoch": 0.24728260869565216, "grad_norm": 0.09918546445053485, "learning_rate": 3.956521739130435e-05, "loss": 1.0087, "step": 546 }, { "epoch": 0.2477355072463768, "grad_norm": 0.11618970016922121, "learning_rate": 3.963768115942029e-05, "loss": 1.0176, "step": 547 }, { "epoch": 0.24818840579710144, "grad_norm": 0.09290867069958472, "learning_rate": 3.971014492753623e-05, "loss": 0.9731, "step": 548 }, { "epoch": 0.24864130434782608, "grad_norm": 0.0861693598067163, "learning_rate": 3.978260869565218e-05, "loss": 1.0136, "step": 549 }, { "epoch": 0.24909420289855072, "grad_norm": 0.11450104953742819, "learning_rate": 3.985507246376812e-05, "loss": 0.995, "step": 550 }, { "epoch": 0.24954710144927536, "grad_norm": 0.11513650718972092, "learning_rate": 3.9927536231884064e-05, "loss": 1.0252, "step": 551 }, { "epoch": 0.25, "grad_norm": 0.09065764612195233, "learning_rate": 4e-05, "loss": 1.0061, "step": 552 }, { "epoch": 0.2504528985507246, "grad_norm": 0.09420245484883498, "learning_rate": 4.007246376811594e-05, "loss": 0.9735, "step": 553 }, { "epoch": 0.2509057971014493, "grad_norm": 0.11536246762770594, "learning_rate": 4.014492753623189e-05, "loss": 1.0217, "step": 554 }, { "epoch": 0.2513586956521739, "grad_norm": 0.11567349465799985, "learning_rate": 4.021739130434782e-05, "loss": 0.975, "step": 555 }, { "epoch": 0.25181159420289856, "grad_norm": 0.09507273242092246, "learning_rate": 4.0289855072463776e-05, "loss": 1.0055, "step": 556 }, { "epoch": 0.25226449275362317, "grad_norm": 0.19929728250293277, "learning_rate": 4.036231884057971e-05, "loss": 1.0, "step": 557 }, { "epoch": 0.25271739130434784, "grad_norm": 0.11590559685006674, "learning_rate": 4.0434782608695655e-05, "loss": 1.0068, "step": 558 }, { "epoch": 0.25317028985507245, "grad_norm": 0.11160658007243639, "learning_rate": 4.0507246376811595e-05, "loss": 0.9902, "step": 559 }, { "epoch": 0.2536231884057971, "grad_norm": 0.10316226364573776, "learning_rate": 4.057971014492754e-05, "loss": 0.9972, "step": 560 }, { "epoch": 0.25407608695652173, "grad_norm": 0.09952602535335812, "learning_rate": 4.065217391304348e-05, "loss": 0.992, "step": 561 }, { "epoch": 0.2545289855072464, "grad_norm": 0.1242281524235226, "learning_rate": 4.072463768115943e-05, "loss": 1.0109, "step": 562 }, { "epoch": 0.254981884057971, "grad_norm": 0.09151177074569986, "learning_rate": 4.079710144927537e-05, "loss": 1.0118, "step": 563 }, { "epoch": 0.2554347826086957, "grad_norm": 0.11698943555606725, "learning_rate": 4.0869565217391314e-05, "loss": 0.9771, "step": 564 }, { "epoch": 0.2558876811594203, "grad_norm": 0.10428739406397561, "learning_rate": 4.094202898550725e-05, "loss": 1.0178, "step": 565 }, { "epoch": 0.2563405797101449, "grad_norm": 0.10721711837516386, "learning_rate": 4.101449275362319e-05, "loss": 0.9999, "step": 566 }, { "epoch": 0.25679347826086957, "grad_norm": 0.1006922332961643, "learning_rate": 4.1086956521739134e-05, "loss": 0.9931, "step": 567 }, { "epoch": 0.2572463768115942, "grad_norm": 0.09627602142407905, "learning_rate": 4.115942028985507e-05, "loss": 0.9939, "step": 568 }, { "epoch": 0.25769927536231885, "grad_norm": 0.08625357872327147, "learning_rate": 4.123188405797102e-05, "loss": 0.9973, "step": 569 }, { "epoch": 0.25815217391304346, "grad_norm": 0.0957365285357928, "learning_rate": 4.130434782608696e-05, "loss": 1.0169, "step": 570 }, { "epoch": 0.2586050724637681, "grad_norm": 0.08463530246617733, "learning_rate": 4.1376811594202906e-05, "loss": 1.017, "step": 571 }, { "epoch": 0.25905797101449274, "grad_norm": 0.08589672911880936, "learning_rate": 4.144927536231884e-05, "loss": 1.0064, "step": 572 }, { "epoch": 0.2595108695652174, "grad_norm": 0.10360375106077042, "learning_rate": 4.1521739130434786e-05, "loss": 1.0033, "step": 573 }, { "epoch": 0.259963768115942, "grad_norm": 0.10689584256799937, "learning_rate": 4.1594202898550726e-05, "loss": 0.999, "step": 574 }, { "epoch": 0.2604166666666667, "grad_norm": 0.10773861071064368, "learning_rate": 4.166666666666667e-05, "loss": 1.0018, "step": 575 }, { "epoch": 0.2608695652173913, "grad_norm": 0.09918684945569094, "learning_rate": 4.173913043478261e-05, "loss": 1.0174, "step": 576 }, { "epoch": 0.26132246376811596, "grad_norm": 0.07730608647022333, "learning_rate": 4.181159420289856e-05, "loss": 0.9771, "step": 577 }, { "epoch": 0.2617753623188406, "grad_norm": 0.09451964255619208, "learning_rate": 4.18840579710145e-05, "loss": 1.0145, "step": 578 }, { "epoch": 0.26222826086956524, "grad_norm": 0.08072668579813895, "learning_rate": 4.1956521739130445e-05, "loss": 0.9837, "step": 579 }, { "epoch": 0.26268115942028986, "grad_norm": 0.09449254809856551, "learning_rate": 4.202898550724638e-05, "loss": 0.9937, "step": 580 }, { "epoch": 0.26313405797101447, "grad_norm": 0.11993648765702827, "learning_rate": 4.210144927536232e-05, "loss": 1.0294, "step": 581 }, { "epoch": 0.26358695652173914, "grad_norm": 0.10415795049472655, "learning_rate": 4.2173913043478264e-05, "loss": 0.9881, "step": 582 }, { "epoch": 0.26403985507246375, "grad_norm": 0.09549681530473142, "learning_rate": 4.2246376811594204e-05, "loss": 1.0049, "step": 583 }, { "epoch": 0.2644927536231884, "grad_norm": 0.12290448579393186, "learning_rate": 4.231884057971015e-05, "loss": 1.0034, "step": 584 }, { "epoch": 0.264945652173913, "grad_norm": 0.12058599116554305, "learning_rate": 4.239130434782609e-05, "loss": 0.9979, "step": 585 }, { "epoch": 0.2653985507246377, "grad_norm": 0.11192030047403184, "learning_rate": 4.2463768115942037e-05, "loss": 0.9917, "step": 586 }, { "epoch": 0.2658514492753623, "grad_norm": 0.10928128080795374, "learning_rate": 4.253623188405797e-05, "loss": 1.0101, "step": 587 }, { "epoch": 0.266304347826087, "grad_norm": 0.10340519328384115, "learning_rate": 4.2608695652173916e-05, "loss": 1.0235, "step": 588 }, { "epoch": 0.2667572463768116, "grad_norm": 0.11854353778475371, "learning_rate": 4.2681159420289856e-05, "loss": 1.0209, "step": 589 }, { "epoch": 0.26721014492753625, "grad_norm": 0.08759250086211726, "learning_rate": 4.27536231884058e-05, "loss": 1.0054, "step": 590 }, { "epoch": 0.26766304347826086, "grad_norm": 0.11550328423860963, "learning_rate": 4.282608695652174e-05, "loss": 0.9899, "step": 591 }, { "epoch": 0.26811594202898553, "grad_norm": 0.07713161699570709, "learning_rate": 4.289855072463769e-05, "loss": 1.0085, "step": 592 }, { "epoch": 0.26856884057971014, "grad_norm": 0.09311209399900516, "learning_rate": 4.297101449275363e-05, "loss": 0.9625, "step": 593 }, { "epoch": 0.26902173913043476, "grad_norm": 0.10032006095696952, "learning_rate": 4.304347826086956e-05, "loss": 1.0013, "step": 594 }, { "epoch": 0.2694746376811594, "grad_norm": 0.10366734187126453, "learning_rate": 4.311594202898551e-05, "loss": 0.9873, "step": 595 }, { "epoch": 0.26992753623188404, "grad_norm": 0.10170214322175405, "learning_rate": 4.318840579710145e-05, "loss": 0.9841, "step": 596 }, { "epoch": 0.2703804347826087, "grad_norm": 0.1088343288484781, "learning_rate": 4.3260869565217394e-05, "loss": 0.988, "step": 597 }, { "epoch": 0.2708333333333333, "grad_norm": 0.10673473728570484, "learning_rate": 4.3333333333333334e-05, "loss": 1.0102, "step": 598 }, { "epoch": 0.271286231884058, "grad_norm": 0.11769638282292365, "learning_rate": 4.340579710144928e-05, "loss": 0.9847, "step": 599 }, { "epoch": 0.2717391304347826, "grad_norm": 0.09564111123516798, "learning_rate": 4.347826086956522e-05, "loss": 0.9943, "step": 600 }, { "epoch": 0.27219202898550726, "grad_norm": 0.09499721929380074, "learning_rate": 4.355072463768117e-05, "loss": 1.0007, "step": 601 }, { "epoch": 0.2726449275362319, "grad_norm": 0.09135729261737692, "learning_rate": 4.362318840579711e-05, "loss": 0.9807, "step": 602 }, { "epoch": 0.27309782608695654, "grad_norm": 0.08550684482176352, "learning_rate": 4.369565217391305e-05, "loss": 1.0113, "step": 603 }, { "epoch": 0.27355072463768115, "grad_norm": 0.08597779660017614, "learning_rate": 4.3768115942028986e-05, "loss": 0.986, "step": 604 }, { "epoch": 0.2740036231884058, "grad_norm": 0.11041905940869912, "learning_rate": 4.384057971014493e-05, "loss": 1.0083, "step": 605 }, { "epoch": 0.27445652173913043, "grad_norm": 0.13049565771464047, "learning_rate": 4.391304347826087e-05, "loss": 0.9875, "step": 606 }, { "epoch": 0.2749094202898551, "grad_norm": 0.13608135263278512, "learning_rate": 4.398550724637682e-05, "loss": 1.0089, "step": 607 }, { "epoch": 0.2753623188405797, "grad_norm": 0.14479140892812967, "learning_rate": 4.405797101449276e-05, "loss": 0.9754, "step": 608 }, { "epoch": 0.2758152173913043, "grad_norm": 0.15454218912922357, "learning_rate": 4.41304347826087e-05, "loss": 0.9886, "step": 609 }, { "epoch": 0.276268115942029, "grad_norm": 0.1457430849389804, "learning_rate": 4.4202898550724645e-05, "loss": 1.0022, "step": 610 }, { "epoch": 0.2767210144927536, "grad_norm": 0.11856780800677999, "learning_rate": 4.427536231884058e-05, "loss": 1.0018, "step": 611 }, { "epoch": 0.27717391304347827, "grad_norm": 0.13572464004609502, "learning_rate": 4.4347826086956525e-05, "loss": 1.0042, "step": 612 }, { "epoch": 0.2776268115942029, "grad_norm": 0.18275193502849915, "learning_rate": 4.4420289855072464e-05, "loss": 1.0027, "step": 613 }, { "epoch": 0.27807971014492755, "grad_norm": 0.22059305425259584, "learning_rate": 4.449275362318841e-05, "loss": 1.0066, "step": 614 }, { "epoch": 0.27853260869565216, "grad_norm": 0.21070015883271695, "learning_rate": 4.456521739130435e-05, "loss": 1.018, "step": 615 }, { "epoch": 0.27898550724637683, "grad_norm": 0.17433059835317455, "learning_rate": 4.46376811594203e-05, "loss": 1.0027, "step": 616 }, { "epoch": 0.27943840579710144, "grad_norm": 0.11241097741710905, "learning_rate": 4.471014492753624e-05, "loss": 0.9895, "step": 617 }, { "epoch": 0.2798913043478261, "grad_norm": 0.09076121108257715, "learning_rate": 4.4782608695652184e-05, "loss": 1.0138, "step": 618 }, { "epoch": 0.2803442028985507, "grad_norm": 0.13310206195630772, "learning_rate": 4.4855072463768117e-05, "loss": 0.9775, "step": 619 }, { "epoch": 0.2807971014492754, "grad_norm": 0.14092384798388838, "learning_rate": 4.492753623188406e-05, "loss": 0.9915, "step": 620 }, { "epoch": 0.28125, "grad_norm": 0.1584540865739647, "learning_rate": 4.5e-05, "loss": 0.9953, "step": 621 }, { "epoch": 0.2817028985507246, "grad_norm": 0.1834436464448011, "learning_rate": 4.507246376811594e-05, "loss": 1.0116, "step": 622 }, { "epoch": 0.2821557971014493, "grad_norm": 0.1820081676492916, "learning_rate": 4.514492753623189e-05, "loss": 0.9997, "step": 623 }, { "epoch": 0.2826086956521739, "grad_norm": 0.15111921567510772, "learning_rate": 4.521739130434783e-05, "loss": 0.99, "step": 624 }, { "epoch": 0.28306159420289856, "grad_norm": 0.10715125656332837, "learning_rate": 4.5289855072463775e-05, "loss": 0.9902, "step": 625 }, { "epoch": 0.28351449275362317, "grad_norm": 0.09588326131815135, "learning_rate": 4.536231884057971e-05, "loss": 0.9966, "step": 626 }, { "epoch": 0.28396739130434784, "grad_norm": 0.10756955165501059, "learning_rate": 4.5434782608695655e-05, "loss": 0.9934, "step": 627 }, { "epoch": 0.28442028985507245, "grad_norm": 0.1828190916399338, "learning_rate": 4.5507246376811595e-05, "loss": 1.0086, "step": 628 }, { "epoch": 0.2848731884057971, "grad_norm": 0.254935696892034, "learning_rate": 4.557971014492754e-05, "loss": 1.0052, "step": 629 }, { "epoch": 0.28532608695652173, "grad_norm": 0.27222435288015406, "learning_rate": 4.565217391304348e-05, "loss": 0.9777, "step": 630 }, { "epoch": 0.2857789855072464, "grad_norm": 0.2680803669010136, "learning_rate": 4.572463768115943e-05, "loss": 0.9822, "step": 631 }, { "epoch": 0.286231884057971, "grad_norm": 0.24478179323341243, "learning_rate": 4.579710144927537e-05, "loss": 0.997, "step": 632 }, { "epoch": 0.2866847826086957, "grad_norm": 0.1771675901625526, "learning_rate": 4.5869565217391314e-05, "loss": 1.0268, "step": 633 }, { "epoch": 0.2871376811594203, "grad_norm": 0.15428331392077088, "learning_rate": 4.594202898550725e-05, "loss": 0.9965, "step": 634 }, { "epoch": 0.2875905797101449, "grad_norm": 0.11559267558120961, "learning_rate": 4.601449275362319e-05, "loss": 0.9974, "step": 635 }, { "epoch": 0.28804347826086957, "grad_norm": 0.12976219727181856, "learning_rate": 4.608695652173913e-05, "loss": 0.9751, "step": 636 }, { "epoch": 0.2884963768115942, "grad_norm": 0.12476162003237115, "learning_rate": 4.615942028985507e-05, "loss": 0.9734, "step": 637 }, { "epoch": 0.28894927536231885, "grad_norm": 0.1009279215290163, "learning_rate": 4.623188405797102e-05, "loss": 0.9897, "step": 638 }, { "epoch": 0.28940217391304346, "grad_norm": 0.08587031999588787, "learning_rate": 4.630434782608696e-05, "loss": 0.9902, "step": 639 }, { "epoch": 0.2898550724637681, "grad_norm": 0.10170645353602574, "learning_rate": 4.6376811594202906e-05, "loss": 0.9795, "step": 640 }, { "epoch": 0.29030797101449274, "grad_norm": 0.08899552071142648, "learning_rate": 4.644927536231884e-05, "loss": 0.9842, "step": 641 }, { "epoch": 0.2907608695652174, "grad_norm": 0.0925766767959319, "learning_rate": 4.652173913043479e-05, "loss": 0.9681, "step": 642 }, { "epoch": 0.291213768115942, "grad_norm": 0.12112800248369836, "learning_rate": 4.6594202898550725e-05, "loss": 1.011, "step": 643 }, { "epoch": 0.2916666666666667, "grad_norm": 0.13644219387085327, "learning_rate": 4.666666666666667e-05, "loss": 0.9938, "step": 644 }, { "epoch": 0.2921195652173913, "grad_norm": 0.11365834618295723, "learning_rate": 4.673913043478261e-05, "loss": 1.0027, "step": 645 }, { "epoch": 0.29257246376811596, "grad_norm": 0.09376002366135891, "learning_rate": 4.681159420289856e-05, "loss": 0.9848, "step": 646 }, { "epoch": 0.2930253623188406, "grad_norm": 0.12201508299906141, "learning_rate": 4.68840579710145e-05, "loss": 0.9855, "step": 647 }, { "epoch": 0.29347826086956524, "grad_norm": 0.11055609260984835, "learning_rate": 4.6956521739130444e-05, "loss": 1.0037, "step": 648 }, { "epoch": 0.29393115942028986, "grad_norm": 0.10257440445808362, "learning_rate": 4.7028985507246384e-05, "loss": 0.9866, "step": 649 }, { "epoch": 0.29438405797101447, "grad_norm": 0.10164516539872873, "learning_rate": 4.710144927536232e-05, "loss": 0.9923, "step": 650 }, { "epoch": 0.29483695652173914, "grad_norm": 0.13425707945838494, "learning_rate": 4.7173913043478264e-05, "loss": 0.9876, "step": 651 }, { "epoch": 0.29528985507246375, "grad_norm": 0.18846961185093242, "learning_rate": 4.72463768115942e-05, "loss": 0.9844, "step": 652 }, { "epoch": 0.2957427536231884, "grad_norm": 0.2091578854325559, "learning_rate": 4.731884057971015e-05, "loss": 1.0052, "step": 653 }, { "epoch": 0.296195652173913, "grad_norm": 0.23057854942231337, "learning_rate": 4.739130434782609e-05, "loss": 0.9951, "step": 654 }, { "epoch": 0.2966485507246377, "grad_norm": 0.1788704656489832, "learning_rate": 4.7463768115942036e-05, "loss": 1.0036, "step": 655 }, { "epoch": 0.2971014492753623, "grad_norm": 0.13944646050181717, "learning_rate": 4.7536231884057976e-05, "loss": 1.0038, "step": 656 }, { "epoch": 0.297554347826087, "grad_norm": 0.09208535206336685, "learning_rate": 4.760869565217392e-05, "loss": 1.0059, "step": 657 }, { "epoch": 0.2980072463768116, "grad_norm": 0.4169616892813906, "learning_rate": 4.7681159420289855e-05, "loss": 1.0084, "step": 658 }, { "epoch": 0.29846014492753625, "grad_norm": 0.10556662522064053, "learning_rate": 4.77536231884058e-05, "loss": 1.0025, "step": 659 }, { "epoch": 0.29891304347826086, "grad_norm": 0.11434459528389497, "learning_rate": 4.782608695652174e-05, "loss": 0.9903, "step": 660 }, { "epoch": 0.29936594202898553, "grad_norm": 0.13364429492291416, "learning_rate": 4.789855072463769e-05, "loss": 0.961, "step": 661 }, { "epoch": 0.29981884057971014, "grad_norm": 0.15902792277670208, "learning_rate": 4.797101449275363e-05, "loss": 0.9825, "step": 662 }, { "epoch": 0.30027173913043476, "grad_norm": 0.1613981235406443, "learning_rate": 4.804347826086957e-05, "loss": 0.9923, "step": 663 }, { "epoch": 0.3007246376811594, "grad_norm": 0.12875923500281083, "learning_rate": 4.8115942028985514e-05, "loss": 1.011, "step": 664 }, { "epoch": 0.30117753623188404, "grad_norm": 0.08815073592228989, "learning_rate": 4.818840579710145e-05, "loss": 0.9903, "step": 665 }, { "epoch": 0.3016304347826087, "grad_norm": 0.11429145572274844, "learning_rate": 4.8260869565217394e-05, "loss": 0.99, "step": 666 }, { "epoch": 0.3020833333333333, "grad_norm": 0.12899391158306167, "learning_rate": 4.8333333333333334e-05, "loss": 1.003, "step": 667 }, { "epoch": 0.302536231884058, "grad_norm": 0.12010425653762245, "learning_rate": 4.840579710144928e-05, "loss": 1.008, "step": 668 }, { "epoch": 0.3029891304347826, "grad_norm": 0.09377585207976177, "learning_rate": 4.847826086956522e-05, "loss": 0.9963, "step": 669 }, { "epoch": 0.30344202898550726, "grad_norm": 0.11730500269132754, "learning_rate": 4.8550724637681166e-05, "loss": 0.9736, "step": 670 }, { "epoch": 0.3038949275362319, "grad_norm": 0.14983516909391728, "learning_rate": 4.8623188405797106e-05, "loss": 1.0077, "step": 671 }, { "epoch": 0.30434782608695654, "grad_norm": 0.18964080490270915, "learning_rate": 4.869565217391305e-05, "loss": 0.9928, "step": 672 }, { "epoch": 0.30480072463768115, "grad_norm": 0.21846573529341945, "learning_rate": 4.8768115942028986e-05, "loss": 0.9936, "step": 673 }, { "epoch": 0.3052536231884058, "grad_norm": 0.25380731983216703, "learning_rate": 4.884057971014493e-05, "loss": 0.966, "step": 674 }, { "epoch": 0.30570652173913043, "grad_norm": 0.2919905353373322, "learning_rate": 4.891304347826087e-05, "loss": 1.0152, "step": 675 }, { "epoch": 0.3061594202898551, "grad_norm": 0.2755388243059558, "learning_rate": 4.898550724637682e-05, "loss": 0.9951, "step": 676 }, { "epoch": 0.3066123188405797, "grad_norm": 0.21727470128498255, "learning_rate": 4.905797101449276e-05, "loss": 1.0178, "step": 677 }, { "epoch": 0.3070652173913043, "grad_norm": 0.17676183788725341, "learning_rate": 4.91304347826087e-05, "loss": 1.0131, "step": 678 }, { "epoch": 0.307518115942029, "grad_norm": 0.1795914686962457, "learning_rate": 4.9202898550724645e-05, "loss": 0.9856, "step": 679 }, { "epoch": 0.3079710144927536, "grad_norm": 0.2008541053890237, "learning_rate": 4.927536231884058e-05, "loss": 1.0149, "step": 680 }, { "epoch": 0.30842391304347827, "grad_norm": 0.16707384778070594, "learning_rate": 4.9347826086956524e-05, "loss": 0.9775, "step": 681 }, { "epoch": 0.3088768115942029, "grad_norm": 0.12178536212523451, "learning_rate": 4.9420289855072464e-05, "loss": 1.0033, "step": 682 }, { "epoch": 0.30932971014492755, "grad_norm": 0.10447828289496852, "learning_rate": 4.949275362318841e-05, "loss": 0.9932, "step": 683 }, { "epoch": 0.30978260869565216, "grad_norm": 0.10384979272519106, "learning_rate": 4.956521739130435e-05, "loss": 1.0055, "step": 684 }, { "epoch": 0.31023550724637683, "grad_norm": 0.13978752476121034, "learning_rate": 4.96376811594203e-05, "loss": 1.013, "step": 685 }, { "epoch": 0.31068840579710144, "grad_norm": 0.2321397703073103, "learning_rate": 4.9710144927536237e-05, "loss": 0.9704, "step": 686 }, { "epoch": 0.3111413043478261, "grad_norm": 0.24443162852303046, "learning_rate": 4.978260869565218e-05, "loss": 1.0093, "step": 687 }, { "epoch": 0.3115942028985507, "grad_norm": 0.1791189248945562, "learning_rate": 4.9855072463768116e-05, "loss": 0.9878, "step": 688 }, { "epoch": 0.3120471014492754, "grad_norm": 0.10802116276950553, "learning_rate": 4.992753623188407e-05, "loss": 1.0018, "step": 689 }, { "epoch": 0.3125, "grad_norm": 0.10794971639004008, "learning_rate": 5e-05, "loss": 0.9927, "step": 690 }, { "epoch": 0.3129528985507246, "grad_norm": 0.1087134058001799, "learning_rate": 5.007246376811594e-05, "loss": 0.9844, "step": 691 }, { "epoch": 0.3134057971014493, "grad_norm": 0.11731601612184206, "learning_rate": 5.014492753623189e-05, "loss": 1.0218, "step": 692 }, { "epoch": 0.3138586956521739, "grad_norm": 0.11317180859811185, "learning_rate": 5.021739130434783e-05, "loss": 1.0095, "step": 693 }, { "epoch": 0.31431159420289856, "grad_norm": 0.11536735328683041, "learning_rate": 5.0289855072463775e-05, "loss": 0.9989, "step": 694 }, { "epoch": 0.31476449275362317, "grad_norm": 0.11618858261341625, "learning_rate": 5.0362318840579715e-05, "loss": 0.9813, "step": 695 }, { "epoch": 0.31521739130434784, "grad_norm": 0.1363668862792484, "learning_rate": 5.043478260869566e-05, "loss": 0.9865, "step": 696 }, { "epoch": 0.31567028985507245, "grad_norm": 0.1982660665381474, "learning_rate": 5.0507246376811594e-05, "loss": 1.0067, "step": 697 }, { "epoch": 0.3161231884057971, "grad_norm": 0.2615539761421687, "learning_rate": 5.057971014492754e-05, "loss": 1.0142, "step": 698 }, { "epoch": 0.31657608695652173, "grad_norm": 0.2841898597963357, "learning_rate": 5.065217391304348e-05, "loss": 1.0096, "step": 699 }, { "epoch": 0.3170289855072464, "grad_norm": 0.23656416055749538, "learning_rate": 5.072463768115943e-05, "loss": 1.0141, "step": 700 }, { "epoch": 0.317481884057971, "grad_norm": 0.13746473872646184, "learning_rate": 5.079710144927537e-05, "loss": 0.9992, "step": 701 }, { "epoch": 0.3179347826086957, "grad_norm": 0.15417343455997234, "learning_rate": 5.0869565217391313e-05, "loss": 0.9833, "step": 702 }, { "epoch": 0.3183876811594203, "grad_norm": 0.23104637771400746, "learning_rate": 5.094202898550725e-05, "loss": 0.9861, "step": 703 }, { "epoch": 0.3188405797101449, "grad_norm": 0.23672563035784147, "learning_rate": 5.1014492753623186e-05, "loss": 0.9989, "step": 704 }, { "epoch": 0.31929347826086957, "grad_norm": 0.22667571054961924, "learning_rate": 5.108695652173913e-05, "loss": 0.9754, "step": 705 }, { "epoch": 0.3197463768115942, "grad_norm": 0.23231110878576208, "learning_rate": 5.115942028985507e-05, "loss": 0.9957, "step": 706 }, { "epoch": 0.32019927536231885, "grad_norm": 0.21868555226727715, "learning_rate": 5.123188405797102e-05, "loss": 1.0268, "step": 707 }, { "epoch": 0.32065217391304346, "grad_norm": 0.17615115211061075, "learning_rate": 5.130434782608696e-05, "loss": 0.9953, "step": 708 }, { "epoch": 0.3211050724637681, "grad_norm": 0.10402916590753102, "learning_rate": 5.1376811594202905e-05, "loss": 1.0141, "step": 709 }, { "epoch": 0.32155797101449274, "grad_norm": 0.10635338064502127, "learning_rate": 5.1449275362318845e-05, "loss": 1.0094, "step": 710 }, { "epoch": 0.3220108695652174, "grad_norm": 0.13506040079661757, "learning_rate": 5.152173913043479e-05, "loss": 0.9936, "step": 711 }, { "epoch": 0.322463768115942, "grad_norm": 0.15190081089471297, "learning_rate": 5.1594202898550725e-05, "loss": 0.9931, "step": 712 }, { "epoch": 0.3229166666666667, "grad_norm": 0.1990273068777892, "learning_rate": 5.166666666666667e-05, "loss": 1.0014, "step": 713 }, { "epoch": 0.3233695652173913, "grad_norm": 0.2546373247653369, "learning_rate": 5.173913043478261e-05, "loss": 0.9916, "step": 714 }, { "epoch": 0.32382246376811596, "grad_norm": 0.2597769184811472, "learning_rate": 5.181159420289856e-05, "loss": 0.9961, "step": 715 }, { "epoch": 0.3242753623188406, "grad_norm": 0.19860365439873426, "learning_rate": 5.18840579710145e-05, "loss": 0.9799, "step": 716 }, { "epoch": 0.32472826086956524, "grad_norm": 0.17496456148536793, "learning_rate": 5.1956521739130444e-05, "loss": 0.9662, "step": 717 }, { "epoch": 0.32518115942028986, "grad_norm": 0.17703025444219603, "learning_rate": 5.2028985507246384e-05, "loss": 1.0029, "step": 718 }, { "epoch": 0.32563405797101447, "grad_norm": 0.18474232096539855, "learning_rate": 5.2101449275362317e-05, "loss": 0.9891, "step": 719 }, { "epoch": 0.32608695652173914, "grad_norm": 0.16092620569217334, "learning_rate": 5.217391304347826e-05, "loss": 0.9729, "step": 720 }, { "epoch": 0.32653985507246375, "grad_norm": 0.17506545695737186, "learning_rate": 5.22463768115942e-05, "loss": 0.9857, "step": 721 }, { "epoch": 0.3269927536231884, "grad_norm": 0.17749848680634694, "learning_rate": 5.231884057971015e-05, "loss": 0.9952, "step": 722 }, { "epoch": 0.327445652173913, "grad_norm": 0.20813354923868552, "learning_rate": 5.239130434782609e-05, "loss": 0.9991, "step": 723 }, { "epoch": 0.3278985507246377, "grad_norm": 0.1632912241961635, "learning_rate": 5.2463768115942036e-05, "loss": 0.9718, "step": 724 }, { "epoch": 0.3283514492753623, "grad_norm": 0.1031206574739428, "learning_rate": 5.2536231884057975e-05, "loss": 0.983, "step": 725 }, { "epoch": 0.328804347826087, "grad_norm": 0.14867964108016393, "learning_rate": 5.260869565217392e-05, "loss": 1.0057, "step": 726 }, { "epoch": 0.3292572463768116, "grad_norm": 0.17043435072050814, "learning_rate": 5.2681159420289855e-05, "loss": 0.9896, "step": 727 }, { "epoch": 0.32971014492753625, "grad_norm": 0.12662889175328532, "learning_rate": 5.27536231884058e-05, "loss": 1.0117, "step": 728 }, { "epoch": 0.33016304347826086, "grad_norm": 0.13830682590686438, "learning_rate": 5.282608695652174e-05, "loss": 0.9913, "step": 729 }, { "epoch": 0.33061594202898553, "grad_norm": 0.12422820585002936, "learning_rate": 5.289855072463769e-05, "loss": 0.9837, "step": 730 }, { "epoch": 0.33106884057971014, "grad_norm": 0.14205514403378833, "learning_rate": 5.297101449275363e-05, "loss": 1.0095, "step": 731 }, { "epoch": 0.33152173913043476, "grad_norm": 0.2114661606757739, "learning_rate": 5.304347826086957e-05, "loss": 0.9977, "step": 732 }, { "epoch": 0.3319746376811594, "grad_norm": 0.21840074532519635, "learning_rate": 5.3115942028985514e-05, "loss": 0.9896, "step": 733 }, { "epoch": 0.33242753623188404, "grad_norm": 0.21357109983349887, "learning_rate": 5.318840579710145e-05, "loss": 1.0, "step": 734 }, { "epoch": 0.3328804347826087, "grad_norm": 0.164987614480511, "learning_rate": 5.32608695652174e-05, "loss": 0.9887, "step": 735 }, { "epoch": 0.3333333333333333, "grad_norm": 0.13886743251559888, "learning_rate": 5.333333333333333e-05, "loss": 1.0157, "step": 736 }, { "epoch": 0.333786231884058, "grad_norm": 0.16323205283062045, "learning_rate": 5.340579710144928e-05, "loss": 0.9885, "step": 737 }, { "epoch": 0.3342391304347826, "grad_norm": 0.18051944682249976, "learning_rate": 5.347826086956522e-05, "loss": 1.0145, "step": 738 }, { "epoch": 0.33469202898550726, "grad_norm": 0.18268200272182997, "learning_rate": 5.3550724637681166e-05, "loss": 0.9911, "step": 739 }, { "epoch": 0.3351449275362319, "grad_norm": 0.1479115972551013, "learning_rate": 5.3623188405797106e-05, "loss": 1.0149, "step": 740 }, { "epoch": 0.33559782608695654, "grad_norm": 0.10620918728667819, "learning_rate": 5.369565217391305e-05, "loss": 0.9742, "step": 741 }, { "epoch": 0.33605072463768115, "grad_norm": 0.11039042175968185, "learning_rate": 5.376811594202899e-05, "loss": 0.978, "step": 742 }, { "epoch": 0.3365036231884058, "grad_norm": 0.7304994191135934, "learning_rate": 5.384057971014494e-05, "loss": 0.9853, "step": 743 }, { "epoch": 0.33695652173913043, "grad_norm": 0.11958093011700342, "learning_rate": 5.391304347826087e-05, "loss": 0.9973, "step": 744 }, { "epoch": 0.3374094202898551, "grad_norm": 0.17256960632454438, "learning_rate": 5.398550724637682e-05, "loss": 1.0088, "step": 745 }, { "epoch": 0.3378623188405797, "grad_norm": 0.24199524913528578, "learning_rate": 5.405797101449276e-05, "loss": 1.0033, "step": 746 }, { "epoch": 0.3383152173913043, "grad_norm": 0.25082758138923433, "learning_rate": 5.41304347826087e-05, "loss": 0.9902, "step": 747 }, { "epoch": 0.338768115942029, "grad_norm": 0.2327202320569457, "learning_rate": 5.4202898550724644e-05, "loss": 0.9776, "step": 748 }, { "epoch": 0.3392210144927536, "grad_norm": 0.1927270873940652, "learning_rate": 5.4275362318840584e-05, "loss": 0.9785, "step": 749 }, { "epoch": 0.33967391304347827, "grad_norm": 0.1366043337974577, "learning_rate": 5.434782608695653e-05, "loss": 0.9958, "step": 750 }, { "epoch": 0.3401268115942029, "grad_norm": 0.17350229566023562, "learning_rate": 5.4420289855072464e-05, "loss": 0.9927, "step": 751 }, { "epoch": 0.34057971014492755, "grad_norm": 0.14147561506391504, "learning_rate": 5.449275362318841e-05, "loss": 0.986, "step": 752 }, { "epoch": 0.34103260869565216, "grad_norm": 0.14315854473932776, "learning_rate": 5.456521739130435e-05, "loss": 1.0054, "step": 753 }, { "epoch": 0.34148550724637683, "grad_norm": 0.12549179269249194, "learning_rate": 5.4637681159420296e-05, "loss": 1.0023, "step": 754 }, { "epoch": 0.34193840579710144, "grad_norm": 0.14588367922799175, "learning_rate": 5.4710144927536236e-05, "loss": 0.9813, "step": 755 }, { "epoch": 0.3423913043478261, "grad_norm": 0.16804274178841688, "learning_rate": 5.478260869565218e-05, "loss": 0.9875, "step": 756 }, { "epoch": 0.3428442028985507, "grad_norm": 0.13614361241091982, "learning_rate": 5.485507246376812e-05, "loss": 1.0152, "step": 757 }, { "epoch": 0.3432971014492754, "grad_norm": 0.13265547447672862, "learning_rate": 5.492753623188407e-05, "loss": 0.9759, "step": 758 }, { "epoch": 0.34375, "grad_norm": 0.1364539477711312, "learning_rate": 5.5e-05, "loss": 0.994, "step": 759 }, { "epoch": 0.3442028985507246, "grad_norm": 0.1611849027703029, "learning_rate": 5.507246376811594e-05, "loss": 0.9755, "step": 760 }, { "epoch": 0.3446557971014493, "grad_norm": 0.24984804171288988, "learning_rate": 5.514492753623189e-05, "loss": 0.9823, "step": 761 }, { "epoch": 0.3451086956521739, "grad_norm": 0.25659256471235964, "learning_rate": 5.521739130434783e-05, "loss": 0.9883, "step": 762 }, { "epoch": 0.34556159420289856, "grad_norm": 0.20076340411058066, "learning_rate": 5.5289855072463775e-05, "loss": 0.9937, "step": 763 }, { "epoch": 0.34601449275362317, "grad_norm": 0.19543311324615165, "learning_rate": 5.5362318840579714e-05, "loss": 0.9963, "step": 764 }, { "epoch": 0.34646739130434784, "grad_norm": 0.1525872312479798, "learning_rate": 5.543478260869566e-05, "loss": 0.9805, "step": 765 }, { "epoch": 0.34692028985507245, "grad_norm": 0.16095088726652113, "learning_rate": 5.5507246376811594e-05, "loss": 1.0143, "step": 766 }, { "epoch": 0.3473731884057971, "grad_norm": 0.20485410814596508, "learning_rate": 5.557971014492754e-05, "loss": 0.9878, "step": 767 }, { "epoch": 0.34782608695652173, "grad_norm": 0.3205556492767539, "learning_rate": 5.565217391304348e-05, "loss": 0.9942, "step": 768 }, { "epoch": 0.3482789855072464, "grad_norm": 0.3817168789219102, "learning_rate": 5.572463768115943e-05, "loss": 0.987, "step": 769 }, { "epoch": 0.348731884057971, "grad_norm": 0.3196338352053342, "learning_rate": 5.5797101449275366e-05, "loss": 1.0198, "step": 770 }, { "epoch": 0.3491847826086957, "grad_norm": 0.1827902145437277, "learning_rate": 5.586956521739131e-05, "loss": 0.994, "step": 771 }, { "epoch": 0.3496376811594203, "grad_norm": 0.12684374580611532, "learning_rate": 5.594202898550725e-05, "loss": 0.971, "step": 772 }, { "epoch": 0.3500905797101449, "grad_norm": 0.15544607255545626, "learning_rate": 5.6014492753623186e-05, "loss": 0.9789, "step": 773 }, { "epoch": 0.35054347826086957, "grad_norm": 0.1657874702222299, "learning_rate": 5.608695652173913e-05, "loss": 0.9997, "step": 774 }, { "epoch": 0.3509963768115942, "grad_norm": 0.15714274144199367, "learning_rate": 5.615942028985507e-05, "loss": 0.984, "step": 775 }, { "epoch": 0.35144927536231885, "grad_norm": 0.19092915770918734, "learning_rate": 5.623188405797102e-05, "loss": 0.9985, "step": 776 }, { "epoch": 0.35190217391304346, "grad_norm": 0.17020471630475914, "learning_rate": 5.630434782608696e-05, "loss": 0.9969, "step": 777 }, { "epoch": 0.3523550724637681, "grad_norm": 0.16974068692213806, "learning_rate": 5.6376811594202905e-05, "loss": 0.9923, "step": 778 }, { "epoch": 0.35280797101449274, "grad_norm": 0.14285061840756, "learning_rate": 5.6449275362318845e-05, "loss": 0.9964, "step": 779 }, { "epoch": 0.3532608695652174, "grad_norm": 0.12739415646106134, "learning_rate": 5.652173913043479e-05, "loss": 1.0088, "step": 780 }, { "epoch": 0.353713768115942, "grad_norm": 0.11367070295817798, "learning_rate": 5.6594202898550724e-05, "loss": 0.9848, "step": 781 }, { "epoch": 0.3541666666666667, "grad_norm": 0.15069220920877627, "learning_rate": 5.666666666666668e-05, "loss": 0.9978, "step": 782 }, { "epoch": 0.3546195652173913, "grad_norm": 0.16206701757675374, "learning_rate": 5.673913043478261e-05, "loss": 0.9712, "step": 783 }, { "epoch": 0.35507246376811596, "grad_norm": 0.19069138873706143, "learning_rate": 5.681159420289856e-05, "loss": 0.9942, "step": 784 }, { "epoch": 0.3555253623188406, "grad_norm": 0.1751917256365722, "learning_rate": 5.68840579710145e-05, "loss": 0.9968, "step": 785 }, { "epoch": 0.35597826086956524, "grad_norm": 0.164433370440253, "learning_rate": 5.695652173913044e-05, "loss": 1.0059, "step": 786 }, { "epoch": 0.35643115942028986, "grad_norm": 0.2186968833266557, "learning_rate": 5.702898550724638e-05, "loss": 0.9936, "step": 787 }, { "epoch": 0.35688405797101447, "grad_norm": 0.16892939994781872, "learning_rate": 5.710144927536232e-05, "loss": 0.9909, "step": 788 }, { "epoch": 0.35733695652173914, "grad_norm": 0.12863507350786269, "learning_rate": 5.717391304347827e-05, "loss": 0.9864, "step": 789 }, { "epoch": 0.35778985507246375, "grad_norm": 0.1389167924142289, "learning_rate": 5.72463768115942e-05, "loss": 1.0044, "step": 790 }, { "epoch": 0.3582427536231884, "grad_norm": 0.14777468323427448, "learning_rate": 5.731884057971015e-05, "loss": 0.9909, "step": 791 }, { "epoch": 0.358695652173913, "grad_norm": 0.15294536601725225, "learning_rate": 5.739130434782609e-05, "loss": 1.0122, "step": 792 }, { "epoch": 0.3591485507246377, "grad_norm": 0.2314241925287152, "learning_rate": 5.7463768115942035e-05, "loss": 1.0044, "step": 793 }, { "epoch": 0.3596014492753623, "grad_norm": 0.29998454918149337, "learning_rate": 5.7536231884057975e-05, "loss": 0.9833, "step": 794 }, { "epoch": 0.360054347826087, "grad_norm": 0.3013699143208428, "learning_rate": 5.760869565217392e-05, "loss": 0.995, "step": 795 }, { "epoch": 0.3605072463768116, "grad_norm": 0.24112821277128643, "learning_rate": 5.768115942028986e-05, "loss": 0.9999, "step": 796 }, { "epoch": 0.36096014492753625, "grad_norm": 0.21081940270498403, "learning_rate": 5.775362318840581e-05, "loss": 1.0115, "step": 797 }, { "epoch": 0.36141304347826086, "grad_norm": 0.20613467634683688, "learning_rate": 5.782608695652174e-05, "loss": 0.9895, "step": 798 }, { "epoch": 0.36186594202898553, "grad_norm": 0.20232517489007484, "learning_rate": 5.789855072463769e-05, "loss": 1.0053, "step": 799 }, { "epoch": 0.36231884057971014, "grad_norm": 0.20088262149656227, "learning_rate": 5.797101449275363e-05, "loss": 1.0046, "step": 800 }, { "epoch": 0.36277173913043476, "grad_norm": 0.2399438849430171, "learning_rate": 5.804347826086957e-05, "loss": 1.0151, "step": 801 }, { "epoch": 0.3632246376811594, "grad_norm": 0.2706081043028323, "learning_rate": 5.8115942028985513e-05, "loss": 0.977, "step": 802 }, { "epoch": 0.36367753623188404, "grad_norm": 0.3196664766530456, "learning_rate": 5.818840579710145e-05, "loss": 1.0061, "step": 803 }, { "epoch": 0.3641304347826087, "grad_norm": 0.3325966962763208, "learning_rate": 5.82608695652174e-05, "loss": 0.9866, "step": 804 }, { "epoch": 0.3645833333333333, "grad_norm": 0.3149544400239479, "learning_rate": 5.833333333333333e-05, "loss": 0.9936, "step": 805 }, { "epoch": 0.365036231884058, "grad_norm": 0.2608067480913655, "learning_rate": 5.840579710144928e-05, "loss": 1.0137, "step": 806 }, { "epoch": 0.3654891304347826, "grad_norm": 0.16612110230513347, "learning_rate": 5.847826086956522e-05, "loss": 0.9962, "step": 807 }, { "epoch": 0.36594202898550726, "grad_norm": 0.15659472411507802, "learning_rate": 5.8550724637681166e-05, "loss": 0.9959, "step": 808 }, { "epoch": 0.3663949275362319, "grad_norm": 0.18434710347463898, "learning_rate": 5.8623188405797105e-05, "loss": 0.9868, "step": 809 }, { "epoch": 0.36684782608695654, "grad_norm": 0.2652593804017452, "learning_rate": 5.869565217391305e-05, "loss": 1.0021, "step": 810 }, { "epoch": 0.36730072463768115, "grad_norm": 0.2947327129977933, "learning_rate": 5.876811594202899e-05, "loss": 0.9807, "step": 811 }, { "epoch": 0.3677536231884058, "grad_norm": 0.2767890253972671, "learning_rate": 5.884057971014494e-05, "loss": 1.0109, "step": 812 }, { "epoch": 0.36820652173913043, "grad_norm": 0.2670622386689824, "learning_rate": 5.891304347826087e-05, "loss": 0.9981, "step": 813 }, { "epoch": 0.3686594202898551, "grad_norm": 0.22972090634825543, "learning_rate": 5.898550724637682e-05, "loss": 0.9921, "step": 814 }, { "epoch": 0.3691123188405797, "grad_norm": 0.2019693703618123, "learning_rate": 5.905797101449276e-05, "loss": 1.0015, "step": 815 }, { "epoch": 0.3695652173913043, "grad_norm": 0.16542905962941173, "learning_rate": 5.91304347826087e-05, "loss": 0.9838, "step": 816 }, { "epoch": 0.370018115942029, "grad_norm": 0.15144858981705359, "learning_rate": 5.9202898550724644e-05, "loss": 0.9807, "step": 817 }, { "epoch": 0.3704710144927536, "grad_norm": 0.15111405233305697, "learning_rate": 5.9275362318840584e-05, "loss": 0.9857, "step": 818 }, { "epoch": 0.37092391304347827, "grad_norm": 0.17025745870793066, "learning_rate": 5.934782608695653e-05, "loss": 0.9836, "step": 819 }, { "epoch": 0.3713768115942029, "grad_norm": 0.2638847739629444, "learning_rate": 5.942028985507246e-05, "loss": 1.0071, "step": 820 }, { "epoch": 0.37182971014492755, "grad_norm": 0.2088329232501522, "learning_rate": 5.949275362318841e-05, "loss": 0.9789, "step": 821 }, { "epoch": 0.37228260869565216, "grad_norm": 0.161995525999585, "learning_rate": 5.956521739130435e-05, "loss": 0.9975, "step": 822 }, { "epoch": 0.37273550724637683, "grad_norm": 0.12178971911740079, "learning_rate": 5.9637681159420296e-05, "loss": 1.0017, "step": 823 }, { "epoch": 0.37318840579710144, "grad_norm": 0.13010043527781054, "learning_rate": 5.9710144927536236e-05, "loss": 0.9902, "step": 824 }, { "epoch": 0.3736413043478261, "grad_norm": 0.15367955862044244, "learning_rate": 5.978260869565218e-05, "loss": 0.987, "step": 825 }, { "epoch": 0.3740942028985507, "grad_norm": 0.168530753052105, "learning_rate": 5.985507246376812e-05, "loss": 0.9941, "step": 826 }, { "epoch": 0.3745471014492754, "grad_norm": 0.19293709852693744, "learning_rate": 5.992753623188407e-05, "loss": 0.9729, "step": 827 }, { "epoch": 0.375, "grad_norm": 0.2099121253003547, "learning_rate": 6.000000000000001e-05, "loss": 0.9703, "step": 828 }, { "epoch": 0.3754528985507246, "grad_norm": 0.1862957245476465, "learning_rate": 6.007246376811594e-05, "loss": 0.9909, "step": 829 }, { "epoch": 0.3759057971014493, "grad_norm": 0.15794800009854387, "learning_rate": 6.014492753623189e-05, "loss": 0.9836, "step": 830 }, { "epoch": 0.3763586956521739, "grad_norm": 0.15680039789577888, "learning_rate": 6.021739130434783e-05, "loss": 0.9682, "step": 831 }, { "epoch": 0.37681159420289856, "grad_norm": 0.2060601484597629, "learning_rate": 6.0289855072463774e-05, "loss": 0.9924, "step": 832 }, { "epoch": 0.37726449275362317, "grad_norm": 0.23638076121355814, "learning_rate": 6.0362318840579714e-05, "loss": 1.0105, "step": 833 }, { "epoch": 0.37771739130434784, "grad_norm": 0.196824943897607, "learning_rate": 6.043478260869566e-05, "loss": 0.9778, "step": 834 }, { "epoch": 0.37817028985507245, "grad_norm": 0.16142657896651522, "learning_rate": 6.05072463768116e-05, "loss": 0.9968, "step": 835 }, { "epoch": 0.3786231884057971, "grad_norm": 0.1316151089137682, "learning_rate": 6.057971014492755e-05, "loss": 1.0061, "step": 836 }, { "epoch": 0.37907608695652173, "grad_norm": 0.1535852208829198, "learning_rate": 6.065217391304348e-05, "loss": 1.026, "step": 837 }, { "epoch": 0.3795289855072464, "grad_norm": 0.18215978707708974, "learning_rate": 6.0724637681159426e-05, "loss": 1.0247, "step": 838 }, { "epoch": 0.379981884057971, "grad_norm": 0.18115661746504258, "learning_rate": 6.0797101449275366e-05, "loss": 0.9803, "step": 839 }, { "epoch": 0.3804347826086957, "grad_norm": 0.17571666578389655, "learning_rate": 6.086956521739131e-05, "loss": 1.0111, "step": 840 }, { "epoch": 0.3808876811594203, "grad_norm": 0.14345671192896817, "learning_rate": 6.094202898550725e-05, "loss": 0.9926, "step": 841 }, { "epoch": 0.3813405797101449, "grad_norm": 0.16454577413830213, "learning_rate": 6.101449275362319e-05, "loss": 0.9985, "step": 842 }, { "epoch": 0.38179347826086957, "grad_norm": 0.1405844495306108, "learning_rate": 6.108695652173914e-05, "loss": 1.011, "step": 843 }, { "epoch": 0.3822463768115942, "grad_norm": 0.1881747896431805, "learning_rate": 6.115942028985507e-05, "loss": 1.008, "step": 844 }, { "epoch": 0.38269927536231885, "grad_norm": 0.29099456700455845, "learning_rate": 6.123188405797102e-05, "loss": 0.9921, "step": 845 }, { "epoch": 0.38315217391304346, "grad_norm": 0.3255740058921477, "learning_rate": 6.130434782608696e-05, "loss": 0.979, "step": 846 }, { "epoch": 0.3836050724637681, "grad_norm": 0.2954556081549919, "learning_rate": 6.137681159420291e-05, "loss": 1.0292, "step": 847 }, { "epoch": 0.38405797101449274, "grad_norm": 0.2332824083467019, "learning_rate": 6.144927536231884e-05, "loss": 0.9832, "step": 848 }, { "epoch": 0.3845108695652174, "grad_norm": 0.14841450061667827, "learning_rate": 6.152173913043479e-05, "loss": 0.9937, "step": 849 }, { "epoch": 0.384963768115942, "grad_norm": 0.15468480022473316, "learning_rate": 6.159420289855072e-05, "loss": 1.0068, "step": 850 }, { "epoch": 0.3854166666666667, "grad_norm": 0.19136649449655854, "learning_rate": 6.166666666666667e-05, "loss": 1.0047, "step": 851 }, { "epoch": 0.3858695652173913, "grad_norm": 0.1580720650356017, "learning_rate": 6.173913043478262e-05, "loss": 0.9762, "step": 852 }, { "epoch": 0.38632246376811596, "grad_norm": 0.16350361883906037, "learning_rate": 6.181159420289856e-05, "loss": 1.0222, "step": 853 }, { "epoch": 0.3867753623188406, "grad_norm": 0.25918365082450595, "learning_rate": 6.18840579710145e-05, "loss": 0.9826, "step": 854 }, { "epoch": 0.38722826086956524, "grad_norm": 0.3068428720959896, "learning_rate": 6.195652173913044e-05, "loss": 0.9865, "step": 855 }, { "epoch": 0.38768115942028986, "grad_norm": 0.2757030422324209, "learning_rate": 6.202898550724638e-05, "loss": 0.9678, "step": 856 }, { "epoch": 0.38813405797101447, "grad_norm": 0.2620988378002054, "learning_rate": 6.210144927536232e-05, "loss": 0.9934, "step": 857 }, { "epoch": 0.38858695652173914, "grad_norm": 0.30966073810195416, "learning_rate": 6.217391304347827e-05, "loss": 0.9945, "step": 858 }, { "epoch": 0.38903985507246375, "grad_norm": 0.373774218770936, "learning_rate": 6.22463768115942e-05, "loss": 0.9901, "step": 859 }, { "epoch": 0.3894927536231884, "grad_norm": 0.3984071922597216, "learning_rate": 6.231884057971015e-05, "loss": 0.9909, "step": 860 }, { "epoch": 0.389945652173913, "grad_norm": 0.34948958379237627, "learning_rate": 6.23913043478261e-05, "loss": 0.991, "step": 861 }, { "epoch": 0.3903985507246377, "grad_norm": 0.22588423994381815, "learning_rate": 6.246376811594204e-05, "loss": 1.0063, "step": 862 }, { "epoch": 0.3908514492753623, "grad_norm": 0.15546423324444877, "learning_rate": 6.253623188405797e-05, "loss": 0.9972, "step": 863 }, { "epoch": 0.391304347826087, "grad_norm": 0.14814652333362283, "learning_rate": 6.260869565217392e-05, "loss": 0.9962, "step": 864 }, { "epoch": 0.3917572463768116, "grad_norm": 0.23161903027751285, "learning_rate": 6.268115942028985e-05, "loss": 0.9711, "step": 865 }, { "epoch": 0.39221014492753625, "grad_norm": 0.29320317854441386, "learning_rate": 6.27536231884058e-05, "loss": 0.9804, "step": 866 }, { "epoch": 0.39266304347826086, "grad_norm": 0.3106858090400903, "learning_rate": 6.282608695652175e-05, "loss": 0.9749, "step": 867 }, { "epoch": 0.39311594202898553, "grad_norm": 0.2633657635177838, "learning_rate": 6.28985507246377e-05, "loss": 0.9741, "step": 868 }, { "epoch": 0.39356884057971014, "grad_norm": 0.13982621693239283, "learning_rate": 6.297101449275363e-05, "loss": 0.9883, "step": 869 }, { "epoch": 0.39402173913043476, "grad_norm": 0.17623569611425974, "learning_rate": 6.304347826086956e-05, "loss": 0.9809, "step": 870 }, { "epoch": 0.3944746376811594, "grad_norm": 0.2943124921054578, "learning_rate": 6.31159420289855e-05, "loss": 1.0064, "step": 871 }, { "epoch": 0.39492753623188404, "grad_norm": 0.3528572645235867, "learning_rate": 6.318840579710145e-05, "loss": 0.9903, "step": 872 }, { "epoch": 0.3953804347826087, "grad_norm": 0.3831367617717058, "learning_rate": 6.32608695652174e-05, "loss": 0.991, "step": 873 }, { "epoch": 0.3958333333333333, "grad_norm": 0.31261292254944895, "learning_rate": 6.333333333333333e-05, "loss": 0.9799, "step": 874 }, { "epoch": 0.396286231884058, "grad_norm": 0.17071195563593072, "learning_rate": 6.340579710144928e-05, "loss": 0.9886, "step": 875 }, { "epoch": 0.3967391304347826, "grad_norm": 0.16581406200154666, "learning_rate": 6.347826086956523e-05, "loss": 0.9935, "step": 876 }, { "epoch": 0.39719202898550726, "grad_norm": 0.2820977193853305, "learning_rate": 6.355072463768117e-05, "loss": 0.9893, "step": 877 }, { "epoch": 0.3976449275362319, "grad_norm": 0.3270145633406759, "learning_rate": 6.36231884057971e-05, "loss": 0.9742, "step": 878 }, { "epoch": 0.39809782608695654, "grad_norm": 0.30255971797220094, "learning_rate": 6.369565217391305e-05, "loss": 1.0023, "step": 879 }, { "epoch": 0.39855072463768115, "grad_norm": 0.2304255199135394, "learning_rate": 6.376811594202898e-05, "loss": 0.9853, "step": 880 }, { "epoch": 0.3990036231884058, "grad_norm": 0.14458968662996896, "learning_rate": 6.384057971014493e-05, "loss": 0.9942, "step": 881 }, { "epoch": 0.39945652173913043, "grad_norm": 0.1906514301596453, "learning_rate": 6.391304347826088e-05, "loss": 0.9741, "step": 882 }, { "epoch": 0.3999094202898551, "grad_norm": 0.2686474057182902, "learning_rate": 6.398550724637682e-05, "loss": 0.9845, "step": 883 }, { "epoch": 0.4003623188405797, "grad_norm": 0.2607052545027108, "learning_rate": 6.405797101449276e-05, "loss": 1.0057, "step": 884 }, { "epoch": 0.4008152173913043, "grad_norm": 0.22519807208351222, "learning_rate": 6.41304347826087e-05, "loss": 1.0008, "step": 885 }, { "epoch": 0.401268115942029, "grad_norm": 0.17499111263357348, "learning_rate": 6.420289855072465e-05, "loss": 1.0232, "step": 886 }, { "epoch": 0.4017210144927536, "grad_norm": 0.13323446239394304, "learning_rate": 6.427536231884058e-05, "loss": 0.9719, "step": 887 }, { "epoch": 0.40217391304347827, "grad_norm": 0.15773300333649407, "learning_rate": 6.434782608695653e-05, "loss": 0.9771, "step": 888 }, { "epoch": 0.4026268115942029, "grad_norm": 0.24476271930384222, "learning_rate": 6.442028985507246e-05, "loss": 0.97, "step": 889 }, { "epoch": 0.40307971014492755, "grad_norm": 0.2736889245367642, "learning_rate": 6.449275362318841e-05, "loss": 0.9758, "step": 890 }, { "epoch": 0.40353260869565216, "grad_norm": 0.21320641649286173, "learning_rate": 6.456521739130436e-05, "loss": 0.9905, "step": 891 }, { "epoch": 0.40398550724637683, "grad_norm": 0.17820657902023784, "learning_rate": 6.46376811594203e-05, "loss": 0.9897, "step": 892 }, { "epoch": 0.40443840579710144, "grad_norm": 0.18248599521173325, "learning_rate": 6.471014492753624e-05, "loss": 0.985, "step": 893 }, { "epoch": 0.4048913043478261, "grad_norm": 0.18909801783958124, "learning_rate": 6.478260869565218e-05, "loss": 0.9789, "step": 894 }, { "epoch": 0.4053442028985507, "grad_norm": 0.19992270035735044, "learning_rate": 6.485507246376811e-05, "loss": 0.9829, "step": 895 }, { "epoch": 0.4057971014492754, "grad_norm": 0.18329286962483943, "learning_rate": 6.492753623188406e-05, "loss": 0.9734, "step": 896 }, { "epoch": 0.40625, "grad_norm": 0.1846769018486621, "learning_rate": 6.500000000000001e-05, "loss": 1.0225, "step": 897 }, { "epoch": 0.4067028985507246, "grad_norm": 0.20912568633277467, "learning_rate": 6.507246376811594e-05, "loss": 0.9872, "step": 898 }, { "epoch": 0.4071557971014493, "grad_norm": 0.1462007965291803, "learning_rate": 6.514492753623189e-05, "loss": 0.9819, "step": 899 }, { "epoch": 0.4076086956521739, "grad_norm": 0.19995968280331042, "learning_rate": 6.521739130434783e-05, "loss": 0.9818, "step": 900 }, { "epoch": 0.40806159420289856, "grad_norm": 0.21991621375215953, "learning_rate": 6.528985507246378e-05, "loss": 0.9998, "step": 901 }, { "epoch": 0.40851449275362317, "grad_norm": 0.20865054695889515, "learning_rate": 6.536231884057971e-05, "loss": 0.9732, "step": 902 }, { "epoch": 0.40896739130434784, "grad_norm": 0.18410416339429705, "learning_rate": 6.543478260869566e-05, "loss": 0.9929, "step": 903 }, { "epoch": 0.40942028985507245, "grad_norm": 0.16533149586876222, "learning_rate": 6.550724637681159e-05, "loss": 0.9895, "step": 904 }, { "epoch": 0.4098731884057971, "grad_norm": 0.18778829189506963, "learning_rate": 6.557971014492754e-05, "loss": 1.0034, "step": 905 }, { "epoch": 0.41032608695652173, "grad_norm": 0.2111947009565163, "learning_rate": 6.565217391304349e-05, "loss": 0.9617, "step": 906 }, { "epoch": 0.4107789855072464, "grad_norm": 0.19369658686713162, "learning_rate": 6.572463768115943e-05, "loss": 1.002, "step": 907 }, { "epoch": 0.411231884057971, "grad_norm": 0.1769425412803981, "learning_rate": 6.579710144927537e-05, "loss": 0.9979, "step": 908 }, { "epoch": 0.4116847826086957, "grad_norm": 0.18278644816975134, "learning_rate": 6.586956521739131e-05, "loss": 1.0101, "step": 909 }, { "epoch": 0.4121376811594203, "grad_norm": 0.21593124227611765, "learning_rate": 6.594202898550725e-05, "loss": 0.9704, "step": 910 }, { "epoch": 0.4125905797101449, "grad_norm": 0.18630774073156486, "learning_rate": 6.601449275362319e-05, "loss": 0.9973, "step": 911 }, { "epoch": 0.41304347826086957, "grad_norm": 0.12901301692527645, "learning_rate": 6.608695652173914e-05, "loss": 0.9714, "step": 912 }, { "epoch": 0.4134963768115942, "grad_norm": 0.14968650168503614, "learning_rate": 6.615942028985507e-05, "loss": 0.9936, "step": 913 }, { "epoch": 0.41394927536231885, "grad_norm": 0.25045403586334586, "learning_rate": 6.623188405797102e-05, "loss": 1.0005, "step": 914 }, { "epoch": 0.41440217391304346, "grad_norm": 0.26595837878369, "learning_rate": 6.630434782608696e-05, "loss": 1.0122, "step": 915 }, { "epoch": 0.4148550724637681, "grad_norm": 0.26979308447474215, "learning_rate": 6.637681159420291e-05, "loss": 1.0018, "step": 916 }, { "epoch": 0.41530797101449274, "grad_norm": 0.1898243265061809, "learning_rate": 6.644927536231884e-05, "loss": 0.9759, "step": 917 }, { "epoch": 0.4157608695652174, "grad_norm": 0.13384594602651131, "learning_rate": 6.652173913043479e-05, "loss": 0.9961, "step": 918 }, { "epoch": 0.416213768115942, "grad_norm": 0.17384174991157397, "learning_rate": 6.659420289855072e-05, "loss": 0.9827, "step": 919 }, { "epoch": 0.4166666666666667, "grad_norm": 0.168874170530707, "learning_rate": 6.666666666666667e-05, "loss": 0.9797, "step": 920 }, { "epoch": 0.4171195652173913, "grad_norm": 0.19766836642311844, "learning_rate": 6.673913043478262e-05, "loss": 1.0069, "step": 921 }, { "epoch": 0.41757246376811596, "grad_norm": 0.23071248873518596, "learning_rate": 6.681159420289856e-05, "loss": 0.9791, "step": 922 }, { "epoch": 0.4180253623188406, "grad_norm": 0.2940566199625595, "learning_rate": 6.68840579710145e-05, "loss": 1.0093, "step": 923 }, { "epoch": 0.41847826086956524, "grad_norm": 0.3155241672573196, "learning_rate": 6.695652173913044e-05, "loss": 1.0004, "step": 924 }, { "epoch": 0.41893115942028986, "grad_norm": 0.296022330311441, "learning_rate": 6.702898550724639e-05, "loss": 1.0047, "step": 925 }, { "epoch": 0.41938405797101447, "grad_norm": 0.21976575595653253, "learning_rate": 6.710144927536232e-05, "loss": 0.9759, "step": 926 }, { "epoch": 0.41983695652173914, "grad_norm": 0.16432344586253017, "learning_rate": 6.717391304347827e-05, "loss": 1.0067, "step": 927 }, { "epoch": 0.42028985507246375, "grad_norm": 0.1594961345381402, "learning_rate": 6.72463768115942e-05, "loss": 0.9892, "step": 928 }, { "epoch": 0.4207427536231884, "grad_norm": 0.1528614597497426, "learning_rate": 6.731884057971015e-05, "loss": 1.0113, "step": 929 }, { "epoch": 0.421195652173913, "grad_norm": 0.17039270363302692, "learning_rate": 6.73913043478261e-05, "loss": 0.9934, "step": 930 }, { "epoch": 0.4216485507246377, "grad_norm": 0.19390303455059085, "learning_rate": 6.746376811594204e-05, "loss": 0.9723, "step": 931 }, { "epoch": 0.4221014492753623, "grad_norm": 0.1932914145084164, "learning_rate": 6.753623188405797e-05, "loss": 0.9915, "step": 932 }, { "epoch": 0.422554347826087, "grad_norm": 0.1727792876563812, "learning_rate": 6.760869565217392e-05, "loss": 0.9997, "step": 933 }, { "epoch": 0.4230072463768116, "grad_norm": 0.17903767602313078, "learning_rate": 6.768115942028985e-05, "loss": 1.0162, "step": 934 }, { "epoch": 0.42346014492753625, "grad_norm": 0.14951279306869927, "learning_rate": 6.77536231884058e-05, "loss": 1.0108, "step": 935 }, { "epoch": 0.42391304347826086, "grad_norm": 0.15628407411022094, "learning_rate": 6.782608695652175e-05, "loss": 0.9747, "step": 936 }, { "epoch": 0.42436594202898553, "grad_norm": 0.18028827554233515, "learning_rate": 6.78985507246377e-05, "loss": 0.9689, "step": 937 }, { "epoch": 0.42481884057971014, "grad_norm": 0.21682344719266694, "learning_rate": 6.797101449275363e-05, "loss": 0.989, "step": 938 }, { "epoch": 0.42527173913043476, "grad_norm": 0.22187008989915133, "learning_rate": 6.804347826086957e-05, "loss": 0.9781, "step": 939 }, { "epoch": 0.4257246376811594, "grad_norm": 0.2068125192524301, "learning_rate": 6.811594202898552e-05, "loss": 0.9717, "step": 940 }, { "epoch": 0.42617753623188404, "grad_norm": 0.2472942240347539, "learning_rate": 6.818840579710145e-05, "loss": 0.9765, "step": 941 }, { "epoch": 0.4266304347826087, "grad_norm": 0.25384571640831527, "learning_rate": 6.82608695652174e-05, "loss": 0.9976, "step": 942 }, { "epoch": 0.4270833333333333, "grad_norm": 0.21715759434568102, "learning_rate": 6.833333333333333e-05, "loss": 0.9581, "step": 943 }, { "epoch": 0.427536231884058, "grad_norm": 0.18322553468422184, "learning_rate": 6.840579710144928e-05, "loss": 0.9732, "step": 944 }, { "epoch": 0.4279891304347826, "grad_norm": 0.19655265971601976, "learning_rate": 6.847826086956522e-05, "loss": 0.9608, "step": 945 }, { "epoch": 0.42844202898550726, "grad_norm": 0.17669109127051633, "learning_rate": 6.855072463768117e-05, "loss": 0.9845, "step": 946 }, { "epoch": 0.4288949275362319, "grad_norm": 0.13630645195623786, "learning_rate": 6.86231884057971e-05, "loss": 0.9898, "step": 947 }, { "epoch": 0.42934782608695654, "grad_norm": 0.19910794307665386, "learning_rate": 6.869565217391305e-05, "loss": 0.9838, "step": 948 }, { "epoch": 0.42980072463768115, "grad_norm": 0.2600089266804566, "learning_rate": 6.876811594202898e-05, "loss": 0.9742, "step": 949 }, { "epoch": 0.4302536231884058, "grad_norm": 0.2666125547170936, "learning_rate": 6.884057971014493e-05, "loss": 0.9979, "step": 950 }, { "epoch": 0.43070652173913043, "grad_norm": 0.22048783306722317, "learning_rate": 6.891304347826088e-05, "loss": 0.9805, "step": 951 }, { "epoch": 0.4311594202898551, "grad_norm": 0.19550684357752265, "learning_rate": 6.898550724637682e-05, "loss": 0.9911, "step": 952 }, { "epoch": 0.4316123188405797, "grad_norm": 0.20214780643547164, "learning_rate": 6.905797101449276e-05, "loss": 0.9876, "step": 953 }, { "epoch": 0.4320652173913043, "grad_norm": 0.19018011549853694, "learning_rate": 6.91304347826087e-05, "loss": 0.9887, "step": 954 }, { "epoch": 0.432518115942029, "grad_norm": 0.12666643174053194, "learning_rate": 6.920289855072465e-05, "loss": 0.9879, "step": 955 }, { "epoch": 0.4329710144927536, "grad_norm": 0.15697523591098275, "learning_rate": 6.927536231884058e-05, "loss": 0.988, "step": 956 }, { "epoch": 0.43342391304347827, "grad_norm": 0.1944509524638277, "learning_rate": 6.934782608695653e-05, "loss": 1.0093, "step": 957 }, { "epoch": 0.4338768115942029, "grad_norm": 0.18210346910793032, "learning_rate": 6.942028985507246e-05, "loss": 0.9616, "step": 958 }, { "epoch": 0.43432971014492755, "grad_norm": 0.14390952531707293, "learning_rate": 6.949275362318841e-05, "loss": 0.9833, "step": 959 }, { "epoch": 0.43478260869565216, "grad_norm": 0.14243546487005723, "learning_rate": 6.956521739130436e-05, "loss": 0.9766, "step": 960 }, { "epoch": 0.43523550724637683, "grad_norm": 0.1516464981028082, "learning_rate": 6.96376811594203e-05, "loss": 0.9935, "step": 961 }, { "epoch": 0.43568840579710144, "grad_norm": 0.13300859815227534, "learning_rate": 6.971014492753623e-05, "loss": 0.9785, "step": 962 }, { "epoch": 0.4361413043478261, "grad_norm": 0.19686909441468814, "learning_rate": 6.978260869565218e-05, "loss": 1.01, "step": 963 }, { "epoch": 0.4365942028985507, "grad_norm": 0.19198251297701666, "learning_rate": 6.985507246376811e-05, "loss": 0.995, "step": 964 }, { "epoch": 0.4370471014492754, "grad_norm": 0.21925409576627858, "learning_rate": 6.992753623188407e-05, "loss": 0.9909, "step": 965 }, { "epoch": 0.4375, "grad_norm": 0.24173208253176762, "learning_rate": 7.000000000000001e-05, "loss": 0.9813, "step": 966 }, { "epoch": 0.4379528985507246, "grad_norm": 0.2520704536848573, "learning_rate": 7.007246376811594e-05, "loss": 1.0026, "step": 967 }, { "epoch": 0.4384057971014493, "grad_norm": 0.2815293741127259, "learning_rate": 7.014492753623189e-05, "loss": 1.0032, "step": 968 }, { "epoch": 0.4388586956521739, "grad_norm": 0.29553587350535165, "learning_rate": 7.021739130434783e-05, "loss": 0.9759, "step": 969 }, { "epoch": 0.43931159420289856, "grad_norm": 0.29722806734959156, "learning_rate": 7.028985507246378e-05, "loss": 1.0252, "step": 970 }, { "epoch": 0.43976449275362317, "grad_norm": 0.2717704576317258, "learning_rate": 7.036231884057971e-05, "loss": 0.9799, "step": 971 }, { "epoch": 0.44021739130434784, "grad_norm": 0.21977206729958193, "learning_rate": 7.043478260869566e-05, "loss": 1.0057, "step": 972 }, { "epoch": 0.44067028985507245, "grad_norm": 0.15836383577333635, "learning_rate": 7.050724637681159e-05, "loss": 0.9886, "step": 973 }, { "epoch": 0.4411231884057971, "grad_norm": 0.1336075255269942, "learning_rate": 7.057971014492754e-05, "loss": 0.9779, "step": 974 }, { "epoch": 0.44157608695652173, "grad_norm": 0.19944394750736075, "learning_rate": 7.065217391304349e-05, "loss": 0.9919, "step": 975 }, { "epoch": 0.4420289855072464, "grad_norm": 0.25641624899227367, "learning_rate": 7.072463768115943e-05, "loss": 1.0041, "step": 976 }, { "epoch": 0.442481884057971, "grad_norm": 0.23576320500422154, "learning_rate": 7.079710144927537e-05, "loss": 1.0085, "step": 977 }, { "epoch": 0.4429347826086957, "grad_norm": 0.2148128869983808, "learning_rate": 7.086956521739131e-05, "loss": 0.9668, "step": 978 }, { "epoch": 0.4433876811594203, "grad_norm": 0.21890826639499752, "learning_rate": 7.094202898550726e-05, "loss": 0.9823, "step": 979 }, { "epoch": 0.4438405797101449, "grad_norm": 0.2392136404616325, "learning_rate": 7.101449275362319e-05, "loss": 0.9853, "step": 980 }, { "epoch": 0.44429347826086957, "grad_norm": 0.30360415468256235, "learning_rate": 7.108695652173914e-05, "loss": 0.9899, "step": 981 }, { "epoch": 0.4447463768115942, "grad_norm": 0.4066165694538652, "learning_rate": 7.115942028985507e-05, "loss": 0.9945, "step": 982 }, { "epoch": 0.44519927536231885, "grad_norm": 0.4871360693612673, "learning_rate": 7.123188405797102e-05, "loss": 0.9735, "step": 983 }, { "epoch": 0.44565217391304346, "grad_norm": 0.4898181877389764, "learning_rate": 7.130434782608696e-05, "loss": 0.9806, "step": 984 }, { "epoch": 0.4461050724637681, "grad_norm": 0.40608748528582667, "learning_rate": 7.137681159420291e-05, "loss": 0.9926, "step": 985 }, { "epoch": 0.44655797101449274, "grad_norm": 0.31244768407512613, "learning_rate": 7.144927536231884e-05, "loss": 0.9832, "step": 986 }, { "epoch": 0.4470108695652174, "grad_norm": 0.23898951250801984, "learning_rate": 7.152173913043479e-05, "loss": 1.0001, "step": 987 }, { "epoch": 0.447463768115942, "grad_norm": 0.28394720593216943, "learning_rate": 7.159420289855072e-05, "loss": 1.0007, "step": 988 }, { "epoch": 0.4479166666666667, "grad_norm": 0.3010335690178386, "learning_rate": 7.166666666666667e-05, "loss": 0.9894, "step": 989 }, { "epoch": 0.4483695652173913, "grad_norm": 0.22360973846076448, "learning_rate": 7.173913043478262e-05, "loss": 0.9834, "step": 990 }, { "epoch": 0.44882246376811596, "grad_norm": 0.18642777329629728, "learning_rate": 7.181159420289856e-05, "loss": 0.9781, "step": 991 }, { "epoch": 0.4492753623188406, "grad_norm": 0.20529307162999458, "learning_rate": 7.18840579710145e-05, "loss": 0.9956, "step": 992 }, { "epoch": 0.44972826086956524, "grad_norm": 0.22749520928547115, "learning_rate": 7.195652173913044e-05, "loss": 1.0104, "step": 993 }, { "epoch": 0.45018115942028986, "grad_norm": 0.18820484218650446, "learning_rate": 7.202898550724639e-05, "loss": 0.99, "step": 994 }, { "epoch": 0.45063405797101447, "grad_norm": 0.20286202523618882, "learning_rate": 7.210144927536232e-05, "loss": 0.9896, "step": 995 }, { "epoch": 0.45108695652173914, "grad_norm": 0.22895163314799416, "learning_rate": 7.217391304347827e-05, "loss": 0.9805, "step": 996 }, { "epoch": 0.45153985507246375, "grad_norm": 0.25088916340578404, "learning_rate": 7.22463768115942e-05, "loss": 0.9889, "step": 997 }, { "epoch": 0.4519927536231884, "grad_norm": 0.26664023965346906, "learning_rate": 7.231884057971015e-05, "loss": 0.9819, "step": 998 }, { "epoch": 0.452445652173913, "grad_norm": 0.3089540217787421, "learning_rate": 7.23913043478261e-05, "loss": 0.9973, "step": 999 }, { "epoch": 0.4528985507246377, "grad_norm": 0.26510775676303927, "learning_rate": 7.246376811594204e-05, "loss": 0.9717, "step": 1000 }, { "epoch": 0.4533514492753623, "grad_norm": 0.2083416665711474, "learning_rate": 7.253623188405797e-05, "loss": 0.9837, "step": 1001 }, { "epoch": 0.453804347826087, "grad_norm": 0.15027717308661256, "learning_rate": 7.260869565217392e-05, "loss": 0.9804, "step": 1002 }, { "epoch": 0.4542572463768116, "grad_norm": 0.1494476197337847, "learning_rate": 7.268115942028985e-05, "loss": 0.9857, "step": 1003 }, { "epoch": 0.45471014492753625, "grad_norm": 0.16272073591894012, "learning_rate": 7.27536231884058e-05, "loss": 1.0023, "step": 1004 }, { "epoch": 0.45516304347826086, "grad_norm": 0.1905551102202482, "learning_rate": 7.282608695652175e-05, "loss": 0.9918, "step": 1005 }, { "epoch": 0.45561594202898553, "grad_norm": 0.19173902264040485, "learning_rate": 7.289855072463769e-05, "loss": 0.9944, "step": 1006 }, { "epoch": 0.45606884057971014, "grad_norm": 0.2065758268460471, "learning_rate": 7.297101449275363e-05, "loss": 0.9872, "step": 1007 }, { "epoch": 0.45652173913043476, "grad_norm": 0.23532348740429557, "learning_rate": 7.304347826086957e-05, "loss": 0.9969, "step": 1008 }, { "epoch": 0.4569746376811594, "grad_norm": 0.2616177946812646, "learning_rate": 7.311594202898552e-05, "loss": 1.0053, "step": 1009 }, { "epoch": 0.45742753623188404, "grad_norm": 0.23117444421348157, "learning_rate": 7.318840579710145e-05, "loss": 0.9948, "step": 1010 }, { "epoch": 0.4578804347826087, "grad_norm": 0.166699412894635, "learning_rate": 7.32608695652174e-05, "loss": 0.9821, "step": 1011 }, { "epoch": 0.4583333333333333, "grad_norm": 0.18534391683725301, "learning_rate": 7.333333333333333e-05, "loss": 1.0112, "step": 1012 }, { "epoch": 0.458786231884058, "grad_norm": 0.19858316501285614, "learning_rate": 7.340579710144928e-05, "loss": 0.9834, "step": 1013 }, { "epoch": 0.4592391304347826, "grad_norm": 0.1908047568769865, "learning_rate": 7.347826086956522e-05, "loss": 0.9582, "step": 1014 }, { "epoch": 0.45969202898550726, "grad_norm": 0.12688266989215072, "learning_rate": 7.355072463768117e-05, "loss": 0.9981, "step": 1015 }, { "epoch": 0.4601449275362319, "grad_norm": 0.12786750585151024, "learning_rate": 7.36231884057971e-05, "loss": 0.9794, "step": 1016 }, { "epoch": 0.46059782608695654, "grad_norm": 0.16409693685920976, "learning_rate": 7.369565217391305e-05, "loss": 0.9809, "step": 1017 }, { "epoch": 0.46105072463768115, "grad_norm": 0.19630916430343676, "learning_rate": 7.3768115942029e-05, "loss": 1.0045, "step": 1018 }, { "epoch": 0.4615036231884058, "grad_norm": 0.1916453098825203, "learning_rate": 7.384057971014494e-05, "loss": 0.9954, "step": 1019 }, { "epoch": 0.46195652173913043, "grad_norm": 0.1500749700572157, "learning_rate": 7.391304347826088e-05, "loss": 0.9622, "step": 1020 }, { "epoch": 0.4624094202898551, "grad_norm": 0.1436786812214122, "learning_rate": 7.398550724637682e-05, "loss": 1.0136, "step": 1021 }, { "epoch": 0.4628623188405797, "grad_norm": 0.1219533336529064, "learning_rate": 7.405797101449276e-05, "loss": 0.9801, "step": 1022 }, { "epoch": 0.4633152173913043, "grad_norm": 0.1419823240922593, "learning_rate": 7.41304347826087e-05, "loss": 0.9745, "step": 1023 }, { "epoch": 0.463768115942029, "grad_norm": 0.1979933080480807, "learning_rate": 7.420289855072465e-05, "loss": 0.9831, "step": 1024 }, { "epoch": 0.4642210144927536, "grad_norm": 0.23880737122318224, "learning_rate": 7.427536231884058e-05, "loss": 0.9909, "step": 1025 }, { "epoch": 0.46467391304347827, "grad_norm": 0.31690624654163263, "learning_rate": 7.434782608695653e-05, "loss": 0.9814, "step": 1026 }, { "epoch": 0.4651268115942029, "grad_norm": 0.36085901451241503, "learning_rate": 7.442028985507246e-05, "loss": 0.9912, "step": 1027 }, { "epoch": 0.46557971014492755, "grad_norm": 0.4224496693570546, "learning_rate": 7.449275362318841e-05, "loss": 0.9852, "step": 1028 }, { "epoch": 0.46603260869565216, "grad_norm": 0.46058391984908614, "learning_rate": 7.456521739130435e-05, "loss": 1.0108, "step": 1029 }, { "epoch": 0.46648550724637683, "grad_norm": 0.5182325493243221, "learning_rate": 7.46376811594203e-05, "loss": 0.9654, "step": 1030 }, { "epoch": 0.46693840579710144, "grad_norm": 0.5241428194095304, "learning_rate": 7.471014492753623e-05, "loss": 1.006, "step": 1031 }, { "epoch": 0.4673913043478261, "grad_norm": 0.4537708898454745, "learning_rate": 7.478260869565218e-05, "loss": 0.9974, "step": 1032 }, { "epoch": 0.4678442028985507, "grad_norm": 0.364537088756873, "learning_rate": 7.485507246376813e-05, "loss": 1.0086, "step": 1033 }, { "epoch": 0.4682971014492754, "grad_norm": 0.28742099334229615, "learning_rate": 7.492753623188407e-05, "loss": 0.9851, "step": 1034 }, { "epoch": 0.46875, "grad_norm": 0.2975851887296509, "learning_rate": 7.500000000000001e-05, "loss": 1.0014, "step": 1035 }, { "epoch": 0.4692028985507246, "grad_norm": 0.2335926102479201, "learning_rate": 7.507246376811594e-05, "loss": 0.9664, "step": 1036 }, { "epoch": 0.4696557971014493, "grad_norm": 0.18543977604987205, "learning_rate": 7.514492753623189e-05, "loss": 0.9847, "step": 1037 }, { "epoch": 0.4701086956521739, "grad_norm": 0.16721980992540622, "learning_rate": 7.521739130434783e-05, "loss": 0.9782, "step": 1038 }, { "epoch": 0.47056159420289856, "grad_norm": 0.196491773359573, "learning_rate": 7.528985507246378e-05, "loss": 0.9867, "step": 1039 }, { "epoch": 0.47101449275362317, "grad_norm": 0.16495002525302718, "learning_rate": 7.536231884057971e-05, "loss": 0.9881, "step": 1040 }, { "epoch": 0.47146739130434784, "grad_norm": 0.15976445531443081, "learning_rate": 7.543478260869566e-05, "loss": 1.0, "step": 1041 }, { "epoch": 0.47192028985507245, "grad_norm": 0.2133740625300016, "learning_rate": 7.550724637681159e-05, "loss": 1.0032, "step": 1042 }, { "epoch": 0.4723731884057971, "grad_norm": 0.24633940612775196, "learning_rate": 7.557971014492754e-05, "loss": 1.0036, "step": 1043 }, { "epoch": 0.47282608695652173, "grad_norm": 0.21487483076164474, "learning_rate": 7.565217391304349e-05, "loss": 0.9844, "step": 1044 }, { "epoch": 0.4732789855072464, "grad_norm": 0.1732365169143759, "learning_rate": 7.572463768115943e-05, "loss": 0.9908, "step": 1045 }, { "epoch": 0.473731884057971, "grad_norm": 0.22286210022361458, "learning_rate": 7.579710144927536e-05, "loss": 1.0233, "step": 1046 }, { "epoch": 0.4741847826086957, "grad_norm": 0.19050796994553945, "learning_rate": 7.586956521739131e-05, "loss": 0.9891, "step": 1047 }, { "epoch": 0.4746376811594203, "grad_norm": 0.18974198601724043, "learning_rate": 7.594202898550726e-05, "loss": 0.9862, "step": 1048 }, { "epoch": 0.4750905797101449, "grad_norm": 0.2280486718403394, "learning_rate": 7.601449275362319e-05, "loss": 0.9975, "step": 1049 }, { "epoch": 0.47554347826086957, "grad_norm": 0.20865459802077863, "learning_rate": 7.608695652173914e-05, "loss": 0.9862, "step": 1050 }, { "epoch": 0.4759963768115942, "grad_norm": 0.1754626453985718, "learning_rate": 7.615942028985507e-05, "loss": 1.0052, "step": 1051 }, { "epoch": 0.47644927536231885, "grad_norm": 0.18585814241145235, "learning_rate": 7.623188405797102e-05, "loss": 0.9832, "step": 1052 }, { "epoch": 0.47690217391304346, "grad_norm": 0.1652708186725455, "learning_rate": 7.630434782608696e-05, "loss": 0.9548, "step": 1053 }, { "epoch": 0.4773550724637681, "grad_norm": 0.18456183148234603, "learning_rate": 7.637681159420291e-05, "loss": 0.995, "step": 1054 }, { "epoch": 0.47780797101449274, "grad_norm": 0.17671493384555106, "learning_rate": 7.644927536231884e-05, "loss": 0.9899, "step": 1055 }, { "epoch": 0.4782608695652174, "grad_norm": 0.1595085306734073, "learning_rate": 7.652173913043479e-05, "loss": 0.9641, "step": 1056 }, { "epoch": 0.478713768115942, "grad_norm": 0.17106344840981658, "learning_rate": 7.659420289855072e-05, "loss": 0.9973, "step": 1057 }, { "epoch": 0.4791666666666667, "grad_norm": 0.21138509550548443, "learning_rate": 7.666666666666668e-05, "loss": 0.9876, "step": 1058 }, { "epoch": 0.4796195652173913, "grad_norm": 0.24444321377364078, "learning_rate": 7.673913043478262e-05, "loss": 0.9788, "step": 1059 }, { "epoch": 0.48007246376811596, "grad_norm": 0.20684396567424354, "learning_rate": 7.681159420289856e-05, "loss": 0.9913, "step": 1060 }, { "epoch": 0.4805253623188406, "grad_norm": 0.18032645573877837, "learning_rate": 7.68840579710145e-05, "loss": 0.9768, "step": 1061 }, { "epoch": 0.48097826086956524, "grad_norm": 0.3289681919030814, "learning_rate": 7.695652173913044e-05, "loss": 0.9806, "step": 1062 }, { "epoch": 0.48143115942028986, "grad_norm": 0.17496576105798395, "learning_rate": 7.702898550724639e-05, "loss": 0.9765, "step": 1063 }, { "epoch": 0.48188405797101447, "grad_norm": 0.1797941771110797, "learning_rate": 7.710144927536232e-05, "loss": 0.9902, "step": 1064 }, { "epoch": 0.48233695652173914, "grad_norm": 0.1919661117586851, "learning_rate": 7.717391304347827e-05, "loss": 0.9856, "step": 1065 }, { "epoch": 0.48278985507246375, "grad_norm": 0.17474875351896296, "learning_rate": 7.72463768115942e-05, "loss": 0.9973, "step": 1066 }, { "epoch": 0.4832427536231884, "grad_norm": 0.1459722452011687, "learning_rate": 7.731884057971015e-05, "loss": 0.9942, "step": 1067 }, { "epoch": 0.483695652173913, "grad_norm": 0.14561710048879142, "learning_rate": 7.73913043478261e-05, "loss": 0.9995, "step": 1068 }, { "epoch": 0.4841485507246377, "grad_norm": 0.17859806985953988, "learning_rate": 7.746376811594204e-05, "loss": 0.9696, "step": 1069 }, { "epoch": 0.4846014492753623, "grad_norm": 0.1964160514876371, "learning_rate": 7.753623188405797e-05, "loss": 0.9897, "step": 1070 }, { "epoch": 0.485054347826087, "grad_norm": 0.208326198331013, "learning_rate": 7.760869565217392e-05, "loss": 1.0055, "step": 1071 }, { "epoch": 0.4855072463768116, "grad_norm": 0.26489155457815433, "learning_rate": 7.768115942028987e-05, "loss": 0.9595, "step": 1072 }, { "epoch": 0.48596014492753625, "grad_norm": 0.2985296528030204, "learning_rate": 7.775362318840581e-05, "loss": 1.0018, "step": 1073 }, { "epoch": 0.48641304347826086, "grad_norm": 0.3190439349434315, "learning_rate": 7.782608695652175e-05, "loss": 0.9786, "step": 1074 }, { "epoch": 0.48686594202898553, "grad_norm": 0.27255864405387586, "learning_rate": 7.789855072463769e-05, "loss": 0.9693, "step": 1075 }, { "epoch": 0.48731884057971014, "grad_norm": 0.24203956428517326, "learning_rate": 7.797101449275363e-05, "loss": 0.9819, "step": 1076 }, { "epoch": 0.48777173913043476, "grad_norm": 0.1980347135716694, "learning_rate": 7.804347826086957e-05, "loss": 0.9867, "step": 1077 }, { "epoch": 0.4882246376811594, "grad_norm": 0.18807135968394867, "learning_rate": 7.811594202898552e-05, "loss": 0.9861, "step": 1078 }, { "epoch": 0.48867753623188404, "grad_norm": 0.1689974319997666, "learning_rate": 7.818840579710145e-05, "loss": 0.9877, "step": 1079 }, { "epoch": 0.4891304347826087, "grad_norm": 0.2219090320178802, "learning_rate": 7.82608695652174e-05, "loss": 0.9514, "step": 1080 }, { "epoch": 0.4895833333333333, "grad_norm": 0.24043422796376754, "learning_rate": 7.833333333333333e-05, "loss": 0.9774, "step": 1081 }, { "epoch": 0.490036231884058, "grad_norm": 0.20552576263290642, "learning_rate": 7.840579710144928e-05, "loss": 0.9817, "step": 1082 }, { "epoch": 0.4904891304347826, "grad_norm": 0.1943727143586816, "learning_rate": 7.847826086956522e-05, "loss": 0.995, "step": 1083 }, { "epoch": 0.49094202898550726, "grad_norm": 0.20146170657272747, "learning_rate": 7.855072463768117e-05, "loss": 0.9958, "step": 1084 }, { "epoch": 0.4913949275362319, "grad_norm": 0.19367644925196123, "learning_rate": 7.86231884057971e-05, "loss": 1.0026, "step": 1085 }, { "epoch": 0.49184782608695654, "grad_norm": 0.2065210921740518, "learning_rate": 7.869565217391305e-05, "loss": 0.9875, "step": 1086 }, { "epoch": 0.49230072463768115, "grad_norm": 0.22318730694199232, "learning_rate": 7.8768115942029e-05, "loss": 0.9703, "step": 1087 }, { "epoch": 0.4927536231884058, "grad_norm": 0.2375368767097045, "learning_rate": 7.884057971014494e-05, "loss": 0.9821, "step": 1088 }, { "epoch": 0.49320652173913043, "grad_norm": 0.20799266519066467, "learning_rate": 7.891304347826088e-05, "loss": 0.9864, "step": 1089 }, { "epoch": 0.4936594202898551, "grad_norm": 0.1677334054395399, "learning_rate": 7.898550724637682e-05, "loss": 0.989, "step": 1090 }, { "epoch": 0.4941123188405797, "grad_norm": 0.1780220553012331, "learning_rate": 7.905797101449276e-05, "loss": 0.9783, "step": 1091 }, { "epoch": 0.4945652173913043, "grad_norm": 0.18709224304769326, "learning_rate": 7.91304347826087e-05, "loss": 0.9929, "step": 1092 }, { "epoch": 0.495018115942029, "grad_norm": 0.20841414260642777, "learning_rate": 7.920289855072465e-05, "loss": 0.9838, "step": 1093 }, { "epoch": 0.4954710144927536, "grad_norm": 0.2865210484980916, "learning_rate": 7.927536231884058e-05, "loss": 0.9681, "step": 1094 }, { "epoch": 0.49592391304347827, "grad_norm": 0.3824102285812833, "learning_rate": 7.934782608695653e-05, "loss": 0.9728, "step": 1095 }, { "epoch": 0.4963768115942029, "grad_norm": 0.5093891594675539, "learning_rate": 7.942028985507246e-05, "loss": 1.0172, "step": 1096 }, { "epoch": 0.49682971014492755, "grad_norm": 0.5874569415175036, "learning_rate": 7.949275362318841e-05, "loss": 0.9644, "step": 1097 }, { "epoch": 0.49728260869565216, "grad_norm": 0.5174905850298839, "learning_rate": 7.956521739130435e-05, "loss": 0.9929, "step": 1098 }, { "epoch": 0.49773550724637683, "grad_norm": 0.3309743086865451, "learning_rate": 7.96376811594203e-05, "loss": 0.9889, "step": 1099 }, { "epoch": 0.49818840579710144, "grad_norm": 0.25597134074711575, "learning_rate": 7.971014492753623e-05, "loss": 0.9798, "step": 1100 }, { "epoch": 0.4986413043478261, "grad_norm": 0.3423914729920874, "learning_rate": 7.978260869565218e-05, "loss": 0.999, "step": 1101 }, { "epoch": 0.4990942028985507, "grad_norm": 0.3364358327769493, "learning_rate": 7.985507246376813e-05, "loss": 0.9931, "step": 1102 }, { "epoch": 0.4995471014492754, "grad_norm": 0.3077298449743297, "learning_rate": 7.992753623188407e-05, "loss": 0.963, "step": 1103 }, { "epoch": 0.5, "grad_norm": 0.2585981087638313, "learning_rate": 8e-05, "loss": 0.9887, "step": 1104 }, { "epoch": 0.5004528985507246, "grad_norm": 0.20924106377195423, "learning_rate": 7.999999800056831e-05, "loss": 1.0044, "step": 1105 }, { "epoch": 0.5009057971014492, "grad_norm": 0.18392780920840288, "learning_rate": 7.999999200227343e-05, "loss": 0.968, "step": 1106 }, { "epoch": 0.501358695652174, "grad_norm": 0.17383995242349543, "learning_rate": 7.999998200511597e-05, "loss": 0.9681, "step": 1107 }, { "epoch": 0.5018115942028986, "grad_norm": 0.17264066457889948, "learning_rate": 7.999996800909693e-05, "loss": 1.0038, "step": 1108 }, { "epoch": 0.5022644927536232, "grad_norm": 0.18779555459290945, "learning_rate": 7.999995001421769e-05, "loss": 0.9956, "step": 1109 }, { "epoch": 0.5027173913043478, "grad_norm": 0.1681091937540805, "learning_rate": 7.999992802048006e-05, "loss": 0.9668, "step": 1110 }, { "epoch": 0.5031702898550725, "grad_norm": 0.2063593636787065, "learning_rate": 7.999990202788624e-05, "loss": 0.9903, "step": 1111 }, { "epoch": 0.5036231884057971, "grad_norm": 0.1981952025380767, "learning_rate": 7.999987203643883e-05, "loss": 0.9978, "step": 1112 }, { "epoch": 0.5040760869565217, "grad_norm": 0.17791740071655307, "learning_rate": 7.999983804614083e-05, "loss": 0.9933, "step": 1113 }, { "epoch": 0.5045289855072463, "grad_norm": 0.11759989260719518, "learning_rate": 7.999980005699563e-05, "loss": 0.9875, "step": 1114 }, { "epoch": 0.5049818840579711, "grad_norm": 0.11730943333874391, "learning_rate": 7.999975806900705e-05, "loss": 0.9817, "step": 1115 }, { "epoch": 0.5054347826086957, "grad_norm": 0.1903728769433857, "learning_rate": 7.999971208217926e-05, "loss": 0.9666, "step": 1116 }, { "epoch": 0.5058876811594203, "grad_norm": 0.21839264657003643, "learning_rate": 7.999966209651687e-05, "loss": 0.9518, "step": 1117 }, { "epoch": 0.5063405797101449, "grad_norm": 0.21281284063733621, "learning_rate": 7.999960811202487e-05, "loss": 0.9684, "step": 1118 }, { "epoch": 0.5067934782608695, "grad_norm": 0.22106107650355764, "learning_rate": 7.999955012870866e-05, "loss": 1.0112, "step": 1119 }, { "epoch": 0.5072463768115942, "grad_norm": 0.2397240781196963, "learning_rate": 7.999948814657406e-05, "loss": 1.0047, "step": 1120 }, { "epoch": 0.5076992753623188, "grad_norm": 0.23508075740244322, "learning_rate": 7.999942216562722e-05, "loss": 1.0152, "step": 1121 }, { "epoch": 0.5081521739130435, "grad_norm": 0.21419107499349724, "learning_rate": 7.999935218587477e-05, "loss": 0.9996, "step": 1122 }, { "epoch": 0.5086050724637681, "grad_norm": 0.1887850955689022, "learning_rate": 7.99992782073237e-05, "loss": 0.9856, "step": 1123 }, { "epoch": 0.5090579710144928, "grad_norm": 0.20541494352338274, "learning_rate": 7.99992002299814e-05, "loss": 0.9697, "step": 1124 }, { "epoch": 0.5095108695652174, "grad_norm": 0.20010946811765878, "learning_rate": 7.999911825385567e-05, "loss": 0.9672, "step": 1125 }, { "epoch": 0.509963768115942, "grad_norm": 0.21693571110817259, "learning_rate": 7.999903227895471e-05, "loss": 0.9818, "step": 1126 }, { "epoch": 0.5104166666666666, "grad_norm": 0.19917599704583874, "learning_rate": 7.99989423052871e-05, "loss": 0.9785, "step": 1127 }, { "epoch": 0.5108695652173914, "grad_norm": 0.16610115448760207, "learning_rate": 7.999884833286186e-05, "loss": 0.9717, "step": 1128 }, { "epoch": 0.511322463768116, "grad_norm": 0.21030975526631382, "learning_rate": 7.999875036168835e-05, "loss": 0.9651, "step": 1129 }, { "epoch": 0.5117753623188406, "grad_norm": 0.1886161317110839, "learning_rate": 7.999864839177639e-05, "loss": 0.9749, "step": 1130 }, { "epoch": 0.5122282608695652, "grad_norm": 0.15663382101693496, "learning_rate": 7.999854242313617e-05, "loss": 0.9978, "step": 1131 }, { "epoch": 0.5126811594202898, "grad_norm": 0.15809111517291832, "learning_rate": 7.999843245577828e-05, "loss": 1.0005, "step": 1132 }, { "epoch": 0.5131340579710145, "grad_norm": 0.11861555821081016, "learning_rate": 7.99983184897137e-05, "loss": 0.9808, "step": 1133 }, { "epoch": 0.5135869565217391, "grad_norm": 0.12874955313049696, "learning_rate": 7.999820052495386e-05, "loss": 0.9648, "step": 1134 }, { "epoch": 0.5140398550724637, "grad_norm": 0.154037378520377, "learning_rate": 7.999807856151052e-05, "loss": 0.9827, "step": 1135 }, { "epoch": 0.5144927536231884, "grad_norm": 0.2302820368557603, "learning_rate": 7.999795259939589e-05, "loss": 0.957, "step": 1136 }, { "epoch": 0.5149456521739131, "grad_norm": 0.2768736530063063, "learning_rate": 7.999782263862255e-05, "loss": 0.9581, "step": 1137 }, { "epoch": 0.5153985507246377, "grad_norm": 0.28216279880900197, "learning_rate": 7.99976886792035e-05, "loss": 0.9838, "step": 1138 }, { "epoch": 0.5158514492753623, "grad_norm": 0.2369503158890959, "learning_rate": 7.999755072115213e-05, "loss": 0.9807, "step": 1139 }, { "epoch": 0.5163043478260869, "grad_norm": 0.20509121060037405, "learning_rate": 7.999740876448222e-05, "loss": 0.9637, "step": 1140 }, { "epoch": 0.5167572463768116, "grad_norm": 0.20520995892244495, "learning_rate": 7.999726280920799e-05, "loss": 1.004, "step": 1141 }, { "epoch": 0.5172101449275363, "grad_norm": 0.2132589327301504, "learning_rate": 7.999711285534402e-05, "loss": 0.9958, "step": 1142 }, { "epoch": 0.5176630434782609, "grad_norm": 0.21745549239287215, "learning_rate": 7.999695890290529e-05, "loss": 0.9865, "step": 1143 }, { "epoch": 0.5181159420289855, "grad_norm": 0.1837097749519613, "learning_rate": 7.99968009519072e-05, "loss": 0.9668, "step": 1144 }, { "epoch": 0.5185688405797102, "grad_norm": 0.13836768076061673, "learning_rate": 7.999663900236555e-05, "loss": 0.9605, "step": 1145 }, { "epoch": 0.5190217391304348, "grad_norm": 0.12159926680515798, "learning_rate": 7.99964730542965e-05, "loss": 0.9725, "step": 1146 }, { "epoch": 0.5194746376811594, "grad_norm": 0.13592725128857377, "learning_rate": 7.999630310771667e-05, "loss": 0.9681, "step": 1147 }, { "epoch": 0.519927536231884, "grad_norm": 0.1780117714285987, "learning_rate": 7.999612916264302e-05, "loss": 0.9714, "step": 1148 }, { "epoch": 0.5203804347826086, "grad_norm": 0.16613086032041954, "learning_rate": 7.999595121909298e-05, "loss": 0.9871, "step": 1149 }, { "epoch": 0.5208333333333334, "grad_norm": 0.15434623813522233, "learning_rate": 7.99957692770843e-05, "loss": 0.9773, "step": 1150 }, { "epoch": 0.521286231884058, "grad_norm": 0.14055042617933272, "learning_rate": 7.99955833366352e-05, "loss": 0.9734, "step": 1151 }, { "epoch": 0.5217391304347826, "grad_norm": 0.1611331147494066, "learning_rate": 7.999539339776424e-05, "loss": 1.0034, "step": 1152 }, { "epoch": 0.5221920289855072, "grad_norm": 0.17497892619262406, "learning_rate": 7.999519946049045e-05, "loss": 0.9879, "step": 1153 }, { "epoch": 0.5226449275362319, "grad_norm": 0.30728914841097904, "learning_rate": 7.999500152483318e-05, "loss": 0.9975, "step": 1154 }, { "epoch": 0.5230978260869565, "grad_norm": 0.1750305617854668, "learning_rate": 7.999479959081223e-05, "loss": 0.9717, "step": 1155 }, { "epoch": 0.5235507246376812, "grad_norm": 0.18742025351950792, "learning_rate": 7.999459365844779e-05, "loss": 0.9767, "step": 1156 }, { "epoch": 0.5240036231884058, "grad_norm": 0.17660705230877238, "learning_rate": 7.999438372776043e-05, "loss": 0.9761, "step": 1157 }, { "epoch": 0.5244565217391305, "grad_norm": 0.19754998635790094, "learning_rate": 7.999416979877117e-05, "loss": 0.9626, "step": 1158 }, { "epoch": 0.5249094202898551, "grad_norm": 0.2119351739259744, "learning_rate": 7.999395187150137e-05, "loss": 0.9899, "step": 1159 }, { "epoch": 0.5253623188405797, "grad_norm": 0.21455219398076456, "learning_rate": 7.999372994597282e-05, "loss": 0.9745, "step": 1160 }, { "epoch": 0.5258152173913043, "grad_norm": 0.2274808170563173, "learning_rate": 7.999350402220772e-05, "loss": 0.9692, "step": 1161 }, { "epoch": 0.5262681159420289, "grad_norm": 0.21811224840757587, "learning_rate": 7.999327410022866e-05, "loss": 0.975, "step": 1162 }, { "epoch": 0.5267210144927537, "grad_norm": 0.21133984399592531, "learning_rate": 7.99930401800586e-05, "loss": 0.9885, "step": 1163 }, { "epoch": 0.5271739130434783, "grad_norm": 0.2505692839037768, "learning_rate": 7.999280226172094e-05, "loss": 0.9774, "step": 1164 }, { "epoch": 0.5276268115942029, "grad_norm": 0.2501009566890808, "learning_rate": 7.999256034523946e-05, "loss": 1.0026, "step": 1165 }, { "epoch": 0.5280797101449275, "grad_norm": 0.21852559870695412, "learning_rate": 7.999231443063835e-05, "loss": 0.9656, "step": 1166 }, { "epoch": 0.5285326086956522, "grad_norm": 0.1978847858177087, "learning_rate": 7.999206451794219e-05, "loss": 0.9773, "step": 1167 }, { "epoch": 0.5289855072463768, "grad_norm": 0.19063889363073727, "learning_rate": 7.999181060717597e-05, "loss": 1.0014, "step": 1168 }, { "epoch": 0.5294384057971014, "grad_norm": 0.17102943392574985, "learning_rate": 7.999155269836508e-05, "loss": 0.9725, "step": 1169 }, { "epoch": 0.529891304347826, "grad_norm": 0.20771204155775794, "learning_rate": 7.999129079153529e-05, "loss": 0.9949, "step": 1170 }, { "epoch": 0.5303442028985508, "grad_norm": 0.2673288224980219, "learning_rate": 7.99910248867128e-05, "loss": 0.9817, "step": 1171 }, { "epoch": 0.5307971014492754, "grad_norm": 0.2759057424932376, "learning_rate": 7.999075498392416e-05, "loss": 0.973, "step": 1172 }, { "epoch": 0.53125, "grad_norm": 0.24126549223187674, "learning_rate": 7.999048108319636e-05, "loss": 0.9942, "step": 1173 }, { "epoch": 0.5317028985507246, "grad_norm": 0.2335245031378726, "learning_rate": 7.999020318455682e-05, "loss": 0.9752, "step": 1174 }, { "epoch": 0.5321557971014492, "grad_norm": 0.21766853594158286, "learning_rate": 7.99899212880333e-05, "loss": 0.9735, "step": 1175 }, { "epoch": 0.532608695652174, "grad_norm": 0.2209629773212257, "learning_rate": 7.998963539365397e-05, "loss": 0.993, "step": 1176 }, { "epoch": 0.5330615942028986, "grad_norm": 0.20604199812382942, "learning_rate": 7.998934550144743e-05, "loss": 0.988, "step": 1177 }, { "epoch": 0.5335144927536232, "grad_norm": 0.1628247982490746, "learning_rate": 7.998905161144264e-05, "loss": 0.9602, "step": 1178 }, { "epoch": 0.5339673913043478, "grad_norm": 0.1897128291402789, "learning_rate": 7.9988753723669e-05, "loss": 0.9925, "step": 1179 }, { "epoch": 0.5344202898550725, "grad_norm": 0.17988958240262007, "learning_rate": 7.998845183815628e-05, "loss": 0.9483, "step": 1180 }, { "epoch": 0.5348731884057971, "grad_norm": 0.1266072822001214, "learning_rate": 7.998814595493466e-05, "loss": 0.9883, "step": 1181 }, { "epoch": 0.5353260869565217, "grad_norm": 0.15704718258875966, "learning_rate": 7.998783607403473e-05, "loss": 0.998, "step": 1182 }, { "epoch": 0.5357789855072463, "grad_norm": 0.21531079448326892, "learning_rate": 7.998752219548746e-05, "loss": 0.9853, "step": 1183 }, { "epoch": 0.5362318840579711, "grad_norm": 0.23145606420831913, "learning_rate": 7.998720431932423e-05, "loss": 0.9841, "step": 1184 }, { "epoch": 0.5366847826086957, "grad_norm": 0.2449826448852337, "learning_rate": 7.99868824455768e-05, "loss": 0.9895, "step": 1185 }, { "epoch": 0.5371376811594203, "grad_norm": 0.1948450850936934, "learning_rate": 7.998655657427739e-05, "loss": 1.0061, "step": 1186 }, { "epoch": 0.5375905797101449, "grad_norm": 0.15193653550136596, "learning_rate": 7.998622670545855e-05, "loss": 0.9545, "step": 1187 }, { "epoch": 0.5380434782608695, "grad_norm": 0.14666560596052738, "learning_rate": 7.998589283915326e-05, "loss": 0.9693, "step": 1188 }, { "epoch": 0.5384963768115942, "grad_norm": 0.1661145657657045, "learning_rate": 7.99855549753949e-05, "loss": 0.9788, "step": 1189 }, { "epoch": 0.5389492753623188, "grad_norm": 0.16048869118797054, "learning_rate": 7.998521311421725e-05, "loss": 0.9725, "step": 1190 }, { "epoch": 0.5394021739130435, "grad_norm": 0.19109919522090868, "learning_rate": 7.998486725565448e-05, "loss": 0.9968, "step": 1191 }, { "epoch": 0.5398550724637681, "grad_norm": 0.22482790792379886, "learning_rate": 7.998451739974117e-05, "loss": 0.9792, "step": 1192 }, { "epoch": 0.5403079710144928, "grad_norm": 0.23584370693709358, "learning_rate": 7.99841635465123e-05, "loss": 0.9786, "step": 1193 }, { "epoch": 0.5407608695652174, "grad_norm": 0.24922402479906797, "learning_rate": 7.998380569600324e-05, "loss": 0.9856, "step": 1194 }, { "epoch": 0.541213768115942, "grad_norm": 0.23155407055862934, "learning_rate": 7.998344384824975e-05, "loss": 0.9725, "step": 1195 }, { "epoch": 0.5416666666666666, "grad_norm": 0.21674852921362478, "learning_rate": 7.998307800328803e-05, "loss": 0.9659, "step": 1196 }, { "epoch": 0.5421195652173914, "grad_norm": 0.2183748260973661, "learning_rate": 7.998270816115464e-05, "loss": 0.9668, "step": 1197 }, { "epoch": 0.542572463768116, "grad_norm": 0.24780390252612464, "learning_rate": 7.998233432188656e-05, "loss": 0.9781, "step": 1198 }, { "epoch": 0.5430253623188406, "grad_norm": 0.2920041637012515, "learning_rate": 7.998195648552116e-05, "loss": 0.9834, "step": 1199 }, { "epoch": 0.5434782608695652, "grad_norm": 0.2840406514403322, "learning_rate": 7.998157465209622e-05, "loss": 0.9825, "step": 1200 }, { "epoch": 0.5439311594202898, "grad_norm": 0.2229882822561638, "learning_rate": 7.998118882164989e-05, "loss": 0.9831, "step": 1201 }, { "epoch": 0.5443840579710145, "grad_norm": 0.26216556676331115, "learning_rate": 7.998079899422076e-05, "loss": 0.9848, "step": 1202 }, { "epoch": 0.5448369565217391, "grad_norm": 0.3160968550303011, "learning_rate": 7.998040516984782e-05, "loss": 0.9902, "step": 1203 }, { "epoch": 0.5452898550724637, "grad_norm": 0.33320069847436856, "learning_rate": 7.99800073485704e-05, "loss": 0.9837, "step": 1204 }, { "epoch": 0.5457427536231884, "grad_norm": 0.3206465313424093, "learning_rate": 7.99796055304283e-05, "loss": 0.9825, "step": 1205 }, { "epoch": 0.5461956521739131, "grad_norm": 0.27779519954472665, "learning_rate": 7.997919971546169e-05, "loss": 0.9875, "step": 1206 }, { "epoch": 0.5466485507246377, "grad_norm": 0.16955245467255656, "learning_rate": 7.997878990371111e-05, "loss": 0.9828, "step": 1207 }, { "epoch": 0.5471014492753623, "grad_norm": 0.11730218964135476, "learning_rate": 7.997837609521756e-05, "loss": 0.9615, "step": 1208 }, { "epoch": 0.5475543478260869, "grad_norm": 0.1905799446080018, "learning_rate": 7.99779582900224e-05, "loss": 0.9781, "step": 1209 }, { "epoch": 0.5480072463768116, "grad_norm": 0.22844675137328455, "learning_rate": 7.99775364881674e-05, "loss": 0.9685, "step": 1210 }, { "epoch": 0.5484601449275363, "grad_norm": 0.17867667891488234, "learning_rate": 7.997711068969474e-05, "loss": 0.9929, "step": 1211 }, { "epoch": 0.5489130434782609, "grad_norm": 0.1685580892885212, "learning_rate": 7.997668089464696e-05, "loss": 0.9794, "step": 1212 }, { "epoch": 0.5493659420289855, "grad_norm": 0.18317960608408662, "learning_rate": 7.997624710306706e-05, "loss": 0.9981, "step": 1213 }, { "epoch": 0.5498188405797102, "grad_norm": 0.15704545695343386, "learning_rate": 7.997580931499837e-05, "loss": 0.9777, "step": 1214 }, { "epoch": 0.5502717391304348, "grad_norm": 0.10999097352301104, "learning_rate": 7.997536753048469e-05, "loss": 0.9616, "step": 1215 }, { "epoch": 0.5507246376811594, "grad_norm": 0.13286904620249967, "learning_rate": 7.997492174957016e-05, "loss": 1.008, "step": 1216 }, { "epoch": 0.551177536231884, "grad_norm": 0.14907366048888412, "learning_rate": 7.997447197229937e-05, "loss": 0.977, "step": 1217 }, { "epoch": 0.5516304347826086, "grad_norm": 0.16684925704749629, "learning_rate": 7.997401819871726e-05, "loss": 0.9935, "step": 1218 }, { "epoch": 0.5520833333333334, "grad_norm": 0.17605496934793208, "learning_rate": 7.997356042886921e-05, "loss": 0.9763, "step": 1219 }, { "epoch": 0.552536231884058, "grad_norm": 0.16683050866578025, "learning_rate": 7.997309866280098e-05, "loss": 0.9678, "step": 1220 }, { "epoch": 0.5529891304347826, "grad_norm": 0.18424412643614724, "learning_rate": 7.997263290055873e-05, "loss": 0.958, "step": 1221 }, { "epoch": 0.5534420289855072, "grad_norm": 0.2295860842169903, "learning_rate": 7.997216314218903e-05, "loss": 0.9833, "step": 1222 }, { "epoch": 0.5538949275362319, "grad_norm": 0.2553937216306973, "learning_rate": 7.997168938773884e-05, "loss": 0.9818, "step": 1223 }, { "epoch": 0.5543478260869565, "grad_norm": 0.2513577486061325, "learning_rate": 7.997121163725553e-05, "loss": 0.9395, "step": 1224 }, { "epoch": 0.5548007246376812, "grad_norm": 0.23124957545038025, "learning_rate": 7.997072989078684e-05, "loss": 0.9598, "step": 1225 }, { "epoch": 0.5552536231884058, "grad_norm": 0.2577163002459174, "learning_rate": 7.997024414838095e-05, "loss": 0.965, "step": 1226 }, { "epoch": 0.5557065217391305, "grad_norm": 0.29061969397208814, "learning_rate": 7.996975441008642e-05, "loss": 0.9891, "step": 1227 }, { "epoch": 0.5561594202898551, "grad_norm": 0.32730869279797076, "learning_rate": 7.99692606759522e-05, "loss": 0.9762, "step": 1228 }, { "epoch": 0.5566123188405797, "grad_norm": 0.32312603839767584, "learning_rate": 7.996876294602765e-05, "loss": 0.9915, "step": 1229 }, { "epoch": 0.5570652173913043, "grad_norm": 0.27264056380719837, "learning_rate": 7.996826122036252e-05, "loss": 0.9831, "step": 1230 }, { "epoch": 0.5575181159420289, "grad_norm": 0.23189877204807438, "learning_rate": 7.9967755499007e-05, "loss": 1.0036, "step": 1231 }, { "epoch": 0.5579710144927537, "grad_norm": 0.20935946691872293, "learning_rate": 7.996724578201162e-05, "loss": 0.984, "step": 1232 }, { "epoch": 0.5584239130434783, "grad_norm": 0.21860717856203601, "learning_rate": 7.996673206942735e-05, "loss": 0.995, "step": 1233 }, { "epoch": 0.5588768115942029, "grad_norm": 0.23207652517189928, "learning_rate": 7.996621436130553e-05, "loss": 0.9776, "step": 1234 }, { "epoch": 0.5593297101449275, "grad_norm": 0.2190241978507779, "learning_rate": 7.996569265769795e-05, "loss": 0.9782, "step": 1235 }, { "epoch": 0.5597826086956522, "grad_norm": 0.2188123639830773, "learning_rate": 7.996516695865674e-05, "loss": 0.9988, "step": 1236 }, { "epoch": 0.5602355072463768, "grad_norm": 0.23551797796230645, "learning_rate": 7.996463726423445e-05, "loss": 0.9769, "step": 1237 }, { "epoch": 0.5606884057971014, "grad_norm": 0.19416327202317635, "learning_rate": 7.996410357448404e-05, "loss": 0.9789, "step": 1238 }, { "epoch": 0.561141304347826, "grad_norm": 0.20653958676071807, "learning_rate": 7.996356588945887e-05, "loss": 0.9969, "step": 1239 }, { "epoch": 0.5615942028985508, "grad_norm": 0.2892614007513807, "learning_rate": 7.99630242092127e-05, "loss": 0.9843, "step": 1240 }, { "epoch": 0.5620471014492754, "grad_norm": 0.3465513712379516, "learning_rate": 7.996247853379965e-05, "loss": 0.9764, "step": 1241 }, { "epoch": 0.5625, "grad_norm": 0.3594710147448849, "learning_rate": 7.996192886327432e-05, "loss": 0.9677, "step": 1242 }, { "epoch": 0.5629528985507246, "grad_norm": 0.31910770605178923, "learning_rate": 7.996137519769163e-05, "loss": 0.9918, "step": 1243 }, { "epoch": 0.5634057971014492, "grad_norm": 0.24698635441842048, "learning_rate": 7.996081753710693e-05, "loss": 0.9702, "step": 1244 }, { "epoch": 0.563858695652174, "grad_norm": 0.21894604933378242, "learning_rate": 7.996025588157598e-05, "loss": 0.9817, "step": 1245 }, { "epoch": 0.5643115942028986, "grad_norm": 0.2258417968712589, "learning_rate": 7.995969023115493e-05, "loss": 0.9911, "step": 1246 }, { "epoch": 0.5647644927536232, "grad_norm": 0.20877343801305423, "learning_rate": 7.995912058590031e-05, "loss": 0.9715, "step": 1247 }, { "epoch": 0.5652173913043478, "grad_norm": 0.21518954454578004, "learning_rate": 7.995854694586911e-05, "loss": 0.9748, "step": 1248 }, { "epoch": 0.5656702898550725, "grad_norm": 0.18817786648589183, "learning_rate": 7.995796931111864e-05, "loss": 0.9888, "step": 1249 }, { "epoch": 0.5661231884057971, "grad_norm": 0.15142086280887254, "learning_rate": 7.995738768170666e-05, "loss": 1.0014, "step": 1250 }, { "epoch": 0.5665760869565217, "grad_norm": 0.14874052228761409, "learning_rate": 7.995680205769132e-05, "loss": 0.9808, "step": 1251 }, { "epoch": 0.5670289855072463, "grad_norm": 0.19306701689315117, "learning_rate": 7.995621243913115e-05, "loss": 0.9771, "step": 1252 }, { "epoch": 0.5674818840579711, "grad_norm": 0.21774409555526728, "learning_rate": 7.995561882608512e-05, "loss": 0.9882, "step": 1253 }, { "epoch": 0.5679347826086957, "grad_norm": 0.15988395480199996, "learning_rate": 7.995502121861255e-05, "loss": 0.985, "step": 1254 }, { "epoch": 0.5683876811594203, "grad_norm": 0.1279093458033311, "learning_rate": 7.995441961677319e-05, "loss": 0.9603, "step": 1255 }, { "epoch": 0.5688405797101449, "grad_norm": 0.16467198209765674, "learning_rate": 7.99538140206272e-05, "loss": 0.972, "step": 1256 }, { "epoch": 0.5692934782608695, "grad_norm": 0.17033729301230452, "learning_rate": 7.995320443023511e-05, "loss": 0.9779, "step": 1257 }, { "epoch": 0.5697463768115942, "grad_norm": 0.15233155076158295, "learning_rate": 7.995259084565786e-05, "loss": 0.9824, "step": 1258 }, { "epoch": 0.5701992753623188, "grad_norm": 0.14737657950938815, "learning_rate": 7.995197326695678e-05, "loss": 0.9765, "step": 1259 }, { "epoch": 0.5706521739130435, "grad_norm": 0.18475487883370378, "learning_rate": 7.995135169419365e-05, "loss": 0.9752, "step": 1260 }, { "epoch": 0.5711050724637681, "grad_norm": 0.1977056899216566, "learning_rate": 7.995072612743055e-05, "loss": 0.9768, "step": 1261 }, { "epoch": 0.5715579710144928, "grad_norm": 0.19000073122524488, "learning_rate": 7.995009656673008e-05, "loss": 0.9688, "step": 1262 }, { "epoch": 0.5720108695652174, "grad_norm": 0.18232353680164226, "learning_rate": 7.994946301215514e-05, "loss": 0.9856, "step": 1263 }, { "epoch": 0.572463768115942, "grad_norm": 0.19128887910837022, "learning_rate": 7.994882546376908e-05, "loss": 0.9835, "step": 1264 }, { "epoch": 0.5729166666666666, "grad_norm": 0.1990181126292675, "learning_rate": 7.994818392163563e-05, "loss": 0.9772, "step": 1265 }, { "epoch": 0.5733695652173914, "grad_norm": 0.17930028186927716, "learning_rate": 7.994753838581892e-05, "loss": 0.9966, "step": 1266 }, { "epoch": 0.573822463768116, "grad_norm": 0.18773922617921965, "learning_rate": 7.994688885638351e-05, "loss": 0.9904, "step": 1267 }, { "epoch": 0.5742753623188406, "grad_norm": 0.1893365910655687, "learning_rate": 7.99462353333943e-05, "loss": 0.9536, "step": 1268 }, { "epoch": 0.5747282608695652, "grad_norm": 0.21930370905471544, "learning_rate": 7.994557781691668e-05, "loss": 0.9654, "step": 1269 }, { "epoch": 0.5751811594202898, "grad_norm": 0.15713694218956192, "learning_rate": 7.994491630701633e-05, "loss": 0.9716, "step": 1270 }, { "epoch": 0.5756340579710145, "grad_norm": 0.16036022229030414, "learning_rate": 7.994425080375939e-05, "loss": 1.0129, "step": 1271 }, { "epoch": 0.5760869565217391, "grad_norm": 0.16739207399169434, "learning_rate": 7.994358130721238e-05, "loss": 0.9991, "step": 1272 }, { "epoch": 0.5765398550724637, "grad_norm": 0.20220038329271825, "learning_rate": 7.994290781744229e-05, "loss": 0.9756, "step": 1273 }, { "epoch": 0.5769927536231884, "grad_norm": 0.22727750513153536, "learning_rate": 7.99422303345164e-05, "loss": 0.9727, "step": 1274 }, { "epoch": 0.5774456521739131, "grad_norm": 0.2298532454226194, "learning_rate": 7.994154885850244e-05, "loss": 0.9866, "step": 1275 }, { "epoch": 0.5778985507246377, "grad_norm": 0.24048399938468273, "learning_rate": 7.994086338946855e-05, "loss": 0.9924, "step": 1276 }, { "epoch": 0.5783514492753623, "grad_norm": 0.3629325754104891, "learning_rate": 7.994017392748325e-05, "loss": 0.9871, "step": 1277 }, { "epoch": 0.5788043478260869, "grad_norm": 0.17116677456659365, "learning_rate": 7.993948047261549e-05, "loss": 0.9794, "step": 1278 }, { "epoch": 0.5792572463768116, "grad_norm": 0.1652858481762181, "learning_rate": 7.993878302493456e-05, "loss": 0.9785, "step": 1279 }, { "epoch": 0.5797101449275363, "grad_norm": 0.21660579383114853, "learning_rate": 7.993808158451022e-05, "loss": 0.9812, "step": 1280 }, { "epoch": 0.5801630434782609, "grad_norm": 0.17698700194242564, "learning_rate": 7.993737615141255e-05, "loss": 0.9517, "step": 1281 }, { "epoch": 0.5806159420289855, "grad_norm": 0.1974746116019036, "learning_rate": 7.993666672571213e-05, "loss": 0.9768, "step": 1282 }, { "epoch": 0.5810688405797102, "grad_norm": 0.2596708087224647, "learning_rate": 7.993595330747983e-05, "loss": 0.9818, "step": 1283 }, { "epoch": 0.5815217391304348, "grad_norm": 0.2438760609651509, "learning_rate": 7.993523589678701e-05, "loss": 0.9693, "step": 1284 }, { "epoch": 0.5819746376811594, "grad_norm": 0.2278091647487771, "learning_rate": 7.993451449370537e-05, "loss": 0.9823, "step": 1285 }, { "epoch": 0.582427536231884, "grad_norm": 0.23322353164250015, "learning_rate": 7.993378909830702e-05, "loss": 0.9739, "step": 1286 }, { "epoch": 0.5828804347826086, "grad_norm": 0.18927673613485255, "learning_rate": 7.993305971066451e-05, "loss": 0.9765, "step": 1287 }, { "epoch": 0.5833333333333334, "grad_norm": 0.2215052406700798, "learning_rate": 7.993232633085074e-05, "loss": 1.0133, "step": 1288 }, { "epoch": 0.583786231884058, "grad_norm": 0.1582384355240669, "learning_rate": 7.993158895893902e-05, "loss": 0.9593, "step": 1289 }, { "epoch": 0.5842391304347826, "grad_norm": 0.1370738615200656, "learning_rate": 7.993084759500308e-05, "loss": 1.0042, "step": 1290 }, { "epoch": 0.5846920289855072, "grad_norm": 0.12840463105847727, "learning_rate": 7.993010223911702e-05, "loss": 1.0054, "step": 1291 }, { "epoch": 0.5851449275362319, "grad_norm": 0.1270366179779952, "learning_rate": 7.992935289135538e-05, "loss": 0.9908, "step": 1292 }, { "epoch": 0.5855978260869565, "grad_norm": 0.14582984505055166, "learning_rate": 7.992859955179306e-05, "loss": 0.988, "step": 1293 }, { "epoch": 0.5860507246376812, "grad_norm": 0.1685953665493537, "learning_rate": 7.992784222050536e-05, "loss": 0.9841, "step": 1294 }, { "epoch": 0.5865036231884058, "grad_norm": 0.2067163245710263, "learning_rate": 7.9927080897568e-05, "loss": 0.952, "step": 1295 }, { "epoch": 0.5869565217391305, "grad_norm": 0.20052149708720796, "learning_rate": 7.992631558305709e-05, "loss": 0.9964, "step": 1296 }, { "epoch": 0.5874094202898551, "grad_norm": 0.16114014937372795, "learning_rate": 7.992554627704915e-05, "loss": 0.9801, "step": 1297 }, { "epoch": 0.5878623188405797, "grad_norm": 0.1737521518208235, "learning_rate": 7.992477297962108e-05, "loss": 0.9696, "step": 1298 }, { "epoch": 0.5883152173913043, "grad_norm": 0.22108093349408278, "learning_rate": 7.99239956908502e-05, "loss": 0.9739, "step": 1299 }, { "epoch": 0.5887681159420289, "grad_norm": 0.3072219838222703, "learning_rate": 7.992321441081419e-05, "loss": 0.9625, "step": 1300 }, { "epoch": 0.5892210144927537, "grad_norm": 0.3343043914633944, "learning_rate": 7.992242913959118e-05, "loss": 1.0055, "step": 1301 }, { "epoch": 0.5896739130434783, "grad_norm": 0.36684451716990607, "learning_rate": 7.992163987725967e-05, "loss": 0.9547, "step": 1302 }, { "epoch": 0.5901268115942029, "grad_norm": 0.3948286608519826, "learning_rate": 7.992084662389854e-05, "loss": 1.0066, "step": 1303 }, { "epoch": 0.5905797101449275, "grad_norm": 0.3646894239426363, "learning_rate": 7.992004937958714e-05, "loss": 0.9943, "step": 1304 }, { "epoch": 0.5910326086956522, "grad_norm": 0.3074188048292729, "learning_rate": 7.991924814440513e-05, "loss": 0.9793, "step": 1305 }, { "epoch": 0.5914855072463768, "grad_norm": 0.16017236486709935, "learning_rate": 7.991844291843263e-05, "loss": 0.9789, "step": 1306 }, { "epoch": 0.5919384057971014, "grad_norm": 0.16133260203032726, "learning_rate": 7.991763370175013e-05, "loss": 0.9855, "step": 1307 }, { "epoch": 0.592391304347826, "grad_norm": 0.2800716877449172, "learning_rate": 7.991682049443853e-05, "loss": 0.9744, "step": 1308 }, { "epoch": 0.5928442028985508, "grad_norm": 0.30509438044564247, "learning_rate": 7.991600329657914e-05, "loss": 0.9718, "step": 1309 }, { "epoch": 0.5932971014492754, "grad_norm": 0.26177301722035673, "learning_rate": 7.991518210825364e-05, "loss": 1.0033, "step": 1310 }, { "epoch": 0.59375, "grad_norm": 0.23912809451644015, "learning_rate": 7.991435692954414e-05, "loss": 0.9932, "step": 1311 }, { "epoch": 0.5942028985507246, "grad_norm": 0.1684973040117503, "learning_rate": 7.991352776053314e-05, "loss": 0.9938, "step": 1312 }, { "epoch": 0.5946557971014492, "grad_norm": 0.13464131395329626, "learning_rate": 7.991269460130351e-05, "loss": 0.9752, "step": 1313 }, { "epoch": 0.595108695652174, "grad_norm": 0.15392880098776188, "learning_rate": 7.991185745193854e-05, "loss": 1.0027, "step": 1314 }, { "epoch": 0.5955615942028986, "grad_norm": 0.14739423887890202, "learning_rate": 7.991101631252195e-05, "loss": 0.9692, "step": 1315 }, { "epoch": 0.5960144927536232, "grad_norm": 0.16198232490026368, "learning_rate": 7.991017118313781e-05, "loss": 0.9888, "step": 1316 }, { "epoch": 0.5964673913043478, "grad_norm": 0.19858523876142958, "learning_rate": 7.990932206387062e-05, "loss": 1.0082, "step": 1317 }, { "epoch": 0.5969202898550725, "grad_norm": 0.2266085237788067, "learning_rate": 7.990846895480525e-05, "loss": 0.9933, "step": 1318 }, { "epoch": 0.5973731884057971, "grad_norm": 0.22980059214396698, "learning_rate": 7.9907611856027e-05, "loss": 0.9655, "step": 1319 }, { "epoch": 0.5978260869565217, "grad_norm": 0.21379085407472143, "learning_rate": 7.990675076762158e-05, "loss": 0.9564, "step": 1320 }, { "epoch": 0.5982789855072463, "grad_norm": 0.16762059579465224, "learning_rate": 7.990588568967502e-05, "loss": 0.9775, "step": 1321 }, { "epoch": 0.5987318840579711, "grad_norm": 0.27211247530003446, "learning_rate": 7.990501662227384e-05, "loss": 0.9792, "step": 1322 }, { "epoch": 0.5991847826086957, "grad_norm": 0.18113863932073146, "learning_rate": 7.990414356550492e-05, "loss": 0.9983, "step": 1323 }, { "epoch": 0.5996376811594203, "grad_norm": 0.1821228874978836, "learning_rate": 7.990326651945552e-05, "loss": 0.9761, "step": 1324 }, { "epoch": 0.6000905797101449, "grad_norm": 0.21184284984107846, "learning_rate": 7.990238548421334e-05, "loss": 0.9801, "step": 1325 }, { "epoch": 0.6005434782608695, "grad_norm": 0.21204215889970107, "learning_rate": 7.990150045986645e-05, "loss": 0.9451, "step": 1326 }, { "epoch": 0.6009963768115942, "grad_norm": 0.2168416229162866, "learning_rate": 7.990061144650334e-05, "loss": 0.986, "step": 1327 }, { "epoch": 0.6014492753623188, "grad_norm": 0.1936545626659387, "learning_rate": 7.989971844421286e-05, "loss": 0.9614, "step": 1328 }, { "epoch": 0.6019021739130435, "grad_norm": 0.1494952736095799, "learning_rate": 7.98988214530843e-05, "loss": 0.9802, "step": 1329 }, { "epoch": 0.6023550724637681, "grad_norm": 0.15210193847406406, "learning_rate": 7.989792047320736e-05, "loss": 0.9783, "step": 1330 }, { "epoch": 0.6028079710144928, "grad_norm": 0.13998959144803033, "learning_rate": 7.989701550467207e-05, "loss": 0.9802, "step": 1331 }, { "epoch": 0.6032608695652174, "grad_norm": 0.16850649683760616, "learning_rate": 7.989610654756891e-05, "loss": 0.979, "step": 1332 }, { "epoch": 0.603713768115942, "grad_norm": 0.15389489972331138, "learning_rate": 7.989519360198878e-05, "loss": 0.9759, "step": 1333 }, { "epoch": 0.6041666666666666, "grad_norm": 0.16276556250735016, "learning_rate": 7.98942766680229e-05, "loss": 0.9692, "step": 1334 }, { "epoch": 0.6046195652173914, "grad_norm": 0.12959966473019496, "learning_rate": 7.989335574576297e-05, "loss": 0.9545, "step": 1335 }, { "epoch": 0.605072463768116, "grad_norm": 0.12161568196414031, "learning_rate": 7.989243083530106e-05, "loss": 0.9717, "step": 1336 }, { "epoch": 0.6055253623188406, "grad_norm": 0.1449679582980524, "learning_rate": 7.989150193672962e-05, "loss": 0.9795, "step": 1337 }, { "epoch": 0.6059782608695652, "grad_norm": 0.1391445899826951, "learning_rate": 7.989056905014152e-05, "loss": 0.9841, "step": 1338 }, { "epoch": 0.6064311594202898, "grad_norm": 0.1701878036610053, "learning_rate": 7.988963217563001e-05, "loss": 0.9641, "step": 1339 }, { "epoch": 0.6068840579710145, "grad_norm": 0.16160389422103863, "learning_rate": 7.988869131328876e-05, "loss": 0.9845, "step": 1340 }, { "epoch": 0.6073369565217391, "grad_norm": 0.12276962088880518, "learning_rate": 7.988774646321186e-05, "loss": 0.9792, "step": 1341 }, { "epoch": 0.6077898550724637, "grad_norm": 0.12931886146167265, "learning_rate": 7.988679762549369e-05, "loss": 0.9788, "step": 1342 }, { "epoch": 0.6082427536231884, "grad_norm": 0.12884598006113923, "learning_rate": 7.988584480022919e-05, "loss": 0.9557, "step": 1343 }, { "epoch": 0.6086956521739131, "grad_norm": 0.15635695206646677, "learning_rate": 7.988488798751358e-05, "loss": 0.9627, "step": 1344 }, { "epoch": 0.6091485507246377, "grad_norm": 0.17606692338060137, "learning_rate": 7.98839271874425e-05, "loss": 0.9868, "step": 1345 }, { "epoch": 0.6096014492753623, "grad_norm": 0.16402629592708542, "learning_rate": 7.988296240011203e-05, "loss": 0.9622, "step": 1346 }, { "epoch": 0.6100543478260869, "grad_norm": 0.12473761881085917, "learning_rate": 7.98819936256186e-05, "loss": 0.987, "step": 1347 }, { "epoch": 0.6105072463768116, "grad_norm": 0.12413828212076651, "learning_rate": 7.988102086405907e-05, "loss": 0.9732, "step": 1348 }, { "epoch": 0.6109601449275363, "grad_norm": 0.14407676081985277, "learning_rate": 7.98800441155307e-05, "loss": 0.9829, "step": 1349 }, { "epoch": 0.6114130434782609, "grad_norm": 0.1669143651796158, "learning_rate": 7.98790633801311e-05, "loss": 0.9588, "step": 1350 }, { "epoch": 0.6118659420289855, "grad_norm": 0.19664417762735673, "learning_rate": 7.987807865795836e-05, "loss": 1.0092, "step": 1351 }, { "epoch": 0.6123188405797102, "grad_norm": 0.16625710001762864, "learning_rate": 7.987708994911091e-05, "loss": 0.9763, "step": 1352 }, { "epoch": 0.6127717391304348, "grad_norm": 0.18210978708770437, "learning_rate": 7.987609725368757e-05, "loss": 1.0067, "step": 1353 }, { "epoch": 0.6132246376811594, "grad_norm": 0.19182038334243, "learning_rate": 7.987510057178761e-05, "loss": 0.9574, "step": 1354 }, { "epoch": 0.613677536231884, "grad_norm": 0.17423095798569516, "learning_rate": 7.987409990351065e-05, "loss": 0.9816, "step": 1355 }, { "epoch": 0.6141304347826086, "grad_norm": 0.15991859180339346, "learning_rate": 7.987309524895674e-05, "loss": 0.9722, "step": 1356 }, { "epoch": 0.6145833333333334, "grad_norm": 0.14247035632524344, "learning_rate": 7.987208660822631e-05, "loss": 0.9571, "step": 1357 }, { "epoch": 0.615036231884058, "grad_norm": 0.15748064393115505, "learning_rate": 7.98710739814202e-05, "loss": 0.9625, "step": 1358 }, { "epoch": 0.6154891304347826, "grad_norm": 0.1644451090591752, "learning_rate": 7.987005736863964e-05, "loss": 0.9766, "step": 1359 }, { "epoch": 0.6159420289855072, "grad_norm": 0.1728982198114495, "learning_rate": 7.986903676998628e-05, "loss": 0.9835, "step": 1360 }, { "epoch": 0.6163949275362319, "grad_norm": 0.2153494998236758, "learning_rate": 7.986801218556212e-05, "loss": 0.9676, "step": 1361 }, { "epoch": 0.6168478260869565, "grad_norm": 0.25839870290910383, "learning_rate": 7.986698361546962e-05, "loss": 0.9772, "step": 1362 }, { "epoch": 0.6173007246376812, "grad_norm": 0.2919573496806207, "learning_rate": 7.986595105981159e-05, "loss": 0.9452, "step": 1363 }, { "epoch": 0.6177536231884058, "grad_norm": 0.3406421491642597, "learning_rate": 7.986491451869125e-05, "loss": 0.9809, "step": 1364 }, { "epoch": 0.6182065217391305, "grad_norm": 0.31288175556626724, "learning_rate": 7.986387399221224e-05, "loss": 0.9757, "step": 1365 }, { "epoch": 0.6186594202898551, "grad_norm": 0.2757061532243715, "learning_rate": 7.986282948047857e-05, "loss": 0.965, "step": 1366 }, { "epoch": 0.6191123188405797, "grad_norm": 0.22772996767590428, "learning_rate": 7.986178098359468e-05, "loss": 0.9886, "step": 1367 }, { "epoch": 0.6195652173913043, "grad_norm": 0.17799152751337652, "learning_rate": 7.986072850166538e-05, "loss": 0.9845, "step": 1368 }, { "epoch": 0.6200181159420289, "grad_norm": 0.16166200477678552, "learning_rate": 7.985967203479588e-05, "loss": 0.9636, "step": 1369 }, { "epoch": 0.6204710144927537, "grad_norm": 0.17340508436959925, "learning_rate": 7.98586115830918e-05, "loss": 0.9603, "step": 1370 }, { "epoch": 0.6209239130434783, "grad_norm": 0.21330206805933102, "learning_rate": 7.985754714665916e-05, "loss": 0.9603, "step": 1371 }, { "epoch": 0.6213768115942029, "grad_norm": 0.22331945100404962, "learning_rate": 7.985647872560437e-05, "loss": 0.9694, "step": 1372 }, { "epoch": 0.6218297101449275, "grad_norm": 0.24065705840262294, "learning_rate": 7.985540632003425e-05, "loss": 0.9814, "step": 1373 }, { "epoch": 0.6222826086956522, "grad_norm": 0.2670132934069037, "learning_rate": 7.985432993005601e-05, "loss": 0.978, "step": 1374 }, { "epoch": 0.6227355072463768, "grad_norm": 0.22483221487618366, "learning_rate": 7.985324955577724e-05, "loss": 0.9799, "step": 1375 }, { "epoch": 0.6231884057971014, "grad_norm": 0.15344251390840258, "learning_rate": 7.985216519730597e-05, "loss": 1.0061, "step": 1376 }, { "epoch": 0.623641304347826, "grad_norm": 0.12905164454128898, "learning_rate": 7.985107685475058e-05, "loss": 0.9833, "step": 1377 }, { "epoch": 0.6240942028985508, "grad_norm": 0.13449904363607748, "learning_rate": 7.98499845282199e-05, "loss": 0.9684, "step": 1378 }, { "epoch": 0.6245471014492754, "grad_norm": 0.1700834316086501, "learning_rate": 7.984888821782312e-05, "loss": 0.9715, "step": 1379 }, { "epoch": 0.625, "grad_norm": 0.19910222845362263, "learning_rate": 7.984778792366983e-05, "loss": 0.9633, "step": 1380 }, { "epoch": 0.6254528985507246, "grad_norm": 0.2271525149565432, "learning_rate": 7.984668364587004e-05, "loss": 0.9592, "step": 1381 }, { "epoch": 0.6259057971014492, "grad_norm": 0.2704576793155046, "learning_rate": 7.984557538453415e-05, "loss": 0.996, "step": 1382 }, { "epoch": 0.626358695652174, "grad_norm": 0.2919946280244541, "learning_rate": 7.984446313977293e-05, "loss": 0.9552, "step": 1383 }, { "epoch": 0.6268115942028986, "grad_norm": 0.26395556484108823, "learning_rate": 7.984334691169761e-05, "loss": 0.9784, "step": 1384 }, { "epoch": 0.6272644927536232, "grad_norm": 0.23561412632011916, "learning_rate": 7.984222670041975e-05, "loss": 0.9744, "step": 1385 }, { "epoch": 0.6277173913043478, "grad_norm": 0.20053443491926665, "learning_rate": 7.984110250605136e-05, "loss": 0.9729, "step": 1386 }, { "epoch": 0.6281702898550725, "grad_norm": 0.18415257748259153, "learning_rate": 7.983997432870481e-05, "loss": 0.9795, "step": 1387 }, { "epoch": 0.6286231884057971, "grad_norm": 0.1431405149158087, "learning_rate": 7.98388421684929e-05, "loss": 0.965, "step": 1388 }, { "epoch": 0.6290760869565217, "grad_norm": 0.13905181975734784, "learning_rate": 7.983770602552881e-05, "loss": 0.976, "step": 1389 }, { "epoch": 0.6295289855072463, "grad_norm": 0.17224880243115298, "learning_rate": 7.983656589992612e-05, "loss": 0.9718, "step": 1390 }, { "epoch": 0.6299818840579711, "grad_norm": 0.17174901461550646, "learning_rate": 7.98354217917988e-05, "loss": 0.9769, "step": 1391 }, { "epoch": 0.6304347826086957, "grad_norm": 0.13528386908191192, "learning_rate": 7.983427370126126e-05, "loss": 0.9562, "step": 1392 }, { "epoch": 0.6308876811594203, "grad_norm": 0.1529703063241725, "learning_rate": 7.983312162842824e-05, "loss": 0.9532, "step": 1393 }, { "epoch": 0.6313405797101449, "grad_norm": 0.15807973901987793, "learning_rate": 7.983196557341493e-05, "loss": 0.9794, "step": 1394 }, { "epoch": 0.6317934782608695, "grad_norm": 0.12048827235362809, "learning_rate": 7.983080553633691e-05, "loss": 0.9519, "step": 1395 }, { "epoch": 0.6322463768115942, "grad_norm": 0.1178653670130735, "learning_rate": 7.982964151731014e-05, "loss": 0.9975, "step": 1396 }, { "epoch": 0.6326992753623188, "grad_norm": 0.154578112671014, "learning_rate": 7.9828473516451e-05, "loss": 0.9598, "step": 1397 }, { "epoch": 0.6331521739130435, "grad_norm": 0.17948468625603642, "learning_rate": 7.982730153387623e-05, "loss": 0.9794, "step": 1398 }, { "epoch": 0.6336050724637681, "grad_norm": 0.20578279854493695, "learning_rate": 7.982612556970302e-05, "loss": 0.9822, "step": 1399 }, { "epoch": 0.6340579710144928, "grad_norm": 0.1707319342332598, "learning_rate": 7.982494562404893e-05, "loss": 0.969, "step": 1400 }, { "epoch": 0.6345108695652174, "grad_norm": 0.1872724571971369, "learning_rate": 7.982376169703192e-05, "loss": 0.9675, "step": 1401 }, { "epoch": 0.634963768115942, "grad_norm": 0.17808487705423784, "learning_rate": 7.982257378877034e-05, "loss": 0.9898, "step": 1402 }, { "epoch": 0.6354166666666666, "grad_norm": 0.12714450153404802, "learning_rate": 7.982138189938296e-05, "loss": 0.987, "step": 1403 }, { "epoch": 0.6358695652173914, "grad_norm": 0.11177012477348132, "learning_rate": 7.982018602898893e-05, "loss": 0.9586, "step": 1404 }, { "epoch": 0.636322463768116, "grad_norm": 0.15617776652601195, "learning_rate": 7.98189861777078e-05, "loss": 0.9707, "step": 1405 }, { "epoch": 0.6367753623188406, "grad_norm": 0.19722082201593255, "learning_rate": 7.981778234565953e-05, "loss": 1.0021, "step": 1406 }, { "epoch": 0.6372282608695652, "grad_norm": 0.16022233340622993, "learning_rate": 7.981657453296445e-05, "loss": 0.968, "step": 1407 }, { "epoch": 0.6376811594202898, "grad_norm": 0.15287914312815118, "learning_rate": 7.981536273974332e-05, "loss": 0.9746, "step": 1408 }, { "epoch": 0.6381340579710145, "grad_norm": 0.20406605041317302, "learning_rate": 7.98141469661173e-05, "loss": 0.9841, "step": 1409 }, { "epoch": 0.6385869565217391, "grad_norm": 0.1989323840521731, "learning_rate": 7.98129272122079e-05, "loss": 0.9858, "step": 1410 }, { "epoch": 0.6390398550724637, "grad_norm": 0.19387667968921157, "learning_rate": 7.981170347813708e-05, "loss": 0.987, "step": 1411 }, { "epoch": 0.6394927536231884, "grad_norm": 0.16955763322722356, "learning_rate": 7.981047576402718e-05, "loss": 0.9682, "step": 1412 }, { "epoch": 0.6399456521739131, "grad_norm": 0.20928235915407986, "learning_rate": 7.980924407000094e-05, "loss": 0.9973, "step": 1413 }, { "epoch": 0.6403985507246377, "grad_norm": 0.15981021302805412, "learning_rate": 7.980800839618149e-05, "loss": 0.9787, "step": 1414 }, { "epoch": 0.6408514492753623, "grad_norm": 0.1502661551594261, "learning_rate": 7.980676874269235e-05, "loss": 0.96, "step": 1415 }, { "epoch": 0.6413043478260869, "grad_norm": 0.15754078469346902, "learning_rate": 7.980552510965744e-05, "loss": 0.9746, "step": 1416 }, { "epoch": 0.6417572463768116, "grad_norm": 0.1331288292660801, "learning_rate": 7.980427749720113e-05, "loss": 0.958, "step": 1417 }, { "epoch": 0.6422101449275363, "grad_norm": 0.16047928104019413, "learning_rate": 7.980302590544812e-05, "loss": 0.9694, "step": 1418 }, { "epoch": 0.6426630434782609, "grad_norm": 0.1990846438029361, "learning_rate": 7.980177033452354e-05, "loss": 0.9802, "step": 1419 }, { "epoch": 0.6431159420289855, "grad_norm": 0.2604027566834796, "learning_rate": 7.980051078455291e-05, "loss": 0.965, "step": 1420 }, { "epoch": 0.6435688405797102, "grad_norm": 0.1764983639109391, "learning_rate": 7.979924725566214e-05, "loss": 0.9903, "step": 1421 }, { "epoch": 0.6440217391304348, "grad_norm": 0.2142795262595466, "learning_rate": 7.979797974797755e-05, "loss": 0.9813, "step": 1422 }, { "epoch": 0.6444746376811594, "grad_norm": 0.20467551049456337, "learning_rate": 7.979670826162587e-05, "loss": 0.9804, "step": 1423 }, { "epoch": 0.644927536231884, "grad_norm": 0.15652193258611616, "learning_rate": 7.979543279673419e-05, "loss": 0.9922, "step": 1424 }, { "epoch": 0.6453804347826086, "grad_norm": 0.15524137220846526, "learning_rate": 7.979415335343005e-05, "loss": 0.9685, "step": 1425 }, { "epoch": 0.6458333333333334, "grad_norm": 0.15403661244053035, "learning_rate": 7.979286993184134e-05, "loss": 0.9709, "step": 1426 }, { "epoch": 0.646286231884058, "grad_norm": 0.17571087866739932, "learning_rate": 7.979158253209634e-05, "loss": 0.992, "step": 1427 }, { "epoch": 0.6467391304347826, "grad_norm": 0.16231581540547338, "learning_rate": 7.97902911543238e-05, "loss": 0.9645, "step": 1428 }, { "epoch": 0.6471920289855072, "grad_norm": 0.1596306850657105, "learning_rate": 7.97889957986528e-05, "loss": 0.9851, "step": 1429 }, { "epoch": 0.6476449275362319, "grad_norm": 0.14996823286289473, "learning_rate": 7.978769646521283e-05, "loss": 0.9736, "step": 1430 }, { "epoch": 0.6480978260869565, "grad_norm": 0.16244520664231954, "learning_rate": 7.97863931541338e-05, "loss": 0.9902, "step": 1431 }, { "epoch": 0.6485507246376812, "grad_norm": 0.16548768196092298, "learning_rate": 7.978508586554599e-05, "loss": 0.9764, "step": 1432 }, { "epoch": 0.6490036231884058, "grad_norm": 0.16721687811739536, "learning_rate": 7.978377459958012e-05, "loss": 0.9617, "step": 1433 }, { "epoch": 0.6494565217391305, "grad_norm": 0.16381128135157766, "learning_rate": 7.978245935636725e-05, "loss": 0.9721, "step": 1434 }, { "epoch": 0.6499094202898551, "grad_norm": 0.12651739140908974, "learning_rate": 7.978114013603888e-05, "loss": 0.9716, "step": 1435 }, { "epoch": 0.6503623188405797, "grad_norm": 0.13797839480999285, "learning_rate": 7.977981693872688e-05, "loss": 0.9763, "step": 1436 }, { "epoch": 0.6508152173913043, "grad_norm": 0.12453439033096673, "learning_rate": 7.977848976456356e-05, "loss": 0.9591, "step": 1437 }, { "epoch": 0.6512681159420289, "grad_norm": 0.126211345301636, "learning_rate": 7.977715861368157e-05, "loss": 0.9689, "step": 1438 }, { "epoch": 0.6517210144927537, "grad_norm": 0.1607104127336661, "learning_rate": 7.977582348621402e-05, "loss": 0.9639, "step": 1439 }, { "epoch": 0.6521739130434783, "grad_norm": 0.22820467515607634, "learning_rate": 7.977448438229435e-05, "loss": 1.0029, "step": 1440 }, { "epoch": 0.6526268115942029, "grad_norm": 0.13915792148606743, "learning_rate": 7.977314130205646e-05, "loss": 0.9784, "step": 1441 }, { "epoch": 0.6530797101449275, "grad_norm": 0.13871019696579873, "learning_rate": 7.97717942456346e-05, "loss": 0.9714, "step": 1442 }, { "epoch": 0.6535326086956522, "grad_norm": 0.14060376608254818, "learning_rate": 7.977044321316344e-05, "loss": 0.9825, "step": 1443 }, { "epoch": 0.6539855072463768, "grad_norm": 0.1526021755547854, "learning_rate": 7.976908820477805e-05, "loss": 0.9788, "step": 1444 }, { "epoch": 0.6544384057971014, "grad_norm": 0.15805366500831833, "learning_rate": 7.97677292206139e-05, "loss": 0.9784, "step": 1445 }, { "epoch": 0.654891304347826, "grad_norm": 0.12635494915898784, "learning_rate": 7.976636626080685e-05, "loss": 0.9576, "step": 1446 }, { "epoch": 0.6553442028985508, "grad_norm": 0.1380698128883278, "learning_rate": 7.976499932549314e-05, "loss": 1.0127, "step": 1447 }, { "epoch": 0.6557971014492754, "grad_norm": 0.22489907563611605, "learning_rate": 7.976362841480944e-05, "loss": 0.9799, "step": 1448 }, { "epoch": 0.65625, "grad_norm": 0.16426024969551165, "learning_rate": 7.976225352889278e-05, "loss": 0.9871, "step": 1449 }, { "epoch": 0.6567028985507246, "grad_norm": 0.19134117014942492, "learning_rate": 7.976087466788064e-05, "loss": 0.9976, "step": 1450 }, { "epoch": 0.6571557971014492, "grad_norm": 0.2133119773325539, "learning_rate": 7.975949183191086e-05, "loss": 0.9739, "step": 1451 }, { "epoch": 0.657608695652174, "grad_norm": 0.26113435097213306, "learning_rate": 7.975810502112167e-05, "loss": 0.9886, "step": 1452 }, { "epoch": 0.6580615942028986, "grad_norm": 0.3384780527646442, "learning_rate": 7.975671423565171e-05, "loss": 0.9777, "step": 1453 }, { "epoch": 0.6585144927536232, "grad_norm": 0.42006207911811705, "learning_rate": 7.975531947564005e-05, "loss": 0.9835, "step": 1454 }, { "epoch": 0.6589673913043478, "grad_norm": 0.4595524891969281, "learning_rate": 7.975392074122608e-05, "loss": 0.9576, "step": 1455 }, { "epoch": 0.6594202898550725, "grad_norm": 0.45818175741357564, "learning_rate": 7.975251803254967e-05, "loss": 0.9726, "step": 1456 }, { "epoch": 0.6598731884057971, "grad_norm": 0.46128519089323156, "learning_rate": 7.975111134975104e-05, "loss": 0.9825, "step": 1457 }, { "epoch": 0.6603260869565217, "grad_norm": 0.4562495922782116, "learning_rate": 7.974970069297082e-05, "loss": 0.9703, "step": 1458 }, { "epoch": 0.6607789855072463, "grad_norm": 0.40972505269085485, "learning_rate": 7.974828606235002e-05, "loss": 0.9892, "step": 1459 }, { "epoch": 0.6612318840579711, "grad_norm": 0.2868357950107164, "learning_rate": 7.974686745803008e-05, "loss": 0.9706, "step": 1460 }, { "epoch": 0.6616847826086957, "grad_norm": 0.18426466651306558, "learning_rate": 7.974544488015282e-05, "loss": 0.9604, "step": 1461 }, { "epoch": 0.6621376811594203, "grad_norm": 0.21180615189071408, "learning_rate": 7.974401832886046e-05, "loss": 0.9987, "step": 1462 }, { "epoch": 0.6625905797101449, "grad_norm": 0.23790318371769104, "learning_rate": 7.974258780429559e-05, "loss": 0.9785, "step": 1463 }, { "epoch": 0.6630434782608695, "grad_norm": 0.25934584110124426, "learning_rate": 7.974115330660125e-05, "loss": 0.9906, "step": 1464 }, { "epoch": 0.6634963768115942, "grad_norm": 0.24568983705273742, "learning_rate": 7.973971483592085e-05, "loss": 0.9769, "step": 1465 }, { "epoch": 0.6639492753623188, "grad_norm": 0.2059982364026228, "learning_rate": 7.973827239239816e-05, "loss": 0.9959, "step": 1466 }, { "epoch": 0.6644021739130435, "grad_norm": 0.16343520711841103, "learning_rate": 7.973682597617743e-05, "loss": 0.9753, "step": 1467 }, { "epoch": 0.6648550724637681, "grad_norm": 0.1427154764432969, "learning_rate": 7.973537558740323e-05, "loss": 0.9669, "step": 1468 }, { "epoch": 0.6653079710144928, "grad_norm": 0.16091382988259229, "learning_rate": 7.973392122622056e-05, "loss": 0.9413, "step": 1469 }, { "epoch": 0.6657608695652174, "grad_norm": 0.17237851141570928, "learning_rate": 7.973246289277483e-05, "loss": 0.9502, "step": 1470 }, { "epoch": 0.666213768115942, "grad_norm": 0.20762326472776796, "learning_rate": 7.973100058721182e-05, "loss": 0.9762, "step": 1471 }, { "epoch": 0.6666666666666666, "grad_norm": 0.21225926456850844, "learning_rate": 7.972953430967773e-05, "loss": 0.9706, "step": 1472 }, { "epoch": 0.6671195652173914, "grad_norm": 0.17936376539205454, "learning_rate": 7.972806406031914e-05, "loss": 0.9616, "step": 1473 }, { "epoch": 0.667572463768116, "grad_norm": 0.17487283849453605, "learning_rate": 7.972658983928301e-05, "loss": 0.9967, "step": 1474 }, { "epoch": 0.6680253623188406, "grad_norm": 0.1682176955867142, "learning_rate": 7.972511164671675e-05, "loss": 0.9774, "step": 1475 }, { "epoch": 0.6684782608695652, "grad_norm": 0.13671588662441417, "learning_rate": 7.972362948276814e-05, "loss": 0.9606, "step": 1476 }, { "epoch": 0.6689311594202898, "grad_norm": 0.1646010060489425, "learning_rate": 7.972214334758534e-05, "loss": 0.9729, "step": 1477 }, { "epoch": 0.6693840579710145, "grad_norm": 0.21115336525820302, "learning_rate": 7.972065324131692e-05, "loss": 0.9845, "step": 1478 }, { "epoch": 0.6698369565217391, "grad_norm": 0.15686697516092268, "learning_rate": 7.971915916411185e-05, "loss": 0.9719, "step": 1479 }, { "epoch": 0.6702898550724637, "grad_norm": 0.13008678450288652, "learning_rate": 7.97176611161195e-05, "loss": 0.9632, "step": 1480 }, { "epoch": 0.6707427536231884, "grad_norm": 0.18942647865077356, "learning_rate": 7.971615909748963e-05, "loss": 0.9805, "step": 1481 }, { "epoch": 0.6711956521739131, "grad_norm": 0.21325960356047904, "learning_rate": 7.97146531083724e-05, "loss": 0.9967, "step": 1482 }, { "epoch": 0.6716485507246377, "grad_norm": 0.17866493566278122, "learning_rate": 7.971314314891838e-05, "loss": 0.9875, "step": 1483 }, { "epoch": 0.6721014492753623, "grad_norm": 0.14986860454028614, "learning_rate": 7.971162921927848e-05, "loss": 0.9862, "step": 1484 }, { "epoch": 0.6725543478260869, "grad_norm": 0.15778601210043083, "learning_rate": 7.97101113196041e-05, "loss": 0.9746, "step": 1485 }, { "epoch": 0.6730072463768116, "grad_norm": 0.16223771462483483, "learning_rate": 7.970858945004695e-05, "loss": 0.9793, "step": 1486 }, { "epoch": 0.6734601449275363, "grad_norm": 0.13594821793876882, "learning_rate": 7.97070636107592e-05, "loss": 0.9746, "step": 1487 }, { "epoch": 0.6739130434782609, "grad_norm": 0.14237810084455513, "learning_rate": 7.970553380189338e-05, "loss": 0.9675, "step": 1488 }, { "epoch": 0.6743659420289855, "grad_norm": 0.1819380205230972, "learning_rate": 7.970400002360243e-05, "loss": 0.9933, "step": 1489 }, { "epoch": 0.6748188405797102, "grad_norm": 0.16950334066577194, "learning_rate": 7.970246227603966e-05, "loss": 0.9828, "step": 1490 }, { "epoch": 0.6752717391304348, "grad_norm": 0.12289804471524922, "learning_rate": 7.970092055935886e-05, "loss": 0.989, "step": 1491 }, { "epoch": 0.6757246376811594, "grad_norm": 0.18999462351651566, "learning_rate": 7.969937487371409e-05, "loss": 0.9383, "step": 1492 }, { "epoch": 0.676177536231884, "grad_norm": 0.19110617547389616, "learning_rate": 7.969782521925992e-05, "loss": 0.9782, "step": 1493 }, { "epoch": 0.6766304347826086, "grad_norm": 0.1359226669959392, "learning_rate": 7.969627159615124e-05, "loss": 0.9654, "step": 1494 }, { "epoch": 0.6770833333333334, "grad_norm": 0.14473306761499488, "learning_rate": 7.96947140045434e-05, "loss": 0.9834, "step": 1495 }, { "epoch": 0.677536231884058, "grad_norm": 0.16911717777606047, "learning_rate": 7.969315244459209e-05, "loss": 0.9589, "step": 1496 }, { "epoch": 0.6779891304347826, "grad_norm": 0.16300535288403817, "learning_rate": 7.969158691645343e-05, "loss": 0.9828, "step": 1497 }, { "epoch": 0.6784420289855072, "grad_norm": 0.1853355857653066, "learning_rate": 7.969001742028394e-05, "loss": 1.0024, "step": 1498 }, { "epoch": 0.6788949275362319, "grad_norm": 0.18461553908549994, "learning_rate": 7.96884439562405e-05, "loss": 0.9429, "step": 1499 }, { "epoch": 0.6793478260869565, "grad_norm": 0.17959294804955986, "learning_rate": 7.968686652448043e-05, "loss": 0.9673, "step": 1500 }, { "epoch": 0.6798007246376812, "grad_norm": 0.14264713694459177, "learning_rate": 7.968528512516143e-05, "loss": 0.9815, "step": 1501 }, { "epoch": 0.6802536231884058, "grad_norm": 0.10533557799824689, "learning_rate": 7.96836997584416e-05, "loss": 0.9687, "step": 1502 }, { "epoch": 0.6807065217391305, "grad_norm": 0.14053941888015906, "learning_rate": 7.968211042447939e-05, "loss": 0.9595, "step": 1503 }, { "epoch": 0.6811594202898551, "grad_norm": 0.17699637476666316, "learning_rate": 7.968051712343374e-05, "loss": 0.9721, "step": 1504 }, { "epoch": 0.6816123188405797, "grad_norm": 0.15399824622673336, "learning_rate": 7.967891985546391e-05, "loss": 0.943, "step": 1505 }, { "epoch": 0.6820652173913043, "grad_norm": 0.10443761489351212, "learning_rate": 7.967731862072959e-05, "loss": 0.9655, "step": 1506 }, { "epoch": 0.6825181159420289, "grad_norm": 0.13942073091947546, "learning_rate": 7.967571341939085e-05, "loss": 0.9936, "step": 1507 }, { "epoch": 0.6829710144927537, "grad_norm": 0.14465871129498328, "learning_rate": 7.967410425160817e-05, "loss": 0.9937, "step": 1508 }, { "epoch": 0.6834239130434783, "grad_norm": 0.11902431241531428, "learning_rate": 7.96724911175424e-05, "loss": 0.9764, "step": 1509 }, { "epoch": 0.6838768115942029, "grad_norm": 0.11950263553864858, "learning_rate": 7.967087401735486e-05, "loss": 0.9875, "step": 1510 }, { "epoch": 0.6843297101449275, "grad_norm": 0.12525535735383891, "learning_rate": 7.966925295120715e-05, "loss": 0.9623, "step": 1511 }, { "epoch": 0.6847826086956522, "grad_norm": 0.15164563400106398, "learning_rate": 7.966762791926138e-05, "loss": 0.9753, "step": 1512 }, { "epoch": 0.6852355072463768, "grad_norm": 0.17876761153988122, "learning_rate": 7.966599892168e-05, "loss": 0.9937, "step": 1513 }, { "epoch": 0.6856884057971014, "grad_norm": 0.1627160705380191, "learning_rate": 7.966436595862583e-05, "loss": 0.9692, "step": 1514 }, { "epoch": 0.686141304347826, "grad_norm": 0.12184515919402869, "learning_rate": 7.966272903026214e-05, "loss": 0.9792, "step": 1515 }, { "epoch": 0.6865942028985508, "grad_norm": 0.12284722988987992, "learning_rate": 7.966108813675258e-05, "loss": 0.9727, "step": 1516 }, { "epoch": 0.6870471014492754, "grad_norm": 0.12614448289196029, "learning_rate": 7.965944327826119e-05, "loss": 0.9678, "step": 1517 }, { "epoch": 0.6875, "grad_norm": 0.15282951018474067, "learning_rate": 7.965779445495243e-05, "loss": 0.983, "step": 1518 }, { "epoch": 0.6879528985507246, "grad_norm": 0.1670037158179363, "learning_rate": 7.965614166699108e-05, "loss": 0.9882, "step": 1519 }, { "epoch": 0.6884057971014492, "grad_norm": 0.13822199381251257, "learning_rate": 7.965448491454243e-05, "loss": 0.981, "step": 1520 }, { "epoch": 0.688858695652174, "grad_norm": 0.1615874287976248, "learning_rate": 7.96528241977721e-05, "loss": 0.9797, "step": 1521 }, { "epoch": 0.6893115942028986, "grad_norm": 0.2094486070204406, "learning_rate": 7.965115951684608e-05, "loss": 0.9562, "step": 1522 }, { "epoch": 0.6897644927536232, "grad_norm": 0.17482455816258727, "learning_rate": 7.964949087193081e-05, "loss": 0.9859, "step": 1523 }, { "epoch": 0.6902173913043478, "grad_norm": 0.13373285629933598, "learning_rate": 7.96478182631931e-05, "loss": 0.9711, "step": 1524 }, { "epoch": 0.6906702898550725, "grad_norm": 0.14810145168451824, "learning_rate": 7.964614169080018e-05, "loss": 0.9443, "step": 1525 }, { "epoch": 0.6911231884057971, "grad_norm": 0.1771156861921344, "learning_rate": 7.964446115491966e-05, "loss": 0.973, "step": 1526 }, { "epoch": 0.6915760869565217, "grad_norm": 0.16864431476567698, "learning_rate": 7.964277665571952e-05, "loss": 0.9725, "step": 1527 }, { "epoch": 0.6920289855072463, "grad_norm": 0.18486920396714407, "learning_rate": 7.96410881933682e-05, "loss": 0.9763, "step": 1528 }, { "epoch": 0.6924818840579711, "grad_norm": 0.17799842529296964, "learning_rate": 7.963939576803445e-05, "loss": 0.9774, "step": 1529 }, { "epoch": 0.6929347826086957, "grad_norm": 0.13965384094039424, "learning_rate": 7.96376993798875e-05, "loss": 0.9524, "step": 1530 }, { "epoch": 0.6933876811594203, "grad_norm": 0.1357162745927602, "learning_rate": 7.963599902909694e-05, "loss": 0.9651, "step": 1531 }, { "epoch": 0.6938405797101449, "grad_norm": 0.16528316027382908, "learning_rate": 7.963429471583274e-05, "loss": 0.9732, "step": 1532 }, { "epoch": 0.6942934782608695, "grad_norm": 0.22789020857874123, "learning_rate": 7.96325864402653e-05, "loss": 0.9659, "step": 1533 }, { "epoch": 0.6947463768115942, "grad_norm": 0.27114917686957896, "learning_rate": 7.963087420256538e-05, "loss": 0.9717, "step": 1534 }, { "epoch": 0.6951992753623188, "grad_norm": 0.2828423363774735, "learning_rate": 7.962915800290417e-05, "loss": 0.9778, "step": 1535 }, { "epoch": 0.6956521739130435, "grad_norm": 0.30602899614675033, "learning_rate": 7.962743784145323e-05, "loss": 0.9475, "step": 1536 }, { "epoch": 0.6961050724637681, "grad_norm": 0.330297799035602, "learning_rate": 7.962571371838455e-05, "loss": 0.9778, "step": 1537 }, { "epoch": 0.6965579710144928, "grad_norm": 0.33303770253985937, "learning_rate": 7.962398563387047e-05, "loss": 0.9686, "step": 1538 }, { "epoch": 0.6970108695652174, "grad_norm": 0.2802788365218471, "learning_rate": 7.962225358808375e-05, "loss": 0.9798, "step": 1539 }, { "epoch": 0.697463768115942, "grad_norm": 0.19989431690788317, "learning_rate": 7.962051758119756e-05, "loss": 0.9647, "step": 1540 }, { "epoch": 0.6979166666666666, "grad_norm": 0.17038089229113945, "learning_rate": 7.961877761338545e-05, "loss": 0.9619, "step": 1541 }, { "epoch": 0.6983695652173914, "grad_norm": 0.19024587225041747, "learning_rate": 7.961703368482134e-05, "loss": 0.9955, "step": 1542 }, { "epoch": 0.698822463768116, "grad_norm": 0.15971036911145894, "learning_rate": 7.96152857956796e-05, "loss": 0.9809, "step": 1543 }, { "epoch": 0.6992753623188406, "grad_norm": 0.14814409669982642, "learning_rate": 7.961353394613497e-05, "loss": 0.9777, "step": 1544 }, { "epoch": 0.6997282608695652, "grad_norm": 0.12974376863493922, "learning_rate": 7.961177813636258e-05, "loss": 0.9659, "step": 1545 }, { "epoch": 0.7001811594202898, "grad_norm": 0.15941785660150073, "learning_rate": 7.961001836653795e-05, "loss": 0.9733, "step": 1546 }, { "epoch": 0.7006340579710145, "grad_norm": 0.1777127385969093, "learning_rate": 7.960825463683702e-05, "loss": 0.9713, "step": 1547 }, { "epoch": 0.7010869565217391, "grad_norm": 0.16927895776056515, "learning_rate": 7.960648694743612e-05, "loss": 0.9665, "step": 1548 }, { "epoch": 0.7015398550724637, "grad_norm": 0.13840983322380357, "learning_rate": 7.960471529851194e-05, "loss": 0.9762, "step": 1549 }, { "epoch": 0.7019927536231884, "grad_norm": 0.1377146098400078, "learning_rate": 7.960293969024164e-05, "loss": 0.9993, "step": 1550 }, { "epoch": 0.7024456521739131, "grad_norm": 0.1762961173775715, "learning_rate": 7.960116012280267e-05, "loss": 0.9598, "step": 1551 }, { "epoch": 0.7028985507246377, "grad_norm": 0.16927476005872885, "learning_rate": 7.959937659637298e-05, "loss": 0.9756, "step": 1552 }, { "epoch": 0.7033514492753623, "grad_norm": 0.19009823718406793, "learning_rate": 7.959758911113087e-05, "loss": 0.9692, "step": 1553 }, { "epoch": 0.7038043478260869, "grad_norm": 0.1855441916329149, "learning_rate": 7.959579766725503e-05, "loss": 0.9636, "step": 1554 }, { "epoch": 0.7042572463768116, "grad_norm": 0.16971540032894225, "learning_rate": 7.959400226492455e-05, "loss": 0.9881, "step": 1555 }, { "epoch": 0.7047101449275363, "grad_norm": 0.15940265866406958, "learning_rate": 7.959220290431891e-05, "loss": 0.9715, "step": 1556 }, { "epoch": 0.7051630434782609, "grad_norm": 0.14451764865672886, "learning_rate": 7.959039958561801e-05, "loss": 0.9674, "step": 1557 }, { "epoch": 0.7056159420289855, "grad_norm": 0.13840528294818177, "learning_rate": 7.958859230900214e-05, "loss": 0.983, "step": 1558 }, { "epoch": 0.7060688405797102, "grad_norm": 0.13761026117632325, "learning_rate": 7.958678107465194e-05, "loss": 0.9591, "step": 1559 }, { "epoch": 0.7065217391304348, "grad_norm": 0.1099517483170332, "learning_rate": 7.958496588274853e-05, "loss": 0.9856, "step": 1560 }, { "epoch": 0.7069746376811594, "grad_norm": 0.13949661401388266, "learning_rate": 7.958314673347334e-05, "loss": 0.9723, "step": 1561 }, { "epoch": 0.707427536231884, "grad_norm": 0.2084942850897627, "learning_rate": 7.958132362700826e-05, "loss": 0.9876, "step": 1562 }, { "epoch": 0.7078804347826086, "grad_norm": 0.167629931687637, "learning_rate": 7.957949656353551e-05, "loss": 1.0034, "step": 1563 }, { "epoch": 0.7083333333333334, "grad_norm": 0.14497259514112434, "learning_rate": 7.957766554323778e-05, "loss": 0.974, "step": 1564 }, { "epoch": 0.708786231884058, "grad_norm": 0.14131417922048314, "learning_rate": 7.957583056629812e-05, "loss": 0.9702, "step": 1565 }, { "epoch": 0.7092391304347826, "grad_norm": 0.15159512141873874, "learning_rate": 7.957399163289996e-05, "loss": 0.9638, "step": 1566 }, { "epoch": 0.7096920289855072, "grad_norm": 0.17802323150940616, "learning_rate": 7.957214874322714e-05, "loss": 0.9661, "step": 1567 }, { "epoch": 0.7101449275362319, "grad_norm": 0.21571437357806286, "learning_rate": 7.957030189746391e-05, "loss": 0.9574, "step": 1568 }, { "epoch": 0.7105978260869565, "grad_norm": 0.25821667913894153, "learning_rate": 7.956845109579488e-05, "loss": 0.9947, "step": 1569 }, { "epoch": 0.7110507246376812, "grad_norm": 0.2578266109149883, "learning_rate": 7.956659633840512e-05, "loss": 0.9468, "step": 1570 }, { "epoch": 0.7115036231884058, "grad_norm": 0.20926318719525086, "learning_rate": 7.956473762548001e-05, "loss": 0.9632, "step": 1571 }, { "epoch": 0.7119565217391305, "grad_norm": 0.15459673362749354, "learning_rate": 7.956287495720538e-05, "loss": 0.9679, "step": 1572 }, { "epoch": 0.7124094202898551, "grad_norm": 0.1527771003911009, "learning_rate": 7.956100833376745e-05, "loss": 0.9592, "step": 1573 }, { "epoch": 0.7128623188405797, "grad_norm": 0.13950939534093962, "learning_rate": 7.955913775535282e-05, "loss": 0.9733, "step": 1574 }, { "epoch": 0.7133152173913043, "grad_norm": 0.1348428121454017, "learning_rate": 7.955726322214852e-05, "loss": 0.987, "step": 1575 }, { "epoch": 0.7137681159420289, "grad_norm": 0.16362500518256362, "learning_rate": 7.955538473434191e-05, "loss": 0.9771, "step": 1576 }, { "epoch": 0.7142210144927537, "grad_norm": 0.23408916985996647, "learning_rate": 7.955350229212082e-05, "loss": 0.9667, "step": 1577 }, { "epoch": 0.7146739130434783, "grad_norm": 0.20884323651617034, "learning_rate": 7.955161589567343e-05, "loss": 0.9744, "step": 1578 }, { "epoch": 0.7151268115942029, "grad_norm": 0.19808757862889148, "learning_rate": 7.954972554518832e-05, "loss": 0.9755, "step": 1579 }, { "epoch": 0.7155797101449275, "grad_norm": 0.1599691581934039, "learning_rate": 7.954783124085447e-05, "loss": 0.9683, "step": 1580 }, { "epoch": 0.7160326086956522, "grad_norm": 0.13206611999277035, "learning_rate": 7.954593298286128e-05, "loss": 0.9603, "step": 1581 }, { "epoch": 0.7164855072463768, "grad_norm": 0.12440979484190981, "learning_rate": 7.954403077139849e-05, "loss": 0.9632, "step": 1582 }, { "epoch": 0.7169384057971014, "grad_norm": 0.16737351489552807, "learning_rate": 7.954212460665628e-05, "loss": 0.9601, "step": 1583 }, { "epoch": 0.717391304347826, "grad_norm": 0.2136641946357139, "learning_rate": 7.954021448882522e-05, "loss": 0.9871, "step": 1584 }, { "epoch": 0.7178442028985508, "grad_norm": 0.23673233420664577, "learning_rate": 7.953830041809625e-05, "loss": 0.9593, "step": 1585 }, { "epoch": 0.7182971014492754, "grad_norm": 0.25301120959675094, "learning_rate": 7.953638239466074e-05, "loss": 0.9599, "step": 1586 }, { "epoch": 0.71875, "grad_norm": 0.24202209926765064, "learning_rate": 7.953446041871044e-05, "loss": 0.9722, "step": 1587 }, { "epoch": 0.7192028985507246, "grad_norm": 0.22922122140601317, "learning_rate": 7.953253449043747e-05, "loss": 0.9605, "step": 1588 }, { "epoch": 0.7196557971014492, "grad_norm": 0.2152577203709274, "learning_rate": 7.95306046100344e-05, "loss": 0.9656, "step": 1589 }, { "epoch": 0.720108695652174, "grad_norm": 0.24259130342829094, "learning_rate": 7.952867077769412e-05, "loss": 1.001, "step": 1590 }, { "epoch": 0.7205615942028986, "grad_norm": 0.159344051009704, "learning_rate": 7.952673299360999e-05, "loss": 0.9839, "step": 1591 }, { "epoch": 0.7210144927536232, "grad_norm": 0.13979402073480007, "learning_rate": 7.952479125797575e-05, "loss": 0.9593, "step": 1592 }, { "epoch": 0.7214673913043478, "grad_norm": 0.1574512539446016, "learning_rate": 7.952284557098547e-05, "loss": 0.9845, "step": 1593 }, { "epoch": 0.7219202898550725, "grad_norm": 0.16669215329268264, "learning_rate": 7.952089593283371e-05, "loss": 0.9687, "step": 1594 }, { "epoch": 0.7223731884057971, "grad_norm": 0.14748874024496322, "learning_rate": 7.951894234371534e-05, "loss": 0.9773, "step": 1595 }, { "epoch": 0.7228260869565217, "grad_norm": 0.1536173323519664, "learning_rate": 7.951698480382569e-05, "loss": 0.9672, "step": 1596 }, { "epoch": 0.7232789855072463, "grad_norm": 0.15730849444060233, "learning_rate": 7.951502331336044e-05, "loss": 0.9712, "step": 1597 }, { "epoch": 0.7237318840579711, "grad_norm": 0.1734317430174259, "learning_rate": 7.95130578725157e-05, "loss": 0.9878, "step": 1598 }, { "epoch": 0.7241847826086957, "grad_norm": 0.18835811847382092, "learning_rate": 7.951108848148794e-05, "loss": 0.9821, "step": 1599 }, { "epoch": 0.7246376811594203, "grad_norm": 0.1771948625494446, "learning_rate": 7.950911514047407e-05, "loss": 0.9635, "step": 1600 }, { "epoch": 0.7250905797101449, "grad_norm": 0.1863669354773521, "learning_rate": 7.950713784967134e-05, "loss": 0.9763, "step": 1601 }, { "epoch": 0.7255434782608695, "grad_norm": 0.14695415611092774, "learning_rate": 7.950515660927745e-05, "loss": 0.9791, "step": 1602 }, { "epoch": 0.7259963768115942, "grad_norm": 0.11881772882395777, "learning_rate": 7.950317141949043e-05, "loss": 0.9827, "step": 1603 }, { "epoch": 0.7264492753623188, "grad_norm": 0.5700528913846858, "learning_rate": 7.950118228050879e-05, "loss": 0.9825, "step": 1604 }, { "epoch": 0.7269021739130435, "grad_norm": 0.15352598761043745, "learning_rate": 7.949918919253134e-05, "loss": 0.9616, "step": 1605 }, { "epoch": 0.7273550724637681, "grad_norm": 0.15531491515178464, "learning_rate": 7.949719215575737e-05, "loss": 0.9877, "step": 1606 }, { "epoch": 0.7278079710144928, "grad_norm": 0.15903233052917015, "learning_rate": 7.94951911703865e-05, "loss": 0.9946, "step": 1607 }, { "epoch": 0.7282608695652174, "grad_norm": 0.18354400123936557, "learning_rate": 7.94931862366188e-05, "loss": 0.9824, "step": 1608 }, { "epoch": 0.728713768115942, "grad_norm": 0.19483917171798343, "learning_rate": 7.949117735465467e-05, "loss": 0.9578, "step": 1609 }, { "epoch": 0.7291666666666666, "grad_norm": 0.17706813888826925, "learning_rate": 7.948916452469497e-05, "loss": 0.9611, "step": 1610 }, { "epoch": 0.7296195652173914, "grad_norm": 0.14883547373484496, "learning_rate": 7.948714774694092e-05, "loss": 0.9556, "step": 1611 }, { "epoch": 0.730072463768116, "grad_norm": 0.16200818548125154, "learning_rate": 7.948512702159413e-05, "loss": 1.0036, "step": 1612 }, { "epoch": 0.7305253623188406, "grad_norm": 0.1885322131505913, "learning_rate": 7.948310234885661e-05, "loss": 0.9773, "step": 1613 }, { "epoch": 0.7309782608695652, "grad_norm": 0.20205366173116907, "learning_rate": 7.94810737289308e-05, "loss": 0.9814, "step": 1614 }, { "epoch": 0.7314311594202898, "grad_norm": 0.12182688631678822, "learning_rate": 7.94790411620195e-05, "loss": 0.9781, "step": 1615 }, { "epoch": 0.7318840579710145, "grad_norm": 0.13690507311260636, "learning_rate": 7.947700464832587e-05, "loss": 0.9509, "step": 1616 }, { "epoch": 0.7323369565217391, "grad_norm": 0.17251064479468323, "learning_rate": 7.947496418805353e-05, "loss": 0.9727, "step": 1617 }, { "epoch": 0.7327898550724637, "grad_norm": 0.1823699426619357, "learning_rate": 7.947291978140648e-05, "loss": 0.9611, "step": 1618 }, { "epoch": 0.7332427536231884, "grad_norm": 0.17558823350897793, "learning_rate": 7.947087142858909e-05, "loss": 0.9664, "step": 1619 }, { "epoch": 0.7336956521739131, "grad_norm": 0.21382917771738366, "learning_rate": 7.946881912980612e-05, "loss": 0.975, "step": 1620 }, { "epoch": 0.7341485507246377, "grad_norm": 0.26213404514693056, "learning_rate": 7.946676288526276e-05, "loss": 0.9994, "step": 1621 }, { "epoch": 0.7346014492753623, "grad_norm": 0.25729602390387546, "learning_rate": 7.94647026951646e-05, "loss": 0.9764, "step": 1622 }, { "epoch": 0.7350543478260869, "grad_norm": 0.2578496015629996, "learning_rate": 7.946263855971755e-05, "loss": 0.9788, "step": 1623 }, { "epoch": 0.7355072463768116, "grad_norm": 0.23766835692141183, "learning_rate": 7.946057047912798e-05, "loss": 0.9766, "step": 1624 }, { "epoch": 0.7359601449275363, "grad_norm": 0.22092481636808398, "learning_rate": 7.945849845360266e-05, "loss": 0.9639, "step": 1625 }, { "epoch": 0.7364130434782609, "grad_norm": 0.17401297883848488, "learning_rate": 7.945642248334873e-05, "loss": 0.9809, "step": 1626 }, { "epoch": 0.7368659420289855, "grad_norm": 0.13072050224705672, "learning_rate": 7.945434256857372e-05, "loss": 1.0033, "step": 1627 }, { "epoch": 0.7373188405797102, "grad_norm": 0.13863464296974737, "learning_rate": 7.945225870948555e-05, "loss": 0.9739, "step": 1628 }, { "epoch": 0.7377717391304348, "grad_norm": 0.1436772522543605, "learning_rate": 7.945017090629257e-05, "loss": 0.9614, "step": 1629 }, { "epoch": 0.7382246376811594, "grad_norm": 0.15688757036818402, "learning_rate": 7.944807915920349e-05, "loss": 0.988, "step": 1630 }, { "epoch": 0.738677536231884, "grad_norm": 0.19267627193503734, "learning_rate": 7.944598346842743e-05, "loss": 0.9879, "step": 1631 }, { "epoch": 0.7391304347826086, "grad_norm": 0.18133263942898517, "learning_rate": 7.944388383417388e-05, "loss": 0.967, "step": 1632 }, { "epoch": 0.7395833333333334, "grad_norm": 0.17096739131356758, "learning_rate": 7.944178025665277e-05, "loss": 0.9748, "step": 1633 }, { "epoch": 0.740036231884058, "grad_norm": 0.20303606758206047, "learning_rate": 7.943967273607439e-05, "loss": 0.9555, "step": 1634 }, { "epoch": 0.7404891304347826, "grad_norm": 0.22581373732618693, "learning_rate": 7.943756127264942e-05, "loss": 0.9686, "step": 1635 }, { "epoch": 0.7409420289855072, "grad_norm": 0.22135164637495527, "learning_rate": 7.943544586658896e-05, "loss": 0.9744, "step": 1636 }, { "epoch": 0.7413949275362319, "grad_norm": 0.13657807622675194, "learning_rate": 7.943332651810448e-05, "loss": 0.9511, "step": 1637 }, { "epoch": 0.7418478260869565, "grad_norm": 0.1416318714041975, "learning_rate": 7.943120322740786e-05, "loss": 0.9711, "step": 1638 }, { "epoch": 0.7423007246376812, "grad_norm": 0.12643303507939085, "learning_rate": 7.942907599471138e-05, "loss": 0.9708, "step": 1639 }, { "epoch": 0.7427536231884058, "grad_norm": 0.14741543554964456, "learning_rate": 7.942694482022768e-05, "loss": 0.9832, "step": 1640 }, { "epoch": 0.7432065217391305, "grad_norm": 0.1563610761400183, "learning_rate": 7.942480970416985e-05, "loss": 0.9809, "step": 1641 }, { "epoch": 0.7436594202898551, "grad_norm": 0.156139504735042, "learning_rate": 7.94226706467513e-05, "loss": 1.0066, "step": 1642 }, { "epoch": 0.7441123188405797, "grad_norm": 0.15391336860832538, "learning_rate": 7.94205276481859e-05, "loss": 0.9851, "step": 1643 }, { "epoch": 0.7445652173913043, "grad_norm": 0.16225382122812895, "learning_rate": 7.941838070868787e-05, "loss": 0.9729, "step": 1644 }, { "epoch": 0.7450181159420289, "grad_norm": 0.17480971405808707, "learning_rate": 7.941622982847189e-05, "loss": 0.9995, "step": 1645 }, { "epoch": 0.7454710144927537, "grad_norm": 0.1623281041323118, "learning_rate": 7.941407500775292e-05, "loss": 0.9747, "step": 1646 }, { "epoch": 0.7459239130434783, "grad_norm": 0.19035642482034612, "learning_rate": 7.941191624674643e-05, "loss": 0.9405, "step": 1647 }, { "epoch": 0.7463768115942029, "grad_norm": 0.20326522495968902, "learning_rate": 7.940975354566823e-05, "loss": 0.9869, "step": 1648 }, { "epoch": 0.7468297101449275, "grad_norm": 0.2152633494640094, "learning_rate": 7.940758690473451e-05, "loss": 0.9836, "step": 1649 }, { "epoch": 0.7472826086956522, "grad_norm": 0.22003359789732743, "learning_rate": 7.940541632416187e-05, "loss": 0.9643, "step": 1650 }, { "epoch": 0.7477355072463768, "grad_norm": 0.18999536817960747, "learning_rate": 7.940324180416732e-05, "loss": 0.9822, "step": 1651 }, { "epoch": 0.7481884057971014, "grad_norm": 0.21571319136473585, "learning_rate": 7.940106334496825e-05, "loss": 0.9688, "step": 1652 }, { "epoch": 0.748641304347826, "grad_norm": 0.2492684721138993, "learning_rate": 7.939888094678244e-05, "loss": 0.9884, "step": 1653 }, { "epoch": 0.7490942028985508, "grad_norm": 0.30935899899191655, "learning_rate": 7.939669460982807e-05, "loss": 0.9778, "step": 1654 }, { "epoch": 0.7495471014492754, "grad_norm": 0.35860573994239187, "learning_rate": 7.939450433432371e-05, "loss": 0.9691, "step": 1655 }, { "epoch": 0.75, "grad_norm": 0.34460734841109103, "learning_rate": 7.939231012048833e-05, "loss": 0.9556, "step": 1656 }, { "epoch": 0.7504528985507246, "grad_norm": 0.2645738414072756, "learning_rate": 7.939011196854129e-05, "loss": 0.9721, "step": 1657 }, { "epoch": 0.7509057971014492, "grad_norm": 0.23429574233117856, "learning_rate": 7.938790987870232e-05, "loss": 0.9575, "step": 1658 }, { "epoch": 0.751358695652174, "grad_norm": 0.21427696435873175, "learning_rate": 7.938570385119159e-05, "loss": 0.9998, "step": 1659 }, { "epoch": 0.7518115942028986, "grad_norm": 0.18543990361265528, "learning_rate": 7.938349388622964e-05, "loss": 0.9667, "step": 1660 }, { "epoch": 0.7522644927536232, "grad_norm": 0.16303061856331805, "learning_rate": 7.938127998403738e-05, "loss": 0.9437, "step": 1661 }, { "epoch": 0.7527173913043478, "grad_norm": 0.18799229622216673, "learning_rate": 7.937906214483618e-05, "loss": 0.9833, "step": 1662 }, { "epoch": 0.7531702898550725, "grad_norm": 0.17933672521184824, "learning_rate": 7.937684036884772e-05, "loss": 0.9753, "step": 1663 }, { "epoch": 0.7536231884057971, "grad_norm": 0.1573143115845205, "learning_rate": 7.937461465629414e-05, "loss": 0.98, "step": 1664 }, { "epoch": 0.7540760869565217, "grad_norm": 0.15462081224123736, "learning_rate": 7.937238500739794e-05, "loss": 0.9512, "step": 1665 }, { "epoch": 0.7545289855072463, "grad_norm": 0.16281275346390803, "learning_rate": 7.9370151422382e-05, "loss": 0.9632, "step": 1666 }, { "epoch": 0.7549818840579711, "grad_norm": 0.1361401465568143, "learning_rate": 7.936791390146965e-05, "loss": 0.9906, "step": 1667 }, { "epoch": 0.7554347826086957, "grad_norm": 0.14111200800575815, "learning_rate": 7.936567244488458e-05, "loss": 0.9648, "step": 1668 }, { "epoch": 0.7558876811594203, "grad_norm": 0.12354975319868478, "learning_rate": 7.936342705285084e-05, "loss": 0.9997, "step": 1669 }, { "epoch": 0.7563405797101449, "grad_norm": 0.11343589564333612, "learning_rate": 7.936117772559291e-05, "loss": 0.9631, "step": 1670 }, { "epoch": 0.7567934782608695, "grad_norm": 0.11276367492132186, "learning_rate": 7.935892446333568e-05, "loss": 0.9834, "step": 1671 }, { "epoch": 0.7572463768115942, "grad_norm": 0.13502477148507824, "learning_rate": 7.935666726630439e-05, "loss": 0.9761, "step": 1672 }, { "epoch": 0.7576992753623188, "grad_norm": 0.1258049619162128, "learning_rate": 7.935440613472473e-05, "loss": 0.9712, "step": 1673 }, { "epoch": 0.7581521739130435, "grad_norm": 0.14990850975967365, "learning_rate": 7.935214106882272e-05, "loss": 0.9707, "step": 1674 }, { "epoch": 0.7586050724637681, "grad_norm": 0.14322116901714174, "learning_rate": 7.934987206882479e-05, "loss": 0.9804, "step": 1675 }, { "epoch": 0.7590579710144928, "grad_norm": 0.1290908660257245, "learning_rate": 7.93475991349578e-05, "loss": 0.9882, "step": 1676 }, { "epoch": 0.7595108695652174, "grad_norm": 0.1467774596924648, "learning_rate": 7.9345322267449e-05, "loss": 0.9774, "step": 1677 }, { "epoch": 0.759963768115942, "grad_norm": 0.15602637689862484, "learning_rate": 7.934304146652596e-05, "loss": 0.9876, "step": 1678 }, { "epoch": 0.7604166666666666, "grad_norm": 0.17420362854271568, "learning_rate": 7.934075673241672e-05, "loss": 0.9684, "step": 1679 }, { "epoch": 0.7608695652173914, "grad_norm": 0.15339588117013478, "learning_rate": 7.93384680653497e-05, "loss": 0.963, "step": 1680 }, { "epoch": 0.761322463768116, "grad_norm": 0.17267342828438256, "learning_rate": 7.933617546555367e-05, "loss": 0.9874, "step": 1681 }, { "epoch": 0.7617753623188406, "grad_norm": 0.13204213366041845, "learning_rate": 7.933387893325786e-05, "loss": 0.9665, "step": 1682 }, { "epoch": 0.7622282608695652, "grad_norm": 0.15111755433129814, "learning_rate": 7.933157846869184e-05, "loss": 0.9763, "step": 1683 }, { "epoch": 0.7626811594202898, "grad_norm": 0.1376863955743325, "learning_rate": 7.932927407208561e-05, "loss": 0.9527, "step": 1684 }, { "epoch": 0.7631340579710145, "grad_norm": 0.10864604578926454, "learning_rate": 7.932696574366952e-05, "loss": 0.9691, "step": 1685 }, { "epoch": 0.7635869565217391, "grad_norm": 0.11990603209671315, "learning_rate": 7.932465348367432e-05, "loss": 0.9493, "step": 1686 }, { "epoch": 0.7640398550724637, "grad_norm": 0.11450668699843018, "learning_rate": 7.932233729233123e-05, "loss": 0.9717, "step": 1687 }, { "epoch": 0.7644927536231884, "grad_norm": 0.13716592491314283, "learning_rate": 7.932001716987175e-05, "loss": 0.9922, "step": 1688 }, { "epoch": 0.7649456521739131, "grad_norm": 0.17368384848317037, "learning_rate": 7.931769311652785e-05, "loss": 0.9472, "step": 1689 }, { "epoch": 0.7653985507246377, "grad_norm": 0.15324324926928942, "learning_rate": 7.931536513253186e-05, "loss": 0.9464, "step": 1690 }, { "epoch": 0.7658514492753623, "grad_norm": 0.11709138833195915, "learning_rate": 7.931303321811654e-05, "loss": 0.9764, "step": 1691 }, { "epoch": 0.7663043478260869, "grad_norm": 0.1465161112088723, "learning_rate": 7.931069737351497e-05, "loss": 0.9952, "step": 1692 }, { "epoch": 0.7667572463768116, "grad_norm": 0.14121122639471545, "learning_rate": 7.930835759896069e-05, "loss": 0.9727, "step": 1693 }, { "epoch": 0.7672101449275363, "grad_norm": 0.1362901223240435, "learning_rate": 7.930601389468761e-05, "loss": 0.973, "step": 1694 }, { "epoch": 0.7676630434782609, "grad_norm": 0.15840556666836603, "learning_rate": 7.930366626093005e-05, "loss": 0.9373, "step": 1695 }, { "epoch": 0.7681159420289855, "grad_norm": 0.17216009571095256, "learning_rate": 7.930131469792269e-05, "loss": 0.9772, "step": 1696 }, { "epoch": 0.7685688405797102, "grad_norm": 0.17159437066475422, "learning_rate": 7.92989592059006e-05, "loss": 0.9634, "step": 1697 }, { "epoch": 0.7690217391304348, "grad_norm": 0.16070313882038278, "learning_rate": 7.929659978509931e-05, "loss": 0.9756, "step": 1698 }, { "epoch": 0.7694746376811594, "grad_norm": 0.18773810206894473, "learning_rate": 7.929423643575466e-05, "loss": 0.9592, "step": 1699 }, { "epoch": 0.769927536231884, "grad_norm": 0.22477773944369195, "learning_rate": 7.929186915810292e-05, "loss": 0.9577, "step": 1700 }, { "epoch": 0.7703804347826086, "grad_norm": 0.2036484823705362, "learning_rate": 7.928949795238076e-05, "loss": 0.9651, "step": 1701 }, { "epoch": 0.7708333333333334, "grad_norm": 0.1877149635034864, "learning_rate": 7.928712281882523e-05, "loss": 0.9535, "step": 1702 }, { "epoch": 0.771286231884058, "grad_norm": 0.19592281170654052, "learning_rate": 7.928474375767378e-05, "loss": 0.9919, "step": 1703 }, { "epoch": 0.7717391304347826, "grad_norm": 0.18181886163248803, "learning_rate": 7.928236076916423e-05, "loss": 0.9799, "step": 1704 }, { "epoch": 0.7721920289855072, "grad_norm": 0.15579937743210923, "learning_rate": 7.927997385353485e-05, "loss": 0.9517, "step": 1705 }, { "epoch": 0.7726449275362319, "grad_norm": 0.16082159042411118, "learning_rate": 7.927758301102423e-05, "loss": 0.9846, "step": 1706 }, { "epoch": 0.7730978260869565, "grad_norm": 0.18753375987750248, "learning_rate": 7.927518824187139e-05, "loss": 0.9549, "step": 1707 }, { "epoch": 0.7735507246376812, "grad_norm": 0.2092927900977582, "learning_rate": 7.927278954631576e-05, "loss": 0.9715, "step": 1708 }, { "epoch": 0.7740036231884058, "grad_norm": 0.2812326253009748, "learning_rate": 7.927038692459712e-05, "loss": 0.9904, "step": 1709 }, { "epoch": 0.7744565217391305, "grad_norm": 0.37211034392071757, "learning_rate": 7.926798037695564e-05, "loss": 0.9821, "step": 1710 }, { "epoch": 0.7749094202898551, "grad_norm": 0.4257104035394328, "learning_rate": 7.926556990363196e-05, "loss": 0.9694, "step": 1711 }, { "epoch": 0.7753623188405797, "grad_norm": 0.38386400644339386, "learning_rate": 7.926315550486704e-05, "loss": 0.9697, "step": 1712 }, { "epoch": 0.7758152173913043, "grad_norm": 0.33247432778952807, "learning_rate": 7.926073718090226e-05, "loss": 0.9607, "step": 1713 }, { "epoch": 0.7762681159420289, "grad_norm": 0.2639198099614502, "learning_rate": 7.925831493197933e-05, "loss": 0.9511, "step": 1714 }, { "epoch": 0.7767210144927537, "grad_norm": 0.1803857804282703, "learning_rate": 7.925588875834047e-05, "loss": 0.9761, "step": 1715 }, { "epoch": 0.7771739130434783, "grad_norm": 0.13619901141250546, "learning_rate": 7.92534586602282e-05, "loss": 0.9787, "step": 1716 }, { "epoch": 0.7776268115942029, "grad_norm": 0.14855128941519344, "learning_rate": 7.925102463788547e-05, "loss": 0.9752, "step": 1717 }, { "epoch": 0.7780797101449275, "grad_norm": 0.1707301003601964, "learning_rate": 7.92485866915556e-05, "loss": 0.9733, "step": 1718 }, { "epoch": 0.7785326086956522, "grad_norm": 0.16573773565772787, "learning_rate": 7.924614482148233e-05, "loss": 0.9535, "step": 1719 }, { "epoch": 0.7789855072463768, "grad_norm": 0.1464146225291217, "learning_rate": 7.924369902790978e-05, "loss": 0.9929, "step": 1720 }, { "epoch": 0.7794384057971014, "grad_norm": 0.10946884418120022, "learning_rate": 7.924124931108245e-05, "loss": 0.9572, "step": 1721 }, { "epoch": 0.779891304347826, "grad_norm": 0.14979214288773784, "learning_rate": 7.923879567124523e-05, "loss": 0.9798, "step": 1722 }, { "epoch": 0.7803442028985508, "grad_norm": 0.158034145108402, "learning_rate": 7.923633810864342e-05, "loss": 0.9505, "step": 1723 }, { "epoch": 0.7807971014492754, "grad_norm": 0.14551952062977663, "learning_rate": 7.923387662352273e-05, "loss": 0.9553, "step": 1724 }, { "epoch": 0.78125, "grad_norm": 0.2921592571747958, "learning_rate": 7.923141121612922e-05, "loss": 0.9689, "step": 1725 }, { "epoch": 0.7817028985507246, "grad_norm": 0.17169550814087328, "learning_rate": 7.922894188670937e-05, "loss": 0.9488, "step": 1726 }, { "epoch": 0.7821557971014492, "grad_norm": 0.18976838011295236, "learning_rate": 7.922646863551003e-05, "loss": 0.9798, "step": 1727 }, { "epoch": 0.782608695652174, "grad_norm": 0.1966573547828347, "learning_rate": 7.922399146277846e-05, "loss": 0.9544, "step": 1728 }, { "epoch": 0.7830615942028986, "grad_norm": 0.2211235234880721, "learning_rate": 7.92215103687623e-05, "loss": 0.9807, "step": 1729 }, { "epoch": 0.7835144927536232, "grad_norm": 0.26172003813514705, "learning_rate": 7.92190253537096e-05, "loss": 0.9699, "step": 1730 }, { "epoch": 0.7839673913043478, "grad_norm": 0.26635166033707575, "learning_rate": 7.921653641786879e-05, "loss": 0.9789, "step": 1731 }, { "epoch": 0.7844202898550725, "grad_norm": 0.22290925668629322, "learning_rate": 7.92140435614887e-05, "loss": 0.9704, "step": 1732 }, { "epoch": 0.7848731884057971, "grad_norm": 0.18455968560605773, "learning_rate": 7.921154678481852e-05, "loss": 0.9378, "step": 1733 }, { "epoch": 0.7853260869565217, "grad_norm": 0.19061158915224904, "learning_rate": 7.920904608810788e-05, "loss": 0.9613, "step": 1734 }, { "epoch": 0.7857789855072463, "grad_norm": 0.1840774103170461, "learning_rate": 7.920654147160678e-05, "loss": 0.9778, "step": 1735 }, { "epoch": 0.7862318840579711, "grad_norm": 0.19937603631135745, "learning_rate": 7.920403293556558e-05, "loss": 0.9618, "step": 1736 }, { "epoch": 0.7866847826086957, "grad_norm": 0.2277783265134647, "learning_rate": 7.920152048023509e-05, "loss": 0.9856, "step": 1737 }, { "epoch": 0.7871376811594203, "grad_norm": 0.20962198548742586, "learning_rate": 7.919900410586648e-05, "loss": 0.9729, "step": 1738 }, { "epoch": 0.7875905797101449, "grad_norm": 0.20817224718561794, "learning_rate": 7.919648381271132e-05, "loss": 0.969, "step": 1739 }, { "epoch": 0.7880434782608695, "grad_norm": 0.25511325325065415, "learning_rate": 7.919395960102155e-05, "loss": 0.97, "step": 1740 }, { "epoch": 0.7884963768115942, "grad_norm": 0.2486998046180534, "learning_rate": 7.919143147104954e-05, "loss": 0.9794, "step": 1741 }, { "epoch": 0.7889492753623188, "grad_norm": 0.18361918217256143, "learning_rate": 7.918889942304802e-05, "loss": 0.9595, "step": 1742 }, { "epoch": 0.7894021739130435, "grad_norm": 0.13974975829350614, "learning_rate": 7.918636345727012e-05, "loss": 0.9746, "step": 1743 }, { "epoch": 0.7898550724637681, "grad_norm": 0.13429550912796973, "learning_rate": 7.918382357396937e-05, "loss": 0.9762, "step": 1744 }, { "epoch": 0.7903079710144928, "grad_norm": 0.11906771895857551, "learning_rate": 7.918127977339969e-05, "loss": 0.9579, "step": 1745 }, { "epoch": 0.7907608695652174, "grad_norm": 0.14427340940076805, "learning_rate": 7.917873205581539e-05, "loss": 0.9732, "step": 1746 }, { "epoch": 0.791213768115942, "grad_norm": 0.16020216836033335, "learning_rate": 7.917618042147114e-05, "loss": 0.9429, "step": 1747 }, { "epoch": 0.7916666666666666, "grad_norm": 0.1705276499597196, "learning_rate": 7.917362487062207e-05, "loss": 0.9836, "step": 1748 }, { "epoch": 0.7921195652173914, "grad_norm": 0.19375031297155657, "learning_rate": 7.917106540352364e-05, "loss": 0.9698, "step": 1749 }, { "epoch": 0.792572463768116, "grad_norm": 0.19745813356347036, "learning_rate": 7.916850202043172e-05, "loss": 0.9597, "step": 1750 }, { "epoch": 0.7930253623188406, "grad_norm": 0.18686485722378493, "learning_rate": 7.91659347216026e-05, "loss": 0.9699, "step": 1751 }, { "epoch": 0.7934782608695652, "grad_norm": 0.1535264738703784, "learning_rate": 7.916336350729293e-05, "loss": 0.9957, "step": 1752 }, { "epoch": 0.7939311594202898, "grad_norm": 0.13729052775219955, "learning_rate": 7.916078837775973e-05, "loss": 0.9756, "step": 1753 }, { "epoch": 0.7943840579710145, "grad_norm": 0.13256894907174438, "learning_rate": 7.915820933326046e-05, "loss": 0.9931, "step": 1754 }, { "epoch": 0.7948369565217391, "grad_norm": 0.11498769867701597, "learning_rate": 7.915562637405296e-05, "loss": 1.0039, "step": 1755 }, { "epoch": 0.7952898550724637, "grad_norm": 0.1150231064192473, "learning_rate": 7.915303950039545e-05, "loss": 0.9804, "step": 1756 }, { "epoch": 0.7957427536231884, "grad_norm": 0.12968769789969994, "learning_rate": 7.915044871254653e-05, "loss": 0.9706, "step": 1757 }, { "epoch": 0.7961956521739131, "grad_norm": 0.1676627547648544, "learning_rate": 7.914785401076522e-05, "loss": 0.9698, "step": 1758 }, { "epoch": 0.7966485507246377, "grad_norm": 0.20671571289463603, "learning_rate": 7.91452553953109e-05, "loss": 0.9352, "step": 1759 }, { "epoch": 0.7971014492753623, "grad_norm": 0.2174522129296922, "learning_rate": 7.914265286644337e-05, "loss": 0.9516, "step": 1760 }, { "epoch": 0.7975543478260869, "grad_norm": 0.2530019161277246, "learning_rate": 7.914004642442281e-05, "loss": 0.9935, "step": 1761 }, { "epoch": 0.7980072463768116, "grad_norm": 0.25722221619625846, "learning_rate": 7.913743606950979e-05, "loss": 0.977, "step": 1762 }, { "epoch": 0.7984601449275363, "grad_norm": 0.22129392804170095, "learning_rate": 7.913482180196526e-05, "loss": 0.9769, "step": 1763 }, { "epoch": 0.7989130434782609, "grad_norm": 0.15994667296281834, "learning_rate": 7.913220362205059e-05, "loss": 0.9789, "step": 1764 }, { "epoch": 0.7993659420289855, "grad_norm": 0.116062447591797, "learning_rate": 7.91295815300275e-05, "loss": 0.9625, "step": 1765 }, { "epoch": 0.7998188405797102, "grad_norm": 0.16558374446029256, "learning_rate": 7.912695552615814e-05, "loss": 0.9737, "step": 1766 }, { "epoch": 0.8002717391304348, "grad_norm": 0.18153923181594775, "learning_rate": 7.912432561070503e-05, "loss": 0.9628, "step": 1767 }, { "epoch": 0.8007246376811594, "grad_norm": 0.20349490756694077, "learning_rate": 7.91216917839311e-05, "loss": 0.9493, "step": 1768 }, { "epoch": 0.801177536231884, "grad_norm": 0.18980790766369482, "learning_rate": 7.911905404609965e-05, "loss": 0.9869, "step": 1769 }, { "epoch": 0.8016304347826086, "grad_norm": 0.1355732795015463, "learning_rate": 7.911641239747438e-05, "loss": 0.9799, "step": 1770 }, { "epoch": 0.8020833333333334, "grad_norm": 0.13898661451673003, "learning_rate": 7.911376683831937e-05, "loss": 0.9738, "step": 1771 }, { "epoch": 0.802536231884058, "grad_norm": 0.1591478844245774, "learning_rate": 7.91111173688991e-05, "loss": 0.9698, "step": 1772 }, { "epoch": 0.8029891304347826, "grad_norm": 0.37271211956236316, "learning_rate": 7.910846398947847e-05, "loss": 0.9828, "step": 1773 }, { "epoch": 0.8034420289855072, "grad_norm": 0.1811925511405665, "learning_rate": 7.910580670032272e-05, "loss": 0.9791, "step": 1774 }, { "epoch": 0.8038949275362319, "grad_norm": 0.19188841426657444, "learning_rate": 7.910314550169749e-05, "loss": 0.9697, "step": 1775 }, { "epoch": 0.8043478260869565, "grad_norm": 0.21176662359558224, "learning_rate": 7.910048039386885e-05, "loss": 0.956, "step": 1776 }, { "epoch": 0.8048007246376812, "grad_norm": 0.213039489378937, "learning_rate": 7.909781137710322e-05, "loss": 0.9775, "step": 1777 }, { "epoch": 0.8052536231884058, "grad_norm": 0.18613086789343877, "learning_rate": 7.909513845166744e-05, "loss": 0.9712, "step": 1778 }, { "epoch": 0.8057065217391305, "grad_norm": 0.18510176944832446, "learning_rate": 7.90924616178287e-05, "loss": 0.9617, "step": 1779 }, { "epoch": 0.8061594202898551, "grad_norm": 0.17499062089193582, "learning_rate": 7.908978087585465e-05, "loss": 0.9918, "step": 1780 }, { "epoch": 0.8066123188405797, "grad_norm": 0.16319343114654594, "learning_rate": 7.908709622601324e-05, "loss": 0.9686, "step": 1781 }, { "epoch": 0.8070652173913043, "grad_norm": 0.16227579917612245, "learning_rate": 7.908440766857288e-05, "loss": 0.9736, "step": 1782 }, { "epoch": 0.8075181159420289, "grad_norm": 0.1429173450672748, "learning_rate": 7.908171520380236e-05, "loss": 0.9556, "step": 1783 }, { "epoch": 0.8079710144927537, "grad_norm": 0.16849985977165013, "learning_rate": 7.907901883197085e-05, "loss": 0.9728, "step": 1784 }, { "epoch": 0.8084239130434783, "grad_norm": 0.14593716619279282, "learning_rate": 7.907631855334788e-05, "loss": 1.002, "step": 1785 }, { "epoch": 0.8088768115942029, "grad_norm": 0.20470635572683737, "learning_rate": 7.907361436820345e-05, "loss": 0.9562, "step": 1786 }, { "epoch": 0.8093297101449275, "grad_norm": 0.2932545283747906, "learning_rate": 7.907090627680786e-05, "loss": 0.9506, "step": 1787 }, { "epoch": 0.8097826086956522, "grad_norm": 0.32159063481773265, "learning_rate": 7.906819427943185e-05, "loss": 0.9543, "step": 1788 }, { "epoch": 0.8102355072463768, "grad_norm": 0.32947576083863483, "learning_rate": 7.906547837634655e-05, "loss": 0.9772, "step": 1789 }, { "epoch": 0.8106884057971014, "grad_norm": 0.34243605351113965, "learning_rate": 7.906275856782347e-05, "loss": 0.9439, "step": 1790 }, { "epoch": 0.811141304347826, "grad_norm": 0.2986850352619435, "learning_rate": 7.906003485413454e-05, "loss": 0.9495, "step": 1791 }, { "epoch": 0.8115942028985508, "grad_norm": 0.2612562771984127, "learning_rate": 7.905730723555201e-05, "loss": 0.9768, "step": 1792 }, { "epoch": 0.8120471014492754, "grad_norm": 0.17117243592375958, "learning_rate": 7.905457571234858e-05, "loss": 0.9644, "step": 1793 }, { "epoch": 0.8125, "grad_norm": 0.10787356186836129, "learning_rate": 7.905184028479734e-05, "loss": 0.9938, "step": 1794 }, { "epoch": 0.8129528985507246, "grad_norm": 0.20412118327762224, "learning_rate": 7.904910095317175e-05, "loss": 0.9502, "step": 1795 }, { "epoch": 0.8134057971014492, "grad_norm": 0.266862997271615, "learning_rate": 7.904635771774564e-05, "loss": 0.9945, "step": 1796 }, { "epoch": 0.813858695652174, "grad_norm": 0.2805816493323791, "learning_rate": 7.904361057879329e-05, "loss": 0.9728, "step": 1797 }, { "epoch": 0.8143115942028986, "grad_norm": 0.2834453418631539, "learning_rate": 7.90408595365893e-05, "loss": 0.9702, "step": 1798 }, { "epoch": 0.8147644927536232, "grad_norm": 0.26029809420044236, "learning_rate": 7.903810459140873e-05, "loss": 0.9857, "step": 1799 }, { "epoch": 0.8152173913043478, "grad_norm": 0.18090417796170313, "learning_rate": 7.903534574352698e-05, "loss": 0.9709, "step": 1800 }, { "epoch": 0.8156702898550725, "grad_norm": 0.12333772116016042, "learning_rate": 7.903258299321987e-05, "loss": 0.9864, "step": 1801 }, { "epoch": 0.8161231884057971, "grad_norm": 0.20151957165245543, "learning_rate": 7.902981634076357e-05, "loss": 0.9785, "step": 1802 }, { "epoch": 0.8165760869565217, "grad_norm": 0.2814255749475942, "learning_rate": 7.90270457864347e-05, "loss": 0.9833, "step": 1803 }, { "epoch": 0.8170289855072463, "grad_norm": 0.2864005618873278, "learning_rate": 7.90242713305102e-05, "loss": 0.9694, "step": 1804 }, { "epoch": 0.8174818840579711, "grad_norm": 0.20199087338066857, "learning_rate": 7.902149297326746e-05, "loss": 0.9736, "step": 1805 }, { "epoch": 0.8179347826086957, "grad_norm": 0.13740977887217387, "learning_rate": 7.901871071498425e-05, "loss": 0.965, "step": 1806 }, { "epoch": 0.8183876811594203, "grad_norm": 0.13609340130409328, "learning_rate": 7.901592455593869e-05, "loss": 0.9614, "step": 1807 }, { "epoch": 0.8188405797101449, "grad_norm": 0.15897845091359653, "learning_rate": 7.901313449640931e-05, "loss": 0.9638, "step": 1808 }, { "epoch": 0.8192934782608695, "grad_norm": 0.1656225099413582, "learning_rate": 7.901034053667508e-05, "loss": 0.9639, "step": 1809 }, { "epoch": 0.8197463768115942, "grad_norm": 0.17418333267134406, "learning_rate": 7.900754267701526e-05, "loss": 0.9717, "step": 1810 }, { "epoch": 0.8201992753623188, "grad_norm": 0.18721830306068843, "learning_rate": 7.900474091770961e-05, "loss": 0.994, "step": 1811 }, { "epoch": 0.8206521739130435, "grad_norm": 0.22605881874212097, "learning_rate": 7.90019352590382e-05, "loss": 0.9987, "step": 1812 }, { "epoch": 0.8211050724637681, "grad_norm": 0.1484596597525255, "learning_rate": 7.89991257012815e-05, "loss": 0.9557, "step": 1813 }, { "epoch": 0.8215579710144928, "grad_norm": 0.13341457318014127, "learning_rate": 7.899631224472042e-05, "loss": 0.9817, "step": 1814 }, { "epoch": 0.8220108695652174, "grad_norm": 0.14741081163397088, "learning_rate": 7.899349488963621e-05, "loss": 0.9582, "step": 1815 }, { "epoch": 0.822463768115942, "grad_norm": 0.11513848379166819, "learning_rate": 7.899067363631051e-05, "loss": 0.9645, "step": 1816 }, { "epoch": 0.8229166666666666, "grad_norm": 0.13241983465799786, "learning_rate": 7.89878484850254e-05, "loss": 0.9689, "step": 1817 }, { "epoch": 0.8233695652173914, "grad_norm": 0.18817168788645955, "learning_rate": 7.898501943606329e-05, "loss": 0.9453, "step": 1818 }, { "epoch": 0.823822463768116, "grad_norm": 0.20912687043381153, "learning_rate": 7.8982186489707e-05, "loss": 0.9508, "step": 1819 }, { "epoch": 0.8242753623188406, "grad_norm": 0.20574152624434056, "learning_rate": 7.897934964623977e-05, "loss": 0.9808, "step": 1820 }, { "epoch": 0.8247282608695652, "grad_norm": 0.14913446806886918, "learning_rate": 7.897650890594517e-05, "loss": 0.9479, "step": 1821 }, { "epoch": 0.8251811594202898, "grad_norm": 0.1283046617914809, "learning_rate": 7.897366426910723e-05, "loss": 0.9573, "step": 1822 }, { "epoch": 0.8256340579710145, "grad_norm": 0.10614485081222999, "learning_rate": 7.89708157360103e-05, "loss": 0.9513, "step": 1823 }, { "epoch": 0.8260869565217391, "grad_norm": 0.14509228119898712, "learning_rate": 7.896796330693917e-05, "loss": 0.9412, "step": 1824 }, { "epoch": 0.8265398550724637, "grad_norm": 0.1907688513068941, "learning_rate": 7.896510698217899e-05, "loss": 0.9709, "step": 1825 }, { "epoch": 0.8269927536231884, "grad_norm": 0.19410001249123712, "learning_rate": 7.896224676201533e-05, "loss": 0.9683, "step": 1826 }, { "epoch": 0.8274456521739131, "grad_norm": 0.20941957941295883, "learning_rate": 7.895938264673412e-05, "loss": 0.9668, "step": 1827 }, { "epoch": 0.8278985507246377, "grad_norm": 0.2035520222030403, "learning_rate": 7.89565146366217e-05, "loss": 0.9794, "step": 1828 }, { "epoch": 0.8283514492753623, "grad_norm": 0.17418202131872146, "learning_rate": 7.895364273196475e-05, "loss": 0.9521, "step": 1829 }, { "epoch": 0.8288043478260869, "grad_norm": 0.12429503322987913, "learning_rate": 7.895076693305044e-05, "loss": 0.982, "step": 1830 }, { "epoch": 0.8292572463768116, "grad_norm": 0.10704186284372527, "learning_rate": 7.894788724016621e-05, "loss": 0.9447, "step": 1831 }, { "epoch": 0.8297101449275363, "grad_norm": 0.128071743371891, "learning_rate": 7.89450036536e-05, "loss": 0.982, "step": 1832 }, { "epoch": 0.8301630434782609, "grad_norm": 0.12089950991521832, "learning_rate": 7.894211617364005e-05, "loss": 0.9723, "step": 1833 }, { "epoch": 0.8306159420289855, "grad_norm": 0.13575878091125165, "learning_rate": 7.893922480057503e-05, "loss": 0.9751, "step": 1834 }, { "epoch": 0.8310688405797102, "grad_norm": 0.11460220140587359, "learning_rate": 7.893632953469401e-05, "loss": 0.9656, "step": 1835 }, { "epoch": 0.8315217391304348, "grad_norm": 0.12236007396353489, "learning_rate": 7.893343037628643e-05, "loss": 0.9677, "step": 1836 }, { "epoch": 0.8319746376811594, "grad_norm": 0.119169768910706, "learning_rate": 7.893052732564211e-05, "loss": 0.9691, "step": 1837 }, { "epoch": 0.832427536231884, "grad_norm": 0.14870606337892003, "learning_rate": 7.89276203830513e-05, "loss": 0.9652, "step": 1838 }, { "epoch": 0.8328804347826086, "grad_norm": 0.13635679887628324, "learning_rate": 7.892470954880458e-05, "loss": 0.9842, "step": 1839 }, { "epoch": 0.8333333333333334, "grad_norm": 0.11297993003919497, "learning_rate": 7.892179482319297e-05, "loss": 0.964, "step": 1840 }, { "epoch": 0.833786231884058, "grad_norm": 0.11029746592947288, "learning_rate": 7.891887620650784e-05, "loss": 0.9702, "step": 1841 }, { "epoch": 0.8342391304347826, "grad_norm": 0.13792379968243537, "learning_rate": 7.8915953699041e-05, "loss": 0.9726, "step": 1842 }, { "epoch": 0.8346920289855072, "grad_norm": 0.3460600479673167, "learning_rate": 7.891302730108459e-05, "loss": 0.9845, "step": 1843 }, { "epoch": 0.8351449275362319, "grad_norm": 0.17948995597353692, "learning_rate": 7.891009701293119e-05, "loss": 0.9837, "step": 1844 }, { "epoch": 0.8355978260869565, "grad_norm": 0.14992797373811997, "learning_rate": 7.890716283487373e-05, "loss": 0.9719, "step": 1845 }, { "epoch": 0.8360507246376812, "grad_norm": 0.201354843641337, "learning_rate": 7.890422476720554e-05, "loss": 0.9984, "step": 1846 }, { "epoch": 0.8365036231884058, "grad_norm": 0.1877539068426676, "learning_rate": 7.890128281022036e-05, "loss": 0.9644, "step": 1847 }, { "epoch": 0.8369565217391305, "grad_norm": 0.12079255695019787, "learning_rate": 7.889833696421229e-05, "loss": 0.9775, "step": 1848 }, { "epoch": 0.8374094202898551, "grad_norm": 0.13378422487756428, "learning_rate": 7.889538722947583e-05, "loss": 0.9606, "step": 1849 }, { "epoch": 0.8378623188405797, "grad_norm": 0.18509072083965925, "learning_rate": 7.889243360630587e-05, "loss": 0.9733, "step": 1850 }, { "epoch": 0.8383152173913043, "grad_norm": 0.16467658656206483, "learning_rate": 7.88894760949977e-05, "loss": 0.9677, "step": 1851 }, { "epoch": 0.8387681159420289, "grad_norm": 0.13959194933740326, "learning_rate": 7.888651469584698e-05, "loss": 0.9702, "step": 1852 }, { "epoch": 0.8392210144927537, "grad_norm": 0.16316072415781882, "learning_rate": 7.888354940914976e-05, "loss": 0.963, "step": 1853 }, { "epoch": 0.8396739130434783, "grad_norm": 0.1902935870885702, "learning_rate": 7.888058023520249e-05, "loss": 0.9758, "step": 1854 }, { "epoch": 0.8401268115942029, "grad_norm": 0.21156473371439163, "learning_rate": 7.8877607174302e-05, "loss": 0.9829, "step": 1855 }, { "epoch": 0.8405797101449275, "grad_norm": 0.24313428170641818, "learning_rate": 7.887463022674552e-05, "loss": 0.9607, "step": 1856 }, { "epoch": 0.8410326086956522, "grad_norm": 0.23355467546630793, "learning_rate": 7.887164939283066e-05, "loss": 0.971, "step": 1857 }, { "epoch": 0.8414855072463768, "grad_norm": 0.18355626539297532, "learning_rate": 7.88686646728554e-05, "loss": 0.9521, "step": 1858 }, { "epoch": 0.8419384057971014, "grad_norm": 0.20430952494419127, "learning_rate": 7.886567606711815e-05, "loss": 0.9588, "step": 1859 }, { "epoch": 0.842391304347826, "grad_norm": 0.21944241512015472, "learning_rate": 7.886268357591766e-05, "loss": 0.9514, "step": 1860 }, { "epoch": 0.8428442028985508, "grad_norm": 0.1802950875733452, "learning_rate": 7.885968719955312e-05, "loss": 0.9702, "step": 1861 }, { "epoch": 0.8432971014492754, "grad_norm": 0.161047919593799, "learning_rate": 7.885668693832409e-05, "loss": 0.963, "step": 1862 }, { "epoch": 0.84375, "grad_norm": 0.21867720892596837, "learning_rate": 7.885368279253045e-05, "loss": 0.9543, "step": 1863 }, { "epoch": 0.8442028985507246, "grad_norm": 0.19649394493596203, "learning_rate": 7.885067476247261e-05, "loss": 0.9637, "step": 1864 }, { "epoch": 0.8446557971014492, "grad_norm": 0.13232458219777019, "learning_rate": 7.884766284845124e-05, "loss": 0.9795, "step": 1865 }, { "epoch": 0.845108695652174, "grad_norm": 0.1321333097521591, "learning_rate": 7.884464705076745e-05, "loss": 0.9716, "step": 1866 }, { "epoch": 0.8455615942028986, "grad_norm": 0.12883316618964263, "learning_rate": 7.884162736972274e-05, "loss": 0.9656, "step": 1867 }, { "epoch": 0.8460144927536232, "grad_norm": 0.14710357933718254, "learning_rate": 7.883860380561901e-05, "loss": 0.9781, "step": 1868 }, { "epoch": 0.8464673913043478, "grad_norm": 0.1381836767423924, "learning_rate": 7.883557635875849e-05, "loss": 0.9753, "step": 1869 }, { "epoch": 0.8469202898550725, "grad_norm": 0.17450613841355134, "learning_rate": 7.883254502944387e-05, "loss": 0.9526, "step": 1870 }, { "epoch": 0.8473731884057971, "grad_norm": 0.19094715457050754, "learning_rate": 7.882950981797819e-05, "loss": 0.9946, "step": 1871 }, { "epoch": 0.8478260869565217, "grad_norm": 0.20348840604541174, "learning_rate": 7.882647072466488e-05, "loss": 0.958, "step": 1872 }, { "epoch": 0.8482789855072463, "grad_norm": 0.19940289497586838, "learning_rate": 7.882342774980776e-05, "loss": 0.9536, "step": 1873 }, { "epoch": 0.8487318840579711, "grad_norm": 0.23706788108616686, "learning_rate": 7.882038089371104e-05, "loss": 0.9958, "step": 1874 }, { "epoch": 0.8491847826086957, "grad_norm": 0.26672595084454287, "learning_rate": 7.881733015667935e-05, "loss": 0.9458, "step": 1875 }, { "epoch": 0.8496376811594203, "grad_norm": 0.20426045667438822, "learning_rate": 7.881427553901764e-05, "loss": 0.9457, "step": 1876 }, { "epoch": 0.8500905797101449, "grad_norm": 0.1402245592278005, "learning_rate": 7.88112170410313e-05, "loss": 0.9786, "step": 1877 }, { "epoch": 0.8505434782608695, "grad_norm": 0.17129252344841328, "learning_rate": 7.880815466302609e-05, "loss": 0.9756, "step": 1878 }, { "epoch": 0.8509963768115942, "grad_norm": 0.143061364980634, "learning_rate": 7.880508840530815e-05, "loss": 0.962, "step": 1879 }, { "epoch": 0.8514492753623188, "grad_norm": 0.11126858730002291, "learning_rate": 7.880201826818405e-05, "loss": 0.9851, "step": 1880 }, { "epoch": 0.8519021739130435, "grad_norm": 0.16030398166445706, "learning_rate": 7.879894425196069e-05, "loss": 0.9599, "step": 1881 }, { "epoch": 0.8523550724637681, "grad_norm": 0.18390873524286636, "learning_rate": 7.879586635694539e-05, "loss": 1.006, "step": 1882 }, { "epoch": 0.8528079710144928, "grad_norm": 0.1652850143516579, "learning_rate": 7.879278458344585e-05, "loss": 0.9564, "step": 1883 }, { "epoch": 0.8532608695652174, "grad_norm": 0.21260382977140518, "learning_rate": 7.878969893177016e-05, "loss": 0.956, "step": 1884 }, { "epoch": 0.853713768115942, "grad_norm": 0.2449247466757762, "learning_rate": 7.87866094022268e-05, "loss": 0.9454, "step": 1885 }, { "epoch": 0.8541666666666666, "grad_norm": 0.2564117619729713, "learning_rate": 7.878351599512465e-05, "loss": 0.9606, "step": 1886 }, { "epoch": 0.8546195652173914, "grad_norm": 0.2511147616828091, "learning_rate": 7.878041871077294e-05, "loss": 0.9635, "step": 1887 }, { "epoch": 0.855072463768116, "grad_norm": 0.21730104537404088, "learning_rate": 7.87773175494813e-05, "loss": 0.949, "step": 1888 }, { "epoch": 0.8555253623188406, "grad_norm": 0.20366553670031745, "learning_rate": 7.87742125115598e-05, "loss": 0.9675, "step": 1889 }, { "epoch": 0.8559782608695652, "grad_norm": 0.15399584391169935, "learning_rate": 7.877110359731883e-05, "loss": 0.9673, "step": 1890 }, { "epoch": 0.8564311594202898, "grad_norm": 0.12211018188862569, "learning_rate": 7.876799080706919e-05, "loss": 0.965, "step": 1891 }, { "epoch": 0.8568840579710145, "grad_norm": 0.1400633951675724, "learning_rate": 7.876487414112208e-05, "loss": 0.9829, "step": 1892 }, { "epoch": 0.8573369565217391, "grad_norm": 0.13614619509982073, "learning_rate": 7.876175359978907e-05, "loss": 0.9532, "step": 1893 }, { "epoch": 0.8577898550724637, "grad_norm": 0.1583283830377524, "learning_rate": 7.875862918338213e-05, "loss": 0.9468, "step": 1894 }, { "epoch": 0.8582427536231884, "grad_norm": 0.2901723781292373, "learning_rate": 7.875550089221363e-05, "loss": 0.9671, "step": 1895 }, { "epoch": 0.8586956521739131, "grad_norm": 0.18842355341436992, "learning_rate": 7.875236872659626e-05, "loss": 0.9682, "step": 1896 }, { "epoch": 0.8591485507246377, "grad_norm": 0.1976065160819294, "learning_rate": 7.87492326868432e-05, "loss": 0.9507, "step": 1897 }, { "epoch": 0.8596014492753623, "grad_norm": 0.15633477918942268, "learning_rate": 7.874609277326793e-05, "loss": 0.9507, "step": 1898 }, { "epoch": 0.8600543478260869, "grad_norm": 0.13149789002553303, "learning_rate": 7.874294898618438e-05, "loss": 0.9536, "step": 1899 }, { "epoch": 0.8605072463768116, "grad_norm": 0.16638506380709026, "learning_rate": 7.873980132590683e-05, "loss": 0.9682, "step": 1900 }, { "epoch": 0.8609601449275363, "grad_norm": 0.18190698397395835, "learning_rate": 7.873664979274993e-05, "loss": 0.9753, "step": 1901 }, { "epoch": 0.8614130434782609, "grad_norm": 0.1730481239060641, "learning_rate": 7.87334943870288e-05, "loss": 0.96, "step": 1902 }, { "epoch": 0.8618659420289855, "grad_norm": 0.16795893051435898, "learning_rate": 7.873033510905883e-05, "loss": 0.972, "step": 1903 }, { "epoch": 0.8623188405797102, "grad_norm": 0.18653536105274515, "learning_rate": 7.872717195915589e-05, "loss": 0.9709, "step": 1904 }, { "epoch": 0.8627717391304348, "grad_norm": 0.19080403078394706, "learning_rate": 7.872400493763621e-05, "loss": 0.9692, "step": 1905 }, { "epoch": 0.8632246376811594, "grad_norm": 0.17922015310304162, "learning_rate": 7.87208340448164e-05, "loss": 0.9578, "step": 1906 }, { "epoch": 0.863677536231884, "grad_norm": 0.17759500872521855, "learning_rate": 7.871765928101343e-05, "loss": 0.9721, "step": 1907 }, { "epoch": 0.8641304347826086, "grad_norm": 0.20305161150693032, "learning_rate": 7.871448064654473e-05, "loss": 0.9638, "step": 1908 }, { "epoch": 0.8645833333333334, "grad_norm": 0.3644853586274534, "learning_rate": 7.871129814172805e-05, "loss": 1.0096, "step": 1909 }, { "epoch": 0.865036231884058, "grad_norm": 0.1497420202400724, "learning_rate": 7.870811176688155e-05, "loss": 0.9772, "step": 1910 }, { "epoch": 0.8654891304347826, "grad_norm": 0.17856498148971903, "learning_rate": 7.870492152232375e-05, "loss": 0.9583, "step": 1911 }, { "epoch": 0.8659420289855072, "grad_norm": 0.17759759016816767, "learning_rate": 7.870172740837364e-05, "loss": 0.9761, "step": 1912 }, { "epoch": 0.8663949275362319, "grad_norm": 0.1656651475170593, "learning_rate": 7.869852942535051e-05, "loss": 0.9637, "step": 1913 }, { "epoch": 0.8668478260869565, "grad_norm": 0.1819326373351737, "learning_rate": 7.869532757357408e-05, "loss": 0.9647, "step": 1914 }, { "epoch": 0.8673007246376812, "grad_norm": 0.23864261347972945, "learning_rate": 7.86921218533644e-05, "loss": 0.9645, "step": 1915 }, { "epoch": 0.8677536231884058, "grad_norm": 0.29893116842481193, "learning_rate": 7.868891226504202e-05, "loss": 0.9834, "step": 1916 }, { "epoch": 0.8682065217391305, "grad_norm": 0.31189211298004, "learning_rate": 7.868569880892775e-05, "loss": 0.9769, "step": 1917 }, { "epoch": 0.8686594202898551, "grad_norm": 0.26658583624718607, "learning_rate": 7.868248148534288e-05, "loss": 0.9831, "step": 1918 }, { "epoch": 0.8691123188405797, "grad_norm": 0.22103940787189266, "learning_rate": 7.867926029460904e-05, "loss": 0.9831, "step": 1919 }, { "epoch": 0.8695652173913043, "grad_norm": 0.20207646191634512, "learning_rate": 7.867603523704825e-05, "loss": 0.9593, "step": 1920 }, { "epoch": 0.8700181159420289, "grad_norm": 0.1909491841689262, "learning_rate": 7.867280631298293e-05, "loss": 0.9684, "step": 1921 }, { "epoch": 0.8704710144927537, "grad_norm": 0.1559966119424484, "learning_rate": 7.866957352273588e-05, "loss": 0.9735, "step": 1922 }, { "epoch": 0.8709239130434783, "grad_norm": 0.1381890863509169, "learning_rate": 7.866633686663029e-05, "loss": 0.9339, "step": 1923 }, { "epoch": 0.8713768115942029, "grad_norm": 0.2024610891006061, "learning_rate": 7.866309634498973e-05, "loss": 0.9651, "step": 1924 }, { "epoch": 0.8718297101449275, "grad_norm": 0.25961125333403096, "learning_rate": 7.865985195813816e-05, "loss": 0.9701, "step": 1925 }, { "epoch": 0.8722826086956522, "grad_norm": 0.22903374906155796, "learning_rate": 7.865660370639993e-05, "loss": 0.961, "step": 1926 }, { "epoch": 0.8727355072463768, "grad_norm": 0.19466573004866378, "learning_rate": 7.865335159009977e-05, "loss": 0.9681, "step": 1927 }, { "epoch": 0.8731884057971014, "grad_norm": 0.17483881450566038, "learning_rate": 7.86500956095628e-05, "loss": 0.9836, "step": 1928 }, { "epoch": 0.873641304347826, "grad_norm": 0.14237273807217563, "learning_rate": 7.864683576511454e-05, "loss": 0.9562, "step": 1929 }, { "epoch": 0.8740942028985508, "grad_norm": 0.18622698521673917, "learning_rate": 7.864357205708085e-05, "loss": 0.9773, "step": 1930 }, { "epoch": 0.8745471014492754, "grad_norm": 0.15265808057172417, "learning_rate": 7.864030448578803e-05, "loss": 0.959, "step": 1931 }, { "epoch": 0.875, "grad_norm": 0.16237079382555072, "learning_rate": 7.863703305156273e-05, "loss": 0.9516, "step": 1932 }, { "epoch": 0.8754528985507246, "grad_norm": 0.1521367965184492, "learning_rate": 7.863375775473203e-05, "loss": 0.9747, "step": 1933 }, { "epoch": 0.8759057971014492, "grad_norm": 0.1481943769818431, "learning_rate": 7.863047859562333e-05, "loss": 0.9634, "step": 1934 }, { "epoch": 0.876358695652174, "grad_norm": 0.17945479797727998, "learning_rate": 7.862719557456448e-05, "loss": 0.9712, "step": 1935 }, { "epoch": 0.8768115942028986, "grad_norm": 0.10031949509668689, "learning_rate": 7.862390869188367e-05, "loss": 0.9696, "step": 1936 }, { "epoch": 0.8772644927536232, "grad_norm": 0.17632861710353914, "learning_rate": 7.86206179479095e-05, "loss": 0.9825, "step": 1937 }, { "epoch": 0.8777173913043478, "grad_norm": 0.1955470930325886, "learning_rate": 7.861732334297098e-05, "loss": 0.9566, "step": 1938 }, { "epoch": 0.8781702898550725, "grad_norm": 0.17700315852321677, "learning_rate": 7.861402487739743e-05, "loss": 0.9678, "step": 1939 }, { "epoch": 0.8786231884057971, "grad_norm": 0.17933512373513982, "learning_rate": 7.861072255151863e-05, "loss": 0.9716, "step": 1940 }, { "epoch": 0.8790760869565217, "grad_norm": 0.14566433917686614, "learning_rate": 7.86074163656647e-05, "loss": 0.9547, "step": 1941 }, { "epoch": 0.8795289855072463, "grad_norm": 0.1347523645811738, "learning_rate": 7.86041063201662e-05, "loss": 0.9746, "step": 1942 }, { "epoch": 0.8799818840579711, "grad_norm": 0.18163834952827346, "learning_rate": 7.8600792415354e-05, "loss": 0.9797, "step": 1943 }, { "epoch": 0.8804347826086957, "grad_norm": 0.16534400247053468, "learning_rate": 7.859747465155941e-05, "loss": 0.97, "step": 1944 }, { "epoch": 0.8808876811594203, "grad_norm": 0.12392762390991953, "learning_rate": 7.859415302911412e-05, "loss": 0.9849, "step": 1945 }, { "epoch": 0.8813405797101449, "grad_norm": 0.1282420523806662, "learning_rate": 7.85908275483502e-05, "loss": 0.9671, "step": 1946 }, { "epoch": 0.8817934782608695, "grad_norm": 0.17094244481510903, "learning_rate": 7.858749820960009e-05, "loss": 0.9681, "step": 1947 }, { "epoch": 0.8822463768115942, "grad_norm": 0.21587968100309413, "learning_rate": 7.858416501319666e-05, "loss": 0.9649, "step": 1948 }, { "epoch": 0.8826992753623188, "grad_norm": 0.25921979612515705, "learning_rate": 7.858082795947309e-05, "loss": 0.9986, "step": 1949 }, { "epoch": 0.8831521739130435, "grad_norm": 0.29446327333977257, "learning_rate": 7.8577487048763e-05, "loss": 0.977, "step": 1950 }, { "epoch": 0.8836050724637681, "grad_norm": 0.2930789478309897, "learning_rate": 7.857414228140043e-05, "loss": 0.966, "step": 1951 }, { "epoch": 0.8840579710144928, "grad_norm": 0.2744785326660344, "learning_rate": 7.857079365771972e-05, "loss": 0.9644, "step": 1952 }, { "epoch": 0.8845108695652174, "grad_norm": 0.2373021271791748, "learning_rate": 7.856744117805564e-05, "loss": 0.9732, "step": 1953 }, { "epoch": 0.884963768115942, "grad_norm": 0.16328500570447885, "learning_rate": 7.856408484274335e-05, "loss": 0.9575, "step": 1954 }, { "epoch": 0.8854166666666666, "grad_norm": 0.17449524952013648, "learning_rate": 7.856072465211839e-05, "loss": 0.9665, "step": 1955 }, { "epoch": 0.8858695652173914, "grad_norm": 0.24359769573822088, "learning_rate": 7.855736060651668e-05, "loss": 0.9745, "step": 1956 }, { "epoch": 0.886322463768116, "grad_norm": 0.29299773104896054, "learning_rate": 7.855399270627454e-05, "loss": 0.9758, "step": 1957 }, { "epoch": 0.8867753623188406, "grad_norm": 0.28969280468719577, "learning_rate": 7.855062095172865e-05, "loss": 0.9641, "step": 1958 }, { "epoch": 0.8872282608695652, "grad_norm": 0.2565122594900033, "learning_rate": 7.854724534321609e-05, "loss": 0.97, "step": 1959 }, { "epoch": 0.8876811594202898, "grad_norm": 0.21204729870386424, "learning_rate": 7.854386588107433e-05, "loss": 0.9757, "step": 1960 }, { "epoch": 0.8881340579710145, "grad_norm": 0.17413567415629527, "learning_rate": 7.854048256564122e-05, "loss": 0.9566, "step": 1961 }, { "epoch": 0.8885869565217391, "grad_norm": 0.15960162969893452, "learning_rate": 7.853709539725501e-05, "loss": 0.9943, "step": 1962 }, { "epoch": 0.8890398550724637, "grad_norm": 0.23620106720764153, "learning_rate": 7.853370437625429e-05, "loss": 0.9658, "step": 1963 }, { "epoch": 0.8894927536231884, "grad_norm": 0.30907367918341017, "learning_rate": 7.853030950297809e-05, "loss": 0.9795, "step": 1964 }, { "epoch": 0.8899456521739131, "grad_norm": 0.3174632797228026, "learning_rate": 7.852691077776579e-05, "loss": 0.953, "step": 1965 }, { "epoch": 0.8903985507246377, "grad_norm": 0.2758122162329761, "learning_rate": 7.852350820095715e-05, "loss": 0.9711, "step": 1966 }, { "epoch": 0.8908514492753623, "grad_norm": 0.1735611628778856, "learning_rate": 7.852010177289239e-05, "loss": 0.9786, "step": 1967 }, { "epoch": 0.8913043478260869, "grad_norm": 0.11898722573657235, "learning_rate": 7.851669149391198e-05, "loss": 0.9704, "step": 1968 }, { "epoch": 0.8917572463768116, "grad_norm": 0.1289067765172888, "learning_rate": 7.851327736435689e-05, "loss": 0.9671, "step": 1969 }, { "epoch": 0.8922101449275363, "grad_norm": 0.18605468385705012, "learning_rate": 7.850985938456845e-05, "loss": 0.9816, "step": 1970 }, { "epoch": 0.8926630434782609, "grad_norm": 0.19663956269002195, "learning_rate": 7.850643755488833e-05, "loss": 0.9699, "step": 1971 }, { "epoch": 0.8931159420289855, "grad_norm": 0.14687941045728922, "learning_rate": 7.850301187565864e-05, "loss": 0.9537, "step": 1972 }, { "epoch": 0.8935688405797102, "grad_norm": 0.12312578688463877, "learning_rate": 7.849958234722183e-05, "loss": 0.9483, "step": 1973 }, { "epoch": 0.8940217391304348, "grad_norm": 0.1335144185939817, "learning_rate": 7.849614896992077e-05, "loss": 0.961, "step": 1974 }, { "epoch": 0.8944746376811594, "grad_norm": 0.1683063368732806, "learning_rate": 7.849271174409868e-05, "loss": 0.9834, "step": 1975 }, { "epoch": 0.894927536231884, "grad_norm": 0.315862969760638, "learning_rate": 7.848927067009922e-05, "loss": 0.995, "step": 1976 }, { "epoch": 0.8953804347826086, "grad_norm": 0.14435481449007212, "learning_rate": 7.848582574826637e-05, "loss": 0.9763, "step": 1977 }, { "epoch": 0.8958333333333334, "grad_norm": 0.18156480421296983, "learning_rate": 7.848237697894453e-05, "loss": 0.9697, "step": 1978 }, { "epoch": 0.896286231884058, "grad_norm": 0.22614602649028404, "learning_rate": 7.847892436247849e-05, "loss": 0.958, "step": 1979 }, { "epoch": 0.8967391304347826, "grad_norm": 0.19838157106580362, "learning_rate": 7.84754678992134e-05, "loss": 0.9678, "step": 1980 }, { "epoch": 0.8971920289855072, "grad_norm": 0.15588064194097603, "learning_rate": 7.847200758949482e-05, "loss": 1.0009, "step": 1981 }, { "epoch": 0.8976449275362319, "grad_norm": 0.14542504217797705, "learning_rate": 7.846854343366868e-05, "loss": 0.9447, "step": 1982 }, { "epoch": 0.8980978260869565, "grad_norm": 0.1508243119654757, "learning_rate": 7.846507543208129e-05, "loss": 0.9672, "step": 1983 }, { "epoch": 0.8985507246376812, "grad_norm": 0.13580158566751013, "learning_rate": 7.846160358507936e-05, "loss": 0.9712, "step": 1984 }, { "epoch": 0.8990036231884058, "grad_norm": 0.1500372629554894, "learning_rate": 7.845812789300997e-05, "loss": 0.9832, "step": 1985 }, { "epoch": 0.8994565217391305, "grad_norm": 0.13941644936094677, "learning_rate": 7.845464835622059e-05, "loss": 0.9732, "step": 1986 }, { "epoch": 0.8999094202898551, "grad_norm": 0.12262700092097771, "learning_rate": 7.845116497505908e-05, "loss": 0.9762, "step": 1987 }, { "epoch": 0.9003623188405797, "grad_norm": 0.15039328569623572, "learning_rate": 7.844767774987367e-05, "loss": 0.9657, "step": 1988 }, { "epoch": 0.9008152173913043, "grad_norm": 0.14365034278272623, "learning_rate": 7.8444186681013e-05, "loss": 0.955, "step": 1989 }, { "epoch": 0.9012681159420289, "grad_norm": 0.1305813464971235, "learning_rate": 7.844069176882606e-05, "loss": 0.9601, "step": 1990 }, { "epoch": 0.9017210144927537, "grad_norm": 0.12542717807221812, "learning_rate": 7.843719301366226e-05, "loss": 0.9736, "step": 1991 }, { "epoch": 0.9021739130434783, "grad_norm": 0.15232698385913462, "learning_rate": 7.843369041587135e-05, "loss": 0.9633, "step": 1992 }, { "epoch": 0.9026268115942029, "grad_norm": 0.1463856170977692, "learning_rate": 7.843018397580351e-05, "loss": 0.9819, "step": 1993 }, { "epoch": 0.9030797101449275, "grad_norm": 0.11917914659748283, "learning_rate": 7.842667369380929e-05, "loss": 0.946, "step": 1994 }, { "epoch": 0.9035326086956522, "grad_norm": 0.1371808612422895, "learning_rate": 7.842315957023961e-05, "loss": 0.9862, "step": 1995 }, { "epoch": 0.9039855072463768, "grad_norm": 0.150322872539075, "learning_rate": 7.841964160544577e-05, "loss": 0.9706, "step": 1996 }, { "epoch": 0.9044384057971014, "grad_norm": 0.13331369744776203, "learning_rate": 7.841611979977947e-05, "loss": 0.9786, "step": 1997 }, { "epoch": 0.904891304347826, "grad_norm": 0.11820803891692591, "learning_rate": 7.841259415359282e-05, "loss": 0.9482, "step": 1998 }, { "epoch": 0.9053442028985508, "grad_norm": 0.139261164897671, "learning_rate": 7.840906466723823e-05, "loss": 0.9761, "step": 1999 }, { "epoch": 0.9057971014492754, "grad_norm": 0.1522565703617625, "learning_rate": 7.840553134106862e-05, "loss": 0.9256, "step": 2000 }, { "epoch": 0.90625, "grad_norm": 0.17247555375948884, "learning_rate": 7.840199417543716e-05, "loss": 0.9507, "step": 2001 }, { "epoch": 0.9067028985507246, "grad_norm": 0.20214889000789735, "learning_rate": 7.839845317069749e-05, "loss": 0.9855, "step": 2002 }, { "epoch": 0.9071557971014492, "grad_norm": 0.20564842312858964, "learning_rate": 7.839490832720361e-05, "loss": 0.9618, "step": 2003 }, { "epoch": 0.907608695652174, "grad_norm": 0.21126975825993283, "learning_rate": 7.83913596453099e-05, "loss": 0.9541, "step": 2004 }, { "epoch": 0.9080615942028986, "grad_norm": 0.22493972220508254, "learning_rate": 7.838780712537114e-05, "loss": 0.9802, "step": 2005 }, { "epoch": 0.9085144927536232, "grad_norm": 0.2501961202828881, "learning_rate": 7.838425076774247e-05, "loss": 0.95, "step": 2006 }, { "epoch": 0.9089673913043478, "grad_norm": 0.2435093689267719, "learning_rate": 7.838069057277943e-05, "loss": 0.9737, "step": 2007 }, { "epoch": 0.9094202898550725, "grad_norm": 0.2436413404394071, "learning_rate": 7.83771265408379e-05, "loss": 0.9567, "step": 2008 }, { "epoch": 0.9098731884057971, "grad_norm": 0.22112681463811779, "learning_rate": 7.837355867227426e-05, "loss": 0.9508, "step": 2009 }, { "epoch": 0.9103260869565217, "grad_norm": 0.1886817717293824, "learning_rate": 7.836998696744513e-05, "loss": 0.9469, "step": 2010 }, { "epoch": 0.9107789855072463, "grad_norm": 0.13787125787643026, "learning_rate": 7.836641142670759e-05, "loss": 0.9816, "step": 2011 }, { "epoch": 0.9112318840579711, "grad_norm": 0.14011822367285243, "learning_rate": 7.836283205041913e-05, "loss": 0.9641, "step": 2012 }, { "epoch": 0.9116847826086957, "grad_norm": 0.14930907192364745, "learning_rate": 7.835924883893753e-05, "loss": 0.9739, "step": 2013 }, { "epoch": 0.9121376811594203, "grad_norm": 0.16171176538491874, "learning_rate": 7.835566179262106e-05, "loss": 0.9655, "step": 2014 }, { "epoch": 0.9125905797101449, "grad_norm": 0.14997749702069843, "learning_rate": 7.83520709118283e-05, "loss": 0.9562, "step": 2015 }, { "epoch": 0.9130434782608695, "grad_norm": 0.10297950337397621, "learning_rate": 7.834847619691823e-05, "loss": 0.973, "step": 2016 }, { "epoch": 0.9134963768115942, "grad_norm": 0.16842928478275696, "learning_rate": 7.834487764825022e-05, "loss": 0.9486, "step": 2017 }, { "epoch": 0.9139492753623188, "grad_norm": 0.20165756732023807, "learning_rate": 7.834127526618405e-05, "loss": 0.9885, "step": 2018 }, { "epoch": 0.9144021739130435, "grad_norm": 0.1945459257540249, "learning_rate": 7.833766905107982e-05, "loss": 0.9716, "step": 2019 }, { "epoch": 0.9148550724637681, "grad_norm": 0.18732514216905044, "learning_rate": 7.833405900329806e-05, "loss": 0.9715, "step": 2020 }, { "epoch": 0.9153079710144928, "grad_norm": 0.15754595734859622, "learning_rate": 7.833044512319968e-05, "loss": 0.9485, "step": 2021 }, { "epoch": 0.9157608695652174, "grad_norm": 0.13934520198630793, "learning_rate": 7.832682741114597e-05, "loss": 0.9655, "step": 2022 }, { "epoch": 0.916213768115942, "grad_norm": 0.15301509946389166, "learning_rate": 7.832320586749857e-05, "loss": 0.9672, "step": 2023 }, { "epoch": 0.9166666666666666, "grad_norm": 0.1655014157278245, "learning_rate": 7.831958049261956e-05, "loss": 0.9675, "step": 2024 }, { "epoch": 0.9171195652173914, "grad_norm": 0.2119594870935029, "learning_rate": 7.831595128687136e-05, "loss": 0.9484, "step": 2025 }, { "epoch": 0.917572463768116, "grad_norm": 0.22243667437067813, "learning_rate": 7.831231825061681e-05, "loss": 0.9639, "step": 2026 }, { "epoch": 0.9180253623188406, "grad_norm": 0.2147114268848287, "learning_rate": 7.830868138421908e-05, "loss": 0.9737, "step": 2027 }, { "epoch": 0.9184782608695652, "grad_norm": 0.21347808768286375, "learning_rate": 7.830504068804178e-05, "loss": 0.9532, "step": 2028 }, { "epoch": 0.9189311594202898, "grad_norm": 0.1782385195556866, "learning_rate": 7.830139616244886e-05, "loss": 0.9655, "step": 2029 }, { "epoch": 0.9193840579710145, "grad_norm": 0.11336066730000872, "learning_rate": 7.829774780780465e-05, "loss": 0.9499, "step": 2030 }, { "epoch": 0.9198369565217391, "grad_norm": 0.10342285778537395, "learning_rate": 7.829409562447392e-05, "loss": 0.9663, "step": 2031 }, { "epoch": 0.9202898550724637, "grad_norm": 0.16996527867293762, "learning_rate": 7.829043961282177e-05, "loss": 0.9606, "step": 2032 }, { "epoch": 0.9207427536231884, "grad_norm": 0.21503180400912394, "learning_rate": 7.82867797732137e-05, "loss": 0.9686, "step": 2033 }, { "epoch": 0.9211956521739131, "grad_norm": 0.21336411345343373, "learning_rate": 7.828311610601557e-05, "loss": 0.9263, "step": 2034 }, { "epoch": 0.9216485507246377, "grad_norm": 0.14886380187906711, "learning_rate": 7.827944861159367e-05, "loss": 0.9596, "step": 2035 }, { "epoch": 0.9221014492753623, "grad_norm": 0.09890038046623524, "learning_rate": 7.827577729031464e-05, "loss": 0.9608, "step": 2036 }, { "epoch": 0.9225543478260869, "grad_norm": 0.13389826048265877, "learning_rate": 7.827210214254548e-05, "loss": 0.9731, "step": 2037 }, { "epoch": 0.9230072463768116, "grad_norm": 0.19694100802482434, "learning_rate": 7.826842316865363e-05, "loss": 0.9482, "step": 2038 }, { "epoch": 0.9234601449275363, "grad_norm": 0.16810599539155782, "learning_rate": 7.826474036900687e-05, "loss": 0.9531, "step": 2039 }, { "epoch": 0.9239130434782609, "grad_norm": 0.1424928476701714, "learning_rate": 7.82610537439734e-05, "loss": 0.9602, "step": 2040 }, { "epoch": 0.9243659420289855, "grad_norm": 0.11507467088862239, "learning_rate": 7.825736329392174e-05, "loss": 0.963, "step": 2041 }, { "epoch": 0.9248188405797102, "grad_norm": 0.11740880099924207, "learning_rate": 7.825366901922085e-05, "loss": 0.9512, "step": 2042 }, { "epoch": 0.9252717391304348, "grad_norm": 0.12323529763828373, "learning_rate": 7.824997092024004e-05, "loss": 0.9465, "step": 2043 }, { "epoch": 0.9257246376811594, "grad_norm": 0.10838294578589004, "learning_rate": 7.824626899734905e-05, "loss": 0.9674, "step": 2044 }, { "epoch": 0.926177536231884, "grad_norm": 0.1301511539839685, "learning_rate": 7.824256325091793e-05, "loss": 0.9585, "step": 2045 }, { "epoch": 0.9266304347826086, "grad_norm": 0.12760054434745355, "learning_rate": 7.823885368131716e-05, "loss": 0.9662, "step": 2046 }, { "epoch": 0.9270833333333334, "grad_norm": 0.1470592104597635, "learning_rate": 7.823514028891758e-05, "loss": 0.9672, "step": 2047 }, { "epoch": 0.927536231884058, "grad_norm": 0.1784743803286157, "learning_rate": 7.823142307409045e-05, "loss": 0.9639, "step": 2048 }, { "epoch": 0.9279891304347826, "grad_norm": 0.18379689215204575, "learning_rate": 7.822770203720737e-05, "loss": 0.9598, "step": 2049 }, { "epoch": 0.9284420289855072, "grad_norm": 0.17235831072502097, "learning_rate": 7.822397717864035e-05, "loss": 0.967, "step": 2050 }, { "epoch": 0.9288949275362319, "grad_norm": 0.16527766465291016, "learning_rate": 7.822024849876175e-05, "loss": 0.98, "step": 2051 }, { "epoch": 0.9293478260869565, "grad_norm": 0.16622255526044047, "learning_rate": 7.821651599794435e-05, "loss": 0.9728, "step": 2052 }, { "epoch": 0.9298007246376812, "grad_norm": 0.17180320155719958, "learning_rate": 7.821277967656127e-05, "loss": 0.9718, "step": 2053 }, { "epoch": 0.9302536231884058, "grad_norm": 0.20281300882902561, "learning_rate": 7.820903953498607e-05, "loss": 0.9674, "step": 2054 }, { "epoch": 0.9307065217391305, "grad_norm": 0.22240866200543086, "learning_rate": 7.820529557359263e-05, "loss": 0.9737, "step": 2055 }, { "epoch": 0.9311594202898551, "grad_norm": 0.22471036023332858, "learning_rate": 7.820154779275525e-05, "loss": 0.9516, "step": 2056 }, { "epoch": 0.9316123188405797, "grad_norm": 0.2202442215868543, "learning_rate": 7.819779619284861e-05, "loss": 0.9507, "step": 2057 }, { "epoch": 0.9320652173913043, "grad_norm": 0.2004956650487379, "learning_rate": 7.819404077424775e-05, "loss": 0.9914, "step": 2058 }, { "epoch": 0.9325181159420289, "grad_norm": 0.18035528850578417, "learning_rate": 7.819028153732811e-05, "loss": 0.9578, "step": 2059 }, { "epoch": 0.9329710144927537, "grad_norm": 0.15041133617592414, "learning_rate": 7.81865184824655e-05, "loss": 0.9514, "step": 2060 }, { "epoch": 0.9334239130434783, "grad_norm": 0.12708825159839426, "learning_rate": 7.818275161003615e-05, "loss": 0.9583, "step": 2061 }, { "epoch": 0.9338768115942029, "grad_norm": 0.11087055724112223, "learning_rate": 7.817898092041659e-05, "loss": 0.9481, "step": 2062 }, { "epoch": 0.9343297101449275, "grad_norm": 0.12569112130205312, "learning_rate": 7.817520641398381e-05, "loss": 0.9698, "step": 2063 }, { "epoch": 0.9347826086956522, "grad_norm": 0.15114705592332958, "learning_rate": 7.817142809111516e-05, "loss": 0.9621, "step": 2064 }, { "epoch": 0.9352355072463768, "grad_norm": 0.20176377035259918, "learning_rate": 7.816764595218836e-05, "loss": 0.952, "step": 2065 }, { "epoch": 0.9356884057971014, "grad_norm": 0.2620769970366514, "learning_rate": 7.816385999758151e-05, "loss": 0.992, "step": 2066 }, { "epoch": 0.936141304347826, "grad_norm": 0.24687451338121597, "learning_rate": 7.816007022767308e-05, "loss": 0.9889, "step": 2067 }, { "epoch": 0.9365942028985508, "grad_norm": 0.19944109713332253, "learning_rate": 7.815627664284198e-05, "loss": 0.9546, "step": 2068 }, { "epoch": 0.9370471014492754, "grad_norm": 0.17672054687920644, "learning_rate": 7.815247924346743e-05, "loss": 0.977, "step": 2069 }, { "epoch": 0.9375, "grad_norm": 0.1810891088372116, "learning_rate": 7.814867802992907e-05, "loss": 0.9732, "step": 2070 }, { "epoch": 0.9379528985507246, "grad_norm": 0.15647128840906455, "learning_rate": 7.814487300260693e-05, "loss": 0.9563, "step": 2071 }, { "epoch": 0.9384057971014492, "grad_norm": 0.12069029334534431, "learning_rate": 7.81410641618814e-05, "loss": 0.9664, "step": 2072 }, { "epoch": 0.938858695652174, "grad_norm": 0.0956964051132491, "learning_rate": 7.813725150813321e-05, "loss": 0.9691, "step": 2073 }, { "epoch": 0.9393115942028986, "grad_norm": 0.1262245977812174, "learning_rate": 7.813343504174357e-05, "loss": 0.9638, "step": 2074 }, { "epoch": 0.9397644927536232, "grad_norm": 0.18204262475192504, "learning_rate": 7.812961476309401e-05, "loss": 0.9488, "step": 2075 }, { "epoch": 0.9402173913043478, "grad_norm": 0.17863460227726355, "learning_rate": 7.812579067256644e-05, "loss": 0.9658, "step": 2076 }, { "epoch": 0.9406702898550725, "grad_norm": 0.1428509419727445, "learning_rate": 7.812196277054316e-05, "loss": 0.9848, "step": 2077 }, { "epoch": 0.9411231884057971, "grad_norm": 0.11571594177797802, "learning_rate": 7.811813105740684e-05, "loss": 0.955, "step": 2078 }, { "epoch": 0.9415760869565217, "grad_norm": 0.1591012279118768, "learning_rate": 7.811429553354057e-05, "loss": 0.9738, "step": 2079 }, { "epoch": 0.9420289855072463, "grad_norm": 0.14093100980592416, "learning_rate": 7.811045619932775e-05, "loss": 0.9614, "step": 2080 }, { "epoch": 0.9424818840579711, "grad_norm": 0.12943003319984417, "learning_rate": 7.810661305515227e-05, "loss": 0.9675, "step": 2081 }, { "epoch": 0.9429347826086957, "grad_norm": 0.15608313130564228, "learning_rate": 7.810276610139827e-05, "loss": 0.9728, "step": 2082 }, { "epoch": 0.9433876811594203, "grad_norm": 0.15837787218456326, "learning_rate": 7.809891533845039e-05, "loss": 0.9712, "step": 2083 }, { "epoch": 0.9438405797101449, "grad_norm": 0.15992744219034694, "learning_rate": 7.809506076669355e-05, "loss": 0.9753, "step": 2084 }, { "epoch": 0.9442934782608695, "grad_norm": 0.14432273011722657, "learning_rate": 7.809120238651313e-05, "loss": 0.9798, "step": 2085 }, { "epoch": 0.9447463768115942, "grad_norm": 0.16313609474594512, "learning_rate": 7.808734019829484e-05, "loss": 0.9632, "step": 2086 }, { "epoch": 0.9451992753623188, "grad_norm": 0.18374774496891494, "learning_rate": 7.80834742024248e-05, "loss": 0.9554, "step": 2087 }, { "epoch": 0.9456521739130435, "grad_norm": 0.21568082200299463, "learning_rate": 7.807960439928948e-05, "loss": 0.962, "step": 2088 }, { "epoch": 0.9461050724637681, "grad_norm": 0.1911760849220139, "learning_rate": 7.807573078927577e-05, "loss": 0.9745, "step": 2089 }, { "epoch": 0.9465579710144928, "grad_norm": 0.14991633087642528, "learning_rate": 7.807185337277093e-05, "loss": 0.9973, "step": 2090 }, { "epoch": 0.9470108695652174, "grad_norm": 0.12416457603950763, "learning_rate": 7.806797215016257e-05, "loss": 0.946, "step": 2091 }, { "epoch": 0.947463768115942, "grad_norm": 0.1422923128413782, "learning_rate": 7.806408712183873e-05, "loss": 0.964, "step": 2092 }, { "epoch": 0.9479166666666666, "grad_norm": 0.13521461563567397, "learning_rate": 7.806019828818776e-05, "loss": 0.9454, "step": 2093 }, { "epoch": 0.9483695652173914, "grad_norm": 0.12586772712264824, "learning_rate": 7.805630564959846e-05, "loss": 0.964, "step": 2094 }, { "epoch": 0.948822463768116, "grad_norm": 0.13855648819540323, "learning_rate": 7.805240920645998e-05, "loss": 0.9732, "step": 2095 }, { "epoch": 0.9492753623188406, "grad_norm": 0.10743671933892276, "learning_rate": 7.804850895916185e-05, "loss": 0.9636, "step": 2096 }, { "epoch": 0.9497282608695652, "grad_norm": 0.09992377583101231, "learning_rate": 7.804460490809399e-05, "loss": 0.9722, "step": 2097 }, { "epoch": 0.9501811594202898, "grad_norm": 0.12125460936873192, "learning_rate": 7.804069705364669e-05, "loss": 0.9718, "step": 2098 }, { "epoch": 0.9506340579710145, "grad_norm": 0.12178712435439325, "learning_rate": 7.803678539621063e-05, "loss": 0.973, "step": 2099 }, { "epoch": 0.9510869565217391, "grad_norm": 0.1167189091459139, "learning_rate": 7.803286993617685e-05, "loss": 0.9631, "step": 2100 }, { "epoch": 0.9515398550724637, "grad_norm": 0.13406266053771806, "learning_rate": 7.80289506739368e-05, "loss": 0.9693, "step": 2101 }, { "epoch": 0.9519927536231884, "grad_norm": 0.15629801316223105, "learning_rate": 7.802502760988229e-05, "loss": 0.9812, "step": 2102 }, { "epoch": 0.9524456521739131, "grad_norm": 0.16706453729156692, "learning_rate": 7.802110074440551e-05, "loss": 0.9609, "step": 2103 }, { "epoch": 0.9528985507246377, "grad_norm": 0.15234106875450093, "learning_rate": 7.801717007789904e-05, "loss": 0.9708, "step": 2104 }, { "epoch": 0.9533514492753623, "grad_norm": 0.1674125483554769, "learning_rate": 7.801323561075583e-05, "loss": 0.9588, "step": 2105 }, { "epoch": 0.9538043478260869, "grad_norm": 0.21168006365093422, "learning_rate": 7.800929734336922e-05, "loss": 0.9494, "step": 2106 }, { "epoch": 0.9542572463768116, "grad_norm": 0.2045028075209965, "learning_rate": 7.800535527613293e-05, "loss": 0.9698, "step": 2107 }, { "epoch": 0.9547101449275363, "grad_norm": 0.1786335534247547, "learning_rate": 7.800140940944104e-05, "loss": 0.957, "step": 2108 }, { "epoch": 0.9551630434782609, "grad_norm": 0.18177881367383622, "learning_rate": 7.799745974368804e-05, "loss": 0.9591, "step": 2109 }, { "epoch": 0.9556159420289855, "grad_norm": 0.17376157177018095, "learning_rate": 7.799350627926878e-05, "loss": 0.9538, "step": 2110 }, { "epoch": 0.9560688405797102, "grad_norm": 0.15264477188832584, "learning_rate": 7.798954901657848e-05, "loss": 0.961, "step": 2111 }, { "epoch": 0.9565217391304348, "grad_norm": 0.1526367207161691, "learning_rate": 7.798558795601278e-05, "loss": 0.9385, "step": 2112 }, { "epoch": 0.9569746376811594, "grad_norm": 0.1346416778858968, "learning_rate": 7.798162309796764e-05, "loss": 0.9697, "step": 2113 }, { "epoch": 0.957427536231884, "grad_norm": 0.13984059216090686, "learning_rate": 7.797765444283947e-05, "loss": 0.9585, "step": 2114 }, { "epoch": 0.9578804347826086, "grad_norm": 0.1996893462177859, "learning_rate": 7.797368199102501e-05, "loss": 0.9524, "step": 2115 }, { "epoch": 0.9583333333333334, "grad_norm": 0.22612305279223133, "learning_rate": 7.796970574292136e-05, "loss": 0.9508, "step": 2116 }, { "epoch": 0.958786231884058, "grad_norm": 0.20323775660190768, "learning_rate": 7.796572569892607e-05, "loss": 0.9777, "step": 2117 }, { "epoch": 0.9592391304347826, "grad_norm": 0.18961024436450957, "learning_rate": 7.796174185943702e-05, "loss": 0.9626, "step": 2118 }, { "epoch": 0.9596920289855072, "grad_norm": 0.19890062247738013, "learning_rate": 7.79577542248525e-05, "loss": 0.9732, "step": 2119 }, { "epoch": 0.9601449275362319, "grad_norm": 0.19585005483506074, "learning_rate": 7.795376279557111e-05, "loss": 0.9937, "step": 2120 }, { "epoch": 0.9605978260869565, "grad_norm": 0.1662827398169114, "learning_rate": 7.794976757199192e-05, "loss": 0.9516, "step": 2121 }, { "epoch": 0.9610507246376812, "grad_norm": 0.16901295868448005, "learning_rate": 7.794576855451434e-05, "loss": 0.9616, "step": 2122 }, { "epoch": 0.9615036231884058, "grad_norm": 0.1493147462784743, "learning_rate": 7.794176574353815e-05, "loss": 0.9346, "step": 2123 }, { "epoch": 0.9619565217391305, "grad_norm": 0.1855842745587568, "learning_rate": 7.79377591394635e-05, "loss": 0.9835, "step": 2124 }, { "epoch": 0.9624094202898551, "grad_norm": 0.19630690772246742, "learning_rate": 7.793374874269096e-05, "loss": 0.9689, "step": 2125 }, { "epoch": 0.9628623188405797, "grad_norm": 0.17447358270865218, "learning_rate": 7.792973455362145e-05, "loss": 0.9684, "step": 2126 }, { "epoch": 0.9633152173913043, "grad_norm": 0.14539949991729056, "learning_rate": 7.792571657265626e-05, "loss": 0.9316, "step": 2127 }, { "epoch": 0.9637681159420289, "grad_norm": 0.12697200586644553, "learning_rate": 7.79216948001971e-05, "loss": 0.966, "step": 2128 }, { "epoch": 0.9642210144927537, "grad_norm": 0.12603695741203794, "learning_rate": 7.7917669236646e-05, "loss": 0.9677, "step": 2129 }, { "epoch": 0.9646739130434783, "grad_norm": 0.15096504245333736, "learning_rate": 7.791363988240544e-05, "loss": 0.9703, "step": 2130 }, { "epoch": 0.9651268115942029, "grad_norm": 0.1592642692077975, "learning_rate": 7.79096067378782e-05, "loss": 0.9783, "step": 2131 }, { "epoch": 0.9655797101449275, "grad_norm": 0.1457119055183816, "learning_rate": 7.790556980346752e-05, "loss": 0.9683, "step": 2132 }, { "epoch": 0.9660326086956522, "grad_norm": 0.1664500014461271, "learning_rate": 7.790152907957697e-05, "loss": 0.9758, "step": 2133 }, { "epoch": 0.9664855072463768, "grad_norm": 0.14674441702720284, "learning_rate": 7.789748456661047e-05, "loss": 0.9454, "step": 2134 }, { "epoch": 0.9669384057971014, "grad_norm": 0.11356392371851412, "learning_rate": 7.78934362649724e-05, "loss": 0.9666, "step": 2135 }, { "epoch": 0.967391304347826, "grad_norm": 0.1108158948708917, "learning_rate": 7.788938417506746e-05, "loss": 0.96, "step": 2136 }, { "epoch": 0.9678442028985508, "grad_norm": 0.1249756600641402, "learning_rate": 7.788532829730074e-05, "loss": 0.9639, "step": 2137 }, { "epoch": 0.9682971014492754, "grad_norm": 0.10774819220746742, "learning_rate": 7.78812686320777e-05, "loss": 0.9637, "step": 2138 }, { "epoch": 0.96875, "grad_norm": 0.10595769408749599, "learning_rate": 7.787720517980424e-05, "loss": 0.9836, "step": 2139 }, { "epoch": 0.9692028985507246, "grad_norm": 0.11594604067878109, "learning_rate": 7.787313794088653e-05, "loss": 0.9364, "step": 2140 }, { "epoch": 0.9696557971014492, "grad_norm": 0.1113559905332668, "learning_rate": 7.786906691573121e-05, "loss": 0.9588, "step": 2141 }, { "epoch": 0.970108695652174, "grad_norm": 0.12729945421188865, "learning_rate": 7.786499210474527e-05, "loss": 0.9563, "step": 2142 }, { "epoch": 0.9705615942028986, "grad_norm": 0.15509103591105688, "learning_rate": 7.786091350833606e-05, "loss": 0.9727, "step": 2143 }, { "epoch": 0.9710144927536232, "grad_norm": 0.192857459683204, "learning_rate": 7.785683112691135e-05, "loss": 0.9913, "step": 2144 }, { "epoch": 0.9714673913043478, "grad_norm": 0.1447187657378833, "learning_rate": 7.785274496087921e-05, "loss": 0.9492, "step": 2145 }, { "epoch": 0.9719202898550725, "grad_norm": 0.14258930132848485, "learning_rate": 7.784865501064818e-05, "loss": 0.9621, "step": 2146 }, { "epoch": 0.9723731884057971, "grad_norm": 0.24266889830507216, "learning_rate": 7.784456127662715e-05, "loss": 0.9821, "step": 2147 }, { "epoch": 0.9728260869565217, "grad_norm": 0.2577383545123014, "learning_rate": 7.784046375922536e-05, "loss": 0.9654, "step": 2148 }, { "epoch": 0.9732789855072463, "grad_norm": 0.33119231048603803, "learning_rate": 7.783636245885243e-05, "loss": 0.9916, "step": 2149 }, { "epoch": 0.9737318840579711, "grad_norm": 0.217767871751981, "learning_rate": 7.783225737591838e-05, "loss": 0.9679, "step": 2150 }, { "epoch": 0.9741847826086957, "grad_norm": 0.21979465236888643, "learning_rate": 7.782814851083364e-05, "loss": 0.975, "step": 2151 }, { "epoch": 0.9746376811594203, "grad_norm": 0.26525973423524235, "learning_rate": 7.782403586400894e-05, "loss": 0.9711, "step": 2152 }, { "epoch": 0.9750905797101449, "grad_norm": 0.2501350198904765, "learning_rate": 7.781991943585542e-05, "loss": 0.9377, "step": 2153 }, { "epoch": 0.9755434782608695, "grad_norm": 0.21749194961459664, "learning_rate": 7.781579922678463e-05, "loss": 0.9826, "step": 2154 }, { "epoch": 0.9759963768115942, "grad_norm": 0.20976870750673815, "learning_rate": 7.781167523720848e-05, "loss": 0.9788, "step": 2155 }, { "epoch": 0.9764492753623188, "grad_norm": 0.1851903541824173, "learning_rate": 7.780754746753922e-05, "loss": 0.9811, "step": 2156 }, { "epoch": 0.9769021739130435, "grad_norm": 0.4744801727898341, "learning_rate": 7.780341591818954e-05, "loss": 0.985, "step": 2157 }, { "epoch": 0.9773550724637681, "grad_norm": 0.18752978008719567, "learning_rate": 7.779928058957247e-05, "loss": 0.9485, "step": 2158 }, { "epoch": 0.9778079710144928, "grad_norm": 0.191283028383583, "learning_rate": 7.779514148210141e-05, "loss": 0.9864, "step": 2159 }, { "epoch": 0.9782608695652174, "grad_norm": 0.18411080805973634, "learning_rate": 7.779099859619017e-05, "loss": 0.9739, "step": 2160 }, { "epoch": 0.978713768115942, "grad_norm": 0.16480241113219865, "learning_rate": 7.778685193225292e-05, "loss": 0.965, "step": 2161 }, { "epoch": 0.9791666666666666, "grad_norm": 0.14009379153805898, "learning_rate": 7.77827014907042e-05, "loss": 0.9856, "step": 2162 }, { "epoch": 0.9796195652173914, "grad_norm": 0.16879215242314602, "learning_rate": 7.777854727195892e-05, "loss": 0.9645, "step": 2163 }, { "epoch": 0.980072463768116, "grad_norm": 0.19651474862141097, "learning_rate": 7.777438927643244e-05, "loss": 0.9679, "step": 2164 }, { "epoch": 0.9805253623188406, "grad_norm": 0.17009978560233788, "learning_rate": 7.777022750454039e-05, "loss": 0.9819, "step": 2165 }, { "epoch": 0.9809782608695652, "grad_norm": 0.1623129019871757, "learning_rate": 7.776606195669883e-05, "loss": 0.9527, "step": 2166 }, { "epoch": 0.9814311594202898, "grad_norm": 0.18997960121392135, "learning_rate": 7.776189263332423e-05, "loss": 0.9666, "step": 2167 }, { "epoch": 0.9818840579710145, "grad_norm": 0.18704131103785795, "learning_rate": 7.775771953483338e-05, "loss": 0.9509, "step": 2168 }, { "epoch": 0.9823369565217391, "grad_norm": 0.1488897412729749, "learning_rate": 7.775354266164348e-05, "loss": 0.9782, "step": 2169 }, { "epoch": 0.9827898550724637, "grad_norm": 0.17499187191879134, "learning_rate": 7.774936201417209e-05, "loss": 0.9811, "step": 2170 }, { "epoch": 0.9832427536231884, "grad_norm": 0.15053077254152503, "learning_rate": 7.774517759283715e-05, "loss": 0.9542, "step": 2171 }, { "epoch": 0.9836956521739131, "grad_norm": 0.3179150001200943, "learning_rate": 7.774098939805701e-05, "loss": 0.9811, "step": 2172 }, { "epoch": 0.9841485507246377, "grad_norm": 0.10706913006464125, "learning_rate": 7.773679743025033e-05, "loss": 0.9562, "step": 2173 }, { "epoch": 0.9846014492753623, "grad_norm": 0.13985657387099074, "learning_rate": 7.773260168983624e-05, "loss": 0.9502, "step": 2174 }, { "epoch": 0.9850543478260869, "grad_norm": 0.16349544421387813, "learning_rate": 7.772840217723415e-05, "loss": 0.9683, "step": 2175 }, { "epoch": 0.9855072463768116, "grad_norm": 0.14058168758723036, "learning_rate": 7.772419889286392e-05, "loss": 0.9857, "step": 2176 }, { "epoch": 0.9859601449275363, "grad_norm": 0.18062003061800275, "learning_rate": 7.771999183714575e-05, "loss": 0.9666, "step": 2177 }, { "epoch": 0.9864130434782609, "grad_norm": 0.19691615214017658, "learning_rate": 7.771578101050021e-05, "loss": 0.9693, "step": 2178 }, { "epoch": 0.9868659420289855, "grad_norm": 0.19569816047948996, "learning_rate": 7.771156641334829e-05, "loss": 0.985, "step": 2179 }, { "epoch": 0.9873188405797102, "grad_norm": 0.17671904111799563, "learning_rate": 7.770734804611129e-05, "loss": 0.9649, "step": 2180 }, { "epoch": 0.9877717391304348, "grad_norm": 0.14562223940512736, "learning_rate": 7.770312590921098e-05, "loss": 0.9614, "step": 2181 }, { "epoch": 0.9882246376811594, "grad_norm": 0.1375108887678883, "learning_rate": 7.769890000306943e-05, "loss": 1.008, "step": 2182 }, { "epoch": 0.988677536231884, "grad_norm": 0.10761003093259917, "learning_rate": 7.769467032810909e-05, "loss": 0.9737, "step": 2183 }, { "epoch": 0.9891304347826086, "grad_norm": 0.1467152225130261, "learning_rate": 7.769043688475283e-05, "loss": 0.9765, "step": 2184 }, { "epoch": 0.9895833333333334, "grad_norm": 0.17017946420429492, "learning_rate": 7.768619967342386e-05, "loss": 0.9759, "step": 2185 }, { "epoch": 0.990036231884058, "grad_norm": 0.1674513724529522, "learning_rate": 7.76819586945458e-05, "loss": 0.9682, "step": 2186 }, { "epoch": 0.9904891304347826, "grad_norm": 0.16204006377758826, "learning_rate": 7.767771394854263e-05, "loss": 0.9439, "step": 2187 }, { "epoch": 0.9909420289855072, "grad_norm": 0.1788112779323956, "learning_rate": 7.767346543583867e-05, "loss": 0.9777, "step": 2188 }, { "epoch": 0.9913949275362319, "grad_norm": 0.19603624154985302, "learning_rate": 7.766921315685868e-05, "loss": 0.9582, "step": 2189 }, { "epoch": 0.9918478260869565, "grad_norm": 0.2423217514378535, "learning_rate": 7.766495711202777e-05, "loss": 0.9524, "step": 2190 }, { "epoch": 0.9923007246376812, "grad_norm": 0.2813756425343392, "learning_rate": 7.766069730177139e-05, "loss": 0.9448, "step": 2191 }, { "epoch": 0.9927536231884058, "grad_norm": 0.2855227132294126, "learning_rate": 7.765643372651543e-05, "loss": 0.9745, "step": 2192 }, { "epoch": 0.9932065217391305, "grad_norm": 0.2885320170002305, "learning_rate": 7.765216638668612e-05, "loss": 0.9736, "step": 2193 }, { "epoch": 0.9936594202898551, "grad_norm": 0.24963647586206023, "learning_rate": 7.764789528271008e-05, "loss": 0.9584, "step": 2194 }, { "epoch": 0.9941123188405797, "grad_norm": 0.15826246107494463, "learning_rate": 7.764362041501429e-05, "loss": 0.9657, "step": 2195 }, { "epoch": 0.9945652173913043, "grad_norm": 0.1226230315922524, "learning_rate": 7.763934178402611e-05, "loss": 0.9753, "step": 2196 }, { "epoch": 0.9950181159420289, "grad_norm": 0.14231546274330137, "learning_rate": 7.763505939017328e-05, "loss": 0.9748, "step": 2197 }, { "epoch": 0.9954710144927537, "grad_norm": 0.22110768292032987, "learning_rate": 7.763077323388393e-05, "loss": 0.9774, "step": 2198 }, { "epoch": 0.9959239130434783, "grad_norm": 0.27297152966645577, "learning_rate": 7.762648331558656e-05, "loss": 0.9809, "step": 2199 }, { "epoch": 0.9963768115942029, "grad_norm": 0.2754920088446749, "learning_rate": 7.762218963571002e-05, "loss": 0.9594, "step": 2200 }, { "epoch": 0.9968297101449275, "grad_norm": 0.22820979472743255, "learning_rate": 7.761789219468356e-05, "loss": 0.9607, "step": 2201 }, { "epoch": 0.9972826086956522, "grad_norm": 0.20983811054417476, "learning_rate": 7.76135909929368e-05, "loss": 0.9633, "step": 2202 }, { "epoch": 0.9977355072463768, "grad_norm": 0.2090704435910062, "learning_rate": 7.760928603089975e-05, "loss": 0.954, "step": 2203 }, { "epoch": 0.9981884057971014, "grad_norm": 0.20910216144396826, "learning_rate": 7.760497730900279e-05, "loss": 0.9626, "step": 2204 }, { "epoch": 0.998641304347826, "grad_norm": 0.17765365642652545, "learning_rate": 7.760066482767665e-05, "loss": 0.9386, "step": 2205 }, { "epoch": 0.9990942028985508, "grad_norm": 0.1298906073434766, "learning_rate": 7.759634858735247e-05, "loss": 0.9521, "step": 2206 }, { "epoch": 0.9995471014492754, "grad_norm": 0.14664440067979953, "learning_rate": 7.759202858846174e-05, "loss": 0.9415, "step": 2207 }, { "epoch": 1.0, "grad_norm": 0.18811593907167923, "learning_rate": 7.758770483143634e-05, "loss": 0.9577, "step": 2208 }, { "epoch": 1.0004528985507246, "grad_norm": 0.2072899959157164, "learning_rate": 7.758337731670854e-05, "loss": 0.9394, "step": 2209 }, { "epoch": 1.0009057971014492, "grad_norm": 0.2126511864880672, "learning_rate": 7.757904604471093e-05, "loss": 0.9622, "step": 2210 }, { "epoch": 1.0013586956521738, "grad_norm": 0.1755607667702462, "learning_rate": 7.757471101587655e-05, "loss": 0.9384, "step": 2211 }, { "epoch": 1.0018115942028984, "grad_norm": 0.1683471295343361, "learning_rate": 7.757037223063877e-05, "loss": 0.9612, "step": 2212 }, { "epoch": 1.0022644927536233, "grad_norm": 0.16412724674491247, "learning_rate": 7.756602968943133e-05, "loss": 0.9576, "step": 2213 }, { "epoch": 1.002717391304348, "grad_norm": 0.14767787335387123, "learning_rate": 7.75616833926884e-05, "loss": 0.9536, "step": 2214 }, { "epoch": 1.0031702898550725, "grad_norm": 0.1151355512514775, "learning_rate": 7.755733334084443e-05, "loss": 0.9549, "step": 2215 }, { "epoch": 1.0036231884057971, "grad_norm": 0.12490931261475088, "learning_rate": 7.755297953433435e-05, "loss": 0.9848, "step": 2216 }, { "epoch": 1.0040760869565217, "grad_norm": 0.12123320060930465, "learning_rate": 7.75486219735934e-05, "loss": 0.9726, "step": 2217 }, { "epoch": 1.0045289855072463, "grad_norm": 0.09997417180242399, "learning_rate": 7.754426065905719e-05, "loss": 0.9464, "step": 2218 }, { "epoch": 1.004981884057971, "grad_norm": 0.09582936690889308, "learning_rate": 7.753989559116177e-05, "loss": 0.9431, "step": 2219 }, { "epoch": 1.0054347826086956, "grad_norm": 0.11279444149793602, "learning_rate": 7.753552677034349e-05, "loss": 0.9368, "step": 2220 }, { "epoch": 1.0058876811594204, "grad_norm": 0.13579228323372877, "learning_rate": 7.75311541970391e-05, "loss": 0.9617, "step": 2221 }, { "epoch": 1.006340579710145, "grad_norm": 0.1302603328073141, "learning_rate": 7.752677787168578e-05, "loss": 0.9489, "step": 2222 }, { "epoch": 1.0067934782608696, "grad_norm": 0.10811425330994417, "learning_rate": 7.752239779472101e-05, "loss": 0.943, "step": 2223 }, { "epoch": 1.0072463768115942, "grad_norm": 0.11826865419861929, "learning_rate": 7.751801396658266e-05, "loss": 0.9499, "step": 2224 }, { "epoch": 1.0076992753623188, "grad_norm": 0.1167718501642738, "learning_rate": 7.7513626387709e-05, "loss": 0.9497, "step": 2225 }, { "epoch": 1.0081521739130435, "grad_norm": 0.11879197697286803, "learning_rate": 7.750923505853868e-05, "loss": 0.952, "step": 2226 }, { "epoch": 1.008605072463768, "grad_norm": 0.42065179428256144, "learning_rate": 7.750483997951068e-05, "loss": 0.99, "step": 2227 }, { "epoch": 1.0090579710144927, "grad_norm": 0.1047939721145808, "learning_rate": 7.75004411510644e-05, "loss": 0.9502, "step": 2228 }, { "epoch": 1.0095108695652173, "grad_norm": 0.11631318051018574, "learning_rate": 7.749603857363958e-05, "loss": 0.9663, "step": 2229 }, { "epoch": 1.0099637681159421, "grad_norm": 0.11053043402006212, "learning_rate": 7.74916322476764e-05, "loss": 0.9682, "step": 2230 }, { "epoch": 1.0104166666666667, "grad_norm": 0.1364067933021783, "learning_rate": 7.74872221736153e-05, "loss": 0.9482, "step": 2231 }, { "epoch": 1.0108695652173914, "grad_norm": 0.1581332192219068, "learning_rate": 7.748280835189721e-05, "loss": 0.9681, "step": 2232 }, { "epoch": 1.011322463768116, "grad_norm": 0.17504249450890919, "learning_rate": 7.747839078296337e-05, "loss": 0.9559, "step": 2233 }, { "epoch": 1.0117753623188406, "grad_norm": 0.21645100986761565, "learning_rate": 7.747396946725541e-05, "loss": 0.9626, "step": 2234 }, { "epoch": 1.0122282608695652, "grad_norm": 0.167599093309333, "learning_rate": 7.746954440521535e-05, "loss": 0.9528, "step": 2235 }, { "epoch": 1.0126811594202898, "grad_norm": 0.18106903411492836, "learning_rate": 7.746511559728556e-05, "loss": 0.9412, "step": 2236 }, { "epoch": 1.0131340579710144, "grad_norm": 0.16499319832156994, "learning_rate": 7.746068304390878e-05, "loss": 0.9657, "step": 2237 }, { "epoch": 1.013586956521739, "grad_norm": 0.12046175084482308, "learning_rate": 7.745624674552817e-05, "loss": 0.9532, "step": 2238 }, { "epoch": 1.0140398550724639, "grad_norm": 0.13325649939533124, "learning_rate": 7.745180670258722e-05, "loss": 0.9579, "step": 2239 }, { "epoch": 1.0144927536231885, "grad_norm": 0.14593416322510833, "learning_rate": 7.744736291552979e-05, "loss": 0.9544, "step": 2240 }, { "epoch": 1.014945652173913, "grad_norm": 0.15979594549977996, "learning_rate": 7.744291538480015e-05, "loss": 0.986, "step": 2241 }, { "epoch": 1.0153985507246377, "grad_norm": 0.14000700521224188, "learning_rate": 7.743846411084294e-05, "loss": 0.9518, "step": 2242 }, { "epoch": 1.0158514492753623, "grad_norm": 0.15633995064671574, "learning_rate": 7.743400909410315e-05, "loss": 0.9612, "step": 2243 }, { "epoch": 1.016304347826087, "grad_norm": 0.18375150798324869, "learning_rate": 7.742955033502613e-05, "loss": 0.9536, "step": 2244 }, { "epoch": 1.0167572463768115, "grad_norm": 0.20359971448501163, "learning_rate": 7.742508783405767e-05, "loss": 0.9604, "step": 2245 }, { "epoch": 1.0172101449275361, "grad_norm": 0.1929816023659054, "learning_rate": 7.742062159164387e-05, "loss": 0.9377, "step": 2246 }, { "epoch": 1.017663043478261, "grad_norm": 0.1460190966173737, "learning_rate": 7.741615160823122e-05, "loss": 0.9766, "step": 2247 }, { "epoch": 1.0181159420289856, "grad_norm": 0.12041556069192189, "learning_rate": 7.74116778842666e-05, "loss": 0.9551, "step": 2248 }, { "epoch": 1.0185688405797102, "grad_norm": 0.11628317585736307, "learning_rate": 7.740720042019727e-05, "loss": 0.962, "step": 2249 }, { "epoch": 1.0190217391304348, "grad_norm": 0.10676332318663428, "learning_rate": 7.740271921647085e-05, "loss": 0.9406, "step": 2250 }, { "epoch": 1.0194746376811594, "grad_norm": 0.12579982458933003, "learning_rate": 7.73982342735353e-05, "loss": 0.957, "step": 2251 }, { "epoch": 1.019927536231884, "grad_norm": 0.13191925666813614, "learning_rate": 7.739374559183901e-05, "loss": 0.9796, "step": 2252 }, { "epoch": 1.0203804347826086, "grad_norm": 0.12413385983116323, "learning_rate": 7.738925317183072e-05, "loss": 0.9607, "step": 2253 }, { "epoch": 1.0208333333333333, "grad_norm": 0.14093585240474205, "learning_rate": 7.738475701395955e-05, "loss": 0.9731, "step": 2254 }, { "epoch": 1.0212862318840579, "grad_norm": 0.15619570084161763, "learning_rate": 7.738025711867496e-05, "loss": 0.9608, "step": 2255 }, { "epoch": 1.0217391304347827, "grad_norm": 0.14814327059233307, "learning_rate": 7.737575348642685e-05, "loss": 0.9409, "step": 2256 }, { "epoch": 1.0221920289855073, "grad_norm": 0.12956042614257565, "learning_rate": 7.737124611766543e-05, "loss": 0.9497, "step": 2257 }, { "epoch": 1.022644927536232, "grad_norm": 0.1352386859220854, "learning_rate": 7.736673501284131e-05, "loss": 0.9603, "step": 2258 }, { "epoch": 1.0230978260869565, "grad_norm": 0.14642088797265979, "learning_rate": 7.736222017240549e-05, "loss": 0.9662, "step": 2259 }, { "epoch": 1.0235507246376812, "grad_norm": 0.16156232669392245, "learning_rate": 7.735770159680931e-05, "loss": 0.9449, "step": 2260 }, { "epoch": 1.0240036231884058, "grad_norm": 0.1636195711596718, "learning_rate": 7.735317928650451e-05, "loss": 0.9756, "step": 2261 }, { "epoch": 1.0244565217391304, "grad_norm": 0.2074928547660634, "learning_rate": 7.734865324194317e-05, "loss": 0.9664, "step": 2262 }, { "epoch": 1.024909420289855, "grad_norm": 0.25457654630952486, "learning_rate": 7.734412346357779e-05, "loss": 0.9567, "step": 2263 }, { "epoch": 1.0253623188405796, "grad_norm": 0.29757791664965055, "learning_rate": 7.73395899518612e-05, "loss": 0.9727, "step": 2264 }, { "epoch": 1.0258152173913044, "grad_norm": 0.291135258349426, "learning_rate": 7.733505270724665e-05, "loss": 0.9627, "step": 2265 }, { "epoch": 1.026268115942029, "grad_norm": 0.25800045989637294, "learning_rate": 7.733051173018772e-05, "loss": 0.9632, "step": 2266 }, { "epoch": 1.0267210144927537, "grad_norm": 0.18493602934746206, "learning_rate": 7.732596702113837e-05, "loss": 0.952, "step": 2267 }, { "epoch": 1.0271739130434783, "grad_norm": 0.12146090924238681, "learning_rate": 7.732141858055294e-05, "loss": 0.9781, "step": 2268 }, { "epoch": 1.0276268115942029, "grad_norm": 0.15169948722409043, "learning_rate": 7.731686640888617e-05, "loss": 0.9714, "step": 2269 }, { "epoch": 1.0280797101449275, "grad_norm": 0.1818807933199238, "learning_rate": 7.731231050659312e-05, "loss": 0.9552, "step": 2270 }, { "epoch": 1.028532608695652, "grad_norm": 0.21391272777346562, "learning_rate": 7.730775087412926e-05, "loss": 0.9334, "step": 2271 }, { "epoch": 1.0289855072463767, "grad_norm": 0.2698971039020036, "learning_rate": 7.730318751195043e-05, "loss": 0.9573, "step": 2272 }, { "epoch": 1.0294384057971016, "grad_norm": 0.2967462879105615, "learning_rate": 7.729862042051283e-05, "loss": 0.9443, "step": 2273 }, { "epoch": 1.0298913043478262, "grad_norm": 0.2602082893499258, "learning_rate": 7.729404960027305e-05, "loss": 0.9592, "step": 2274 }, { "epoch": 1.0303442028985508, "grad_norm": 0.1984447184646275, "learning_rate": 7.728947505168803e-05, "loss": 0.9723, "step": 2275 }, { "epoch": 1.0307971014492754, "grad_norm": 0.15107447679917266, "learning_rate": 7.72848967752151e-05, "loss": 0.9229, "step": 2276 }, { "epoch": 1.03125, "grad_norm": 0.16004797500918838, "learning_rate": 7.728031477131195e-05, "loss": 0.9654, "step": 2277 }, { "epoch": 1.0317028985507246, "grad_norm": 0.1773566229324036, "learning_rate": 7.727572904043666e-05, "loss": 0.9538, "step": 2278 }, { "epoch": 1.0321557971014492, "grad_norm": 0.21741864092638843, "learning_rate": 7.727113958304767e-05, "loss": 0.961, "step": 2279 }, { "epoch": 1.0326086956521738, "grad_norm": 0.22931506008170432, "learning_rate": 7.726654639960379e-05, "loss": 0.9726, "step": 2280 }, { "epoch": 1.0330615942028984, "grad_norm": 0.2248047536767838, "learning_rate": 7.726194949056421e-05, "loss": 0.9678, "step": 2281 }, { "epoch": 1.0335144927536233, "grad_norm": 0.22510037269820682, "learning_rate": 7.725734885638852e-05, "loss": 0.9476, "step": 2282 }, { "epoch": 1.033967391304348, "grad_norm": 0.18461879047093785, "learning_rate": 7.725274449753659e-05, "loss": 0.9603, "step": 2283 }, { "epoch": 1.0344202898550725, "grad_norm": 0.13416401213371518, "learning_rate": 7.724813641446878e-05, "loss": 0.9636, "step": 2284 }, { "epoch": 1.0348731884057971, "grad_norm": 0.10279972532576621, "learning_rate": 7.724352460764574e-05, "loss": 0.9714, "step": 2285 }, { "epoch": 1.0353260869565217, "grad_norm": 0.11232815627405092, "learning_rate": 7.723890907752854e-05, "loss": 0.9428, "step": 2286 }, { "epoch": 1.0357789855072463, "grad_norm": 0.10091985708211415, "learning_rate": 7.723428982457857e-05, "loss": 0.9827, "step": 2287 }, { "epoch": 1.036231884057971, "grad_norm": 0.11705270517757312, "learning_rate": 7.722966684925766e-05, "loss": 0.9529, "step": 2288 }, { "epoch": 1.0366847826086956, "grad_norm": 0.15367296538042885, "learning_rate": 7.722504015202795e-05, "loss": 0.9418, "step": 2289 }, { "epoch": 1.0371376811594204, "grad_norm": 0.1672794130311585, "learning_rate": 7.722040973335199e-05, "loss": 0.971, "step": 2290 }, { "epoch": 1.037590579710145, "grad_norm": 0.1393728526950105, "learning_rate": 7.72157755936927e-05, "loss": 0.9996, "step": 2291 }, { "epoch": 1.0380434782608696, "grad_norm": 0.11219483562148898, "learning_rate": 7.721113773351333e-05, "loss": 0.9719, "step": 2292 }, { "epoch": 1.0384963768115942, "grad_norm": 0.12970663406243937, "learning_rate": 7.720649615327757e-05, "loss": 0.9589, "step": 2293 }, { "epoch": 1.0389492753623188, "grad_norm": 0.13062919869284192, "learning_rate": 7.720185085344942e-05, "loss": 0.9593, "step": 2294 }, { "epoch": 1.0394021739130435, "grad_norm": 0.16134227609772597, "learning_rate": 7.71972018344933e-05, "loss": 0.9447, "step": 2295 }, { "epoch": 1.039855072463768, "grad_norm": 0.18372837427507258, "learning_rate": 7.719254909687396e-05, "loss": 0.9603, "step": 2296 }, { "epoch": 1.0403079710144927, "grad_norm": 0.16079326434904406, "learning_rate": 7.718789264105655e-05, "loss": 0.9662, "step": 2297 }, { "epoch": 1.0407608695652173, "grad_norm": 0.1561153666820718, "learning_rate": 7.71832324675066e-05, "loss": 0.971, "step": 2298 }, { "epoch": 1.0412137681159421, "grad_norm": 0.1630272878357507, "learning_rate": 7.717856857668996e-05, "loss": 0.97, "step": 2299 }, { "epoch": 1.0416666666666667, "grad_norm": 0.13705424270465888, "learning_rate": 7.71739009690729e-05, "loss": 0.9602, "step": 2300 }, { "epoch": 1.0421195652173914, "grad_norm": 0.13799734098420574, "learning_rate": 7.716922964512206e-05, "loss": 0.9345, "step": 2301 }, { "epoch": 1.042572463768116, "grad_norm": 0.12189957861681999, "learning_rate": 7.716455460530444e-05, "loss": 0.9543, "step": 2302 }, { "epoch": 1.0430253623188406, "grad_norm": 0.11897683776903129, "learning_rate": 7.71598758500874e-05, "loss": 0.9539, "step": 2303 }, { "epoch": 1.0434782608695652, "grad_norm": 0.14918660150425034, "learning_rate": 7.715519337993869e-05, "loss": 0.97, "step": 2304 }, { "epoch": 1.0439311594202898, "grad_norm": 0.12805302689624196, "learning_rate": 7.715050719532642e-05, "loss": 0.9424, "step": 2305 }, { "epoch": 1.0443840579710144, "grad_norm": 0.17230731759001952, "learning_rate": 7.714581729671907e-05, "loss": 0.974, "step": 2306 }, { "epoch": 1.044836956521739, "grad_norm": 0.14436281905556048, "learning_rate": 7.71411236845855e-05, "loss": 0.955, "step": 2307 }, { "epoch": 1.0452898550724639, "grad_norm": 0.1351593402351763, "learning_rate": 7.713642635939493e-05, "loss": 0.9532, "step": 2308 }, { "epoch": 1.0457427536231885, "grad_norm": 0.15182980378429375, "learning_rate": 7.7131725321617e-05, "loss": 0.954, "step": 2309 }, { "epoch": 1.046195652173913, "grad_norm": 0.11303025508419161, "learning_rate": 7.712702057172164e-05, "loss": 0.9546, "step": 2310 }, { "epoch": 1.0466485507246377, "grad_norm": 0.10460906399940414, "learning_rate": 7.712231211017917e-05, "loss": 0.9344, "step": 2311 }, { "epoch": 1.0471014492753623, "grad_norm": 0.12930793455057374, "learning_rate": 7.711759993746037e-05, "loss": 0.9378, "step": 2312 }, { "epoch": 1.047554347826087, "grad_norm": 0.12568036215730463, "learning_rate": 7.711288405403628e-05, "loss": 0.9545, "step": 2313 }, { "epoch": 1.0480072463768115, "grad_norm": 0.15359890055933895, "learning_rate": 7.710816446037835e-05, "loss": 0.9544, "step": 2314 }, { "epoch": 1.0484601449275361, "grad_norm": 0.18880173793736765, "learning_rate": 7.710344115695841e-05, "loss": 0.9722, "step": 2315 }, { "epoch": 1.048913043478261, "grad_norm": 0.2164984045047379, "learning_rate": 7.709871414424869e-05, "loss": 0.9496, "step": 2316 }, { "epoch": 1.0493659420289856, "grad_norm": 0.18402211221715992, "learning_rate": 7.709398342272169e-05, "loss": 0.9687, "step": 2317 }, { "epoch": 1.0498188405797102, "grad_norm": 0.14147931964948418, "learning_rate": 7.708924899285042e-05, "loss": 0.9312, "step": 2318 }, { "epoch": 1.0502717391304348, "grad_norm": 0.12279294726477041, "learning_rate": 7.708451085510813e-05, "loss": 0.9671, "step": 2319 }, { "epoch": 1.0507246376811594, "grad_norm": 0.10589769878868016, "learning_rate": 7.707976900996854e-05, "loss": 0.9417, "step": 2320 }, { "epoch": 1.051177536231884, "grad_norm": 0.13522230212402797, "learning_rate": 7.707502345790568e-05, "loss": 0.9586, "step": 2321 }, { "epoch": 1.0516304347826086, "grad_norm": 0.16557472475277202, "learning_rate": 7.707027419939398e-05, "loss": 0.9558, "step": 2322 }, { "epoch": 1.0520833333333333, "grad_norm": 0.16735866510926908, "learning_rate": 7.706552123490822e-05, "loss": 0.9641, "step": 2323 }, { "epoch": 1.0525362318840579, "grad_norm": 0.17154450959355164, "learning_rate": 7.706076456492356e-05, "loss": 0.9503, "step": 2324 }, { "epoch": 1.0529891304347827, "grad_norm": 0.15082451776533254, "learning_rate": 7.705600418991556e-05, "loss": 0.9667, "step": 2325 }, { "epoch": 1.0534420289855073, "grad_norm": 0.13080418184430684, "learning_rate": 7.705124011036008e-05, "loss": 0.9745, "step": 2326 }, { "epoch": 1.053894927536232, "grad_norm": 0.11817402603074241, "learning_rate": 7.704647232673342e-05, "loss": 0.9635, "step": 2327 }, { "epoch": 1.0543478260869565, "grad_norm": 0.11597444487208651, "learning_rate": 7.704170083951221e-05, "loss": 0.9631, "step": 2328 }, { "epoch": 1.0548007246376812, "grad_norm": 0.13865801173591225, "learning_rate": 7.703692564917348e-05, "loss": 0.9584, "step": 2329 }, { "epoch": 1.0552536231884058, "grad_norm": 0.143470586768955, "learning_rate": 7.70321467561946e-05, "loss": 0.9408, "step": 2330 }, { "epoch": 1.0557065217391304, "grad_norm": 0.1532663052324727, "learning_rate": 7.702736416105333e-05, "loss": 0.9669, "step": 2331 }, { "epoch": 1.056159420289855, "grad_norm": 0.22105434920215547, "learning_rate": 7.702257786422778e-05, "loss": 0.9336, "step": 2332 }, { "epoch": 1.0566123188405796, "grad_norm": 0.24789429988523953, "learning_rate": 7.701778786619646e-05, "loss": 0.9677, "step": 2333 }, { "epoch": 1.0570652173913044, "grad_norm": 0.23505622801418868, "learning_rate": 7.701299416743823e-05, "loss": 0.972, "step": 2334 }, { "epoch": 1.057518115942029, "grad_norm": 0.22559414626239835, "learning_rate": 7.700819676843232e-05, "loss": 0.9692, "step": 2335 }, { "epoch": 1.0579710144927537, "grad_norm": 0.22689017483209425, "learning_rate": 7.700339566965833e-05, "loss": 0.9677, "step": 2336 }, { "epoch": 1.0584239130434783, "grad_norm": 0.38261478665816784, "learning_rate": 7.699859087159624e-05, "loss": 0.9961, "step": 2337 }, { "epoch": 1.0588768115942029, "grad_norm": 0.13044238106631448, "learning_rate": 7.699378237472639e-05, "loss": 0.9769, "step": 2338 }, { "epoch": 1.0593297101449275, "grad_norm": 0.11283562158239425, "learning_rate": 7.698897017952949e-05, "loss": 0.9674, "step": 2339 }, { "epoch": 1.059782608695652, "grad_norm": 0.10181173451951372, "learning_rate": 7.698415428648663e-05, "loss": 0.9457, "step": 2340 }, { "epoch": 1.0602355072463767, "grad_norm": 0.1176655235754415, "learning_rate": 7.697933469607927e-05, "loss": 0.9645, "step": 2341 }, { "epoch": 1.0606884057971016, "grad_norm": 0.1425577021888508, "learning_rate": 7.69745114087892e-05, "loss": 0.9591, "step": 2342 }, { "epoch": 1.0611413043478262, "grad_norm": 0.14138810797971987, "learning_rate": 7.696968442509866e-05, "loss": 0.9208, "step": 2343 }, { "epoch": 1.0615942028985508, "grad_norm": 0.12289591383578594, "learning_rate": 7.696485374549017e-05, "loss": 0.9497, "step": 2344 }, { "epoch": 1.0620471014492754, "grad_norm": 0.10322931156292904, "learning_rate": 7.696001937044668e-05, "loss": 0.9497, "step": 2345 }, { "epoch": 1.0625, "grad_norm": 0.11873810013523522, "learning_rate": 7.695518130045147e-05, "loss": 0.9683, "step": 2346 }, { "epoch": 1.0629528985507246, "grad_norm": 0.15289839645810263, "learning_rate": 7.695033953598825e-05, "loss": 0.9495, "step": 2347 }, { "epoch": 1.0634057971014492, "grad_norm": 0.18038962883351217, "learning_rate": 7.694549407754102e-05, "loss": 0.9765, "step": 2348 }, { "epoch": 1.0638586956521738, "grad_norm": 0.19359153483032554, "learning_rate": 7.69406449255942e-05, "loss": 0.962, "step": 2349 }, { "epoch": 1.0643115942028984, "grad_norm": 0.20918853099719403, "learning_rate": 7.693579208063256e-05, "loss": 0.9421, "step": 2350 }, { "epoch": 1.0647644927536233, "grad_norm": 0.18333331030602257, "learning_rate": 7.693093554314128e-05, "loss": 0.9534, "step": 2351 }, { "epoch": 1.065217391304348, "grad_norm": 0.15602486530680565, "learning_rate": 7.692607531360583e-05, "loss": 0.9553, "step": 2352 }, { "epoch": 1.0656702898550725, "grad_norm": 0.13202990934409284, "learning_rate": 7.692121139251213e-05, "loss": 0.9702, "step": 2353 }, { "epoch": 1.0661231884057971, "grad_norm": 0.13200799895343301, "learning_rate": 7.691634378034641e-05, "loss": 0.941, "step": 2354 }, { "epoch": 1.0665760869565217, "grad_norm": 0.17765464305950848, "learning_rate": 7.691147247759532e-05, "loss": 0.9406, "step": 2355 }, { "epoch": 1.0670289855072463, "grad_norm": 0.18296414494776816, "learning_rate": 7.690659748474583e-05, "loss": 0.9516, "step": 2356 }, { "epoch": 1.067481884057971, "grad_norm": 0.21722356384727995, "learning_rate": 7.690171880228529e-05, "loss": 0.953, "step": 2357 }, { "epoch": 1.0679347826086956, "grad_norm": 0.19685516772280803, "learning_rate": 7.689683643070145e-05, "loss": 0.9586, "step": 2358 }, { "epoch": 1.0683876811594204, "grad_norm": 0.16985163079064342, "learning_rate": 7.689195037048242e-05, "loss": 0.9517, "step": 2359 }, { "epoch": 1.068840579710145, "grad_norm": 0.16732528965975646, "learning_rate": 7.688706062211664e-05, "loss": 0.9715, "step": 2360 }, { "epoch": 1.0692934782608696, "grad_norm": 0.16926409900454328, "learning_rate": 7.688216718609296e-05, "loss": 0.9359, "step": 2361 }, { "epoch": 1.0697463768115942, "grad_norm": 0.16765506304301436, "learning_rate": 7.687727006290058e-05, "loss": 0.9738, "step": 2362 }, { "epoch": 1.0701992753623188, "grad_norm": 0.1545211733328081, "learning_rate": 7.687236925302908e-05, "loss": 0.9536, "step": 2363 }, { "epoch": 1.0706521739130435, "grad_norm": 0.13089193458481974, "learning_rate": 7.686746475696838e-05, "loss": 0.9626, "step": 2364 }, { "epoch": 1.071105072463768, "grad_norm": 0.16614560246672186, "learning_rate": 7.686255657520882e-05, "loss": 0.9507, "step": 2365 }, { "epoch": 1.0715579710144927, "grad_norm": 0.14640971203492054, "learning_rate": 7.685764470824106e-05, "loss": 0.9532, "step": 2366 }, { "epoch": 1.0720108695652173, "grad_norm": 0.14401251671950877, "learning_rate": 7.685272915655617e-05, "loss": 0.9613, "step": 2367 }, { "epoch": 1.0724637681159421, "grad_norm": 0.1413220310225703, "learning_rate": 7.684780992064553e-05, "loss": 0.9697, "step": 2368 }, { "epoch": 1.0729166666666667, "grad_norm": 0.1304669726406633, "learning_rate": 7.684288700100095e-05, "loss": 0.9954, "step": 2369 }, { "epoch": 1.0733695652173914, "grad_norm": 0.1449450856975981, "learning_rate": 7.683796039811457e-05, "loss": 0.9616, "step": 2370 }, { "epoch": 1.073822463768116, "grad_norm": 0.1585987782276191, "learning_rate": 7.68330301124789e-05, "loss": 0.9567, "step": 2371 }, { "epoch": 1.0742753623188406, "grad_norm": 0.14339020846675346, "learning_rate": 7.682809614458686e-05, "loss": 0.9696, "step": 2372 }, { "epoch": 1.0747282608695652, "grad_norm": 0.13357727814276862, "learning_rate": 7.682315849493169e-05, "loss": 0.9558, "step": 2373 }, { "epoch": 1.0751811594202898, "grad_norm": 0.12981060850206796, "learning_rate": 7.681821716400701e-05, "loss": 0.9695, "step": 2374 }, { "epoch": 1.0756340579710144, "grad_norm": 0.12120208086526021, "learning_rate": 7.681327215230682e-05, "loss": 0.9716, "step": 2375 }, { "epoch": 1.0760869565217392, "grad_norm": 0.11901906811852879, "learning_rate": 7.680832346032547e-05, "loss": 0.96, "step": 2376 }, { "epoch": 1.0765398550724639, "grad_norm": 0.1379279952026839, "learning_rate": 7.680337108855771e-05, "loss": 0.9372, "step": 2377 }, { "epoch": 1.0769927536231885, "grad_norm": 0.14038366977532143, "learning_rate": 7.679841503749861e-05, "loss": 0.9828, "step": 2378 }, { "epoch": 1.077445652173913, "grad_norm": 0.13829923162655744, "learning_rate": 7.679345530764365e-05, "loss": 0.9401, "step": 2379 }, { "epoch": 1.0778985507246377, "grad_norm": 0.1771409924692337, "learning_rate": 7.678849189948865e-05, "loss": 0.9391, "step": 2380 }, { "epoch": 1.0783514492753623, "grad_norm": 0.24247700319890378, "learning_rate": 7.678352481352983e-05, "loss": 0.9755, "step": 2381 }, { "epoch": 1.078804347826087, "grad_norm": 0.23105764286433594, "learning_rate": 7.677855405026375e-05, "loss": 0.955, "step": 2382 }, { "epoch": 1.0792572463768115, "grad_norm": 0.23845817400652178, "learning_rate": 7.677357961018732e-05, "loss": 0.9533, "step": 2383 }, { "epoch": 1.0797101449275361, "grad_norm": 0.23980104542628738, "learning_rate": 7.676860149379789e-05, "loss": 0.9565, "step": 2384 }, { "epoch": 1.0801630434782608, "grad_norm": 0.2143337913443182, "learning_rate": 7.676361970159308e-05, "loss": 0.9569, "step": 2385 }, { "epoch": 1.0806159420289856, "grad_norm": 0.17430591278358523, "learning_rate": 7.675863423407097e-05, "loss": 0.9486, "step": 2386 }, { "epoch": 1.0810688405797102, "grad_norm": 0.12373090846735119, "learning_rate": 7.675364509172993e-05, "loss": 0.9481, "step": 2387 }, { "epoch": 1.0815217391304348, "grad_norm": 0.10001364611149667, "learning_rate": 7.674865227506876e-05, "loss": 0.9663, "step": 2388 }, { "epoch": 1.0819746376811594, "grad_norm": 0.11101475070388342, "learning_rate": 7.67436557845866e-05, "loss": 0.9743, "step": 2389 }, { "epoch": 1.082427536231884, "grad_norm": 0.13528551236572037, "learning_rate": 7.673865562078292e-05, "loss": 0.9579, "step": 2390 }, { "epoch": 1.0828804347826086, "grad_norm": 0.15759323801448205, "learning_rate": 7.673365178415764e-05, "loss": 0.966, "step": 2391 }, { "epoch": 1.0833333333333333, "grad_norm": 0.14708353518228473, "learning_rate": 7.672864427521097e-05, "loss": 0.9654, "step": 2392 }, { "epoch": 1.0837862318840579, "grad_norm": 0.15879089823647197, "learning_rate": 7.672363309444353e-05, "loss": 0.9747, "step": 2393 }, { "epoch": 1.0842391304347827, "grad_norm": 0.13357473874988013, "learning_rate": 7.671861824235631e-05, "loss": 0.9254, "step": 2394 }, { "epoch": 1.0846920289855073, "grad_norm": 0.12629283742540842, "learning_rate": 7.671359971945062e-05, "loss": 0.9645, "step": 2395 }, { "epoch": 1.085144927536232, "grad_norm": 0.15089650079660782, "learning_rate": 7.67085775262282e-05, "loss": 0.9507, "step": 2396 }, { "epoch": 1.0855978260869565, "grad_norm": 0.18769892883960293, "learning_rate": 7.670355166319111e-05, "loss": 0.9831, "step": 2397 }, { "epoch": 1.0860507246376812, "grad_norm": 0.21660922804037974, "learning_rate": 7.669852213084181e-05, "loss": 0.9504, "step": 2398 }, { "epoch": 1.0865036231884058, "grad_norm": 0.20164953819702602, "learning_rate": 7.669348892968308e-05, "loss": 0.9531, "step": 2399 }, { "epoch": 1.0869565217391304, "grad_norm": 0.13424654832258684, "learning_rate": 7.668845206021812e-05, "loss": 0.9611, "step": 2400 }, { "epoch": 1.087409420289855, "grad_norm": 0.12792141704527238, "learning_rate": 7.668341152295048e-05, "loss": 0.957, "step": 2401 }, { "epoch": 1.0878623188405796, "grad_norm": 0.1237045437827422, "learning_rate": 7.667836731838406e-05, "loss": 0.9549, "step": 2402 }, { "epoch": 1.0883152173913044, "grad_norm": 0.10046015344470108, "learning_rate": 7.667331944702313e-05, "loss": 0.945, "step": 2403 }, { "epoch": 1.088768115942029, "grad_norm": 0.12511370088118648, "learning_rate": 7.666826790937235e-05, "loss": 0.9412, "step": 2404 }, { "epoch": 1.0892210144927537, "grad_norm": 0.13436493916279013, "learning_rate": 7.666321270593671e-05, "loss": 0.9414, "step": 2405 }, { "epoch": 1.0896739130434783, "grad_norm": 0.14574584816844208, "learning_rate": 7.66581538372216e-05, "loss": 0.9398, "step": 2406 }, { "epoch": 1.0901268115942029, "grad_norm": 0.14982681505891826, "learning_rate": 7.665309130373276e-05, "loss": 0.9338, "step": 2407 }, { "epoch": 1.0905797101449275, "grad_norm": 0.13788602114822968, "learning_rate": 7.664802510597631e-05, "loss": 0.9461, "step": 2408 }, { "epoch": 1.091032608695652, "grad_norm": 0.1592293397375054, "learning_rate": 7.664295524445872e-05, "loss": 0.9728, "step": 2409 }, { "epoch": 1.0914855072463767, "grad_norm": 0.17401822568700132, "learning_rate": 7.663788171968681e-05, "loss": 0.9668, "step": 2410 }, { "epoch": 1.0919384057971016, "grad_norm": 0.17153210891094525, "learning_rate": 7.663280453216782e-05, "loss": 0.9667, "step": 2411 }, { "epoch": 1.0923913043478262, "grad_norm": 0.166166531830347, "learning_rate": 7.66277236824093e-05, "loss": 0.9495, "step": 2412 }, { "epoch": 1.0928442028985508, "grad_norm": 0.15249279272058247, "learning_rate": 7.662263917091922e-05, "loss": 0.9667, "step": 2413 }, { "epoch": 1.0932971014492754, "grad_norm": 0.12758108649514477, "learning_rate": 7.661755099820586e-05, "loss": 0.931, "step": 2414 }, { "epoch": 1.09375, "grad_norm": 0.10358676700220341, "learning_rate": 7.66124591647779e-05, "loss": 0.9492, "step": 2415 }, { "epoch": 1.0942028985507246, "grad_norm": 0.1666989827193987, "learning_rate": 7.660736367114437e-05, "loss": 0.9425, "step": 2416 }, { "epoch": 1.0946557971014492, "grad_norm": 0.2283677976750808, "learning_rate": 7.66022645178147e-05, "loss": 0.9619, "step": 2417 }, { "epoch": 1.0951086956521738, "grad_norm": 0.26996634508588924, "learning_rate": 7.659716170529863e-05, "loss": 0.9624, "step": 2418 }, { "epoch": 1.0955615942028984, "grad_norm": 0.3078437693397308, "learning_rate": 7.659205523410633e-05, "loss": 0.944, "step": 2419 }, { "epoch": 1.0960144927536233, "grad_norm": 0.32359709510899454, "learning_rate": 7.658694510474826e-05, "loss": 0.9468, "step": 2420 }, { "epoch": 1.096467391304348, "grad_norm": 0.29096455228743806, "learning_rate": 7.658183131773533e-05, "loss": 0.9587, "step": 2421 }, { "epoch": 1.0969202898550725, "grad_norm": 0.25332780772379115, "learning_rate": 7.657671387357875e-05, "loss": 0.9558, "step": 2422 }, { "epoch": 1.0973731884057971, "grad_norm": 0.17375778909799772, "learning_rate": 7.65715927727901e-05, "loss": 0.9585, "step": 2423 }, { "epoch": 1.0978260869565217, "grad_norm": 0.14698400062339514, "learning_rate": 7.65664680158814e-05, "loss": 0.9532, "step": 2424 }, { "epoch": 1.0982789855072463, "grad_norm": 0.19877011361234462, "learning_rate": 7.656133960336495e-05, "loss": 0.9782, "step": 2425 }, { "epoch": 1.098731884057971, "grad_norm": 0.24086039217757976, "learning_rate": 7.655620753575342e-05, "loss": 0.9623, "step": 2426 }, { "epoch": 1.0991847826086956, "grad_norm": 0.2101913497708654, "learning_rate": 7.655107181355992e-05, "loss": 0.9628, "step": 2427 }, { "epoch": 1.0996376811594204, "grad_norm": 0.13914184942129545, "learning_rate": 7.654593243729782e-05, "loss": 0.9605, "step": 2428 }, { "epoch": 1.100090579710145, "grad_norm": 0.12725075101745542, "learning_rate": 7.654078940748096e-05, "loss": 0.9672, "step": 2429 }, { "epoch": 1.1005434782608696, "grad_norm": 0.17413743814785532, "learning_rate": 7.653564272462348e-05, "loss": 0.9411, "step": 2430 }, { "epoch": 1.1009963768115942, "grad_norm": 0.1851812549293689, "learning_rate": 7.65304923892399e-05, "loss": 0.9656, "step": 2431 }, { "epoch": 1.1014492753623188, "grad_norm": 0.18142776893973261, "learning_rate": 7.65253384018451e-05, "loss": 0.9258, "step": 2432 }, { "epoch": 1.1019021739130435, "grad_norm": 0.1434748391156394, "learning_rate": 7.652018076295435e-05, "loss": 0.9739, "step": 2433 }, { "epoch": 1.102355072463768, "grad_norm": 0.1631560508317579, "learning_rate": 7.651501947308326e-05, "loss": 0.9681, "step": 2434 }, { "epoch": 1.1028079710144927, "grad_norm": 0.18976212001488982, "learning_rate": 7.65098545327478e-05, "loss": 0.9327, "step": 2435 }, { "epoch": 1.1032608695652173, "grad_norm": 0.1711228023589692, "learning_rate": 7.650468594246433e-05, "loss": 0.9663, "step": 2436 }, { "epoch": 1.1037137681159421, "grad_norm": 0.12791772183249922, "learning_rate": 7.649951370274956e-05, "loss": 0.9614, "step": 2437 }, { "epoch": 1.1041666666666667, "grad_norm": 0.11606329345210582, "learning_rate": 7.649433781412058e-05, "loss": 0.95, "step": 2438 }, { "epoch": 1.1046195652173914, "grad_norm": 0.13635419082299619, "learning_rate": 7.64891582770948e-05, "loss": 0.9725, "step": 2439 }, { "epoch": 1.105072463768116, "grad_norm": 0.15833265683220174, "learning_rate": 7.648397509219005e-05, "loss": 0.9678, "step": 2440 }, { "epoch": 1.1055253623188406, "grad_norm": 0.1608099988204254, "learning_rate": 7.64787882599245e-05, "loss": 0.9479, "step": 2441 }, { "epoch": 1.1059782608695652, "grad_norm": 0.13688334712189898, "learning_rate": 7.647359778081669e-05, "loss": 0.9995, "step": 2442 }, { "epoch": 1.1064311594202898, "grad_norm": 0.12710830920894295, "learning_rate": 7.64684036553855e-05, "loss": 0.9756, "step": 2443 }, { "epoch": 1.1068840579710144, "grad_norm": 0.7633976643529357, "learning_rate": 7.646320588415021e-05, "loss": 0.9791, "step": 2444 }, { "epoch": 1.1073369565217392, "grad_norm": 0.142801483080266, "learning_rate": 7.645800446763044e-05, "loss": 0.9777, "step": 2445 }, { "epoch": 1.1077898550724639, "grad_norm": 0.14168806561798813, "learning_rate": 7.64527994063462e-05, "loss": 0.9551, "step": 2446 }, { "epoch": 1.1082427536231885, "grad_norm": 0.12486350754600979, "learning_rate": 7.644759070081783e-05, "loss": 0.9619, "step": 2447 }, { "epoch": 1.108695652173913, "grad_norm": 0.13752425137015065, "learning_rate": 7.644237835156608e-05, "loss": 0.9629, "step": 2448 }, { "epoch": 1.1091485507246377, "grad_norm": 0.1534724855702949, "learning_rate": 7.6437162359112e-05, "loss": 0.9813, "step": 2449 }, { "epoch": 1.1096014492753623, "grad_norm": 0.16716859184126454, "learning_rate": 7.643194272397708e-05, "loss": 0.9422, "step": 2450 }, { "epoch": 1.110054347826087, "grad_norm": 0.17802221925867495, "learning_rate": 7.642671944668309e-05, "loss": 0.9618, "step": 2451 }, { "epoch": 1.1105072463768115, "grad_norm": 0.19357525078108953, "learning_rate": 7.642149252775224e-05, "loss": 0.9543, "step": 2452 }, { "epoch": 1.1109601449275361, "grad_norm": 0.20086889231913801, "learning_rate": 7.641626196770707e-05, "loss": 0.9542, "step": 2453 }, { "epoch": 1.1114130434782608, "grad_norm": 0.2804743363023598, "learning_rate": 7.64110277670705e-05, "loss": 0.9424, "step": 2454 }, { "epoch": 1.1118659420289856, "grad_norm": 0.3107231470107125, "learning_rate": 7.640578992636576e-05, "loss": 0.9371, "step": 2455 }, { "epoch": 1.1123188405797102, "grad_norm": 0.3100597992230822, "learning_rate": 7.640054844611654e-05, "loss": 0.9706, "step": 2456 }, { "epoch": 1.1127717391304348, "grad_norm": 0.2833875353126641, "learning_rate": 7.639530332684678e-05, "loss": 0.9461, "step": 2457 }, { "epoch": 1.1132246376811594, "grad_norm": 0.23104543742108039, "learning_rate": 7.63900545690809e-05, "loss": 0.9861, "step": 2458 }, { "epoch": 1.113677536231884, "grad_norm": 0.15371547236085464, "learning_rate": 7.638480217334357e-05, "loss": 0.9516, "step": 2459 }, { "epoch": 1.1141304347826086, "grad_norm": 0.14133265580182636, "learning_rate": 7.637954614015993e-05, "loss": 0.9537, "step": 2460 }, { "epoch": 1.1145833333333333, "grad_norm": 0.19397515304126062, "learning_rate": 7.637428647005541e-05, "loss": 0.9677, "step": 2461 }, { "epoch": 1.1150362318840579, "grad_norm": 0.2630620841223688, "learning_rate": 7.636902316355583e-05, "loss": 0.9704, "step": 2462 }, { "epoch": 1.1154891304347827, "grad_norm": 0.30855612337322774, "learning_rate": 7.636375622118738e-05, "loss": 0.9562, "step": 2463 }, { "epoch": 1.1159420289855073, "grad_norm": 0.2762942692788216, "learning_rate": 7.63584856434766e-05, "loss": 0.9534, "step": 2464 }, { "epoch": 1.116394927536232, "grad_norm": 0.2075290079120544, "learning_rate": 7.635321143095038e-05, "loss": 0.9599, "step": 2465 }, { "epoch": 1.1168478260869565, "grad_norm": 0.14267395642094136, "learning_rate": 7.634793358413602e-05, "loss": 0.9794, "step": 2466 }, { "epoch": 1.1173007246376812, "grad_norm": 0.10012481862896405, "learning_rate": 7.634265210356113e-05, "loss": 0.9462, "step": 2467 }, { "epoch": 1.1177536231884058, "grad_norm": 0.12408088156525819, "learning_rate": 7.633736698975373e-05, "loss": 0.9514, "step": 2468 }, { "epoch": 1.1182065217391304, "grad_norm": 0.17838996404821772, "learning_rate": 7.633207824324215e-05, "loss": 0.9613, "step": 2469 }, { "epoch": 1.118659420289855, "grad_norm": 0.24113312569587556, "learning_rate": 7.632678586455517e-05, "loss": 0.9243, "step": 2470 }, { "epoch": 1.1191123188405796, "grad_norm": 0.2669541079298921, "learning_rate": 7.63214898542218e-05, "loss": 0.9443, "step": 2471 }, { "epoch": 1.1195652173913044, "grad_norm": 0.23544486610144588, "learning_rate": 7.631619021277156e-05, "loss": 0.9696, "step": 2472 }, { "epoch": 1.120018115942029, "grad_norm": 0.24286469266117905, "learning_rate": 7.631088694073423e-05, "loss": 0.9588, "step": 2473 }, { "epoch": 1.1204710144927537, "grad_norm": 0.24372516186039825, "learning_rate": 7.630558003864e-05, "loss": 0.9293, "step": 2474 }, { "epoch": 1.1209239130434783, "grad_norm": 0.1659689857958918, "learning_rate": 7.63002695070194e-05, "loss": 0.9698, "step": 2475 }, { "epoch": 1.1213768115942029, "grad_norm": 0.12259275188674705, "learning_rate": 7.629495534640333e-05, "loss": 0.9709, "step": 2476 }, { "epoch": 1.1218297101449275, "grad_norm": 0.14877357502305283, "learning_rate": 7.628963755732306e-05, "loss": 0.9585, "step": 2477 }, { "epoch": 1.122282608695652, "grad_norm": 0.21035994579172454, "learning_rate": 7.628431614031022e-05, "loss": 0.9603, "step": 2478 }, { "epoch": 1.1227355072463767, "grad_norm": 0.27019597756879327, "learning_rate": 7.627899109589679e-05, "loss": 0.9819, "step": 2479 }, { "epoch": 1.1231884057971016, "grad_norm": 0.26577058755602945, "learning_rate": 7.627366242461515e-05, "loss": 0.9519, "step": 2480 }, { "epoch": 1.1236413043478262, "grad_norm": 0.26759058742950703, "learning_rate": 7.626833012699799e-05, "loss": 0.9588, "step": 2481 }, { "epoch": 1.1240942028985508, "grad_norm": 0.2552743180103815, "learning_rate": 7.626299420357839e-05, "loss": 0.9738, "step": 2482 }, { "epoch": 1.1245471014492754, "grad_norm": 0.20376266931018516, "learning_rate": 7.625765465488979e-05, "loss": 0.965, "step": 2483 }, { "epoch": 1.125, "grad_norm": 0.11756741137857578, "learning_rate": 7.625231148146601e-05, "loss": 0.9539, "step": 2484 }, { "epoch": 1.1254528985507246, "grad_norm": 0.1187509721013285, "learning_rate": 7.624696468384119e-05, "loss": 0.931, "step": 2485 }, { "epoch": 1.1259057971014492, "grad_norm": 0.1634273179249078, "learning_rate": 7.624161426254989e-05, "loss": 0.9771, "step": 2486 }, { "epoch": 1.1263586956521738, "grad_norm": 0.20651811847439858, "learning_rate": 7.623626021812697e-05, "loss": 0.9575, "step": 2487 }, { "epoch": 1.1268115942028984, "grad_norm": 0.16218315478012366, "learning_rate": 7.62309025511077e-05, "loss": 0.9744, "step": 2488 }, { "epoch": 1.1272644927536233, "grad_norm": 0.1101528656694027, "learning_rate": 7.62255412620277e-05, "loss": 0.9671, "step": 2489 }, { "epoch": 1.127717391304348, "grad_norm": 0.10448136805326945, "learning_rate": 7.622017635142291e-05, "loss": 0.9738, "step": 2490 }, { "epoch": 1.1281702898550725, "grad_norm": 0.11735215568290443, "learning_rate": 7.621480781982972e-05, "loss": 0.942, "step": 2491 }, { "epoch": 1.1286231884057971, "grad_norm": 0.11660420541563032, "learning_rate": 7.620943566778478e-05, "loss": 0.9496, "step": 2492 }, { "epoch": 1.1290760869565217, "grad_norm": 0.12639495079922483, "learning_rate": 7.62040598958252e-05, "loss": 0.9614, "step": 2493 }, { "epoch": 1.1295289855072463, "grad_norm": 0.13912551636176906, "learning_rate": 7.619868050448838e-05, "loss": 0.9577, "step": 2494 }, { "epoch": 1.129981884057971, "grad_norm": 0.14461902233932716, "learning_rate": 7.619329749431208e-05, "loss": 0.9337, "step": 2495 }, { "epoch": 1.1304347826086956, "grad_norm": 0.10827910271735593, "learning_rate": 7.618791086583451e-05, "loss": 0.9544, "step": 2496 }, { "epoch": 1.1308876811594204, "grad_norm": 0.11279504395188916, "learning_rate": 7.618252061959413e-05, "loss": 0.9836, "step": 2497 }, { "epoch": 1.131340579710145, "grad_norm": 0.10168522664303137, "learning_rate": 7.617712675612983e-05, "loss": 0.9277, "step": 2498 }, { "epoch": 1.1317934782608696, "grad_norm": 0.14760392314390453, "learning_rate": 7.617172927598084e-05, "loss": 0.9802, "step": 2499 }, { "epoch": 1.1322463768115942, "grad_norm": 0.16675539778093154, "learning_rate": 7.616632817968676e-05, "loss": 0.96, "step": 2500 }, { "epoch": 1.1326992753623188, "grad_norm": 0.13705113887533785, "learning_rate": 7.616092346778754e-05, "loss": 0.9787, "step": 2501 }, { "epoch": 1.1331521739130435, "grad_norm": 0.11063969668447533, "learning_rate": 7.61555151408235e-05, "loss": 0.9325, "step": 2502 }, { "epoch": 1.133605072463768, "grad_norm": 0.1044719237553918, "learning_rate": 7.61501031993353e-05, "loss": 0.9433, "step": 2503 }, { "epoch": 1.1340579710144927, "grad_norm": 0.10899628207138502, "learning_rate": 7.614468764386402e-05, "loss": 0.9494, "step": 2504 }, { "epoch": 1.1345108695652173, "grad_norm": 0.12822391452437198, "learning_rate": 7.613926847495104e-05, "loss": 0.961, "step": 2505 }, { "epoch": 1.134963768115942, "grad_norm": 0.13340225982431933, "learning_rate": 7.613384569313812e-05, "loss": 0.979, "step": 2506 }, { "epoch": 1.1354166666666667, "grad_norm": 0.16821324312648034, "learning_rate": 7.612841929896737e-05, "loss": 0.9446, "step": 2507 }, { "epoch": 1.1358695652173914, "grad_norm": 0.19499380252766751, "learning_rate": 7.612298929298132e-05, "loss": 0.947, "step": 2508 }, { "epoch": 1.136322463768116, "grad_norm": 0.21070457207393603, "learning_rate": 7.611755567572276e-05, "loss": 0.9873, "step": 2509 }, { "epoch": 1.1367753623188406, "grad_norm": 0.208715615285171, "learning_rate": 7.611211844773496e-05, "loss": 0.9328, "step": 2510 }, { "epoch": 1.1372282608695652, "grad_norm": 0.19119869191319758, "learning_rate": 7.610667760956143e-05, "loss": 0.9732, "step": 2511 }, { "epoch": 1.1376811594202898, "grad_norm": 0.1564983256136314, "learning_rate": 7.610123316174612e-05, "loss": 0.9459, "step": 2512 }, { "epoch": 1.1381340579710144, "grad_norm": 0.10193003796096267, "learning_rate": 7.609578510483334e-05, "loss": 0.9726, "step": 2513 }, { "epoch": 1.1385869565217392, "grad_norm": 0.10250423723757944, "learning_rate": 7.609033343936771e-05, "loss": 0.9578, "step": 2514 }, { "epoch": 1.1390398550724639, "grad_norm": 0.13865029505392304, "learning_rate": 7.608487816589426e-05, "loss": 0.9595, "step": 2515 }, { "epoch": 1.1394927536231885, "grad_norm": 0.20299287741193903, "learning_rate": 7.607941928495837e-05, "loss": 0.9648, "step": 2516 }, { "epoch": 1.139945652173913, "grad_norm": 0.1939368488744931, "learning_rate": 7.607395679710575e-05, "loss": 0.9435, "step": 2517 }, { "epoch": 1.1403985507246377, "grad_norm": 0.14367245276941906, "learning_rate": 7.60684907028825e-05, "loss": 0.9636, "step": 2518 }, { "epoch": 1.1408514492753623, "grad_norm": 0.123595604402235, "learning_rate": 7.60630210028351e-05, "loss": 0.9391, "step": 2519 }, { "epoch": 1.141304347826087, "grad_norm": 0.1387632058629785, "learning_rate": 7.605754769751033e-05, "loss": 0.9637, "step": 2520 }, { "epoch": 1.1417572463768115, "grad_norm": 0.12204568716824563, "learning_rate": 7.605207078745537e-05, "loss": 0.9465, "step": 2521 }, { "epoch": 1.1422101449275361, "grad_norm": 0.13498536123555513, "learning_rate": 7.604659027321778e-05, "loss": 0.9684, "step": 2522 }, { "epoch": 1.1426630434782608, "grad_norm": 0.15458902370540262, "learning_rate": 7.604110615534542e-05, "loss": 0.9566, "step": 2523 }, { "epoch": 1.1431159420289856, "grad_norm": 0.13620549685850125, "learning_rate": 7.603561843438657e-05, "loss": 0.9464, "step": 2524 }, { "epoch": 1.1435688405797102, "grad_norm": 0.1367662140598925, "learning_rate": 7.603012711088985e-05, "loss": 0.9448, "step": 2525 }, { "epoch": 1.1440217391304348, "grad_norm": 0.14430590264356416, "learning_rate": 7.602463218540423e-05, "loss": 0.9663, "step": 2526 }, { "epoch": 1.1444746376811594, "grad_norm": 0.11756632453255889, "learning_rate": 7.601913365847905e-05, "loss": 0.9481, "step": 2527 }, { "epoch": 1.144927536231884, "grad_norm": 0.12612076879144846, "learning_rate": 7.601363153066399e-05, "loss": 0.9331, "step": 2528 }, { "epoch": 1.1453804347826086, "grad_norm": 0.16008369238712994, "learning_rate": 7.600812580250912e-05, "loss": 0.9748, "step": 2529 }, { "epoch": 1.1458333333333333, "grad_norm": 0.16462587226754993, "learning_rate": 7.600261647456485e-05, "loss": 0.9487, "step": 2530 }, { "epoch": 1.146286231884058, "grad_norm": 0.14809679155799232, "learning_rate": 7.599710354738197e-05, "loss": 0.9661, "step": 2531 }, { "epoch": 1.1467391304347827, "grad_norm": 0.11008873925311632, "learning_rate": 7.59915870215116e-05, "loss": 0.965, "step": 2532 }, { "epoch": 1.1471920289855073, "grad_norm": 0.11907624879984635, "learning_rate": 7.598606689750524e-05, "loss": 0.9454, "step": 2533 }, { "epoch": 1.147644927536232, "grad_norm": 0.1258273580108384, "learning_rate": 7.598054317591474e-05, "loss": 0.9631, "step": 2534 }, { "epoch": 1.1480978260869565, "grad_norm": 0.159959434550033, "learning_rate": 7.597501585729234e-05, "loss": 0.97, "step": 2535 }, { "epoch": 1.1485507246376812, "grad_norm": 0.16629329974384932, "learning_rate": 7.596948494219059e-05, "loss": 0.9547, "step": 2536 }, { "epoch": 1.1490036231884058, "grad_norm": 0.1680870293379807, "learning_rate": 7.596395043116243e-05, "loss": 0.9767, "step": 2537 }, { "epoch": 1.1494565217391304, "grad_norm": 0.1634824116264041, "learning_rate": 7.595841232476116e-05, "loss": 0.9605, "step": 2538 }, { "epoch": 1.149909420289855, "grad_norm": 0.13483549947542858, "learning_rate": 7.595287062354042e-05, "loss": 0.9801, "step": 2539 }, { "epoch": 1.1503623188405796, "grad_norm": 0.12276808553567539, "learning_rate": 7.594732532805425e-05, "loss": 0.9281, "step": 2540 }, { "epoch": 1.1508152173913044, "grad_norm": 0.11211682788969052, "learning_rate": 7.594177643885697e-05, "loss": 0.9464, "step": 2541 }, { "epoch": 1.151268115942029, "grad_norm": 0.11483325788818516, "learning_rate": 7.593622395650338e-05, "loss": 0.9636, "step": 2542 }, { "epoch": 1.1517210144927537, "grad_norm": 0.12126970074045464, "learning_rate": 7.593066788154851e-05, "loss": 0.9548, "step": 2543 }, { "epoch": 1.1521739130434783, "grad_norm": 0.13294432849832763, "learning_rate": 7.592510821454785e-05, "loss": 0.9503, "step": 2544 }, { "epoch": 1.1526268115942029, "grad_norm": 0.10741569065697082, "learning_rate": 7.59195449560572e-05, "loss": 0.9579, "step": 2545 }, { "epoch": 1.1530797101449275, "grad_norm": 0.12455899710436819, "learning_rate": 7.59139781066327e-05, "loss": 0.9421, "step": 2546 }, { "epoch": 1.153532608695652, "grad_norm": 0.14167990582233364, "learning_rate": 7.590840766683091e-05, "loss": 0.9719, "step": 2547 }, { "epoch": 1.1539855072463767, "grad_norm": 0.1311121258648132, "learning_rate": 7.590283363720871e-05, "loss": 0.9616, "step": 2548 }, { "epoch": 1.1544384057971016, "grad_norm": 0.15502168623213952, "learning_rate": 7.589725601832332e-05, "loss": 0.9709, "step": 2549 }, { "epoch": 1.1548913043478262, "grad_norm": 0.1590197651937459, "learning_rate": 7.589167481073238e-05, "loss": 0.9829, "step": 2550 }, { "epoch": 1.1553442028985508, "grad_norm": 0.19825250115903695, "learning_rate": 7.588609001499383e-05, "loss": 0.9594, "step": 2551 }, { "epoch": 1.1557971014492754, "grad_norm": 0.1313641322612241, "learning_rate": 7.588050163166599e-05, "loss": 0.9389, "step": 2552 }, { "epoch": 1.15625, "grad_norm": 0.13898918987968695, "learning_rate": 7.587490966130754e-05, "loss": 0.9736, "step": 2553 }, { "epoch": 1.1567028985507246, "grad_norm": 0.1329486960646459, "learning_rate": 7.586931410447752e-05, "loss": 0.9478, "step": 2554 }, { "epoch": 1.1571557971014492, "grad_norm": 0.12514671942628716, "learning_rate": 7.586371496173535e-05, "loss": 0.9519, "step": 2555 }, { "epoch": 1.1576086956521738, "grad_norm": 0.12709907065303497, "learning_rate": 7.585811223364075e-05, "loss": 0.9556, "step": 2556 }, { "epoch": 1.1580615942028984, "grad_norm": 0.12229818146167958, "learning_rate": 7.585250592075385e-05, "loss": 0.9442, "step": 2557 }, { "epoch": 1.1585144927536233, "grad_norm": 0.16634806752326767, "learning_rate": 7.584689602363512e-05, "loss": 0.9555, "step": 2558 }, { "epoch": 1.158967391304348, "grad_norm": 0.14479894123236, "learning_rate": 7.584128254284538e-05, "loss": 0.9484, "step": 2559 }, { "epoch": 1.1594202898550725, "grad_norm": 0.1265936667652968, "learning_rate": 7.583566547894584e-05, "loss": 0.9671, "step": 2560 }, { "epoch": 1.1598731884057971, "grad_norm": 0.12206669745039841, "learning_rate": 7.583004483249804e-05, "loss": 0.9427, "step": 2561 }, { "epoch": 1.1603260869565217, "grad_norm": 0.16934568951716528, "learning_rate": 7.582442060406386e-05, "loss": 0.9806, "step": 2562 }, { "epoch": 1.1607789855072463, "grad_norm": 0.18964229250086606, "learning_rate": 7.581879279420561e-05, "loss": 0.95, "step": 2563 }, { "epoch": 1.161231884057971, "grad_norm": 0.1933756634641399, "learning_rate": 7.581316140348587e-05, "loss": 0.9553, "step": 2564 }, { "epoch": 1.1616847826086956, "grad_norm": 0.13975643734078044, "learning_rate": 7.580752643246763e-05, "loss": 0.9679, "step": 2565 }, { "epoch": 1.1621376811594204, "grad_norm": 0.14468578149788178, "learning_rate": 7.580188788171425e-05, "loss": 0.9726, "step": 2566 }, { "epoch": 1.162590579710145, "grad_norm": 0.16378553292536752, "learning_rate": 7.579624575178939e-05, "loss": 0.9673, "step": 2567 }, { "epoch": 1.1630434782608696, "grad_norm": 0.19812128794148448, "learning_rate": 7.579060004325712e-05, "loss": 0.9927, "step": 2568 }, { "epoch": 1.1634963768115942, "grad_norm": 0.21169401841505545, "learning_rate": 7.578495075668186e-05, "loss": 0.971, "step": 2569 }, { "epoch": 1.1639492753623188, "grad_norm": 0.17812274094838387, "learning_rate": 7.577929789262835e-05, "loss": 0.9485, "step": 2570 }, { "epoch": 1.1644021739130435, "grad_norm": 0.13947826022966833, "learning_rate": 7.577364145166175e-05, "loss": 0.9711, "step": 2571 }, { "epoch": 1.164855072463768, "grad_norm": 0.1316655080246967, "learning_rate": 7.57679814343475e-05, "loss": 0.9487, "step": 2572 }, { "epoch": 1.1653079710144927, "grad_norm": 0.1444078251418503, "learning_rate": 7.576231784125147e-05, "loss": 0.9633, "step": 2573 }, { "epoch": 1.1657608695652173, "grad_norm": 0.1709289107290525, "learning_rate": 7.575665067293988e-05, "loss": 0.9663, "step": 2574 }, { "epoch": 1.166213768115942, "grad_norm": 0.20734606160576718, "learning_rate": 7.575097992997924e-05, "loss": 0.9714, "step": 2575 }, { "epoch": 1.1666666666666667, "grad_norm": 0.21487833163609016, "learning_rate": 7.57453056129365e-05, "loss": 0.978, "step": 2576 }, { "epoch": 1.1671195652173914, "grad_norm": 0.1653700617650593, "learning_rate": 7.57396277223789e-05, "loss": 0.9385, "step": 2577 }, { "epoch": 1.167572463768116, "grad_norm": 0.12265390694638563, "learning_rate": 7.573394625887409e-05, "loss": 0.9586, "step": 2578 }, { "epoch": 1.1680253623188406, "grad_norm": 0.11913705856532136, "learning_rate": 7.572826122299005e-05, "loss": 0.963, "step": 2579 }, { "epoch": 1.1684782608695652, "grad_norm": 0.18579739270684487, "learning_rate": 7.572257261529511e-05, "loss": 0.9487, "step": 2580 }, { "epoch": 1.1689311594202898, "grad_norm": 0.2477220282604109, "learning_rate": 7.571688043635799e-05, "loss": 0.9714, "step": 2581 }, { "epoch": 1.1693840579710144, "grad_norm": 0.2747812844527379, "learning_rate": 7.571118468674773e-05, "loss": 0.9609, "step": 2582 }, { "epoch": 1.1698369565217392, "grad_norm": 0.2883596822609474, "learning_rate": 7.570548536703375e-05, "loss": 0.9671, "step": 2583 }, { "epoch": 1.1702898550724639, "grad_norm": 0.24508936600723333, "learning_rate": 7.569978247778581e-05, "loss": 0.9655, "step": 2584 }, { "epoch": 1.1707427536231885, "grad_norm": 0.1740501033405785, "learning_rate": 7.569407601957404e-05, "loss": 0.963, "step": 2585 }, { "epoch": 1.171195652173913, "grad_norm": 0.18369732139701603, "learning_rate": 7.568836599296893e-05, "loss": 0.9435, "step": 2586 }, { "epoch": 1.1716485507246377, "grad_norm": 0.16441935992420884, "learning_rate": 7.568265239854133e-05, "loss": 0.9813, "step": 2587 }, { "epoch": 1.1721014492753623, "grad_norm": 0.13179823116669293, "learning_rate": 7.567693523686241e-05, "loss": 0.9602, "step": 2588 }, { "epoch": 1.172554347826087, "grad_norm": 0.13358259236694253, "learning_rate": 7.567121450850376e-05, "loss": 0.9657, "step": 2589 }, { "epoch": 1.1730072463768115, "grad_norm": 0.13497227153471542, "learning_rate": 7.566549021403725e-05, "loss": 0.9648, "step": 2590 }, { "epoch": 1.1734601449275361, "grad_norm": 0.16142773590948734, "learning_rate": 7.565976235403518e-05, "loss": 0.9511, "step": 2591 }, { "epoch": 1.1739130434782608, "grad_norm": 0.2122590098288983, "learning_rate": 7.565403092907015e-05, "loss": 0.9609, "step": 2592 }, { "epoch": 1.1743659420289856, "grad_norm": 0.19870017968685383, "learning_rate": 7.564829593971516e-05, "loss": 0.9472, "step": 2593 }, { "epoch": 1.1748188405797102, "grad_norm": 0.2675257668514363, "learning_rate": 7.564255738654354e-05, "loss": 0.9686, "step": 2594 }, { "epoch": 1.1752717391304348, "grad_norm": 0.1660942495292838, "learning_rate": 7.563681527012897e-05, "loss": 0.9795, "step": 2595 }, { "epoch": 1.1757246376811594, "grad_norm": 0.1870832580990727, "learning_rate": 7.56310695910455e-05, "loss": 0.9533, "step": 2596 }, { "epoch": 1.176177536231884, "grad_norm": 0.15144878515097787, "learning_rate": 7.562532034986755e-05, "loss": 0.9837, "step": 2597 }, { "epoch": 1.1766304347826086, "grad_norm": 0.10356060877468966, "learning_rate": 7.561956754716986e-05, "loss": 0.9592, "step": 2598 }, { "epoch": 1.1770833333333333, "grad_norm": 0.11189342005815624, "learning_rate": 7.561381118352757e-05, "loss": 0.943, "step": 2599 }, { "epoch": 1.177536231884058, "grad_norm": 0.14770275041076447, "learning_rate": 7.560805125951612e-05, "loss": 0.9685, "step": 2600 }, { "epoch": 1.1779891304347827, "grad_norm": 0.20010857728166745, "learning_rate": 7.560228777571138e-05, "loss": 0.9527, "step": 2601 }, { "epoch": 1.1784420289855073, "grad_norm": 0.17555665244521282, "learning_rate": 7.559652073268951e-05, "loss": 0.9667, "step": 2602 }, { "epoch": 1.178894927536232, "grad_norm": 0.12255403360195197, "learning_rate": 7.559075013102705e-05, "loss": 0.9553, "step": 2603 }, { "epoch": 1.1793478260869565, "grad_norm": 0.1711245183532606, "learning_rate": 7.558497597130091e-05, "loss": 0.9421, "step": 2604 }, { "epoch": 1.1798007246376812, "grad_norm": 0.20533719608648182, "learning_rate": 7.557919825408833e-05, "loss": 0.9551, "step": 2605 }, { "epoch": 1.1802536231884058, "grad_norm": 0.11730396142241477, "learning_rate": 7.557341697996692e-05, "loss": 0.932, "step": 2606 }, { "epoch": 1.1807065217391304, "grad_norm": 0.15288154903422843, "learning_rate": 7.556763214951464e-05, "loss": 0.9671, "step": 2607 }, { "epoch": 1.181159420289855, "grad_norm": 0.20731494539897263, "learning_rate": 7.556184376330981e-05, "loss": 0.9904, "step": 2608 }, { "epoch": 1.1816123188405796, "grad_norm": 0.18392709752693126, "learning_rate": 7.555605182193111e-05, "loss": 0.9413, "step": 2609 }, { "epoch": 1.1820652173913044, "grad_norm": 0.1469080104590913, "learning_rate": 7.555025632595758e-05, "loss": 0.9725, "step": 2610 }, { "epoch": 1.182518115942029, "grad_norm": 0.1501624741528834, "learning_rate": 7.554445727596858e-05, "loss": 0.9633, "step": 2611 }, { "epoch": 1.1829710144927537, "grad_norm": 0.16126958824880513, "learning_rate": 7.553865467254386e-05, "loss": 0.9337, "step": 2612 }, { "epoch": 1.1834239130434783, "grad_norm": 0.1458187463142985, "learning_rate": 7.55328485162635e-05, "loss": 0.945, "step": 2613 }, { "epoch": 1.1838768115942029, "grad_norm": 0.17594761430496011, "learning_rate": 7.552703880770798e-05, "loss": 0.9665, "step": 2614 }, { "epoch": 1.1843297101449275, "grad_norm": 0.20522250341680714, "learning_rate": 7.552122554745811e-05, "loss": 0.9674, "step": 2615 }, { "epoch": 1.184782608695652, "grad_norm": 0.2141767417238105, "learning_rate": 7.551540873609502e-05, "loss": 0.9344, "step": 2616 }, { "epoch": 1.1852355072463767, "grad_norm": 0.16766185714009643, "learning_rate": 7.550958837420023e-05, "loss": 0.9516, "step": 2617 }, { "epoch": 1.1856884057971016, "grad_norm": 0.12147699611658226, "learning_rate": 7.550376446235563e-05, "loss": 0.9596, "step": 2618 }, { "epoch": 1.1861413043478262, "grad_norm": 0.15139312684056241, "learning_rate": 7.549793700114344e-05, "loss": 0.961, "step": 2619 }, { "epoch": 1.1865942028985508, "grad_norm": 0.1932830595008674, "learning_rate": 7.549210599114624e-05, "loss": 0.9661, "step": 2620 }, { "epoch": 1.1870471014492754, "grad_norm": 0.2052729012578867, "learning_rate": 7.548627143294695e-05, "loss": 0.954, "step": 2621 }, { "epoch": 1.1875, "grad_norm": 0.19582820242919294, "learning_rate": 7.548043332712887e-05, "loss": 0.9754, "step": 2622 }, { "epoch": 1.1879528985507246, "grad_norm": 0.19115863827904878, "learning_rate": 7.547459167427566e-05, "loss": 0.9777, "step": 2623 }, { "epoch": 1.1884057971014492, "grad_norm": 0.31764529789104673, "learning_rate": 7.546874647497129e-05, "loss": 0.9856, "step": 2624 }, { "epoch": 1.1888586956521738, "grad_norm": 0.13590401335276542, "learning_rate": 7.546289772980014e-05, "loss": 0.9585, "step": 2625 }, { "epoch": 1.1893115942028984, "grad_norm": 0.13406588072803646, "learning_rate": 7.545704543934691e-05, "loss": 0.9446, "step": 2626 }, { "epoch": 1.1897644927536233, "grad_norm": 0.13306769995160692, "learning_rate": 7.545118960419666e-05, "loss": 0.9433, "step": 2627 }, { "epoch": 1.190217391304348, "grad_norm": 0.1614526015287179, "learning_rate": 7.54453302249348e-05, "loss": 0.955, "step": 2628 }, { "epoch": 1.1906702898550725, "grad_norm": 0.13800255228106142, "learning_rate": 7.543946730214713e-05, "loss": 0.949, "step": 2629 }, { "epoch": 1.1911231884057971, "grad_norm": 0.10423127114780148, "learning_rate": 7.543360083641974e-05, "loss": 0.9543, "step": 2630 }, { "epoch": 1.1915760869565217, "grad_norm": 0.09816211775006568, "learning_rate": 7.542773082833912e-05, "loss": 0.9592, "step": 2631 }, { "epoch": 1.1920289855072463, "grad_norm": 0.1250436242840592, "learning_rate": 7.542185727849211e-05, "loss": 0.9522, "step": 2632 }, { "epoch": 1.192481884057971, "grad_norm": 0.11248470366698343, "learning_rate": 7.541598018746591e-05, "loss": 0.9379, "step": 2633 }, { "epoch": 1.1929347826086956, "grad_norm": 0.11262683693299408, "learning_rate": 7.541009955584803e-05, "loss": 0.9368, "step": 2634 }, { "epoch": 1.1933876811594204, "grad_norm": 0.11235314542989715, "learning_rate": 7.54042153842264e-05, "loss": 0.9573, "step": 2635 }, { "epoch": 1.193840579710145, "grad_norm": 0.0996333743877415, "learning_rate": 7.539832767318925e-05, "loss": 0.9338, "step": 2636 }, { "epoch": 1.1942934782608696, "grad_norm": 0.15062035428473822, "learning_rate": 7.53924364233252e-05, "loss": 0.9573, "step": 2637 }, { "epoch": 1.1947463768115942, "grad_norm": 0.19889787101340947, "learning_rate": 7.538654163522319e-05, "loss": 0.9544, "step": 2638 }, { "epoch": 1.1951992753623188, "grad_norm": 0.18505928316515938, "learning_rate": 7.538064330947253e-05, "loss": 0.9517, "step": 2639 }, { "epoch": 1.1956521739130435, "grad_norm": 0.1748611437336659, "learning_rate": 7.53747414466629e-05, "loss": 0.9532, "step": 2640 }, { "epoch": 1.196105072463768, "grad_norm": 0.1831936967320523, "learning_rate": 7.53688360473843e-05, "loss": 0.9603, "step": 2641 }, { "epoch": 1.1965579710144927, "grad_norm": 0.14487405368593878, "learning_rate": 7.536292711222713e-05, "loss": 0.9661, "step": 2642 }, { "epoch": 1.1970108695652173, "grad_norm": 0.10512877781973295, "learning_rate": 7.53570146417821e-05, "loss": 0.9454, "step": 2643 }, { "epoch": 1.197463768115942, "grad_norm": 0.13605340327639504, "learning_rate": 7.535109863664028e-05, "loss": 0.9868, "step": 2644 }, { "epoch": 1.1979166666666667, "grad_norm": 0.15602093503457862, "learning_rate": 7.534517909739312e-05, "loss": 0.9559, "step": 2645 }, { "epoch": 1.1983695652173914, "grad_norm": 0.16036162838111195, "learning_rate": 7.533925602463237e-05, "loss": 0.9627, "step": 2646 }, { "epoch": 1.198822463768116, "grad_norm": 0.16402727357283928, "learning_rate": 7.533332941895023e-05, "loss": 0.946, "step": 2647 }, { "epoch": 1.1992753623188406, "grad_norm": 0.13144382848880362, "learning_rate": 7.532739928093913e-05, "loss": 0.9696, "step": 2648 }, { "epoch": 1.1997282608695652, "grad_norm": 0.14270973990056626, "learning_rate": 7.532146561119196e-05, "loss": 0.9554, "step": 2649 }, { "epoch": 1.2001811594202898, "grad_norm": 0.16086272820228345, "learning_rate": 7.53155284103019e-05, "loss": 0.9729, "step": 2650 }, { "epoch": 1.2006340579710144, "grad_norm": 0.15197450434192672, "learning_rate": 7.530958767886249e-05, "loss": 0.9533, "step": 2651 }, { "epoch": 1.2010869565217392, "grad_norm": 0.19950872638530828, "learning_rate": 7.530364341746765e-05, "loss": 0.9497, "step": 2652 }, { "epoch": 1.2015398550724639, "grad_norm": 0.22129388544104742, "learning_rate": 7.529769562671164e-05, "loss": 0.9643, "step": 2653 }, { "epoch": 1.2019927536231885, "grad_norm": 0.231423551925209, "learning_rate": 7.529174430718906e-05, "loss": 0.9577, "step": 2654 }, { "epoch": 1.202445652173913, "grad_norm": 0.22444710224046674, "learning_rate": 7.528578945949488e-05, "loss": 0.965, "step": 2655 }, { "epoch": 1.2028985507246377, "grad_norm": 0.20890781621288873, "learning_rate": 7.527983108422442e-05, "loss": 0.9451, "step": 2656 }, { "epoch": 1.2033514492753623, "grad_norm": 0.17826816734227027, "learning_rate": 7.527386918197333e-05, "loss": 0.9469, "step": 2657 }, { "epoch": 1.203804347826087, "grad_norm": 0.14163411684221708, "learning_rate": 7.526790375333765e-05, "loss": 0.9386, "step": 2658 }, { "epoch": 1.2042572463768115, "grad_norm": 0.1435555205910353, "learning_rate": 7.526193479891374e-05, "loss": 0.9543, "step": 2659 }, { "epoch": 1.2047101449275361, "grad_norm": 0.15206483022128087, "learning_rate": 7.525596231929833e-05, "loss": 0.9774, "step": 2660 }, { "epoch": 1.2051630434782608, "grad_norm": 0.1808745695070356, "learning_rate": 7.524998631508849e-05, "loss": 0.9581, "step": 2661 }, { "epoch": 1.2056159420289856, "grad_norm": 0.2650869587735258, "learning_rate": 7.524400678688167e-05, "loss": 0.9809, "step": 2662 }, { "epoch": 1.2060688405797102, "grad_norm": 0.21595896113485846, "learning_rate": 7.523802373527565e-05, "loss": 0.9754, "step": 2663 }, { "epoch": 1.2065217391304348, "grad_norm": 0.18692936740649976, "learning_rate": 7.523203716086855e-05, "loss": 0.9687, "step": 2664 }, { "epoch": 1.2069746376811594, "grad_norm": 0.14991683777764858, "learning_rate": 7.522604706425886e-05, "loss": 0.9427, "step": 2665 }, { "epoch": 1.207427536231884, "grad_norm": 0.11324955811586246, "learning_rate": 7.522005344604543e-05, "loss": 0.9792, "step": 2666 }, { "epoch": 1.2078804347826086, "grad_norm": 0.14582633578886917, "learning_rate": 7.521405630682745e-05, "loss": 0.9571, "step": 2667 }, { "epoch": 1.2083333333333333, "grad_norm": 0.17071104101396495, "learning_rate": 7.520805564720444e-05, "loss": 0.9407, "step": 2668 }, { "epoch": 1.208786231884058, "grad_norm": 0.14404229269503654, "learning_rate": 7.520205146777635e-05, "loss": 0.9755, "step": 2669 }, { "epoch": 1.2092391304347827, "grad_norm": 0.18045962168287716, "learning_rate": 7.519604376914336e-05, "loss": 0.9491, "step": 2670 }, { "epoch": 1.2096920289855073, "grad_norm": 0.19585348841295955, "learning_rate": 7.519003255190612e-05, "loss": 0.9738, "step": 2671 }, { "epoch": 1.210144927536232, "grad_norm": 0.15754092234969355, "learning_rate": 7.518401781666555e-05, "loss": 0.9609, "step": 2672 }, { "epoch": 1.2105978260869565, "grad_norm": 0.1075950299729778, "learning_rate": 7.517799956402297e-05, "loss": 0.939, "step": 2673 }, { "epoch": 1.2110507246376812, "grad_norm": 0.1834840377817972, "learning_rate": 7.517197779458003e-05, "loss": 0.9485, "step": 2674 }, { "epoch": 1.2115036231884058, "grad_norm": 0.13516966155175972, "learning_rate": 7.516595250893874e-05, "loss": 0.951, "step": 2675 }, { "epoch": 1.2119565217391304, "grad_norm": 0.12521768045210718, "learning_rate": 7.515992370770145e-05, "loss": 0.954, "step": 2676 }, { "epoch": 1.212409420289855, "grad_norm": 0.12617544718049611, "learning_rate": 7.515389139147087e-05, "loss": 0.9636, "step": 2677 }, { "epoch": 1.2128623188405796, "grad_norm": 0.1153550577495718, "learning_rate": 7.514785556085006e-05, "loss": 0.9362, "step": 2678 }, { "epoch": 1.2133152173913044, "grad_norm": 0.10638466529163494, "learning_rate": 7.514181621644243e-05, "loss": 0.9695, "step": 2679 }, { "epoch": 1.213768115942029, "grad_norm": 0.09949619238182274, "learning_rate": 7.513577335885175e-05, "loss": 0.9398, "step": 2680 }, { "epoch": 1.2142210144927537, "grad_norm": 0.10772257201541509, "learning_rate": 7.512972698868214e-05, "loss": 0.9517, "step": 2681 }, { "epoch": 1.2146739130434783, "grad_norm": 0.10847124663985191, "learning_rate": 7.512367710653804e-05, "loss": 0.9651, "step": 2682 }, { "epoch": 1.2151268115942029, "grad_norm": 0.09933836389395898, "learning_rate": 7.511762371302428e-05, "loss": 0.9712, "step": 2683 }, { "epoch": 1.2155797101449275, "grad_norm": 0.1031917535242824, "learning_rate": 7.511156680874604e-05, "loss": 0.9777, "step": 2684 }, { "epoch": 1.216032608695652, "grad_norm": 0.09251940920710866, "learning_rate": 7.510550639430883e-05, "loss": 0.9592, "step": 2685 }, { "epoch": 1.2164855072463767, "grad_norm": 0.11449876609715805, "learning_rate": 7.50994424703185e-05, "loss": 0.9414, "step": 2686 }, { "epoch": 1.2169384057971016, "grad_norm": 0.12147458300756844, "learning_rate": 7.50933750373813e-05, "loss": 0.9399, "step": 2687 }, { "epoch": 1.2173913043478262, "grad_norm": 0.10813728815361857, "learning_rate": 7.508730409610377e-05, "loss": 0.9692, "step": 2688 }, { "epoch": 1.2178442028985508, "grad_norm": 0.12751754526715206, "learning_rate": 7.508122964709287e-05, "loss": 0.9204, "step": 2689 }, { "epoch": 1.2182971014492754, "grad_norm": 0.11776963851134986, "learning_rate": 7.507515169095584e-05, "loss": 0.9256, "step": 2690 }, { "epoch": 1.21875, "grad_norm": 0.12710553084949197, "learning_rate": 7.506907022830032e-05, "loss": 0.9454, "step": 2691 }, { "epoch": 1.2192028985507246, "grad_norm": 0.13321228848666322, "learning_rate": 7.506298525973427e-05, "loss": 0.9619, "step": 2692 }, { "epoch": 1.2196557971014492, "grad_norm": 0.14438921835812282, "learning_rate": 7.505689678586603e-05, "loss": 0.9458, "step": 2693 }, { "epoch": 1.2201086956521738, "grad_norm": 0.15346408362355932, "learning_rate": 7.505080480730426e-05, "loss": 0.9542, "step": 2694 }, { "epoch": 1.2205615942028984, "grad_norm": 0.13351013427075334, "learning_rate": 7.504470932465799e-05, "loss": 0.9537, "step": 2695 }, { "epoch": 1.2210144927536233, "grad_norm": 0.10518583532217386, "learning_rate": 7.50386103385366e-05, "loss": 0.9531, "step": 2696 }, { "epoch": 1.221467391304348, "grad_norm": 0.12370066404857444, "learning_rate": 7.50325078495498e-05, "loss": 0.9586, "step": 2697 }, { "epoch": 1.2219202898550725, "grad_norm": 0.15243405155180748, "learning_rate": 7.50264018583077e-05, "loss": 0.9693, "step": 2698 }, { "epoch": 1.2223731884057971, "grad_norm": 0.15024310160579438, "learning_rate": 7.502029236542067e-05, "loss": 0.9266, "step": 2699 }, { "epoch": 1.2228260869565217, "grad_norm": 0.14753257078046852, "learning_rate": 7.501417937149954e-05, "loss": 0.9696, "step": 2700 }, { "epoch": 1.2232789855072463, "grad_norm": 0.15905376940405452, "learning_rate": 7.500806287715541e-05, "loss": 0.9894, "step": 2701 }, { "epoch": 1.223731884057971, "grad_norm": 0.155827673871629, "learning_rate": 7.500194288299975e-05, "loss": 0.9465, "step": 2702 }, { "epoch": 1.2241847826086956, "grad_norm": 0.15076487722249074, "learning_rate": 7.499581938964442e-05, "loss": 0.9627, "step": 2703 }, { "epoch": 1.2246376811594204, "grad_norm": 0.14568058552228394, "learning_rate": 7.498969239770155e-05, "loss": 0.9545, "step": 2704 }, { "epoch": 1.225090579710145, "grad_norm": 0.14192702424143167, "learning_rate": 7.498356190778369e-05, "loss": 0.9656, "step": 2705 }, { "epoch": 1.2255434782608696, "grad_norm": 0.14772177846842116, "learning_rate": 7.497742792050369e-05, "loss": 0.9326, "step": 2706 }, { "epoch": 1.2259963768115942, "grad_norm": 0.14311957972887787, "learning_rate": 7.497129043647482e-05, "loss": 0.9669, "step": 2707 }, { "epoch": 1.2264492753623188, "grad_norm": 0.12423728870726472, "learning_rate": 7.496514945631062e-05, "loss": 0.9462, "step": 2708 }, { "epoch": 1.2269021739130435, "grad_norm": 0.13460366449765676, "learning_rate": 7.495900498062501e-05, "loss": 0.9565, "step": 2709 }, { "epoch": 1.227355072463768, "grad_norm": 0.1862511657936904, "learning_rate": 7.49528570100323e-05, "loss": 0.9696, "step": 2710 }, { "epoch": 1.2278079710144927, "grad_norm": 0.21907982207409335, "learning_rate": 7.494670554514707e-05, "loss": 0.9571, "step": 2711 }, { "epoch": 1.2282608695652173, "grad_norm": 0.24649935544746052, "learning_rate": 7.494055058658431e-05, "loss": 0.9597, "step": 2712 }, { "epoch": 1.228713768115942, "grad_norm": 0.2643202069803128, "learning_rate": 7.493439213495935e-05, "loss": 0.9278, "step": 2713 }, { "epoch": 1.2291666666666667, "grad_norm": 0.2586521232179391, "learning_rate": 7.492823019088785e-05, "loss": 0.9458, "step": 2714 }, { "epoch": 1.2296195652173914, "grad_norm": 0.23090472972700934, "learning_rate": 7.492206475498581e-05, "loss": 0.9479, "step": 2715 }, { "epoch": 1.230072463768116, "grad_norm": 0.18694594712245943, "learning_rate": 7.491589582786963e-05, "loss": 0.9706, "step": 2716 }, { "epoch": 1.2305253623188406, "grad_norm": 0.11917663750010685, "learning_rate": 7.490972341015601e-05, "loss": 0.967, "step": 2717 }, { "epoch": 1.2309782608695652, "grad_norm": 0.1408493480949558, "learning_rate": 7.490354750246203e-05, "loss": 0.9472, "step": 2718 }, { "epoch": 1.2314311594202898, "grad_norm": 0.2038172449889158, "learning_rate": 7.489736810540508e-05, "loss": 0.9734, "step": 2719 }, { "epoch": 1.2318840579710144, "grad_norm": 0.23880582748878637, "learning_rate": 7.489118521960296e-05, "loss": 0.9591, "step": 2720 }, { "epoch": 1.2323369565217392, "grad_norm": 0.22276199350862216, "learning_rate": 7.488499884567376e-05, "loss": 0.9517, "step": 2721 }, { "epoch": 1.2327898550724639, "grad_norm": 0.20051933697222235, "learning_rate": 7.487880898423595e-05, "loss": 0.9499, "step": 2722 }, { "epoch": 1.2332427536231885, "grad_norm": 0.193713919257804, "learning_rate": 7.487261563590831e-05, "loss": 0.9782, "step": 2723 }, { "epoch": 1.233695652173913, "grad_norm": 0.19953354367383686, "learning_rate": 7.486641880131006e-05, "loss": 0.9499, "step": 2724 }, { "epoch": 1.2341485507246377, "grad_norm": 0.19102396595678708, "learning_rate": 7.486021848106065e-05, "loss": 0.9303, "step": 2725 }, { "epoch": 1.2346014492753623, "grad_norm": 0.20215189562135186, "learning_rate": 7.485401467577996e-05, "loss": 0.9719, "step": 2726 }, { "epoch": 1.235054347826087, "grad_norm": 0.14176881463332472, "learning_rate": 7.48478073860882e-05, "loss": 0.9642, "step": 2727 }, { "epoch": 1.2355072463768115, "grad_norm": 0.11712473975389216, "learning_rate": 7.48415966126059e-05, "loss": 0.9648, "step": 2728 }, { "epoch": 1.2359601449275361, "grad_norm": 0.15962171693423344, "learning_rate": 7.4835382355954e-05, "loss": 0.9615, "step": 2729 }, { "epoch": 1.2364130434782608, "grad_norm": 0.22540384558108312, "learning_rate": 7.48291646167537e-05, "loss": 0.9225, "step": 2730 }, { "epoch": 1.2368659420289856, "grad_norm": 0.29122321027351167, "learning_rate": 7.482294339562663e-05, "loss": 0.9449, "step": 2731 }, { "epoch": 1.2373188405797102, "grad_norm": 0.353500369782154, "learning_rate": 7.481671869319473e-05, "loss": 0.9526, "step": 2732 }, { "epoch": 1.2377717391304348, "grad_norm": 0.40638222214560077, "learning_rate": 7.481049051008028e-05, "loss": 0.9731, "step": 2733 }, { "epoch": 1.2382246376811594, "grad_norm": 0.39281109682229787, "learning_rate": 7.480425884690594e-05, "loss": 0.9776, "step": 2734 }, { "epoch": 1.238677536231884, "grad_norm": 0.3029991301999259, "learning_rate": 7.47980237042947e-05, "loss": 0.9612, "step": 2735 }, { "epoch": 1.2391304347826086, "grad_norm": 0.2893218286530701, "learning_rate": 7.479178508286987e-05, "loss": 0.9665, "step": 2736 }, { "epoch": 1.2395833333333333, "grad_norm": 0.15111342830541422, "learning_rate": 7.478554298325517e-05, "loss": 0.949, "step": 2737 }, { "epoch": 1.240036231884058, "grad_norm": 0.2380882975222756, "learning_rate": 7.47792974060746e-05, "loss": 0.9688, "step": 2738 }, { "epoch": 1.2404891304347827, "grad_norm": 0.23668477315108138, "learning_rate": 7.477304835195257e-05, "loss": 0.9801, "step": 2739 }, { "epoch": 1.2409420289855073, "grad_norm": 0.17582485984388202, "learning_rate": 7.476679582151379e-05, "loss": 0.9468, "step": 2740 }, { "epoch": 1.241394927536232, "grad_norm": 0.19199713479597907, "learning_rate": 7.476053981538332e-05, "loss": 0.9801, "step": 2741 }, { "epoch": 1.2418478260869565, "grad_norm": 0.1491710448744397, "learning_rate": 7.475428033418663e-05, "loss": 0.959, "step": 2742 }, { "epoch": 1.2423007246376812, "grad_norm": 0.1396459706707031, "learning_rate": 7.474801737854944e-05, "loss": 0.9605, "step": 2743 }, { "epoch": 1.2427536231884058, "grad_norm": 0.13475724674525455, "learning_rate": 7.47417509490979e-05, "loss": 0.9435, "step": 2744 }, { "epoch": 1.2432065217391304, "grad_norm": 0.13818195796809754, "learning_rate": 7.473548104645847e-05, "loss": 0.9669, "step": 2745 }, { "epoch": 1.243659420289855, "grad_norm": 0.13938381827258048, "learning_rate": 7.472920767125795e-05, "loss": 0.9284, "step": 2746 }, { "epoch": 1.2441123188405796, "grad_norm": 0.13064049912679393, "learning_rate": 7.47229308241235e-05, "loss": 0.9529, "step": 2747 }, { "epoch": 1.2445652173913044, "grad_norm": 0.13997346874315633, "learning_rate": 7.471665050568265e-05, "loss": 0.9509, "step": 2748 }, { "epoch": 1.245018115942029, "grad_norm": 0.15537728744899057, "learning_rate": 7.471036671656322e-05, "loss": 0.9583, "step": 2749 }, { "epoch": 1.2454710144927537, "grad_norm": 0.171231835341623, "learning_rate": 7.470407945739343e-05, "loss": 0.9993, "step": 2750 }, { "epoch": 1.2459239130434783, "grad_norm": 0.18291335585234206, "learning_rate": 7.469778872880183e-05, "loss": 0.9399, "step": 2751 }, { "epoch": 1.2463768115942029, "grad_norm": 0.13003241333816637, "learning_rate": 7.469149453141729e-05, "loss": 0.9743, "step": 2752 }, { "epoch": 1.2468297101449275, "grad_norm": 0.15829912864289447, "learning_rate": 7.468519686586909e-05, "loss": 0.9618, "step": 2753 }, { "epoch": 1.247282608695652, "grad_norm": 0.1573097674160597, "learning_rate": 7.467889573278678e-05, "loss": 0.9508, "step": 2754 }, { "epoch": 1.2477355072463767, "grad_norm": 0.13781339589069158, "learning_rate": 7.467259113280032e-05, "loss": 0.9575, "step": 2755 }, { "epoch": 1.2481884057971016, "grad_norm": 0.13890273548412616, "learning_rate": 7.466628306653997e-05, "loss": 0.9667, "step": 2756 }, { "epoch": 1.2486413043478262, "grad_norm": 0.16203603362168706, "learning_rate": 7.465997153463638e-05, "loss": 0.9762, "step": 2757 }, { "epoch": 1.2490942028985508, "grad_norm": 0.16466820890108946, "learning_rate": 7.46536565377205e-05, "loss": 0.9435, "step": 2758 }, { "epoch": 1.2495471014492754, "grad_norm": 0.10826963690106382, "learning_rate": 7.464733807642367e-05, "loss": 0.9657, "step": 2759 }, { "epoch": 1.25, "grad_norm": 0.11117275672241318, "learning_rate": 7.464101615137756e-05, "loss": 0.9544, "step": 2760 }, { "epoch": 1.2504528985507246, "grad_norm": 0.14656392477660743, "learning_rate": 7.463469076321416e-05, "loss": 0.9425, "step": 2761 }, { "epoch": 1.2509057971014492, "grad_norm": 0.14302588438451105, "learning_rate": 7.462836191256584e-05, "loss": 0.9629, "step": 2762 }, { "epoch": 1.2513586956521738, "grad_norm": 0.13252974255851904, "learning_rate": 7.462202960006531e-05, "loss": 0.9743, "step": 2763 }, { "epoch": 1.2518115942028984, "grad_norm": 0.13699200703792147, "learning_rate": 7.461569382634561e-05, "loss": 0.9626, "step": 2764 }, { "epoch": 1.252264492753623, "grad_norm": 0.15168415477274916, "learning_rate": 7.460935459204015e-05, "loss": 0.9308, "step": 2765 }, { "epoch": 1.252717391304348, "grad_norm": 0.2500751471836116, "learning_rate": 7.460301189778266e-05, "loss": 0.9451, "step": 2766 }, { "epoch": 1.2531702898550725, "grad_norm": 0.17166907866537667, "learning_rate": 7.459666574420726e-05, "loss": 0.9409, "step": 2767 }, { "epoch": 1.2536231884057971, "grad_norm": 0.18021835001005085, "learning_rate": 7.459031613194834e-05, "loss": 0.9839, "step": 2768 }, { "epoch": 1.2540760869565217, "grad_norm": 0.1834987164012684, "learning_rate": 7.458396306164072e-05, "loss": 0.9441, "step": 2769 }, { "epoch": 1.2545289855072463, "grad_norm": 0.17593340552915893, "learning_rate": 7.45776065339195e-05, "loss": 0.9636, "step": 2770 }, { "epoch": 1.254981884057971, "grad_norm": 0.15653219295636694, "learning_rate": 7.457124654942016e-05, "loss": 0.9723, "step": 2771 }, { "epoch": 1.2554347826086958, "grad_norm": 0.15115942226911427, "learning_rate": 7.456488310877853e-05, "loss": 0.9611, "step": 2772 }, { "epoch": 1.2558876811594204, "grad_norm": 0.209441963293977, "learning_rate": 7.455851621263075e-05, "loss": 0.9591, "step": 2773 }, { "epoch": 1.256340579710145, "grad_norm": 0.17923960488597118, "learning_rate": 7.455214586161336e-05, "loss": 0.9356, "step": 2774 }, { "epoch": 1.2567934782608696, "grad_norm": 0.156632879897952, "learning_rate": 7.454577205636318e-05, "loss": 0.948, "step": 2775 }, { "epoch": 1.2572463768115942, "grad_norm": 0.1381604159332152, "learning_rate": 7.453939479751742e-05, "loss": 0.9646, "step": 2776 }, { "epoch": 1.2576992753623188, "grad_norm": 0.11278729843454334, "learning_rate": 7.453301408571363e-05, "loss": 0.9671, "step": 2777 }, { "epoch": 1.2581521739130435, "grad_norm": 0.14105916831173657, "learning_rate": 7.452662992158971e-05, "loss": 0.9482, "step": 2778 }, { "epoch": 1.258605072463768, "grad_norm": 0.11695706487954487, "learning_rate": 7.452024230578389e-05, "loss": 0.9472, "step": 2779 }, { "epoch": 1.2590579710144927, "grad_norm": 0.15466546061482975, "learning_rate": 7.451385123893474e-05, "loss": 0.9457, "step": 2780 }, { "epoch": 1.2595108695652173, "grad_norm": 0.1968632305403577, "learning_rate": 7.450745672168118e-05, "loss": 0.9651, "step": 2781 }, { "epoch": 1.259963768115942, "grad_norm": 0.21659160427360089, "learning_rate": 7.45010587546625e-05, "loss": 0.939, "step": 2782 }, { "epoch": 1.2604166666666667, "grad_norm": 0.21382916514534092, "learning_rate": 7.44946573385183e-05, "loss": 0.9565, "step": 2783 }, { "epoch": 1.2608695652173914, "grad_norm": 0.1939587348258884, "learning_rate": 7.448825247388854e-05, "loss": 0.9565, "step": 2784 }, { "epoch": 1.261322463768116, "grad_norm": 0.1477359984399057, "learning_rate": 7.448184416141355e-05, "loss": 0.9721, "step": 2785 }, { "epoch": 1.2617753623188406, "grad_norm": 0.13132095561632193, "learning_rate": 7.447543240173392e-05, "loss": 0.9505, "step": 2786 }, { "epoch": 1.2622282608695652, "grad_norm": 0.16421876344568007, "learning_rate": 7.446901719549073e-05, "loss": 0.9489, "step": 2787 }, { "epoch": 1.2626811594202898, "grad_norm": 0.21231022906315666, "learning_rate": 7.446259854332523e-05, "loss": 0.9701, "step": 2788 }, { "epoch": 1.2631340579710144, "grad_norm": 0.2537083666606882, "learning_rate": 7.445617644587915e-05, "loss": 0.9639, "step": 2789 }, { "epoch": 1.2635869565217392, "grad_norm": 0.2513316138339832, "learning_rate": 7.444975090379453e-05, "loss": 0.9563, "step": 2790 }, { "epoch": 1.2640398550724639, "grad_norm": 0.15596172518484552, "learning_rate": 7.44433219177137e-05, "loss": 0.9657, "step": 2791 }, { "epoch": 1.2644927536231885, "grad_norm": 0.13189485251701055, "learning_rate": 7.443688948827941e-05, "loss": 0.9444, "step": 2792 }, { "epoch": 1.264945652173913, "grad_norm": 0.17670808011493505, "learning_rate": 7.44304536161347e-05, "loss": 0.9449, "step": 2793 }, { "epoch": 1.2653985507246377, "grad_norm": 0.1982697252433833, "learning_rate": 7.442401430192299e-05, "loss": 0.9576, "step": 2794 }, { "epoch": 1.2658514492753623, "grad_norm": 0.22426398920252782, "learning_rate": 7.441757154628802e-05, "loss": 0.9622, "step": 2795 }, { "epoch": 1.266304347826087, "grad_norm": 0.2153016603652356, "learning_rate": 7.441112534987389e-05, "loss": 0.9677, "step": 2796 }, { "epoch": 1.2667572463768115, "grad_norm": 0.1810770040057603, "learning_rate": 7.440467571332502e-05, "loss": 0.9515, "step": 2797 }, { "epoch": 1.2672101449275361, "grad_norm": 0.14636237428589116, "learning_rate": 7.43982226372862e-05, "loss": 0.9644, "step": 2798 }, { "epoch": 1.2676630434782608, "grad_norm": 0.13651361596523187, "learning_rate": 7.439176612240255e-05, "loss": 0.9703, "step": 2799 }, { "epoch": 1.2681159420289856, "grad_norm": 0.12396033426175344, "learning_rate": 7.438530616931955e-05, "loss": 0.9594, "step": 2800 }, { "epoch": 1.2685688405797102, "grad_norm": 0.1501585894343825, "learning_rate": 7.4378842778683e-05, "loss": 0.96, "step": 2801 }, { "epoch": 1.2690217391304348, "grad_norm": 0.15341257654764578, "learning_rate": 7.437237595113906e-05, "loss": 0.9559, "step": 2802 }, { "epoch": 1.2694746376811594, "grad_norm": 0.15962553001957977, "learning_rate": 7.436590568733422e-05, "loss": 0.9615, "step": 2803 }, { "epoch": 1.269927536231884, "grad_norm": 0.1845249756161174, "learning_rate": 7.435943198791535e-05, "loss": 0.9529, "step": 2804 }, { "epoch": 1.2703804347826086, "grad_norm": 0.19179212057860304, "learning_rate": 7.43529548535296e-05, "loss": 0.9552, "step": 2805 }, { "epoch": 1.2708333333333333, "grad_norm": 0.16301706397732627, "learning_rate": 7.434647428482453e-05, "loss": 0.9522, "step": 2806 }, { "epoch": 1.271286231884058, "grad_norm": 0.11537797470924889, "learning_rate": 7.4339990282448e-05, "loss": 0.9526, "step": 2807 }, { "epoch": 1.2717391304347827, "grad_norm": 0.12815479617455514, "learning_rate": 7.433350284704822e-05, "loss": 0.9641, "step": 2808 }, { "epoch": 1.2721920289855073, "grad_norm": 0.17079952337712775, "learning_rate": 7.432701197927374e-05, "loss": 0.9507, "step": 2809 }, { "epoch": 1.272644927536232, "grad_norm": 0.22031811851299418, "learning_rate": 7.43205176797735e-05, "loss": 0.9462, "step": 2810 }, { "epoch": 1.2730978260869565, "grad_norm": 0.2494011200718019, "learning_rate": 7.431401994919671e-05, "loss": 0.9541, "step": 2811 }, { "epoch": 1.2735507246376812, "grad_norm": 0.2177518324561467, "learning_rate": 7.430751878819297e-05, "loss": 0.9769, "step": 2812 }, { "epoch": 1.2740036231884058, "grad_norm": 0.1746774903663362, "learning_rate": 7.430101419741221e-05, "loss": 0.9473, "step": 2813 }, { "epoch": 1.2744565217391304, "grad_norm": 0.1334770703919954, "learning_rate": 7.429450617750471e-05, "loss": 0.9575, "step": 2814 }, { "epoch": 1.274909420289855, "grad_norm": 0.14496643311445398, "learning_rate": 7.428799472912108e-05, "loss": 0.9471, "step": 2815 }, { "epoch": 1.2753623188405796, "grad_norm": 0.14528191146482944, "learning_rate": 7.428147985291228e-05, "loss": 0.9512, "step": 2816 }, { "epoch": 1.2758152173913042, "grad_norm": 0.14638780997694475, "learning_rate": 7.427496154952963e-05, "loss": 0.9439, "step": 2817 }, { "epoch": 1.276268115942029, "grad_norm": 0.1567552331375233, "learning_rate": 7.426843981962475e-05, "loss": 0.9566, "step": 2818 }, { "epoch": 1.2767210144927537, "grad_norm": 0.15353006290135496, "learning_rate": 7.426191466384964e-05, "loss": 0.9334, "step": 2819 }, { "epoch": 1.2771739130434783, "grad_norm": 0.12610308379646765, "learning_rate": 7.425538608285662e-05, "loss": 0.9741, "step": 2820 }, { "epoch": 1.2776268115942029, "grad_norm": 0.09810268984563078, "learning_rate": 7.424885407729838e-05, "loss": 0.9533, "step": 2821 }, { "epoch": 1.2780797101449275, "grad_norm": 0.10962967174808541, "learning_rate": 7.424231864782793e-05, "loss": 0.975, "step": 2822 }, { "epoch": 1.278532608695652, "grad_norm": 0.10995295896754036, "learning_rate": 7.423577979509862e-05, "loss": 0.948, "step": 2823 }, { "epoch": 1.278985507246377, "grad_norm": 0.16599614191483003, "learning_rate": 7.422923751976415e-05, "loss": 0.9594, "step": 2824 }, { "epoch": 1.2794384057971016, "grad_norm": 0.13985759152042687, "learning_rate": 7.422269182247856e-05, "loss": 0.939, "step": 2825 }, { "epoch": 1.2798913043478262, "grad_norm": 0.11375400830369518, "learning_rate": 7.421614270389623e-05, "loss": 0.9665, "step": 2826 }, { "epoch": 1.2803442028985508, "grad_norm": 0.139079840174927, "learning_rate": 7.420959016467193e-05, "loss": 0.964, "step": 2827 }, { "epoch": 1.2807971014492754, "grad_norm": 0.15055761881517143, "learning_rate": 7.420303420546067e-05, "loss": 0.9418, "step": 2828 }, { "epoch": 1.28125, "grad_norm": 0.1596364968732987, "learning_rate": 7.419647482691788e-05, "loss": 0.9555, "step": 2829 }, { "epoch": 1.2817028985507246, "grad_norm": 0.15161324920528582, "learning_rate": 7.41899120296993e-05, "loss": 0.9672, "step": 2830 }, { "epoch": 1.2821557971014492, "grad_norm": 0.14933805990329427, "learning_rate": 7.418334581446106e-05, "loss": 0.9363, "step": 2831 }, { "epoch": 1.2826086956521738, "grad_norm": 0.12216226984274185, "learning_rate": 7.417677618185955e-05, "loss": 0.9477, "step": 2832 }, { "epoch": 1.2830615942028984, "grad_norm": 0.11971323713832865, "learning_rate": 7.417020313255158e-05, "loss": 0.933, "step": 2833 }, { "epoch": 1.283514492753623, "grad_norm": 0.12338071319144985, "learning_rate": 7.416362666719425e-05, "loss": 0.9604, "step": 2834 }, { "epoch": 1.283967391304348, "grad_norm": 0.1358155251195752, "learning_rate": 7.415704678644501e-05, "loss": 0.9611, "step": 2835 }, { "epoch": 1.2844202898550725, "grad_norm": 0.1534349624579873, "learning_rate": 7.41504634909617e-05, "loss": 0.945, "step": 2836 }, { "epoch": 1.2848731884057971, "grad_norm": 0.12533000036711245, "learning_rate": 7.414387678140244e-05, "loss": 0.959, "step": 2837 }, { "epoch": 1.2853260869565217, "grad_norm": 0.11938086616581535, "learning_rate": 7.41372866584257e-05, "loss": 0.9621, "step": 2838 }, { "epoch": 1.2857789855072463, "grad_norm": 0.11054929238048826, "learning_rate": 7.413069312269033e-05, "loss": 0.9446, "step": 2839 }, { "epoch": 1.286231884057971, "grad_norm": 0.12158287857165029, "learning_rate": 7.412409617485547e-05, "loss": 0.9471, "step": 2840 }, { "epoch": 1.2866847826086958, "grad_norm": 0.13752340446216996, "learning_rate": 7.411749581558064e-05, "loss": 0.9389, "step": 2841 }, { "epoch": 1.2871376811594204, "grad_norm": 0.0996952305253322, "learning_rate": 7.41108920455257e-05, "loss": 0.9667, "step": 2842 }, { "epoch": 1.287590579710145, "grad_norm": 0.11459362057269973, "learning_rate": 7.410428486535083e-05, "loss": 0.945, "step": 2843 }, { "epoch": 1.2880434782608696, "grad_norm": 0.1405232161328161, "learning_rate": 7.409767427571655e-05, "loss": 0.9344, "step": 2844 }, { "epoch": 1.2884963768115942, "grad_norm": 0.1802658755506919, "learning_rate": 7.409106027728374e-05, "loss": 0.9644, "step": 2845 }, { "epoch": 1.2889492753623188, "grad_norm": 0.1771507596500797, "learning_rate": 7.408444287071362e-05, "loss": 0.9278, "step": 2846 }, { "epoch": 1.2894021739130435, "grad_norm": 0.16706642825274504, "learning_rate": 7.407782205666774e-05, "loss": 0.9575, "step": 2847 }, { "epoch": 1.289855072463768, "grad_norm": 0.13769561745986605, "learning_rate": 7.407119783580797e-05, "loss": 0.9483, "step": 2848 }, { "epoch": 1.2903079710144927, "grad_norm": 0.10061486005528533, "learning_rate": 7.406457020879658e-05, "loss": 0.9671, "step": 2849 }, { "epoch": 1.2907608695652173, "grad_norm": 0.1402888013860529, "learning_rate": 7.405793917629611e-05, "loss": 0.9355, "step": 2850 }, { "epoch": 1.291213768115942, "grad_norm": 0.1712139315389259, "learning_rate": 7.40513047389695e-05, "loss": 0.9623, "step": 2851 }, { "epoch": 1.2916666666666667, "grad_norm": 0.1568223555759672, "learning_rate": 7.404466689747999e-05, "loss": 0.9443, "step": 2852 }, { "epoch": 1.2921195652173914, "grad_norm": 0.12835110543771003, "learning_rate": 7.403802565249118e-05, "loss": 0.9371, "step": 2853 }, { "epoch": 1.292572463768116, "grad_norm": 0.12717443870179845, "learning_rate": 7.403138100466702e-05, "loss": 0.9324, "step": 2854 }, { "epoch": 1.2930253623188406, "grad_norm": 0.15856626073956087, "learning_rate": 7.402473295467178e-05, "loss": 0.9481, "step": 2855 }, { "epoch": 1.2934782608695652, "grad_norm": 0.18175680442401007, "learning_rate": 7.401808150317005e-05, "loss": 0.9503, "step": 2856 }, { "epoch": 1.2939311594202898, "grad_norm": 0.2022720337053185, "learning_rate": 7.401142665082682e-05, "loss": 0.9416, "step": 2857 }, { "epoch": 1.2943840579710144, "grad_norm": 0.19277993406984048, "learning_rate": 7.400476839830736e-05, "loss": 0.9839, "step": 2858 }, { "epoch": 1.2948369565217392, "grad_norm": 0.1564122397248343, "learning_rate": 7.399810674627732e-05, "loss": 0.9533, "step": 2859 }, { "epoch": 1.2952898550724639, "grad_norm": 0.09188849399953906, "learning_rate": 7.399144169540268e-05, "loss": 0.9664, "step": 2860 }, { "epoch": 1.2957427536231885, "grad_norm": 0.09593665247818527, "learning_rate": 7.398477324634975e-05, "loss": 0.9262, "step": 2861 }, { "epoch": 1.296195652173913, "grad_norm": 0.13674872275054764, "learning_rate": 7.397810139978518e-05, "loss": 0.9478, "step": 2862 }, { "epoch": 1.2966485507246377, "grad_norm": 0.14184419404754375, "learning_rate": 7.397142615637598e-05, "loss": 0.9279, "step": 2863 }, { "epoch": 1.2971014492753623, "grad_norm": 0.14999169461662626, "learning_rate": 7.396474751678947e-05, "loss": 0.9458, "step": 2864 }, { "epoch": 1.297554347826087, "grad_norm": 0.1125290105106249, "learning_rate": 7.395806548169332e-05, "loss": 0.9498, "step": 2865 }, { "epoch": 1.2980072463768115, "grad_norm": 0.1068463623812047, "learning_rate": 7.395138005175555e-05, "loss": 0.9551, "step": 2866 }, { "epoch": 1.2984601449275361, "grad_norm": 0.12035724498031253, "learning_rate": 7.394469122764454e-05, "loss": 0.9489, "step": 2867 }, { "epoch": 1.2989130434782608, "grad_norm": 0.16549714639884447, "learning_rate": 7.393799901002894e-05, "loss": 0.9331, "step": 2868 }, { "epoch": 1.2993659420289856, "grad_norm": 0.1545742333326592, "learning_rate": 7.39313033995778e-05, "loss": 0.9404, "step": 2869 }, { "epoch": 1.2998188405797102, "grad_norm": 0.17008859717204275, "learning_rate": 7.392460439696049e-05, "loss": 0.9625, "step": 2870 }, { "epoch": 1.3002717391304348, "grad_norm": 0.11621146280480965, "learning_rate": 7.391790200284673e-05, "loss": 0.9536, "step": 2871 }, { "epoch": 1.3007246376811594, "grad_norm": 0.12291564294583907, "learning_rate": 7.391119621790655e-05, "loss": 0.9704, "step": 2872 }, { "epoch": 1.301177536231884, "grad_norm": 0.11211865096109835, "learning_rate": 7.390448704281035e-05, "loss": 0.948, "step": 2873 }, { "epoch": 1.3016304347826086, "grad_norm": 0.10372038174379464, "learning_rate": 7.389777447822885e-05, "loss": 0.9399, "step": 2874 }, { "epoch": 1.3020833333333333, "grad_norm": 0.12517700961524375, "learning_rate": 7.389105852483312e-05, "loss": 0.9767, "step": 2875 }, { "epoch": 1.302536231884058, "grad_norm": 0.10439784324475439, "learning_rate": 7.388433918329457e-05, "loss": 0.9448, "step": 2876 }, { "epoch": 1.3029891304347827, "grad_norm": 0.11640736757277599, "learning_rate": 7.387761645428494e-05, "loss": 0.9395, "step": 2877 }, { "epoch": 1.3034420289855073, "grad_norm": 0.1146026044245019, "learning_rate": 7.387089033847631e-05, "loss": 0.9613, "step": 2878 }, { "epoch": 1.303894927536232, "grad_norm": 0.11497895489009115, "learning_rate": 7.386416083654109e-05, "loss": 0.9521, "step": 2879 }, { "epoch": 1.3043478260869565, "grad_norm": 0.14341494838894447, "learning_rate": 7.385742794915205e-05, "loss": 0.9437, "step": 2880 }, { "epoch": 1.3048007246376812, "grad_norm": 0.10628850813899822, "learning_rate": 7.385069167698228e-05, "loss": 0.9337, "step": 2881 }, { "epoch": 1.3052536231884058, "grad_norm": 0.11415792766309686, "learning_rate": 7.384395202070523e-05, "loss": 0.9472, "step": 2882 }, { "epoch": 1.3057065217391304, "grad_norm": 0.1524224869664857, "learning_rate": 7.383720898099468e-05, "loss": 0.9662, "step": 2883 }, { "epoch": 1.306159420289855, "grad_norm": 0.1609675637877204, "learning_rate": 7.38304625585247e-05, "loss": 0.9672, "step": 2884 }, { "epoch": 1.3066123188405796, "grad_norm": 0.13742905580206805, "learning_rate": 7.382371275396979e-05, "loss": 0.9489, "step": 2885 }, { "epoch": 1.3070652173913042, "grad_norm": 0.14586705480555337, "learning_rate": 7.381695956800473e-05, "loss": 0.9576, "step": 2886 }, { "epoch": 1.307518115942029, "grad_norm": 0.16841499188166464, "learning_rate": 7.381020300130461e-05, "loss": 0.9393, "step": 2887 }, { "epoch": 1.3079710144927537, "grad_norm": 0.16158998900670682, "learning_rate": 7.380344305454493e-05, "loss": 0.9373, "step": 2888 }, { "epoch": 1.3084239130434783, "grad_norm": 0.14934146458350966, "learning_rate": 7.379667972840148e-05, "loss": 0.9309, "step": 2889 }, { "epoch": 1.3088768115942029, "grad_norm": 0.13920280464956702, "learning_rate": 7.378991302355038e-05, "loss": 0.9348, "step": 2890 }, { "epoch": 1.3093297101449275, "grad_norm": 0.11526963049240511, "learning_rate": 7.378314294066817e-05, "loss": 0.9567, "step": 2891 }, { "epoch": 1.309782608695652, "grad_norm": 0.11319996838971248, "learning_rate": 7.37763694804316e-05, "loss": 0.9612, "step": 2892 }, { "epoch": 1.310235507246377, "grad_norm": 0.11125753536472562, "learning_rate": 7.376959264351784e-05, "loss": 0.9623, "step": 2893 }, { "epoch": 1.3106884057971016, "grad_norm": 0.13605349441950096, "learning_rate": 7.376281243060443e-05, "loss": 0.9435, "step": 2894 }, { "epoch": 1.3111413043478262, "grad_norm": 0.11076684408196877, "learning_rate": 7.375602884236913e-05, "loss": 0.9725, "step": 2895 }, { "epoch": 1.3115942028985508, "grad_norm": 0.10445051538351748, "learning_rate": 7.374924187949014e-05, "loss": 0.9355, "step": 2896 }, { "epoch": 1.3120471014492754, "grad_norm": 0.11597954502713008, "learning_rate": 7.374245154264597e-05, "loss": 0.9382, "step": 2897 }, { "epoch": 1.3125, "grad_norm": 0.10563504619135712, "learning_rate": 7.373565783251544e-05, "loss": 0.9504, "step": 2898 }, { "epoch": 1.3129528985507246, "grad_norm": 0.1134071881011605, "learning_rate": 7.372886074977774e-05, "loss": 0.9381, "step": 2899 }, { "epoch": 1.3134057971014492, "grad_norm": 0.12757954344599123, "learning_rate": 7.372206029511236e-05, "loss": 0.935, "step": 2900 }, { "epoch": 1.3138586956521738, "grad_norm": 0.16497963255411321, "learning_rate": 7.371525646919921e-05, "loss": 0.9736, "step": 2901 }, { "epoch": 1.3143115942028984, "grad_norm": 0.17486257523704055, "learning_rate": 7.370844927271844e-05, "loss": 0.9551, "step": 2902 }, { "epoch": 1.314764492753623, "grad_norm": 0.14738865350413496, "learning_rate": 7.370163870635056e-05, "loss": 0.9573, "step": 2903 }, { "epoch": 1.315217391304348, "grad_norm": 0.11960644302842847, "learning_rate": 7.369482477077647e-05, "loss": 0.9393, "step": 2904 }, { "epoch": 1.3156702898550725, "grad_norm": 0.1153163615361573, "learning_rate": 7.368800746667734e-05, "loss": 0.9561, "step": 2905 }, { "epoch": 1.3161231884057971, "grad_norm": 0.14234535514594804, "learning_rate": 7.368118679473475e-05, "loss": 0.9455, "step": 2906 }, { "epoch": 1.3165760869565217, "grad_norm": 0.15923320564977306, "learning_rate": 7.367436275563052e-05, "loss": 0.976, "step": 2907 }, { "epoch": 1.3170289855072463, "grad_norm": 0.1187821256450774, "learning_rate": 7.366753535004687e-05, "loss": 0.9598, "step": 2908 }, { "epoch": 1.317481884057971, "grad_norm": 0.10798178901202224, "learning_rate": 7.366070457866639e-05, "loss": 0.95, "step": 2909 }, { "epoch": 1.3179347826086958, "grad_norm": 0.12401421929400551, "learning_rate": 7.365387044217192e-05, "loss": 0.951, "step": 2910 }, { "epoch": 1.3183876811594204, "grad_norm": 0.15013426896754578, "learning_rate": 7.364703294124669e-05, "loss": 0.9619, "step": 2911 }, { "epoch": 1.318840579710145, "grad_norm": 0.16029354121630568, "learning_rate": 7.364019207657427e-05, "loss": 0.9617, "step": 2912 }, { "epoch": 1.3192934782608696, "grad_norm": 0.1451417108164824, "learning_rate": 7.363334784883852e-05, "loss": 0.9662, "step": 2913 }, { "epoch": 1.3197463768115942, "grad_norm": 0.14390560874241104, "learning_rate": 7.36265002587237e-05, "loss": 0.9568, "step": 2914 }, { "epoch": 1.3201992753623188, "grad_norm": 0.13655223342801667, "learning_rate": 7.361964930691438e-05, "loss": 0.9458, "step": 2915 }, { "epoch": 1.3206521739130435, "grad_norm": 0.122878871533095, "learning_rate": 7.361279499409543e-05, "loss": 0.971, "step": 2916 }, { "epoch": 1.321105072463768, "grad_norm": 0.15726377987736775, "learning_rate": 7.36059373209521e-05, "loss": 0.9473, "step": 2917 }, { "epoch": 1.3215579710144927, "grad_norm": 0.1291329296137232, "learning_rate": 7.359907628816995e-05, "loss": 0.9628, "step": 2918 }, { "epoch": 1.3220108695652173, "grad_norm": 0.12012374584021161, "learning_rate": 7.359221189643491e-05, "loss": 0.94, "step": 2919 }, { "epoch": 1.322463768115942, "grad_norm": 0.14788389777975686, "learning_rate": 7.358534414643323e-05, "loss": 0.9749, "step": 2920 }, { "epoch": 1.3229166666666667, "grad_norm": 0.15278250682336436, "learning_rate": 7.357847303885146e-05, "loss": 0.9623, "step": 2921 }, { "epoch": 1.3233695652173914, "grad_norm": 0.1255972142858825, "learning_rate": 7.357159857437654e-05, "loss": 0.9549, "step": 2922 }, { "epoch": 1.323822463768116, "grad_norm": 0.13513145300720675, "learning_rate": 7.35647207536957e-05, "loss": 0.9766, "step": 2923 }, { "epoch": 1.3242753623188406, "grad_norm": 0.13175942040891703, "learning_rate": 7.355783957749656e-05, "loss": 0.9666, "step": 2924 }, { "epoch": 1.3247282608695652, "grad_norm": 0.10311658563848049, "learning_rate": 7.355095504646699e-05, "loss": 0.9388, "step": 2925 }, { "epoch": 1.3251811594202898, "grad_norm": 0.10299990394080577, "learning_rate": 7.35440671612953e-05, "loss": 0.9559, "step": 2926 }, { "epoch": 1.3256340579710144, "grad_norm": 0.14763471117827195, "learning_rate": 7.353717592267005e-05, "loss": 0.9771, "step": 2927 }, { "epoch": 1.3260869565217392, "grad_norm": 0.17422162919762454, "learning_rate": 7.353028133128018e-05, "loss": 0.9496, "step": 2928 }, { "epoch": 1.3265398550724639, "grad_norm": 0.1573348375901316, "learning_rate": 7.352338338781495e-05, "loss": 0.9881, "step": 2929 }, { "epoch": 1.3269927536231885, "grad_norm": 0.121819123090461, "learning_rate": 7.351648209296397e-05, "loss": 0.9743, "step": 2930 }, { "epoch": 1.327445652173913, "grad_norm": 0.11761118743024808, "learning_rate": 7.350957744741716e-05, "loss": 0.9455, "step": 2931 }, { "epoch": 1.3278985507246377, "grad_norm": 0.14370418900523885, "learning_rate": 7.350266945186478e-05, "loss": 0.941, "step": 2932 }, { "epoch": 1.3283514492753623, "grad_norm": 0.17498875322079013, "learning_rate": 7.349575810699745e-05, "loss": 0.9353, "step": 2933 }, { "epoch": 1.328804347826087, "grad_norm": 0.172278620521896, "learning_rate": 7.348884341350612e-05, "loss": 0.9703, "step": 2934 }, { "epoch": 1.3292572463768115, "grad_norm": 0.16816345400904903, "learning_rate": 7.348192537208204e-05, "loss": 0.9684, "step": 2935 }, { "epoch": 1.3297101449275361, "grad_norm": 0.16562433072052996, "learning_rate": 7.347500398341682e-05, "loss": 0.9505, "step": 2936 }, { "epoch": 1.3301630434782608, "grad_norm": 0.17410795106856872, "learning_rate": 7.346807924820241e-05, "loss": 0.9755, "step": 2937 }, { "epoch": 1.3306159420289856, "grad_norm": 0.16760306376062947, "learning_rate": 7.346115116713108e-05, "loss": 0.9613, "step": 2938 }, { "epoch": 1.3310688405797102, "grad_norm": 0.16953240968069797, "learning_rate": 7.345421974089543e-05, "loss": 0.9654, "step": 2939 }, { "epoch": 1.3315217391304348, "grad_norm": 0.13275621129068887, "learning_rate": 7.344728497018844e-05, "loss": 0.9603, "step": 2940 }, { "epoch": 1.3319746376811594, "grad_norm": 0.10399463825030694, "learning_rate": 7.344034685570335e-05, "loss": 0.9465, "step": 2941 }, { "epoch": 1.332427536231884, "grad_norm": 0.2732354682263599, "learning_rate": 7.343340539813382e-05, "loss": 0.9602, "step": 2942 }, { "epoch": 1.3328804347826086, "grad_norm": 0.09622156972366826, "learning_rate": 7.342646059817376e-05, "loss": 0.933, "step": 2943 }, { "epoch": 1.3333333333333333, "grad_norm": 0.10594191795960295, "learning_rate": 7.341951245651747e-05, "loss": 0.9367, "step": 2944 }, { "epoch": 1.333786231884058, "grad_norm": 0.1019637830600752, "learning_rate": 7.341256097385955e-05, "loss": 0.9185, "step": 2945 }, { "epoch": 1.3342391304347827, "grad_norm": 0.1035106217431446, "learning_rate": 7.340560615089496e-05, "loss": 0.9595, "step": 2946 }, { "epoch": 1.3346920289855073, "grad_norm": 0.1245275728199391, "learning_rate": 7.3398647988319e-05, "loss": 0.9567, "step": 2947 }, { "epoch": 1.335144927536232, "grad_norm": 0.14101049162144919, "learning_rate": 7.339168648682728e-05, "loss": 0.967, "step": 2948 }, { "epoch": 1.3355978260869565, "grad_norm": 0.1539426535490226, "learning_rate": 7.338472164711574e-05, "loss": 0.9819, "step": 2949 }, { "epoch": 1.3360507246376812, "grad_norm": 0.1541978210799217, "learning_rate": 7.337775346988068e-05, "loss": 0.9594, "step": 2950 }, { "epoch": 1.3365036231884058, "grad_norm": 0.16808680423667618, "learning_rate": 7.33707819558187e-05, "loss": 0.9508, "step": 2951 }, { "epoch": 1.3369565217391304, "grad_norm": 0.17645318133184054, "learning_rate": 7.336380710562678e-05, "loss": 0.9576, "step": 2952 }, { "epoch": 1.337409420289855, "grad_norm": 0.14022690458358758, "learning_rate": 7.335682892000218e-05, "loss": 0.9623, "step": 2953 }, { "epoch": 1.3378623188405796, "grad_norm": 0.13783002785844747, "learning_rate": 7.334984739964255e-05, "loss": 1.0114, "step": 2954 }, { "epoch": 1.3383152173913042, "grad_norm": 0.13207113304889825, "learning_rate": 7.334286254524582e-05, "loss": 0.9545, "step": 2955 }, { "epoch": 1.338768115942029, "grad_norm": 0.11962512086396361, "learning_rate": 7.333587435751029e-05, "loss": 0.9433, "step": 2956 }, { "epoch": 1.3392210144927537, "grad_norm": 0.10782798416746447, "learning_rate": 7.332888283713457e-05, "loss": 0.9797, "step": 2957 }, { "epoch": 1.3396739130434783, "grad_norm": 0.11688034978403801, "learning_rate": 7.332188798481763e-05, "loss": 0.9378, "step": 2958 }, { "epoch": 1.3401268115942029, "grad_norm": 0.13582131243672627, "learning_rate": 7.331488980125873e-05, "loss": 0.9428, "step": 2959 }, { "epoch": 1.3405797101449275, "grad_norm": 0.17561333800953777, "learning_rate": 7.330788828715752e-05, "loss": 0.9583, "step": 2960 }, { "epoch": 1.341032608695652, "grad_norm": 0.21437979611306127, "learning_rate": 7.330088344321392e-05, "loss": 0.9639, "step": 2961 }, { "epoch": 1.341485507246377, "grad_norm": 0.21798611246777685, "learning_rate": 7.329387527012824e-05, "loss": 0.9826, "step": 2962 }, { "epoch": 1.3419384057971016, "grad_norm": 0.20850877592372324, "learning_rate": 7.32868637686011e-05, "loss": 0.9646, "step": 2963 }, { "epoch": 1.3423913043478262, "grad_norm": 0.1775828735049553, "learning_rate": 7.327984893933342e-05, "loss": 0.9724, "step": 2964 }, { "epoch": 1.3428442028985508, "grad_norm": 0.13133344461809868, "learning_rate": 7.327283078302652e-05, "loss": 0.9286, "step": 2965 }, { "epoch": 1.3432971014492754, "grad_norm": 0.10741096690675536, "learning_rate": 7.326580930038201e-05, "loss": 0.981, "step": 2966 }, { "epoch": 1.34375, "grad_norm": 0.11591452198854904, "learning_rate": 7.325878449210182e-05, "loss": 0.9529, "step": 2967 }, { "epoch": 1.3442028985507246, "grad_norm": 0.1453975331736013, "learning_rate": 7.325175635888823e-05, "loss": 0.9284, "step": 2968 }, { "epoch": 1.3446557971014492, "grad_norm": 0.21712757543355599, "learning_rate": 7.324472490144388e-05, "loss": 0.9522, "step": 2969 }, { "epoch": 1.3451086956521738, "grad_norm": 0.17330080765320874, "learning_rate": 7.323769012047169e-05, "loss": 0.9638, "step": 2970 }, { "epoch": 1.3455615942028984, "grad_norm": 0.21075983342303034, "learning_rate": 7.323065201667496e-05, "loss": 0.9551, "step": 2971 }, { "epoch": 1.346014492753623, "grad_norm": 0.23101469046494247, "learning_rate": 7.322361059075729e-05, "loss": 0.9375, "step": 2972 }, { "epoch": 1.346467391304348, "grad_norm": 0.19093662058739286, "learning_rate": 7.321656584342261e-05, "loss": 0.9412, "step": 2973 }, { "epoch": 1.3469202898550725, "grad_norm": 0.1539770886580237, "learning_rate": 7.320951777537522e-05, "loss": 0.9508, "step": 2974 }, { "epoch": 1.3473731884057971, "grad_norm": 0.11572540353947858, "learning_rate": 7.32024663873197e-05, "loss": 0.9402, "step": 2975 }, { "epoch": 1.3478260869565217, "grad_norm": 0.10825919718716795, "learning_rate": 7.319541167996102e-05, "loss": 0.9543, "step": 2976 }, { "epoch": 1.3482789855072463, "grad_norm": 0.1279463327487897, "learning_rate": 7.318835365400442e-05, "loss": 0.9653, "step": 2977 }, { "epoch": 1.348731884057971, "grad_norm": 0.18057422415148822, "learning_rate": 7.318129231015551e-05, "loss": 0.965, "step": 2978 }, { "epoch": 1.3491847826086958, "grad_norm": 0.20654616168743137, "learning_rate": 7.317422764912023e-05, "loss": 0.9495, "step": 2979 }, { "epoch": 1.3496376811594204, "grad_norm": 0.2223776296981787, "learning_rate": 7.316715967160484e-05, "loss": 0.9713, "step": 2980 }, { "epoch": 1.350090579710145, "grad_norm": 0.1602197144897112, "learning_rate": 7.316008837831595e-05, "loss": 0.9315, "step": 2981 }, { "epoch": 1.3505434782608696, "grad_norm": 0.13767493181856136, "learning_rate": 7.315301376996047e-05, "loss": 0.969, "step": 2982 }, { "epoch": 1.3509963768115942, "grad_norm": 0.10972458983630792, "learning_rate": 7.314593584724566e-05, "loss": 0.9548, "step": 2983 }, { "epoch": 1.3514492753623188, "grad_norm": 0.11183482768945938, "learning_rate": 7.313885461087912e-05, "loss": 0.9486, "step": 2984 }, { "epoch": 1.3519021739130435, "grad_norm": 0.1419243421054468, "learning_rate": 7.31317700615688e-05, "loss": 0.9541, "step": 2985 }, { "epoch": 1.352355072463768, "grad_norm": 0.1582515613585299, "learning_rate": 7.31246822000229e-05, "loss": 0.9477, "step": 2986 }, { "epoch": 1.3528079710144927, "grad_norm": 0.1645748714216847, "learning_rate": 7.311759102695004e-05, "loss": 0.9391, "step": 2987 }, { "epoch": 1.3532608695652173, "grad_norm": 0.13461478955802977, "learning_rate": 7.311049654305911e-05, "loss": 0.9663, "step": 2988 }, { "epoch": 1.353713768115942, "grad_norm": 0.11168903991551622, "learning_rate": 7.310339874905939e-05, "loss": 0.9494, "step": 2989 }, { "epoch": 1.3541666666666667, "grad_norm": 0.10080964327884631, "learning_rate": 7.309629764566042e-05, "loss": 0.9584, "step": 2990 }, { "epoch": 1.3546195652173914, "grad_norm": 0.11487646458043252, "learning_rate": 7.308919323357215e-05, "loss": 0.9573, "step": 2991 }, { "epoch": 1.355072463768116, "grad_norm": 0.13401844007853714, "learning_rate": 7.30820855135048e-05, "loss": 0.947, "step": 2992 }, { "epoch": 1.3555253623188406, "grad_norm": 0.14255600936962473, "learning_rate": 7.307497448616891e-05, "loss": 0.9775, "step": 2993 }, { "epoch": 1.3559782608695652, "grad_norm": 0.14613177553128753, "learning_rate": 7.306786015227543e-05, "loss": 0.9536, "step": 2994 }, { "epoch": 1.3564311594202898, "grad_norm": 0.1347666023398523, "learning_rate": 7.306074251253558e-05, "loss": 0.9872, "step": 2995 }, { "epoch": 1.3568840579710144, "grad_norm": 0.21358541553418178, "learning_rate": 7.305362156766089e-05, "loss": 0.9512, "step": 2996 }, { "epoch": 1.3573369565217392, "grad_norm": 0.2678695935715833, "learning_rate": 7.304649731836328e-05, "loss": 0.9498, "step": 2997 }, { "epoch": 1.3577898550724639, "grad_norm": 0.27406062747053433, "learning_rate": 7.303936976535498e-05, "loss": 0.9618, "step": 2998 }, { "epoch": 1.3582427536231885, "grad_norm": 0.24419908729093426, "learning_rate": 7.303223890934851e-05, "loss": 0.9479, "step": 2999 }, { "epoch": 1.358695652173913, "grad_norm": 0.21143448837980505, "learning_rate": 7.302510475105679e-05, "loss": 0.9563, "step": 3000 }, { "epoch": 1.3591485507246377, "grad_norm": 0.16579092883246163, "learning_rate": 7.301796729119301e-05, "loss": 0.9589, "step": 3001 }, { "epoch": 1.3596014492753623, "grad_norm": 0.19974564023675523, "learning_rate": 7.30108265304707e-05, "loss": 0.9647, "step": 3002 }, { "epoch": 1.360054347826087, "grad_norm": 0.13017704337238206, "learning_rate": 7.300368246960377e-05, "loss": 0.9317, "step": 3003 }, { "epoch": 1.3605072463768115, "grad_norm": 0.2464131545484559, "learning_rate": 7.29965351093064e-05, "loss": 0.9617, "step": 3004 }, { "epoch": 1.3609601449275361, "grad_norm": 0.2923951589141932, "learning_rate": 7.298938445029314e-05, "loss": 0.9465, "step": 3005 }, { "epoch": 1.3614130434782608, "grad_norm": 0.26040452668169667, "learning_rate": 7.298223049327881e-05, "loss": 0.9721, "step": 3006 }, { "epoch": 1.3618659420289856, "grad_norm": 0.19798241473015193, "learning_rate": 7.297507323897867e-05, "loss": 0.9474, "step": 3007 }, { "epoch": 1.3623188405797102, "grad_norm": 0.15979718542825755, "learning_rate": 7.296791268810818e-05, "loss": 0.957, "step": 3008 }, { "epoch": 1.3627717391304348, "grad_norm": 0.13606740255709296, "learning_rate": 7.296074884138322e-05, "loss": 0.9357, "step": 3009 }, { "epoch": 1.3632246376811594, "grad_norm": 0.19077577475974733, "learning_rate": 7.295358169951998e-05, "loss": 0.9706, "step": 3010 }, { "epoch": 1.363677536231884, "grad_norm": 0.22867650918482735, "learning_rate": 7.294641126323494e-05, "loss": 0.9476, "step": 3011 }, { "epoch": 1.3641304347826086, "grad_norm": 0.2306327329226295, "learning_rate": 7.293923753324496e-05, "loss": 0.954, "step": 3012 }, { "epoch": 1.3645833333333333, "grad_norm": 0.1855249901001467, "learning_rate": 7.293206051026722e-05, "loss": 0.9446, "step": 3013 }, { "epoch": 1.365036231884058, "grad_norm": 0.15030433778874341, "learning_rate": 7.292488019501919e-05, "loss": 0.9326, "step": 3014 }, { "epoch": 1.3654891304347827, "grad_norm": 0.12210205996843533, "learning_rate": 7.291769658821873e-05, "loss": 0.9584, "step": 3015 }, { "epoch": 1.3659420289855073, "grad_norm": 0.13117341256523596, "learning_rate": 7.291050969058397e-05, "loss": 0.9484, "step": 3016 }, { "epoch": 1.366394927536232, "grad_norm": 0.12573283904176843, "learning_rate": 7.290331950283341e-05, "loss": 0.9529, "step": 3017 }, { "epoch": 1.3668478260869565, "grad_norm": 0.10789578830379318, "learning_rate": 7.289612602568586e-05, "loss": 0.9432, "step": 3018 }, { "epoch": 1.3673007246376812, "grad_norm": 0.12920079638148643, "learning_rate": 7.288892925986045e-05, "loss": 0.9626, "step": 3019 }, { "epoch": 1.3677536231884058, "grad_norm": 0.15689759733158495, "learning_rate": 7.288172920607668e-05, "loss": 0.9552, "step": 3020 }, { "epoch": 1.3682065217391304, "grad_norm": 0.1532485999959219, "learning_rate": 7.287452586505434e-05, "loss": 0.945, "step": 3021 }, { "epoch": 1.368659420289855, "grad_norm": 0.11351866497623787, "learning_rate": 7.286731923751353e-05, "loss": 0.94, "step": 3022 }, { "epoch": 1.3691123188405796, "grad_norm": 0.1196999965541066, "learning_rate": 7.286010932417474e-05, "loss": 0.9372, "step": 3023 }, { "epoch": 1.3695652173913042, "grad_norm": 0.16633349195469782, "learning_rate": 7.285289612575876e-05, "loss": 0.9396, "step": 3024 }, { "epoch": 1.370018115942029, "grad_norm": 0.18380899527509195, "learning_rate": 7.284567964298669e-05, "loss": 0.953, "step": 3025 }, { "epoch": 1.3704710144927537, "grad_norm": 0.19205867903630555, "learning_rate": 7.283845987657998e-05, "loss": 0.9627, "step": 3026 }, { "epoch": 1.3709239130434783, "grad_norm": 0.19619616811725046, "learning_rate": 7.28312368272604e-05, "loss": 0.95, "step": 3027 }, { "epoch": 1.3713768115942029, "grad_norm": 0.18765744764881037, "learning_rate": 7.282401049575005e-05, "loss": 0.9467, "step": 3028 }, { "epoch": 1.3718297101449275, "grad_norm": 0.12924421539200634, "learning_rate": 7.281678088277134e-05, "loss": 0.9464, "step": 3029 }, { "epoch": 1.372282608695652, "grad_norm": 0.23179205114353968, "learning_rate": 7.280954798904708e-05, "loss": 0.9582, "step": 3030 }, { "epoch": 1.372735507246377, "grad_norm": 0.1656093897794168, "learning_rate": 7.280231181530029e-05, "loss": 0.9607, "step": 3031 }, { "epoch": 1.3731884057971016, "grad_norm": 0.20329095398164132, "learning_rate": 7.279507236225441e-05, "loss": 0.943, "step": 3032 }, { "epoch": 1.3736413043478262, "grad_norm": 0.2961365461642399, "learning_rate": 7.278782963063317e-05, "loss": 0.9443, "step": 3033 }, { "epoch": 1.3740942028985508, "grad_norm": 0.21599610360752663, "learning_rate": 7.278058362116065e-05, "loss": 0.9751, "step": 3034 }, { "epoch": 1.3745471014492754, "grad_norm": 0.19687639798996245, "learning_rate": 7.277333433456126e-05, "loss": 0.9457, "step": 3035 }, { "epoch": 1.375, "grad_norm": 0.16823184758154625, "learning_rate": 7.276608177155968e-05, "loss": 0.9578, "step": 3036 }, { "epoch": 1.3754528985507246, "grad_norm": 0.14448420037921159, "learning_rate": 7.275882593288099e-05, "loss": 0.9356, "step": 3037 }, { "epoch": 1.3759057971014492, "grad_norm": 0.14246110224820016, "learning_rate": 7.275156681925056e-05, "loss": 0.9753, "step": 3038 }, { "epoch": 1.3763586956521738, "grad_norm": 0.19807583173318433, "learning_rate": 7.27443044313941e-05, "loss": 0.934, "step": 3039 }, { "epoch": 1.3768115942028984, "grad_norm": 0.3094853301095586, "learning_rate": 7.273703877003763e-05, "loss": 0.9689, "step": 3040 }, { "epoch": 1.377264492753623, "grad_norm": 0.21773804822998613, "learning_rate": 7.272976983590752e-05, "loss": 0.9434, "step": 3041 }, { "epoch": 1.377717391304348, "grad_norm": 0.12872097541418984, "learning_rate": 7.272249762973046e-05, "loss": 0.9609, "step": 3042 }, { "epoch": 1.3781702898550725, "grad_norm": 0.13261637767764559, "learning_rate": 7.271522215223344e-05, "loss": 0.9433, "step": 3043 }, { "epoch": 1.3786231884057971, "grad_norm": 0.1609685759789749, "learning_rate": 7.270794340414385e-05, "loss": 0.9504, "step": 3044 }, { "epoch": 1.3790760869565217, "grad_norm": 0.19189732113480762, "learning_rate": 7.270066138618931e-05, "loss": 0.9432, "step": 3045 }, { "epoch": 1.3795289855072463, "grad_norm": 0.19986760849073554, "learning_rate": 7.269337609909783e-05, "loss": 0.9467, "step": 3046 }, { "epoch": 1.379981884057971, "grad_norm": 0.17447753728662868, "learning_rate": 7.268608754359774e-05, "loss": 0.9477, "step": 3047 }, { "epoch": 1.3804347826086958, "grad_norm": 0.134792684349206, "learning_rate": 7.267879572041768e-05, "loss": 0.9679, "step": 3048 }, { "epoch": 1.3808876811594204, "grad_norm": 0.1100336099925265, "learning_rate": 7.267150063028664e-05, "loss": 0.9497, "step": 3049 }, { "epoch": 1.381340579710145, "grad_norm": 0.11067452996334469, "learning_rate": 7.26642022739339e-05, "loss": 0.9525, "step": 3050 }, { "epoch": 1.3817934782608696, "grad_norm": 0.13702952173247593, "learning_rate": 7.26569006520891e-05, "loss": 0.9598, "step": 3051 }, { "epoch": 1.3822463768115942, "grad_norm": 0.1264586428148133, "learning_rate": 7.264959576548218e-05, "loss": 0.9591, "step": 3052 }, { "epoch": 1.3826992753623188, "grad_norm": 0.12443443069746997, "learning_rate": 7.264228761484344e-05, "loss": 0.9334, "step": 3053 }, { "epoch": 1.3831521739130435, "grad_norm": 0.1229264239839802, "learning_rate": 7.263497620090349e-05, "loss": 0.9492, "step": 3054 }, { "epoch": 1.383605072463768, "grad_norm": 0.09898176750820065, "learning_rate": 7.262766152439326e-05, "loss": 0.9508, "step": 3055 }, { "epoch": 1.3840579710144927, "grad_norm": 0.1165169275707175, "learning_rate": 7.262034358604398e-05, "loss": 0.9425, "step": 3056 }, { "epoch": 1.3845108695652173, "grad_norm": 0.11907736752296143, "learning_rate": 7.261302238658727e-05, "loss": 0.9619, "step": 3057 }, { "epoch": 1.384963768115942, "grad_norm": 0.0993556002750929, "learning_rate": 7.260569792675504e-05, "loss": 0.9488, "step": 3058 }, { "epoch": 1.3854166666666667, "grad_norm": 0.1278129533908867, "learning_rate": 7.259837020727953e-05, "loss": 0.9691, "step": 3059 }, { "epoch": 1.3858695652173914, "grad_norm": 0.14070191871936727, "learning_rate": 7.259103922889328e-05, "loss": 0.9532, "step": 3060 }, { "epoch": 1.386322463768116, "grad_norm": 0.12909273377773023, "learning_rate": 7.258370499232918e-05, "loss": 0.9424, "step": 3061 }, { "epoch": 1.3867753623188406, "grad_norm": 0.13793873466485784, "learning_rate": 7.257636749832048e-05, "loss": 0.9222, "step": 3062 }, { "epoch": 1.3872282608695652, "grad_norm": 0.1421531821506935, "learning_rate": 7.256902674760069e-05, "loss": 0.9684, "step": 3063 }, { "epoch": 1.3876811594202898, "grad_norm": 0.15318566855722499, "learning_rate": 7.256168274090369e-05, "loss": 0.9488, "step": 3064 }, { "epoch": 1.3881340579710144, "grad_norm": 0.16678892779170715, "learning_rate": 7.255433547896367e-05, "loss": 0.9603, "step": 3065 }, { "epoch": 1.3885869565217392, "grad_norm": 0.1972942801516565, "learning_rate": 7.254698496251513e-05, "loss": 0.9473, "step": 3066 }, { "epoch": 1.3890398550724639, "grad_norm": 0.16865176987542552, "learning_rate": 7.253963119229295e-05, "loss": 0.9639, "step": 3067 }, { "epoch": 1.3894927536231885, "grad_norm": 0.13845455070490745, "learning_rate": 7.253227416903226e-05, "loss": 0.9558, "step": 3068 }, { "epoch": 1.389945652173913, "grad_norm": 0.1335816870837328, "learning_rate": 7.252491389346858e-05, "loss": 0.9325, "step": 3069 }, { "epoch": 1.3903985507246377, "grad_norm": 0.12922853365752257, "learning_rate": 7.251755036633772e-05, "loss": 0.9348, "step": 3070 }, { "epoch": 1.3908514492753623, "grad_norm": 0.14759736514692137, "learning_rate": 7.25101835883758e-05, "loss": 0.9736, "step": 3071 }, { "epoch": 1.391304347826087, "grad_norm": 0.167360826122632, "learning_rate": 7.250281356031932e-05, "loss": 0.961, "step": 3072 }, { "epoch": 1.3917572463768115, "grad_norm": 0.15791195618535667, "learning_rate": 7.249544028290508e-05, "loss": 0.9535, "step": 3073 }, { "epoch": 1.3922101449275361, "grad_norm": 0.126122536260034, "learning_rate": 7.248806375687016e-05, "loss": 0.9639, "step": 3074 }, { "epoch": 1.3926630434782608, "grad_norm": 0.11838038145709887, "learning_rate": 7.248068398295203e-05, "loss": 0.9678, "step": 3075 }, { "epoch": 1.3931159420289856, "grad_norm": 0.11832549314949704, "learning_rate": 7.247330096188846e-05, "loss": 0.9982, "step": 3076 }, { "epoch": 1.3935688405797102, "grad_norm": 0.12413641729493768, "learning_rate": 7.246591469441754e-05, "loss": 0.9398, "step": 3077 }, { "epoch": 1.3940217391304348, "grad_norm": 0.11638051187116669, "learning_rate": 7.245852518127766e-05, "loss": 0.9579, "step": 3078 }, { "epoch": 1.3944746376811594, "grad_norm": 0.11833267648037708, "learning_rate": 7.24511324232076e-05, "loss": 0.9445, "step": 3079 }, { "epoch": 1.394927536231884, "grad_norm": 0.11822253492266588, "learning_rate": 7.244373642094641e-05, "loss": 0.9724, "step": 3080 }, { "epoch": 1.3953804347826086, "grad_norm": 0.14693598564004992, "learning_rate": 7.243633717523348e-05, "loss": 0.9453, "step": 3081 }, { "epoch": 1.3958333333333333, "grad_norm": 0.15345534610119094, "learning_rate": 7.242893468680849e-05, "loss": 0.953, "step": 3082 }, { "epoch": 1.396286231884058, "grad_norm": 0.1304531546513346, "learning_rate": 7.242152895641154e-05, "loss": 0.9585, "step": 3083 }, { "epoch": 1.3967391304347827, "grad_norm": 0.13374737046644153, "learning_rate": 7.241411998478296e-05, "loss": 0.9575, "step": 3084 }, { "epoch": 1.3971920289855073, "grad_norm": 0.14552805695090226, "learning_rate": 7.240670777266345e-05, "loss": 0.9482, "step": 3085 }, { "epoch": 1.397644927536232, "grad_norm": 0.1373371564283488, "learning_rate": 7.239929232079399e-05, "loss": 0.9507, "step": 3086 }, { "epoch": 1.3980978260869565, "grad_norm": 0.14010380306383408, "learning_rate": 7.239187362991595e-05, "loss": 0.9753, "step": 3087 }, { "epoch": 1.3985507246376812, "grad_norm": 0.16090980653066772, "learning_rate": 7.238445170077096e-05, "loss": 0.9381, "step": 3088 }, { "epoch": 1.3990036231884058, "grad_norm": 0.11299598177570402, "learning_rate": 7.237702653410104e-05, "loss": 0.9666, "step": 3089 }, { "epoch": 1.3994565217391304, "grad_norm": 0.13451121593992754, "learning_rate": 7.236959813064846e-05, "loss": 0.9411, "step": 3090 }, { "epoch": 1.399909420289855, "grad_norm": 0.16137977482347074, "learning_rate": 7.236216649115586e-05, "loss": 0.9389, "step": 3091 }, { "epoch": 1.4003623188405796, "grad_norm": 0.14717577677261853, "learning_rate": 7.235473161636618e-05, "loss": 0.963, "step": 3092 }, { "epoch": 1.4008152173913042, "grad_norm": 0.12755794079551833, "learning_rate": 7.234729350702273e-05, "loss": 0.9437, "step": 3093 }, { "epoch": 1.401268115942029, "grad_norm": 0.12080697252401693, "learning_rate": 7.233985216386908e-05, "loss": 0.9469, "step": 3094 }, { "epoch": 1.4017210144927537, "grad_norm": 0.16309263227827897, "learning_rate": 7.233240758764916e-05, "loss": 0.9604, "step": 3095 }, { "epoch": 1.4021739130434783, "grad_norm": 0.1617766464013761, "learning_rate": 7.232495977910722e-05, "loss": 0.9553, "step": 3096 }, { "epoch": 1.4026268115942029, "grad_norm": 0.1412820089652992, "learning_rate": 7.231750873898783e-05, "loss": 0.9608, "step": 3097 }, { "epoch": 1.4030797101449275, "grad_norm": 0.132954887687444, "learning_rate": 7.231005446803588e-05, "loss": 0.9655, "step": 3098 }, { "epoch": 1.403532608695652, "grad_norm": 0.11816617763289049, "learning_rate": 7.230259696699659e-05, "loss": 0.9645, "step": 3099 }, { "epoch": 1.403985507246377, "grad_norm": 0.12745050309506958, "learning_rate": 7.229513623661549e-05, "loss": 0.9647, "step": 3100 }, { "epoch": 1.4044384057971016, "grad_norm": 0.13249608058611426, "learning_rate": 7.228767227763843e-05, "loss": 0.9481, "step": 3101 }, { "epoch": 1.4048913043478262, "grad_norm": 0.11950546162163064, "learning_rate": 7.228020509081162e-05, "loss": 0.934, "step": 3102 }, { "epoch": 1.4053442028985508, "grad_norm": 0.16193959747755066, "learning_rate": 7.227273467688156e-05, "loss": 0.9712, "step": 3103 }, { "epoch": 1.4057971014492754, "grad_norm": 0.21554590370647778, "learning_rate": 7.226526103659508e-05, "loss": 0.9219, "step": 3104 }, { "epoch": 1.40625, "grad_norm": 0.2377156680360693, "learning_rate": 7.225778417069932e-05, "loss": 0.9522, "step": 3105 }, { "epoch": 1.4067028985507246, "grad_norm": 0.22165838727110648, "learning_rate": 7.225030407994175e-05, "loss": 0.9774, "step": 3106 }, { "epoch": 1.4071557971014492, "grad_norm": 0.18043118858405363, "learning_rate": 7.224282076507017e-05, "loss": 0.9496, "step": 3107 }, { "epoch": 1.4076086956521738, "grad_norm": 0.13785086570087002, "learning_rate": 7.223533422683272e-05, "loss": 0.938, "step": 3108 }, { "epoch": 1.4080615942028984, "grad_norm": 0.13007428855508812, "learning_rate": 7.222784446597783e-05, "loss": 0.9623, "step": 3109 }, { "epoch": 1.408514492753623, "grad_norm": 0.15733412845874453, "learning_rate": 7.222035148325424e-05, "loss": 0.943, "step": 3110 }, { "epoch": 1.408967391304348, "grad_norm": 0.16100254627237578, "learning_rate": 7.221285527941106e-05, "loss": 0.9658, "step": 3111 }, { "epoch": 1.4094202898550725, "grad_norm": 0.1608081431249069, "learning_rate": 7.22053558551977e-05, "loss": 0.9627, "step": 3112 }, { "epoch": 1.4098731884057971, "grad_norm": 0.14684159229724747, "learning_rate": 7.219785321136388e-05, "loss": 0.9685, "step": 3113 }, { "epoch": 1.4103260869565217, "grad_norm": 0.1406068450899538, "learning_rate": 7.219034734865965e-05, "loss": 0.9374, "step": 3114 }, { "epoch": 1.4107789855072463, "grad_norm": 0.12417078998353882, "learning_rate": 7.218283826783539e-05, "loss": 0.947, "step": 3115 }, { "epoch": 1.411231884057971, "grad_norm": 0.11819192090697249, "learning_rate": 7.217532596964179e-05, "loss": 0.9501, "step": 3116 }, { "epoch": 1.4116847826086958, "grad_norm": 0.13601319618695248, "learning_rate": 7.216781045482986e-05, "loss": 0.9388, "step": 3117 }, { "epoch": 1.4121376811594204, "grad_norm": 0.21448201523711835, "learning_rate": 7.216029172415095e-05, "loss": 0.9631, "step": 3118 }, { "epoch": 1.412590579710145, "grad_norm": 0.22896942424680344, "learning_rate": 7.215276977835672e-05, "loss": 0.9513, "step": 3119 }, { "epoch": 1.4130434782608696, "grad_norm": 0.2094724329956416, "learning_rate": 7.214524461819913e-05, "loss": 0.9832, "step": 3120 }, { "epoch": 1.4134963768115942, "grad_norm": 0.20745213078741828, "learning_rate": 7.21377162444305e-05, "loss": 0.9408, "step": 3121 }, { "epoch": 1.4139492753623188, "grad_norm": 0.16787943528845015, "learning_rate": 7.213018465780347e-05, "loss": 0.9746, "step": 3122 }, { "epoch": 1.4144021739130435, "grad_norm": 0.1058159825121261, "learning_rate": 7.212264985907095e-05, "loss": 0.9576, "step": 3123 }, { "epoch": 1.414855072463768, "grad_norm": 0.10783155116460212, "learning_rate": 7.211511184898622e-05, "loss": 0.9561, "step": 3124 }, { "epoch": 1.4153079710144927, "grad_norm": 0.18573826536550356, "learning_rate": 7.210757062830286e-05, "loss": 0.9646, "step": 3125 }, { "epoch": 1.4157608695652173, "grad_norm": 0.23135535809928562, "learning_rate": 7.21000261977748e-05, "loss": 0.9497, "step": 3126 }, { "epoch": 1.416213768115942, "grad_norm": 0.24822713629559193, "learning_rate": 7.209247855815625e-05, "loss": 0.9496, "step": 3127 }, { "epoch": 1.4166666666666667, "grad_norm": 0.2216959621648599, "learning_rate": 7.208492771020176e-05, "loss": 0.9718, "step": 3128 }, { "epoch": 1.4171195652173914, "grad_norm": 0.11835577774274357, "learning_rate": 7.20773736546662e-05, "loss": 0.9414, "step": 3129 }, { "epoch": 1.417572463768116, "grad_norm": 0.13839287330533767, "learning_rate": 7.206981639230478e-05, "loss": 0.9637, "step": 3130 }, { "epoch": 1.4180253623188406, "grad_norm": 0.160492695597912, "learning_rate": 7.206225592387298e-05, "loss": 0.9493, "step": 3131 }, { "epoch": 1.4184782608695652, "grad_norm": 0.21199903058185204, "learning_rate": 7.205469225012666e-05, "loss": 0.9714, "step": 3132 }, { "epoch": 1.4189311594202898, "grad_norm": 0.22636641394634296, "learning_rate": 7.204712537182196e-05, "loss": 0.967, "step": 3133 }, { "epoch": 1.4193840579710144, "grad_norm": 0.18527181389393096, "learning_rate": 7.203955528971534e-05, "loss": 0.9347, "step": 3134 }, { "epoch": 1.4198369565217392, "grad_norm": 0.14987223388156334, "learning_rate": 7.203198200456362e-05, "loss": 0.9567, "step": 3135 }, { "epoch": 1.4202898550724639, "grad_norm": 0.12842746441195163, "learning_rate": 7.20244055171239e-05, "loss": 0.9396, "step": 3136 }, { "epoch": 1.4207427536231885, "grad_norm": 0.12596998602511167, "learning_rate": 7.20168258281536e-05, "loss": 0.9617, "step": 3137 }, { "epoch": 1.421195652173913, "grad_norm": 0.123205599636221, "learning_rate": 7.20092429384105e-05, "loss": 0.9557, "step": 3138 }, { "epoch": 1.4216485507246377, "grad_norm": 0.11165438824492714, "learning_rate": 7.200165684865266e-05, "loss": 0.9558, "step": 3139 }, { "epoch": 1.4221014492753623, "grad_norm": 0.12759453095382675, "learning_rate": 7.199406755963847e-05, "loss": 0.9684, "step": 3140 }, { "epoch": 1.422554347826087, "grad_norm": 0.13328253079338698, "learning_rate": 7.198647507212664e-05, "loss": 0.9416, "step": 3141 }, { "epoch": 1.4230072463768115, "grad_norm": 0.15265726996422574, "learning_rate": 7.197887938687622e-05, "loss": 0.9531, "step": 3142 }, { "epoch": 1.4234601449275361, "grad_norm": 0.1466329712805041, "learning_rate": 7.197128050464655e-05, "loss": 0.9437, "step": 3143 }, { "epoch": 1.4239130434782608, "grad_norm": 0.16740133324890455, "learning_rate": 7.196367842619732e-05, "loss": 0.9684, "step": 3144 }, { "epoch": 1.4243659420289856, "grad_norm": 0.19970997978353203, "learning_rate": 7.19560731522885e-05, "loss": 0.9537, "step": 3145 }, { "epoch": 1.4248188405797102, "grad_norm": 0.2010264809108459, "learning_rate": 7.194846468368038e-05, "loss": 0.9625, "step": 3146 }, { "epoch": 1.4252717391304348, "grad_norm": 0.17000355545661391, "learning_rate": 7.194085302113365e-05, "loss": 0.9526, "step": 3147 }, { "epoch": 1.4257246376811594, "grad_norm": 0.10867681205088872, "learning_rate": 7.193323816540922e-05, "loss": 0.9561, "step": 3148 }, { "epoch": 1.426177536231884, "grad_norm": 0.0988250332074863, "learning_rate": 7.192562011726835e-05, "loss": 0.9525, "step": 3149 }, { "epoch": 1.4266304347826086, "grad_norm": 0.1554760342209734, "learning_rate": 7.191799887747267e-05, "loss": 0.931, "step": 3150 }, { "epoch": 1.4270833333333333, "grad_norm": 0.19319103665108597, "learning_rate": 7.191037444678407e-05, "loss": 0.9724, "step": 3151 }, { "epoch": 1.427536231884058, "grad_norm": 0.18463512488939626, "learning_rate": 7.190274682596475e-05, "loss": 0.9425, "step": 3152 }, { "epoch": 1.4279891304347827, "grad_norm": 0.1500036872968907, "learning_rate": 7.189511601577728e-05, "loss": 0.9503, "step": 3153 }, { "epoch": 1.4284420289855073, "grad_norm": 0.11238728667594311, "learning_rate": 7.188748201698452e-05, "loss": 0.9473, "step": 3154 }, { "epoch": 1.428894927536232, "grad_norm": 0.14763525135275612, "learning_rate": 7.187984483034966e-05, "loss": 0.9497, "step": 3155 }, { "epoch": 1.4293478260869565, "grad_norm": 0.15077883656460073, "learning_rate": 7.187220445663618e-05, "loss": 0.942, "step": 3156 }, { "epoch": 1.4298007246376812, "grad_norm": 0.15291534963241965, "learning_rate": 7.186456089660793e-05, "loss": 0.948, "step": 3157 }, { "epoch": 1.4302536231884058, "grad_norm": 0.16452121781217813, "learning_rate": 7.185691415102903e-05, "loss": 0.96, "step": 3158 }, { "epoch": 1.4307065217391304, "grad_norm": 0.1656775516230119, "learning_rate": 7.184926422066394e-05, "loss": 0.948, "step": 3159 }, { "epoch": 1.431159420289855, "grad_norm": 0.1368757370323547, "learning_rate": 7.184161110627743e-05, "loss": 0.9686, "step": 3160 }, { "epoch": 1.4316123188405796, "grad_norm": 0.11582312091072543, "learning_rate": 7.183395480863461e-05, "loss": 0.943, "step": 3161 }, { "epoch": 1.4320652173913042, "grad_norm": 0.14644614574990766, "learning_rate": 7.182629532850087e-05, "loss": 0.9571, "step": 3162 }, { "epoch": 1.432518115942029, "grad_norm": 0.1490555983334549, "learning_rate": 7.181863266664197e-05, "loss": 0.9546, "step": 3163 }, { "epoch": 1.4329710144927537, "grad_norm": 0.11093470151377988, "learning_rate": 7.181096682382393e-05, "loss": 0.9388, "step": 3164 }, { "epoch": 1.4334239130434783, "grad_norm": 0.12270876416135672, "learning_rate": 7.180329780081313e-05, "loss": 0.9556, "step": 3165 }, { "epoch": 1.4338768115942029, "grad_norm": 0.16391423975657649, "learning_rate": 7.179562559837626e-05, "loss": 0.9555, "step": 3166 }, { "epoch": 1.4343297101449275, "grad_norm": 0.14255412272700335, "learning_rate": 7.17879502172803e-05, "loss": 0.9555, "step": 3167 }, { "epoch": 1.434782608695652, "grad_norm": 0.1048751602789927, "learning_rate": 7.178027165829259e-05, "loss": 0.9773, "step": 3168 }, { "epoch": 1.435235507246377, "grad_norm": 0.1344733952281489, "learning_rate": 7.177258992218077e-05, "loss": 0.949, "step": 3169 }, { "epoch": 1.4356884057971016, "grad_norm": 0.17046376334502877, "learning_rate": 7.176490500971277e-05, "loss": 0.9337, "step": 3170 }, { "epoch": 1.4361413043478262, "grad_norm": 0.14857519461426372, "learning_rate": 7.17572169216569e-05, "loss": 0.9395, "step": 3171 }, { "epoch": 1.4365942028985508, "grad_norm": 0.1342095289660778, "learning_rate": 7.174952565878171e-05, "loss": 0.9321, "step": 3172 }, { "epoch": 1.4370471014492754, "grad_norm": 0.15911870626073296, "learning_rate": 7.174183122185615e-05, "loss": 0.9471, "step": 3173 }, { "epoch": 1.4375, "grad_norm": 0.1460035459348224, "learning_rate": 7.173413361164941e-05, "loss": 0.957, "step": 3174 }, { "epoch": 1.4379528985507246, "grad_norm": 0.1444202208937143, "learning_rate": 7.172643282893106e-05, "loss": 0.9562, "step": 3175 }, { "epoch": 1.4384057971014492, "grad_norm": 0.15566205004072303, "learning_rate": 7.171872887447093e-05, "loss": 0.9242, "step": 3176 }, { "epoch": 1.4388586956521738, "grad_norm": 0.14711556065086173, "learning_rate": 7.171102174903922e-05, "loss": 0.9486, "step": 3177 }, { "epoch": 1.4393115942028984, "grad_norm": 0.14493792354734297, "learning_rate": 7.17033114534064e-05, "loss": 0.9484, "step": 3178 }, { "epoch": 1.439764492753623, "grad_norm": 0.1231112891604427, "learning_rate": 7.169559798834332e-05, "loss": 0.9509, "step": 3179 }, { "epoch": 1.440217391304348, "grad_norm": 0.13105988927996085, "learning_rate": 7.168788135462106e-05, "loss": 0.9473, "step": 3180 }, { "epoch": 1.4406702898550725, "grad_norm": 0.11988933307867995, "learning_rate": 7.16801615530111e-05, "loss": 0.957, "step": 3181 }, { "epoch": 1.4411231884057971, "grad_norm": 0.13565427229503685, "learning_rate": 7.16724385842852e-05, "loss": 0.943, "step": 3182 }, { "epoch": 1.4415760869565217, "grad_norm": 0.1452297961565436, "learning_rate": 7.16647124492154e-05, "loss": 0.9612, "step": 3183 }, { "epoch": 1.4420289855072463, "grad_norm": 0.12666882703706256, "learning_rate": 7.165698314857413e-05, "loss": 0.9678, "step": 3184 }, { "epoch": 1.442481884057971, "grad_norm": 0.10462986364109114, "learning_rate": 7.164925068313409e-05, "loss": 0.949, "step": 3185 }, { "epoch": 1.4429347826086958, "grad_norm": 0.0826474300780281, "learning_rate": 7.164151505366831e-05, "loss": 0.9383, "step": 3186 }, { "epoch": 1.4433876811594204, "grad_norm": 0.09084033005401311, "learning_rate": 7.163377626095013e-05, "loss": 0.972, "step": 3187 }, { "epoch": 1.443840579710145, "grad_norm": 0.10136854554783224, "learning_rate": 7.16260343057532e-05, "loss": 0.9638, "step": 3188 }, { "epoch": 1.4442934782608696, "grad_norm": 0.11923259880660486, "learning_rate": 7.161828918885151e-05, "loss": 0.9702, "step": 3189 }, { "epoch": 1.4447463768115942, "grad_norm": 0.1746816795171855, "learning_rate": 7.161054091101936e-05, "loss": 0.9518, "step": 3190 }, { "epoch": 1.4451992753623188, "grad_norm": 0.1612916448772773, "learning_rate": 7.160278947303132e-05, "loss": 0.9587, "step": 3191 }, { "epoch": 1.4456521739130435, "grad_norm": 0.13024312264459195, "learning_rate": 7.159503487566236e-05, "loss": 0.9694, "step": 3192 }, { "epoch": 1.446105072463768, "grad_norm": 0.14355001015089233, "learning_rate": 7.158727711968768e-05, "loss": 0.9488, "step": 3193 }, { "epoch": 1.4465579710144927, "grad_norm": 0.1229578685520653, "learning_rate": 7.157951620588286e-05, "loss": 0.9527, "step": 3194 }, { "epoch": 1.4470108695652173, "grad_norm": 0.12614141570524812, "learning_rate": 7.157175213502376e-05, "loss": 0.9489, "step": 3195 }, { "epoch": 1.447463768115942, "grad_norm": 0.11796794680221406, "learning_rate": 7.156398490788656e-05, "loss": 0.9411, "step": 3196 }, { "epoch": 1.4479166666666667, "grad_norm": 0.11083508515197864, "learning_rate": 7.155621452524779e-05, "loss": 0.951, "step": 3197 }, { "epoch": 1.4483695652173914, "grad_norm": 0.12685613625196632, "learning_rate": 7.154844098788424e-05, "loss": 0.9614, "step": 3198 }, { "epoch": 1.448822463768116, "grad_norm": 0.12377502886787219, "learning_rate": 7.154066429657304e-05, "loss": 0.9534, "step": 3199 }, { "epoch": 1.4492753623188406, "grad_norm": 0.12486753894700958, "learning_rate": 7.153288445209166e-05, "loss": 0.9489, "step": 3200 }, { "epoch": 1.4497282608695652, "grad_norm": 0.13851671297619944, "learning_rate": 7.152510145521786e-05, "loss": 0.9482, "step": 3201 }, { "epoch": 1.4501811594202898, "grad_norm": 0.15721942601423028, "learning_rate": 7.151731530672971e-05, "loss": 0.9466, "step": 3202 }, { "epoch": 1.4506340579710144, "grad_norm": 0.1845121988355928, "learning_rate": 7.15095260074056e-05, "loss": 0.9567, "step": 3203 }, { "epoch": 1.4510869565217392, "grad_norm": 0.19467136329433565, "learning_rate": 7.150173355802425e-05, "loss": 0.9819, "step": 3204 }, { "epoch": 1.4515398550724639, "grad_norm": 0.18755351964459271, "learning_rate": 7.149393795936468e-05, "loss": 0.921, "step": 3205 }, { "epoch": 1.4519927536231885, "grad_norm": 0.14489357170449435, "learning_rate": 7.148613921220622e-05, "loss": 0.9234, "step": 3206 }, { "epoch": 1.452445652173913, "grad_norm": 0.1406233714187823, "learning_rate": 7.147833731732852e-05, "loss": 0.945, "step": 3207 }, { "epoch": 1.4528985507246377, "grad_norm": 0.13354357871590022, "learning_rate": 7.147053227551157e-05, "loss": 0.9472, "step": 3208 }, { "epoch": 1.4533514492753623, "grad_norm": 0.14196879535423076, "learning_rate": 7.146272408753563e-05, "loss": 0.9725, "step": 3209 }, { "epoch": 1.453804347826087, "grad_norm": 0.14541914718138899, "learning_rate": 7.14549127541813e-05, "loss": 0.9505, "step": 3210 }, { "epoch": 1.4542572463768115, "grad_norm": 0.14085281113791912, "learning_rate": 7.144709827622952e-05, "loss": 0.9678, "step": 3211 }, { "epoch": 1.4547101449275361, "grad_norm": 0.09815783653856072, "learning_rate": 7.143928065446148e-05, "loss": 0.9503, "step": 3212 }, { "epoch": 1.4551630434782608, "grad_norm": 0.09687341386383617, "learning_rate": 7.143145988965872e-05, "loss": 0.9342, "step": 3213 }, { "epoch": 1.4556159420289856, "grad_norm": 0.13568561730502504, "learning_rate": 7.142363598260313e-05, "loss": 0.9331, "step": 3214 }, { "epoch": 1.4560688405797102, "grad_norm": 0.16720565884947097, "learning_rate": 7.141580893407683e-05, "loss": 0.9527, "step": 3215 }, { "epoch": 1.4565217391304348, "grad_norm": 0.19536305938135168, "learning_rate": 7.140797874486237e-05, "loss": 0.9482, "step": 3216 }, { "epoch": 1.4569746376811594, "grad_norm": 0.18026370826399105, "learning_rate": 7.140014541574246e-05, "loss": 0.9693, "step": 3217 }, { "epoch": 1.457427536231884, "grad_norm": 0.2925566489416867, "learning_rate": 7.139230894750028e-05, "loss": 0.9364, "step": 3218 }, { "epoch": 1.4578804347826086, "grad_norm": 0.1435318823726752, "learning_rate": 7.138446934091923e-05, "loss": 0.9212, "step": 3219 }, { "epoch": 1.4583333333333333, "grad_norm": 0.13891761604241865, "learning_rate": 7.137662659678303e-05, "loss": 0.9594, "step": 3220 }, { "epoch": 1.458786231884058, "grad_norm": 0.1140730915927373, "learning_rate": 7.136878071587576e-05, "loss": 0.9345, "step": 3221 }, { "epoch": 1.4592391304347827, "grad_norm": 0.12920348635209342, "learning_rate": 7.136093169898178e-05, "loss": 0.9474, "step": 3222 }, { "epoch": 1.4596920289855073, "grad_norm": 0.1317583072630093, "learning_rate": 7.135307954688575e-05, "loss": 0.9302, "step": 3223 }, { "epoch": 1.460144927536232, "grad_norm": 0.13160300774283565, "learning_rate": 7.134522426037268e-05, "loss": 0.9526, "step": 3224 }, { "epoch": 1.4605978260869565, "grad_norm": 0.10674974958290666, "learning_rate": 7.133736584022787e-05, "loss": 0.9422, "step": 3225 }, { "epoch": 1.4610507246376812, "grad_norm": 0.10462580857223383, "learning_rate": 7.132950428723695e-05, "loss": 0.9902, "step": 3226 }, { "epoch": 1.4615036231884058, "grad_norm": 0.11379618929241697, "learning_rate": 7.132163960218582e-05, "loss": 0.9419, "step": 3227 }, { "epoch": 1.4619565217391304, "grad_norm": 0.10849030306345953, "learning_rate": 7.131377178586075e-05, "loss": 0.9378, "step": 3228 }, { "epoch": 1.462409420289855, "grad_norm": 0.10794102260976128, "learning_rate": 7.13059008390483e-05, "loss": 0.9623, "step": 3229 }, { "epoch": 1.4628623188405796, "grad_norm": 0.11371674505084038, "learning_rate": 7.129802676253534e-05, "loss": 0.9516, "step": 3230 }, { "epoch": 1.4633152173913042, "grad_norm": 0.149418519065975, "learning_rate": 7.129014955710902e-05, "loss": 0.94, "step": 3231 }, { "epoch": 1.463768115942029, "grad_norm": 0.11402790691247118, "learning_rate": 7.128226922355691e-05, "loss": 0.9662, "step": 3232 }, { "epoch": 1.4642210144927537, "grad_norm": 0.0989372758137346, "learning_rate": 7.127438576266673e-05, "loss": 0.9596, "step": 3233 }, { "epoch": 1.4646739130434783, "grad_norm": 0.1303776076476035, "learning_rate": 7.126649917522667e-05, "loss": 0.9685, "step": 3234 }, { "epoch": 1.4651268115942029, "grad_norm": 0.1924544934210199, "learning_rate": 7.125860946202516e-05, "loss": 0.9446, "step": 3235 }, { "epoch": 1.4655797101449275, "grad_norm": 0.22804793507666213, "learning_rate": 7.12507166238509e-05, "loss": 0.9547, "step": 3236 }, { "epoch": 1.466032608695652, "grad_norm": 0.21687032586180136, "learning_rate": 7.124282066149298e-05, "loss": 0.9506, "step": 3237 }, { "epoch": 1.466485507246377, "grad_norm": 0.18803627662204914, "learning_rate": 7.123492157574077e-05, "loss": 0.9562, "step": 3238 }, { "epoch": 1.4669384057971016, "grad_norm": 0.15360680938833132, "learning_rate": 7.122701936738397e-05, "loss": 0.9544, "step": 3239 }, { "epoch": 1.4673913043478262, "grad_norm": 0.12061774739825283, "learning_rate": 7.121911403721254e-05, "loss": 0.9592, "step": 3240 }, { "epoch": 1.4678442028985508, "grad_norm": 0.11284374662393795, "learning_rate": 7.121120558601683e-05, "loss": 0.9393, "step": 3241 }, { "epoch": 1.4682971014492754, "grad_norm": 0.0977573701284954, "learning_rate": 7.120329401458742e-05, "loss": 0.947, "step": 3242 }, { "epoch": 1.46875, "grad_norm": 0.14993176042904788, "learning_rate": 7.119537932371527e-05, "loss": 0.9317, "step": 3243 }, { "epoch": 1.4692028985507246, "grad_norm": 0.15779387174894016, "learning_rate": 7.118746151419162e-05, "loss": 0.9294, "step": 3244 }, { "epoch": 1.4696557971014492, "grad_norm": 0.1520726160055097, "learning_rate": 7.117954058680803e-05, "loss": 0.9379, "step": 3245 }, { "epoch": 1.4701086956521738, "grad_norm": 0.16453155337267739, "learning_rate": 7.117161654235634e-05, "loss": 0.9613, "step": 3246 }, { "epoch": 1.4705615942028984, "grad_norm": 0.19419932379077384, "learning_rate": 7.116368938162875e-05, "loss": 0.9331, "step": 3247 }, { "epoch": 1.471014492753623, "grad_norm": 0.18303991575298045, "learning_rate": 7.115575910541776e-05, "loss": 0.9456, "step": 3248 }, { "epoch": 1.471467391304348, "grad_norm": 0.1473785503071682, "learning_rate": 7.114782571451616e-05, "loss": 0.9703, "step": 3249 }, { "epoch": 1.4719202898550725, "grad_norm": 0.10693942889468637, "learning_rate": 7.113988920971706e-05, "loss": 0.9561, "step": 3250 }, { "epoch": 1.4723731884057971, "grad_norm": 0.10565948608203865, "learning_rate": 7.11319495918139e-05, "loss": 0.9698, "step": 3251 }, { "epoch": 1.4728260869565217, "grad_norm": 0.10133787312855158, "learning_rate": 7.112400686160039e-05, "loss": 0.9461, "step": 3252 }, { "epoch": 1.4732789855072463, "grad_norm": 0.08663118711099975, "learning_rate": 7.11160610198706e-05, "loss": 0.9474, "step": 3253 }, { "epoch": 1.473731884057971, "grad_norm": 0.09580088179998683, "learning_rate": 7.110811206741889e-05, "loss": 0.9486, "step": 3254 }, { "epoch": 1.4741847826086958, "grad_norm": 0.10981367934777604, "learning_rate": 7.11001600050399e-05, "loss": 0.9385, "step": 3255 }, { "epoch": 1.4746376811594204, "grad_norm": 0.13138426038433895, "learning_rate": 7.109220483352867e-05, "loss": 0.9507, "step": 3256 }, { "epoch": 1.475090579710145, "grad_norm": 0.127708935287636, "learning_rate": 7.108424655368043e-05, "loss": 0.9573, "step": 3257 }, { "epoch": 1.4755434782608696, "grad_norm": 0.11259796719429734, "learning_rate": 7.107628516629079e-05, "loss": 0.9393, "step": 3258 }, { "epoch": 1.4759963768115942, "grad_norm": 0.0940359621084622, "learning_rate": 7.106832067215569e-05, "loss": 0.952, "step": 3259 }, { "epoch": 1.4764492753623188, "grad_norm": 0.10651961340922592, "learning_rate": 7.106035307207135e-05, "loss": 0.9667, "step": 3260 }, { "epoch": 1.4769021739130435, "grad_norm": 0.12165619196957529, "learning_rate": 7.105238236683429e-05, "loss": 0.9364, "step": 3261 }, { "epoch": 1.477355072463768, "grad_norm": 0.1207290677982003, "learning_rate": 7.104440855724136e-05, "loss": 0.9483, "step": 3262 }, { "epoch": 1.4778079710144927, "grad_norm": 0.09428447222782811, "learning_rate": 7.103643164408971e-05, "loss": 0.9472, "step": 3263 }, { "epoch": 1.4782608695652173, "grad_norm": 0.10351692964159773, "learning_rate": 7.10284516281768e-05, "loss": 0.981, "step": 3264 }, { "epoch": 1.478713768115942, "grad_norm": 0.09985751254640048, "learning_rate": 7.102046851030041e-05, "loss": 0.9424, "step": 3265 }, { "epoch": 1.4791666666666667, "grad_norm": 0.11972583154142742, "learning_rate": 7.101248229125864e-05, "loss": 0.9633, "step": 3266 }, { "epoch": 1.4796195652173914, "grad_norm": 0.16279818048476913, "learning_rate": 7.100449297184986e-05, "loss": 0.9426, "step": 3267 }, { "epoch": 1.480072463768116, "grad_norm": 0.20550759958600995, "learning_rate": 7.09965005528728e-05, "loss": 0.9513, "step": 3268 }, { "epoch": 1.4805253623188406, "grad_norm": 0.17950813627762327, "learning_rate": 7.098850503512645e-05, "loss": 0.9622, "step": 3269 }, { "epoch": 1.4809782608695652, "grad_norm": 0.1273999856032942, "learning_rate": 7.098050641941015e-05, "loss": 0.9426, "step": 3270 }, { "epoch": 1.4814311594202898, "grad_norm": 0.11412071859539645, "learning_rate": 7.097250470652353e-05, "loss": 0.9392, "step": 3271 }, { "epoch": 1.4818840579710144, "grad_norm": 0.1327288219866518, "learning_rate": 7.096449989726653e-05, "loss": 0.9688, "step": 3272 }, { "epoch": 1.4823369565217392, "grad_norm": 0.13052886782257284, "learning_rate": 7.095649199243943e-05, "loss": 0.9681, "step": 3273 }, { "epoch": 1.4827898550724639, "grad_norm": 0.12225086497685816, "learning_rate": 7.094848099284275e-05, "loss": 0.926, "step": 3274 }, { "epoch": 1.4832427536231885, "grad_norm": 0.11777246734528678, "learning_rate": 7.094046689927739e-05, "loss": 0.9393, "step": 3275 }, { "epoch": 1.483695652173913, "grad_norm": 0.11810101848986877, "learning_rate": 7.093244971254452e-05, "loss": 0.9706, "step": 3276 }, { "epoch": 1.4841485507246377, "grad_norm": 0.11546084108405065, "learning_rate": 7.092442943344563e-05, "loss": 0.9622, "step": 3277 }, { "epoch": 1.4846014492753623, "grad_norm": 0.22047272811623825, "learning_rate": 7.091640606278253e-05, "loss": 0.9565, "step": 3278 }, { "epoch": 1.485054347826087, "grad_norm": 0.1049162176448839, "learning_rate": 7.090837960135734e-05, "loss": 0.9796, "step": 3279 }, { "epoch": 1.4855072463768115, "grad_norm": 0.11594943481226576, "learning_rate": 7.090035004997245e-05, "loss": 0.919, "step": 3280 }, { "epoch": 1.4859601449275361, "grad_norm": 0.18508467418025273, "learning_rate": 7.089231740943059e-05, "loss": 0.9641, "step": 3281 }, { "epoch": 1.4864130434782608, "grad_norm": 0.21794287764283923, "learning_rate": 7.088428168053481e-05, "loss": 0.9388, "step": 3282 }, { "epoch": 1.4868659420289856, "grad_norm": 0.24468806878241642, "learning_rate": 7.087624286408845e-05, "loss": 0.9583, "step": 3283 }, { "epoch": 1.4873188405797102, "grad_norm": 0.23970470068133332, "learning_rate": 7.086820096089518e-05, "loss": 0.9479, "step": 3284 }, { "epoch": 1.4877717391304348, "grad_norm": 0.20886050715623192, "learning_rate": 7.086015597175893e-05, "loss": 0.9333, "step": 3285 }, { "epoch": 1.4882246376811594, "grad_norm": 0.15612151030626892, "learning_rate": 7.085210789748399e-05, "loss": 0.97, "step": 3286 }, { "epoch": 1.488677536231884, "grad_norm": 0.12283726974764281, "learning_rate": 7.084405673887494e-05, "loss": 0.9415, "step": 3287 }, { "epoch": 1.4891304347826086, "grad_norm": 0.10304551892039175, "learning_rate": 7.083600249673665e-05, "loss": 0.964, "step": 3288 }, { "epoch": 1.4895833333333333, "grad_norm": 0.10666461991521484, "learning_rate": 7.082794517187432e-05, "loss": 0.9423, "step": 3289 }, { "epoch": 1.490036231884058, "grad_norm": 0.14269329359880414, "learning_rate": 7.081988476509348e-05, "loss": 0.958, "step": 3290 }, { "epoch": 1.4904891304347827, "grad_norm": 0.16100110651608246, "learning_rate": 7.081182127719992e-05, "loss": 0.9372, "step": 3291 }, { "epoch": 1.4909420289855073, "grad_norm": 0.17003328934605969, "learning_rate": 7.080375470899974e-05, "loss": 0.9458, "step": 3292 }, { "epoch": 1.491394927536232, "grad_norm": 0.16904551851778152, "learning_rate": 7.079568506129941e-05, "loss": 0.9703, "step": 3293 }, { "epoch": 1.4918478260869565, "grad_norm": 0.1877070341961091, "learning_rate": 7.078761233490564e-05, "loss": 0.9777, "step": 3294 }, { "epoch": 1.4923007246376812, "grad_norm": 0.16858299012475098, "learning_rate": 7.077953653062548e-05, "loss": 0.963, "step": 3295 }, { "epoch": 1.4927536231884058, "grad_norm": 0.13530952409621846, "learning_rate": 7.077145764926628e-05, "loss": 0.9528, "step": 3296 }, { "epoch": 1.4932065217391304, "grad_norm": 0.2101398601743832, "learning_rate": 7.076337569163568e-05, "loss": 0.9454, "step": 3297 }, { "epoch": 1.493659420289855, "grad_norm": 0.20854144560687402, "learning_rate": 7.075529065854168e-05, "loss": 0.952, "step": 3298 }, { "epoch": 1.4941123188405796, "grad_norm": 0.21736895431416317, "learning_rate": 7.074720255079252e-05, "loss": 0.9279, "step": 3299 }, { "epoch": 1.4945652173913042, "grad_norm": 0.22747120629108106, "learning_rate": 7.073911136919683e-05, "loss": 0.9817, "step": 3300 }, { "epoch": 1.495018115942029, "grad_norm": 0.25247250596371906, "learning_rate": 7.073101711456343e-05, "loss": 0.9518, "step": 3301 }, { "epoch": 1.4954710144927537, "grad_norm": 0.22962215900794536, "learning_rate": 7.072291978770157e-05, "loss": 0.9497, "step": 3302 }, { "epoch": 1.4959239130434783, "grad_norm": 0.20185883109560532, "learning_rate": 7.071481938942074e-05, "loss": 0.9335, "step": 3303 }, { "epoch": 1.4963768115942029, "grad_norm": 0.15665968608654526, "learning_rate": 7.070671592053073e-05, "loss": 0.9279, "step": 3304 }, { "epoch": 1.4968297101449275, "grad_norm": 0.11814575309965378, "learning_rate": 7.069860938184168e-05, "loss": 0.9552, "step": 3305 }, { "epoch": 1.497282608695652, "grad_norm": 0.10285462068280131, "learning_rate": 7.069049977416399e-05, "loss": 0.9943, "step": 3306 }, { "epoch": 1.497735507246377, "grad_norm": 0.10902912992024193, "learning_rate": 7.06823870983084e-05, "loss": 0.9438, "step": 3307 }, { "epoch": 1.4981884057971016, "grad_norm": 0.10899893495498043, "learning_rate": 7.067427135508598e-05, "loss": 0.9598, "step": 3308 }, { "epoch": 1.4986413043478262, "grad_norm": 0.10481781168851007, "learning_rate": 7.066615254530803e-05, "loss": 0.9485, "step": 3309 }, { "epoch": 1.4990942028985508, "grad_norm": 0.12992595333425475, "learning_rate": 7.065803066978621e-05, "loss": 0.9427, "step": 3310 }, { "epoch": 1.4995471014492754, "grad_norm": 0.14343507741571984, "learning_rate": 7.06499057293325e-05, "loss": 0.9214, "step": 3311 }, { "epoch": 1.5, "grad_norm": 0.1528922286310592, "learning_rate": 7.064177772475912e-05, "loss": 0.9559, "step": 3312 }, { "epoch": 1.5004528985507246, "grad_norm": 0.1131355546972645, "learning_rate": 7.063364665687869e-05, "loss": 0.9463, "step": 3313 }, { "epoch": 1.5009057971014492, "grad_norm": 0.14041793344378647, "learning_rate": 7.062551252650405e-05, "loss": 0.9879, "step": 3314 }, { "epoch": 1.5013586956521738, "grad_norm": 0.16101531433349017, "learning_rate": 7.061737533444838e-05, "loss": 0.9587, "step": 3315 }, { "epoch": 1.5018115942028984, "grad_norm": 0.1227058336281007, "learning_rate": 7.06092350815252e-05, "loss": 0.9572, "step": 3316 }, { "epoch": 1.502264492753623, "grad_norm": 0.09938989575241261, "learning_rate": 7.060109176854827e-05, "loss": 0.947, "step": 3317 }, { "epoch": 1.5027173913043477, "grad_norm": 0.11032723886659909, "learning_rate": 7.059294539633171e-05, "loss": 0.9531, "step": 3318 }, { "epoch": 1.5031702898550725, "grad_norm": 0.13406967478426532, "learning_rate": 7.058479596568993e-05, "loss": 0.9517, "step": 3319 }, { "epoch": 1.5036231884057971, "grad_norm": 0.13675098362585697, "learning_rate": 7.057664347743762e-05, "loss": 0.9524, "step": 3320 }, { "epoch": 1.5040760869565217, "grad_norm": 0.14096928102820533, "learning_rate": 7.056848793238981e-05, "loss": 0.9678, "step": 3321 }, { "epoch": 1.5045289855072463, "grad_norm": 0.11047733970624225, "learning_rate": 7.056032933136183e-05, "loss": 0.9253, "step": 3322 }, { "epoch": 1.5049818840579712, "grad_norm": 0.10218620491624089, "learning_rate": 7.055216767516928e-05, "loss": 0.9623, "step": 3323 }, { "epoch": 1.5054347826086958, "grad_norm": 0.13906329574849674, "learning_rate": 7.054400296462813e-05, "loss": 0.9634, "step": 3324 }, { "epoch": 1.5058876811594204, "grad_norm": 0.11070438962947471, "learning_rate": 7.053583520055459e-05, "loss": 0.9375, "step": 3325 }, { "epoch": 1.506340579710145, "grad_norm": 0.11856975282870792, "learning_rate": 7.052766438376522e-05, "loss": 0.9707, "step": 3326 }, { "epoch": 1.5067934782608696, "grad_norm": 0.12009199683623327, "learning_rate": 7.051949051507689e-05, "loss": 0.9416, "step": 3327 }, { "epoch": 1.5072463768115942, "grad_norm": 0.1225802464219543, "learning_rate": 7.051131359530671e-05, "loss": 0.965, "step": 3328 }, { "epoch": 1.5076992753623188, "grad_norm": 0.1389928685900668, "learning_rate": 7.050313362527216e-05, "loss": 0.9498, "step": 3329 }, { "epoch": 1.5081521739130435, "grad_norm": 0.15547165620480136, "learning_rate": 7.0494950605791e-05, "loss": 0.9636, "step": 3330 }, { "epoch": 1.508605072463768, "grad_norm": 0.1707662558740942, "learning_rate": 7.048676453768131e-05, "loss": 0.9769, "step": 3331 }, { "epoch": 1.5090579710144927, "grad_norm": 0.19188160974552354, "learning_rate": 7.047857542176145e-05, "loss": 0.944, "step": 3332 }, { "epoch": 1.5095108695652173, "grad_norm": 0.1631545539133122, "learning_rate": 7.047038325885012e-05, "loss": 0.9551, "step": 3333 }, { "epoch": 1.509963768115942, "grad_norm": 0.10648065218087303, "learning_rate": 7.04621880497663e-05, "loss": 0.9329, "step": 3334 }, { "epoch": 1.5104166666666665, "grad_norm": 0.11455387986618029, "learning_rate": 7.045398979532925e-05, "loss": 0.9638, "step": 3335 }, { "epoch": 1.5108695652173914, "grad_norm": 0.12311550047352654, "learning_rate": 7.04457884963586e-05, "loss": 0.9454, "step": 3336 }, { "epoch": 1.511322463768116, "grad_norm": 0.17275759376040348, "learning_rate": 7.043758415367422e-05, "loss": 0.9506, "step": 3337 }, { "epoch": 1.5117753623188406, "grad_norm": 0.20854982482181675, "learning_rate": 7.042937676809631e-05, "loss": 0.9371, "step": 3338 }, { "epoch": 1.5122282608695652, "grad_norm": 0.19088206762815121, "learning_rate": 7.04211663404454e-05, "loss": 0.9522, "step": 3339 }, { "epoch": 1.5126811594202898, "grad_norm": 0.15459885045065613, "learning_rate": 7.041295287154228e-05, "loss": 0.9512, "step": 3340 }, { "epoch": 1.5131340579710146, "grad_norm": 0.12147630452758021, "learning_rate": 7.040473636220807e-05, "loss": 0.9358, "step": 3341 }, { "epoch": 1.5135869565217392, "grad_norm": 0.14722249790275055, "learning_rate": 7.039651681326418e-05, "loss": 0.9487, "step": 3342 }, { "epoch": 1.5140398550724639, "grad_norm": 0.17816574974539734, "learning_rate": 7.038829422553234e-05, "loss": 0.9478, "step": 3343 }, { "epoch": 1.5144927536231885, "grad_norm": 0.13730487484711915, "learning_rate": 7.038006859983455e-05, "loss": 0.9634, "step": 3344 }, { "epoch": 1.514945652173913, "grad_norm": 0.10747270838149307, "learning_rate": 7.03718399369932e-05, "loss": 0.9481, "step": 3345 }, { "epoch": 1.5153985507246377, "grad_norm": 0.11067671514933881, "learning_rate": 7.036360823783085e-05, "loss": 0.9532, "step": 3346 }, { "epoch": 1.5158514492753623, "grad_norm": 0.11174191312500581, "learning_rate": 7.035537350317049e-05, "loss": 0.9505, "step": 3347 }, { "epoch": 1.516304347826087, "grad_norm": 0.12958142912370038, "learning_rate": 7.034713573383533e-05, "loss": 0.951, "step": 3348 }, { "epoch": 1.5167572463768115, "grad_norm": 0.15836089149096721, "learning_rate": 7.033889493064893e-05, "loss": 0.9855, "step": 3349 }, { "epoch": 1.5172101449275361, "grad_norm": 0.18228684235761541, "learning_rate": 7.033065109443512e-05, "loss": 0.9422, "step": 3350 }, { "epoch": 1.5176630434782608, "grad_norm": 0.18130643371091357, "learning_rate": 7.032240422601805e-05, "loss": 0.9571, "step": 3351 }, { "epoch": 1.5181159420289854, "grad_norm": 0.19998964238667338, "learning_rate": 7.031415432622219e-05, "loss": 0.949, "step": 3352 }, { "epoch": 1.5185688405797102, "grad_norm": 0.18183472480296117, "learning_rate": 7.030590139587228e-05, "loss": 0.9637, "step": 3353 }, { "epoch": 1.5190217391304348, "grad_norm": 0.16398020938482458, "learning_rate": 7.029764543579339e-05, "loss": 0.946, "step": 3354 }, { "epoch": 1.5194746376811594, "grad_norm": 0.14103804070608816, "learning_rate": 7.028938644681088e-05, "loss": 0.9419, "step": 3355 }, { "epoch": 1.519927536231884, "grad_norm": 0.09879659786966087, "learning_rate": 7.028112442975038e-05, "loss": 0.934, "step": 3356 }, { "epoch": 1.5203804347826086, "grad_norm": 0.1338862855744594, "learning_rate": 7.027285938543791e-05, "loss": 0.9276, "step": 3357 }, { "epoch": 1.5208333333333335, "grad_norm": 0.18050949226638996, "learning_rate": 7.026459131469972e-05, "loss": 0.9493, "step": 3358 }, { "epoch": 1.521286231884058, "grad_norm": 0.22317777722965929, "learning_rate": 7.025632021836238e-05, "loss": 0.9537, "step": 3359 }, { "epoch": 1.5217391304347827, "grad_norm": 0.2403004900356599, "learning_rate": 7.024804609725275e-05, "loss": 0.9622, "step": 3360 }, { "epoch": 1.5221920289855073, "grad_norm": 0.23096508259234733, "learning_rate": 7.023976895219802e-05, "loss": 0.964, "step": 3361 }, { "epoch": 1.522644927536232, "grad_norm": 0.2057306984709974, "learning_rate": 7.023148878402566e-05, "loss": 0.9574, "step": 3362 }, { "epoch": 1.5230978260869565, "grad_norm": 0.16016559487908374, "learning_rate": 7.022320559356348e-05, "loss": 0.9439, "step": 3363 }, { "epoch": 1.5235507246376812, "grad_norm": 0.10459460241735621, "learning_rate": 7.021491938163953e-05, "loss": 0.9701, "step": 3364 }, { "epoch": 1.5240036231884058, "grad_norm": 0.13819825550725576, "learning_rate": 7.020663014908222e-05, "loss": 0.9627, "step": 3365 }, { "epoch": 1.5244565217391304, "grad_norm": 0.17891653966046298, "learning_rate": 7.019833789672023e-05, "loss": 0.9369, "step": 3366 }, { "epoch": 1.524909420289855, "grad_norm": 0.1778604567977752, "learning_rate": 7.019004262538254e-05, "loss": 0.9403, "step": 3367 }, { "epoch": 1.5253623188405796, "grad_norm": 0.20323648108867445, "learning_rate": 7.018174433589844e-05, "loss": 0.9629, "step": 3368 }, { "epoch": 1.5258152173913042, "grad_norm": 0.20295813452960867, "learning_rate": 7.017344302909755e-05, "loss": 0.9395, "step": 3369 }, { "epoch": 1.5262681159420288, "grad_norm": 0.18138203435404898, "learning_rate": 7.016513870580973e-05, "loss": 0.9492, "step": 3370 }, { "epoch": 1.5267210144927537, "grad_norm": 0.13804274693328314, "learning_rate": 7.01568313668652e-05, "loss": 0.9583, "step": 3371 }, { "epoch": 1.5271739130434783, "grad_norm": 0.1309229638712217, "learning_rate": 7.014852101309445e-05, "loss": 0.9486, "step": 3372 }, { "epoch": 1.5276268115942029, "grad_norm": 0.1398625911269056, "learning_rate": 7.014020764532828e-05, "loss": 0.9611, "step": 3373 }, { "epoch": 1.5280797101449275, "grad_norm": 0.1506123320881181, "learning_rate": 7.013189126439778e-05, "loss": 0.9576, "step": 3374 }, { "epoch": 1.5285326086956523, "grad_norm": 0.14059917163450683, "learning_rate": 7.012357187113438e-05, "loss": 0.9285, "step": 3375 }, { "epoch": 1.528985507246377, "grad_norm": 0.13640615137010495, "learning_rate": 7.011524946636976e-05, "loss": 0.9414, "step": 3376 }, { "epoch": 1.5294384057971016, "grad_norm": 0.12797578224909414, "learning_rate": 7.010692405093593e-05, "loss": 0.9465, "step": 3377 }, { "epoch": 1.5298913043478262, "grad_norm": 0.12569409863821654, "learning_rate": 7.009859562566518e-05, "loss": 0.9532, "step": 3378 }, { "epoch": 1.5303442028985508, "grad_norm": 0.13723071406333923, "learning_rate": 7.009026419139015e-05, "loss": 0.9412, "step": 3379 }, { "epoch": 1.5307971014492754, "grad_norm": 0.18583110185864946, "learning_rate": 7.008192974894371e-05, "loss": 0.9473, "step": 3380 }, { "epoch": 1.53125, "grad_norm": 0.17920894583431837, "learning_rate": 7.00735922991591e-05, "loss": 0.9534, "step": 3381 }, { "epoch": 1.5317028985507246, "grad_norm": 0.17450643019718576, "learning_rate": 7.00652518428698e-05, "loss": 0.9346, "step": 3382 }, { "epoch": 1.5321557971014492, "grad_norm": 0.20105577415032588, "learning_rate": 7.005690838090964e-05, "loss": 0.9511, "step": 3383 }, { "epoch": 1.5326086956521738, "grad_norm": 0.18478066505778964, "learning_rate": 7.004856191411271e-05, "loss": 0.9352, "step": 3384 }, { "epoch": 1.5330615942028984, "grad_norm": 0.10520432097602421, "learning_rate": 7.004021244331344e-05, "loss": 0.9425, "step": 3385 }, { "epoch": 1.533514492753623, "grad_norm": 0.12428726191087332, "learning_rate": 7.003185996934653e-05, "loss": 0.9475, "step": 3386 }, { "epoch": 1.5339673913043477, "grad_norm": 0.15076327080556068, "learning_rate": 7.002350449304698e-05, "loss": 0.9173, "step": 3387 }, { "epoch": 1.5344202898550725, "grad_norm": 0.11324697359216325, "learning_rate": 7.001514601525012e-05, "loss": 0.9365, "step": 3388 }, { "epoch": 1.5348731884057971, "grad_norm": 0.1279127973197731, "learning_rate": 7.000678453679154e-05, "loss": 0.9417, "step": 3389 }, { "epoch": 1.5353260869565217, "grad_norm": 0.1448509532467587, "learning_rate": 6.999842005850715e-05, "loss": 0.974, "step": 3390 }, { "epoch": 1.5357789855072463, "grad_norm": 0.12738550621333605, "learning_rate": 6.999005258123319e-05, "loss": 0.9516, "step": 3391 }, { "epoch": 1.5362318840579712, "grad_norm": 0.1154378081768558, "learning_rate": 6.998168210580614e-05, "loss": 0.9454, "step": 3392 }, { "epoch": 1.5366847826086958, "grad_norm": 0.13446976728566576, "learning_rate": 6.997330863306282e-05, "loss": 0.9506, "step": 3393 }, { "epoch": 1.5371376811594204, "grad_norm": 0.12957866177145436, "learning_rate": 6.996493216384034e-05, "loss": 0.964, "step": 3394 }, { "epoch": 1.537590579710145, "grad_norm": 0.10827099148963189, "learning_rate": 6.995655269897611e-05, "loss": 0.9565, "step": 3395 }, { "epoch": 1.5380434782608696, "grad_norm": 0.10425057744149513, "learning_rate": 6.994817023930784e-05, "loss": 0.9415, "step": 3396 }, { "epoch": 1.5384963768115942, "grad_norm": 0.18309305589620134, "learning_rate": 6.993978478567352e-05, "loss": 0.9786, "step": 3397 }, { "epoch": 1.5389492753623188, "grad_norm": 0.13863085298082503, "learning_rate": 6.99313963389115e-05, "loss": 0.9486, "step": 3398 }, { "epoch": 1.5394021739130435, "grad_norm": 0.11662285718905262, "learning_rate": 6.992300489986033e-05, "loss": 0.9542, "step": 3399 }, { "epoch": 1.539855072463768, "grad_norm": 0.09230193633871937, "learning_rate": 6.991461046935895e-05, "loss": 0.9562, "step": 3400 }, { "epoch": 1.5403079710144927, "grad_norm": 0.11333881778942156, "learning_rate": 6.990621304824656e-05, "loss": 0.9365, "step": 3401 }, { "epoch": 1.5407608695652173, "grad_norm": 0.12944110271481107, "learning_rate": 6.989781263736266e-05, "loss": 0.953, "step": 3402 }, { "epoch": 1.541213768115942, "grad_norm": 0.13725729026739025, "learning_rate": 6.988940923754706e-05, "loss": 0.9769, "step": 3403 }, { "epoch": 1.5416666666666665, "grad_norm": 0.1625660028279402, "learning_rate": 6.988100284963985e-05, "loss": 0.9695, "step": 3404 }, { "epoch": 1.5421195652173914, "grad_norm": 0.15961352750732166, "learning_rate": 6.987259347448143e-05, "loss": 0.9584, "step": 3405 }, { "epoch": 1.542572463768116, "grad_norm": 0.21702101480822816, "learning_rate": 6.986418111291252e-05, "loss": 0.9381, "step": 3406 }, { "epoch": 1.5430253623188406, "grad_norm": 0.12480566035612577, "learning_rate": 6.985576576577406e-05, "loss": 0.952, "step": 3407 }, { "epoch": 1.5434782608695652, "grad_norm": 0.14564592009834473, "learning_rate": 6.984734743390743e-05, "loss": 0.9594, "step": 3408 }, { "epoch": 1.5439311594202898, "grad_norm": 0.15393951635204245, "learning_rate": 6.983892611815416e-05, "loss": 0.9565, "step": 3409 }, { "epoch": 1.5443840579710146, "grad_norm": 0.1623762120135755, "learning_rate": 6.983050181935616e-05, "loss": 0.946, "step": 3410 }, { "epoch": 1.5448369565217392, "grad_norm": 0.21420480191324826, "learning_rate": 6.982207453835563e-05, "loss": 0.9495, "step": 3411 }, { "epoch": 1.5452898550724639, "grad_norm": 0.21410859372581922, "learning_rate": 6.981364427599505e-05, "loss": 0.9461, "step": 3412 }, { "epoch": 1.5457427536231885, "grad_norm": 0.19090210310760447, "learning_rate": 6.98052110331172e-05, "loss": 0.9593, "step": 3413 }, { "epoch": 1.546195652173913, "grad_norm": 0.14077687301185224, "learning_rate": 6.979677481056519e-05, "loss": 0.9582, "step": 3414 }, { "epoch": 1.5466485507246377, "grad_norm": 0.09669934395333589, "learning_rate": 6.978833560918236e-05, "loss": 0.9444, "step": 3415 }, { "epoch": 1.5471014492753623, "grad_norm": 0.12278794305376296, "learning_rate": 6.977989342981244e-05, "loss": 0.9618, "step": 3416 }, { "epoch": 1.547554347826087, "grad_norm": 0.17812203345007543, "learning_rate": 6.977144827329937e-05, "loss": 0.9536, "step": 3417 }, { "epoch": 1.5480072463768115, "grad_norm": 0.15388681032914917, "learning_rate": 6.976300014048745e-05, "loss": 0.9507, "step": 3418 }, { "epoch": 1.5484601449275361, "grad_norm": 0.12703952427968032, "learning_rate": 6.975454903222124e-05, "loss": 0.9713, "step": 3419 }, { "epoch": 1.5489130434782608, "grad_norm": 0.13568128292367806, "learning_rate": 6.97460949493456e-05, "loss": 0.9537, "step": 3420 }, { "epoch": 1.5493659420289854, "grad_norm": 0.13165455064252882, "learning_rate": 6.973763789270573e-05, "loss": 0.951, "step": 3421 }, { "epoch": 1.5498188405797102, "grad_norm": 0.12812217390716885, "learning_rate": 6.972917786314706e-05, "loss": 0.9492, "step": 3422 }, { "epoch": 1.5502717391304348, "grad_norm": 0.1372293015620681, "learning_rate": 6.972071486151536e-05, "loss": 0.9283, "step": 3423 }, { "epoch": 1.5507246376811594, "grad_norm": 0.192242645282448, "learning_rate": 6.971224888865673e-05, "loss": 0.9476, "step": 3424 }, { "epoch": 1.551177536231884, "grad_norm": 0.11368626826553246, "learning_rate": 6.970377994541749e-05, "loss": 0.9588, "step": 3425 }, { "epoch": 1.5516304347826086, "grad_norm": 0.13942446336966718, "learning_rate": 6.969530803264428e-05, "loss": 0.9528, "step": 3426 }, { "epoch": 1.5520833333333335, "grad_norm": 0.11036629593006617, "learning_rate": 6.968683315118407e-05, "loss": 0.9855, "step": 3427 }, { "epoch": 1.552536231884058, "grad_norm": 0.11442818238756176, "learning_rate": 6.967835530188413e-05, "loss": 0.9451, "step": 3428 }, { "epoch": 1.5529891304347827, "grad_norm": 0.09019296710486165, "learning_rate": 6.966987448559196e-05, "loss": 0.9663, "step": 3429 }, { "epoch": 1.5534420289855073, "grad_norm": 0.09514127862704372, "learning_rate": 6.966139070315543e-05, "loss": 0.9396, "step": 3430 }, { "epoch": 1.553894927536232, "grad_norm": 0.11983598817725989, "learning_rate": 6.965290395542267e-05, "loss": 0.9836, "step": 3431 }, { "epoch": 1.5543478260869565, "grad_norm": 0.15683304015020244, "learning_rate": 6.964441424324211e-05, "loss": 0.949, "step": 3432 }, { "epoch": 1.5548007246376812, "grad_norm": 0.14161487997461822, "learning_rate": 6.963592156746249e-05, "loss": 0.9197, "step": 3433 }, { "epoch": 1.5552536231884058, "grad_norm": 0.10557194951484497, "learning_rate": 6.96274259289328e-05, "loss": 0.9488, "step": 3434 }, { "epoch": 1.5557065217391304, "grad_norm": 0.100888808949302, "learning_rate": 6.961892732850242e-05, "loss": 0.9604, "step": 3435 }, { "epoch": 1.556159420289855, "grad_norm": 0.1510086190773479, "learning_rate": 6.961042576702094e-05, "loss": 0.9644, "step": 3436 }, { "epoch": 1.5566123188405796, "grad_norm": 0.17025400007392832, "learning_rate": 6.960192124533826e-05, "loss": 0.9556, "step": 3437 }, { "epoch": 1.5570652173913042, "grad_norm": 0.1636916854939981, "learning_rate": 6.959341376430461e-05, "loss": 0.9357, "step": 3438 }, { "epoch": 1.5575181159420288, "grad_norm": 0.15258764054208787, "learning_rate": 6.95849033247705e-05, "loss": 0.9561, "step": 3439 }, { "epoch": 1.5579710144927537, "grad_norm": 0.12975276639528377, "learning_rate": 6.95763899275867e-05, "loss": 0.9483, "step": 3440 }, { "epoch": 1.5584239130434783, "grad_norm": 0.10223123228511471, "learning_rate": 6.956787357360436e-05, "loss": 0.9502, "step": 3441 }, { "epoch": 1.5588768115942029, "grad_norm": 0.10348734531585474, "learning_rate": 6.955935426367482e-05, "loss": 0.9606, "step": 3442 }, { "epoch": 1.5593297101449275, "grad_norm": 0.13869496191083017, "learning_rate": 6.955083199864981e-05, "loss": 0.9434, "step": 3443 }, { "epoch": 1.5597826086956523, "grad_norm": 0.16114939477401297, "learning_rate": 6.95423067793813e-05, "loss": 0.9425, "step": 3444 }, { "epoch": 1.560235507246377, "grad_norm": 0.1403624362223639, "learning_rate": 6.953377860672156e-05, "loss": 0.9552, "step": 3445 }, { "epoch": 1.5606884057971016, "grad_norm": 0.12381203419882274, "learning_rate": 6.952524748152316e-05, "loss": 0.9473, "step": 3446 }, { "epoch": 1.5611413043478262, "grad_norm": 0.10206781871081863, "learning_rate": 6.951671340463901e-05, "loss": 0.9264, "step": 3447 }, { "epoch": 1.5615942028985508, "grad_norm": 0.142574014842494, "learning_rate": 6.950817637692222e-05, "loss": 0.9249, "step": 3448 }, { "epoch": 1.5620471014492754, "grad_norm": 0.15998364988502592, "learning_rate": 6.949963639922629e-05, "loss": 0.9302, "step": 3449 }, { "epoch": 1.5625, "grad_norm": 0.16905917786020963, "learning_rate": 6.949109347240496e-05, "loss": 0.9394, "step": 3450 }, { "epoch": 1.5629528985507246, "grad_norm": 0.19683064365476094, "learning_rate": 6.94825475973123e-05, "loss": 0.937, "step": 3451 }, { "epoch": 1.5634057971014492, "grad_norm": 0.21577475849210503, "learning_rate": 6.947399877480263e-05, "loss": 0.9502, "step": 3452 }, { "epoch": 1.5638586956521738, "grad_norm": 0.20168933541247488, "learning_rate": 6.946544700573058e-05, "loss": 0.9369, "step": 3453 }, { "epoch": 1.5643115942028984, "grad_norm": 0.15268082756423157, "learning_rate": 6.945689229095111e-05, "loss": 0.9341, "step": 3454 }, { "epoch": 1.564764492753623, "grad_norm": 0.13981661226584036, "learning_rate": 6.944833463131944e-05, "loss": 0.9652, "step": 3455 }, { "epoch": 1.5652173913043477, "grad_norm": 0.12975135783390163, "learning_rate": 6.94397740276911e-05, "loss": 0.9806, "step": 3456 }, { "epoch": 1.5656702898550725, "grad_norm": 0.118407194020162, "learning_rate": 6.943121048092188e-05, "loss": 0.937, "step": 3457 }, { "epoch": 1.5661231884057971, "grad_norm": 0.0849417587643178, "learning_rate": 6.942264399186793e-05, "loss": 0.9417, "step": 3458 }, { "epoch": 1.5665760869565217, "grad_norm": 0.10878881715963147, "learning_rate": 6.941407456138563e-05, "loss": 0.963, "step": 3459 }, { "epoch": 1.5670289855072463, "grad_norm": 0.13480960320197038, "learning_rate": 6.940550219033168e-05, "loss": 0.9659, "step": 3460 }, { "epoch": 1.5674818840579712, "grad_norm": 0.15668808709650425, "learning_rate": 6.939692687956309e-05, "loss": 0.9714, "step": 3461 }, { "epoch": 1.5679347826086958, "grad_norm": 0.1853359203005774, "learning_rate": 6.938834862993713e-05, "loss": 0.9489, "step": 3462 }, { "epoch": 1.5683876811594204, "grad_norm": 0.2105902724565574, "learning_rate": 6.937976744231139e-05, "loss": 0.9406, "step": 3463 }, { "epoch": 1.568840579710145, "grad_norm": 0.2296170711719701, "learning_rate": 6.937118331754373e-05, "loss": 0.9744, "step": 3464 }, { "epoch": 1.5692934782608696, "grad_norm": 0.1962992917259857, "learning_rate": 6.936259625649235e-05, "loss": 0.9426, "step": 3465 }, { "epoch": 1.5697463768115942, "grad_norm": 0.1706018659136511, "learning_rate": 6.935400626001567e-05, "loss": 0.9553, "step": 3466 }, { "epoch": 1.5701992753623188, "grad_norm": 0.15645990121171627, "learning_rate": 6.93454133289725e-05, "loss": 0.9634, "step": 3467 }, { "epoch": 1.5706521739130435, "grad_norm": 0.1486550583266763, "learning_rate": 6.933681746422183e-05, "loss": 0.9489, "step": 3468 }, { "epoch": 1.571105072463768, "grad_norm": 0.12538974211890416, "learning_rate": 6.932821866662304e-05, "loss": 0.9314, "step": 3469 }, { "epoch": 1.5715579710144927, "grad_norm": 0.1146243142746021, "learning_rate": 6.931961693703576e-05, "loss": 0.9318, "step": 3470 }, { "epoch": 1.5720108695652173, "grad_norm": 0.09414193226695844, "learning_rate": 6.931101227631991e-05, "loss": 0.9388, "step": 3471 }, { "epoch": 1.572463768115942, "grad_norm": 0.10797690894160852, "learning_rate": 6.93024046853357e-05, "loss": 0.9451, "step": 3472 }, { "epoch": 1.5729166666666665, "grad_norm": 0.10123611644852047, "learning_rate": 6.929379416494369e-05, "loss": 0.9469, "step": 3473 }, { "epoch": 1.5733695652173914, "grad_norm": 0.13211618639258524, "learning_rate": 6.928518071600463e-05, "loss": 0.9405, "step": 3474 }, { "epoch": 1.573822463768116, "grad_norm": 0.14588775399367013, "learning_rate": 6.927656433937968e-05, "loss": 0.9396, "step": 3475 }, { "epoch": 1.5742753623188406, "grad_norm": 0.12158429525246967, "learning_rate": 6.926794503593017e-05, "loss": 0.9252, "step": 3476 }, { "epoch": 1.5747282608695652, "grad_norm": 0.10834070096724191, "learning_rate": 6.925932280651783e-05, "loss": 0.9479, "step": 3477 }, { "epoch": 1.5751811594202898, "grad_norm": 0.12376596069773024, "learning_rate": 6.925069765200461e-05, "loss": 0.9442, "step": 3478 }, { "epoch": 1.5756340579710146, "grad_norm": 0.1506959323940004, "learning_rate": 6.924206957325281e-05, "loss": 0.9567, "step": 3479 }, { "epoch": 1.5760869565217392, "grad_norm": 0.13841000803006723, "learning_rate": 6.923343857112497e-05, "loss": 0.96, "step": 3480 }, { "epoch": 1.5765398550724639, "grad_norm": 0.09789526031760007, "learning_rate": 6.922480464648395e-05, "loss": 0.939, "step": 3481 }, { "epoch": 1.5769927536231885, "grad_norm": 0.10237215709139455, "learning_rate": 6.921616780019291e-05, "loss": 0.9687, "step": 3482 }, { "epoch": 1.577445652173913, "grad_norm": 0.10383853054376493, "learning_rate": 6.920752803311527e-05, "loss": 0.9638, "step": 3483 }, { "epoch": 1.5778985507246377, "grad_norm": 0.08806319699299174, "learning_rate": 6.919888534611476e-05, "loss": 0.9445, "step": 3484 }, { "epoch": 1.5783514492753623, "grad_norm": 0.12609632699636353, "learning_rate": 6.919023974005542e-05, "loss": 0.9583, "step": 3485 }, { "epoch": 1.578804347826087, "grad_norm": 0.12709191591141364, "learning_rate": 6.918159121580155e-05, "loss": 0.9388, "step": 3486 }, { "epoch": 1.5792572463768115, "grad_norm": 0.1522946152921004, "learning_rate": 6.917293977421778e-05, "loss": 0.9669, "step": 3487 }, { "epoch": 1.5797101449275361, "grad_norm": 0.15667101073047157, "learning_rate": 6.916428541616899e-05, "loss": 0.9676, "step": 3488 }, { "epoch": 1.5801630434782608, "grad_norm": 0.12846455572023735, "learning_rate": 6.915562814252036e-05, "loss": 0.9392, "step": 3489 }, { "epoch": 1.5806159420289854, "grad_norm": 0.10514586001468765, "learning_rate": 6.91469679541374e-05, "loss": 0.9461, "step": 3490 }, { "epoch": 1.5810688405797102, "grad_norm": 0.09870000692106154, "learning_rate": 6.913830485188584e-05, "loss": 0.9665, "step": 3491 }, { "epoch": 1.5815217391304348, "grad_norm": 0.12303079531901627, "learning_rate": 6.91296388366318e-05, "loss": 0.9534, "step": 3492 }, { "epoch": 1.5819746376811594, "grad_norm": 0.1375628538877406, "learning_rate": 6.912096990924159e-05, "loss": 0.9384, "step": 3493 }, { "epoch": 1.582427536231884, "grad_norm": 0.12613188238166975, "learning_rate": 6.911229807058187e-05, "loss": 0.9526, "step": 3494 }, { "epoch": 1.5828804347826086, "grad_norm": 0.13831864330660115, "learning_rate": 6.91036233215196e-05, "loss": 0.9737, "step": 3495 }, { "epoch": 1.5833333333333335, "grad_norm": 0.14854445709276065, "learning_rate": 6.909494566292195e-05, "loss": 0.9449, "step": 3496 }, { "epoch": 1.583786231884058, "grad_norm": 0.14307549832869657, "learning_rate": 6.908626509565651e-05, "loss": 0.9452, "step": 3497 }, { "epoch": 1.5842391304347827, "grad_norm": 0.12254731441847692, "learning_rate": 6.907758162059103e-05, "loss": 0.9417, "step": 3498 }, { "epoch": 1.5846920289855073, "grad_norm": 0.09280649126578495, "learning_rate": 6.906889523859366e-05, "loss": 0.9295, "step": 3499 }, { "epoch": 1.585144927536232, "grad_norm": 0.12017200118087962, "learning_rate": 6.906020595053276e-05, "loss": 0.9371, "step": 3500 }, { "epoch": 1.5855978260869565, "grad_norm": 0.15026305341919538, "learning_rate": 6.905151375727704e-05, "loss": 0.9385, "step": 3501 }, { "epoch": 1.5860507246376812, "grad_norm": 0.1533129863387056, "learning_rate": 6.904281865969543e-05, "loss": 0.9484, "step": 3502 }, { "epoch": 1.5865036231884058, "grad_norm": 0.17750876555765255, "learning_rate": 6.903412065865722e-05, "loss": 0.9512, "step": 3503 }, { "epoch": 1.5869565217391304, "grad_norm": 0.18288805728170468, "learning_rate": 6.902541975503196e-05, "loss": 0.9531, "step": 3504 }, { "epoch": 1.587409420289855, "grad_norm": 0.1508496706290462, "learning_rate": 6.901671594968951e-05, "loss": 0.9475, "step": 3505 }, { "epoch": 1.5878623188405796, "grad_norm": 0.12915561302826478, "learning_rate": 6.900800924349997e-05, "loss": 0.9594, "step": 3506 }, { "epoch": 1.5883152173913042, "grad_norm": 0.1021155704362746, "learning_rate": 6.899929963733378e-05, "loss": 0.9519, "step": 3507 }, { "epoch": 1.5887681159420288, "grad_norm": 0.09831969551597151, "learning_rate": 6.899058713206165e-05, "loss": 0.9538, "step": 3508 }, { "epoch": 1.5892210144927537, "grad_norm": 0.08838510686313762, "learning_rate": 6.898187172855458e-05, "loss": 0.9305, "step": 3509 }, { "epoch": 1.5896739130434783, "grad_norm": 0.12924081860292808, "learning_rate": 6.897315342768388e-05, "loss": 0.9546, "step": 3510 }, { "epoch": 1.5901268115942029, "grad_norm": 0.15048548097391468, "learning_rate": 6.896443223032111e-05, "loss": 0.9646, "step": 3511 }, { "epoch": 1.5905797101449275, "grad_norm": 0.1311175302126681, "learning_rate": 6.895570813733816e-05, "loss": 0.9635, "step": 3512 }, { "epoch": 1.5910326086956523, "grad_norm": 0.15208237594877863, "learning_rate": 6.894698114960718e-05, "loss": 0.9597, "step": 3513 }, { "epoch": 1.591485507246377, "grad_norm": 0.1561668720076271, "learning_rate": 6.893825126800062e-05, "loss": 0.9547, "step": 3514 }, { "epoch": 1.5919384057971016, "grad_norm": 0.13106154652035326, "learning_rate": 6.892951849339124e-05, "loss": 0.9349, "step": 3515 }, { "epoch": 1.5923913043478262, "grad_norm": 0.15524065946272003, "learning_rate": 6.892078282665204e-05, "loss": 0.9527, "step": 3516 }, { "epoch": 1.5928442028985508, "grad_norm": 0.1710476285155282, "learning_rate": 6.891204426865635e-05, "loss": 0.9657, "step": 3517 }, { "epoch": 1.5932971014492754, "grad_norm": 0.12943613572192897, "learning_rate": 6.890330282027778e-05, "loss": 0.952, "step": 3518 }, { "epoch": 1.59375, "grad_norm": 0.09437653290957504, "learning_rate": 6.889455848239022e-05, "loss": 0.9555, "step": 3519 }, { "epoch": 1.5942028985507246, "grad_norm": 0.12438175985672045, "learning_rate": 6.888581125586788e-05, "loss": 0.9304, "step": 3520 }, { "epoch": 1.5946557971014492, "grad_norm": 0.15449772430953204, "learning_rate": 6.88770611415852e-05, "loss": 0.9398, "step": 3521 }, { "epoch": 1.5951086956521738, "grad_norm": 0.13281615411621123, "learning_rate": 6.886830814041697e-05, "loss": 0.9532, "step": 3522 }, { "epoch": 1.5955615942028984, "grad_norm": 0.1285842841585347, "learning_rate": 6.885955225323822e-05, "loss": 0.959, "step": 3523 }, { "epoch": 1.596014492753623, "grad_norm": 0.14080052563633338, "learning_rate": 6.88507934809243e-05, "loss": 0.9402, "step": 3524 }, { "epoch": 1.5964673913043477, "grad_norm": 0.1269306101111155, "learning_rate": 6.884203182435084e-05, "loss": 0.9425, "step": 3525 }, { "epoch": 1.5969202898550725, "grad_norm": 0.140078993190414, "learning_rate": 6.883326728439375e-05, "loss": 0.9641, "step": 3526 }, { "epoch": 1.5973731884057971, "grad_norm": 0.16003747625278344, "learning_rate": 6.882449986192925e-05, "loss": 0.928, "step": 3527 }, { "epoch": 1.5978260869565217, "grad_norm": 0.15861279623824, "learning_rate": 6.881572955783382e-05, "loss": 0.9682, "step": 3528 }, { "epoch": 1.5982789855072463, "grad_norm": 0.18300632195771543, "learning_rate": 6.880695637298424e-05, "loss": 0.9075, "step": 3529 }, { "epoch": 1.5987318840579712, "grad_norm": 0.20287643125322166, "learning_rate": 6.879818030825758e-05, "loss": 0.9584, "step": 3530 }, { "epoch": 1.5991847826086958, "grad_norm": 0.21307532204251767, "learning_rate": 6.878940136453119e-05, "loss": 0.9512, "step": 3531 }, { "epoch": 1.5996376811594204, "grad_norm": 0.1755277165965875, "learning_rate": 6.878061954268274e-05, "loss": 0.9412, "step": 3532 }, { "epoch": 1.600090579710145, "grad_norm": 0.1297631260418262, "learning_rate": 6.877183484359016e-05, "loss": 0.9713, "step": 3533 }, { "epoch": 1.6005434782608696, "grad_norm": 0.19008958714704374, "learning_rate": 6.876304726813163e-05, "loss": 0.9511, "step": 3534 }, { "epoch": 1.6009963768115942, "grad_norm": 0.2563993072342934, "learning_rate": 6.87542568171857e-05, "loss": 0.9556, "step": 3535 }, { "epoch": 1.6014492753623188, "grad_norm": 0.2727561961347448, "learning_rate": 6.874546349163115e-05, "loss": 0.9536, "step": 3536 }, { "epoch": 1.6019021739130435, "grad_norm": 0.2332426675017764, "learning_rate": 6.873666729234707e-05, "loss": 0.9414, "step": 3537 }, { "epoch": 1.602355072463768, "grad_norm": 0.1629519399687768, "learning_rate": 6.872786822021281e-05, "loss": 0.9565, "step": 3538 }, { "epoch": 1.6028079710144927, "grad_norm": 0.13168298971343373, "learning_rate": 6.871906627610805e-05, "loss": 0.9436, "step": 3539 }, { "epoch": 1.6032608695652173, "grad_norm": 0.13315602325916592, "learning_rate": 6.871026146091272e-05, "loss": 0.9419, "step": 3540 }, { "epoch": 1.603713768115942, "grad_norm": 0.13702799612253252, "learning_rate": 6.870145377550706e-05, "loss": 0.9442, "step": 3541 }, { "epoch": 1.6041666666666665, "grad_norm": 0.16022758445898802, "learning_rate": 6.869264322077158e-05, "loss": 0.965, "step": 3542 }, { "epoch": 1.6046195652173914, "grad_norm": 0.1655308650893984, "learning_rate": 6.86838297975871e-05, "loss": 0.9477, "step": 3543 }, { "epoch": 1.605072463768116, "grad_norm": 0.1698238466998924, "learning_rate": 6.867501350683469e-05, "loss": 0.9502, "step": 3544 }, { "epoch": 1.6055253623188406, "grad_norm": 0.15986366753891254, "learning_rate": 6.866619434939574e-05, "loss": 0.9482, "step": 3545 }, { "epoch": 1.6059782608695652, "grad_norm": 0.11180473971641203, "learning_rate": 6.865737232615193e-05, "loss": 0.939, "step": 3546 }, { "epoch": 1.6064311594202898, "grad_norm": 0.0913022291903259, "learning_rate": 6.864854743798519e-05, "loss": 0.9522, "step": 3547 }, { "epoch": 1.6068840579710146, "grad_norm": 0.12790458803241078, "learning_rate": 6.863971968577776e-05, "loss": 0.9287, "step": 3548 }, { "epoch": 1.6073369565217392, "grad_norm": 0.14674613022279215, "learning_rate": 6.863088907041218e-05, "loss": 0.9449, "step": 3549 }, { "epoch": 1.6077898550724639, "grad_norm": 0.12821692169550122, "learning_rate": 6.862205559277124e-05, "loss": 0.9557, "step": 3550 }, { "epoch": 1.6082427536231885, "grad_norm": 0.13152454979671516, "learning_rate": 6.861321925373804e-05, "loss": 0.9524, "step": 3551 }, { "epoch": 1.608695652173913, "grad_norm": 0.15421849235481613, "learning_rate": 6.860438005419598e-05, "loss": 0.961, "step": 3552 }, { "epoch": 1.6091485507246377, "grad_norm": 0.1388192917123901, "learning_rate": 6.859553799502871e-05, "loss": 0.9544, "step": 3553 }, { "epoch": 1.6096014492753623, "grad_norm": 0.13129239770885087, "learning_rate": 6.85866930771202e-05, "loss": 0.9462, "step": 3554 }, { "epoch": 1.610054347826087, "grad_norm": 0.14144828749986366, "learning_rate": 6.857784530135469e-05, "loss": 0.9361, "step": 3555 }, { "epoch": 1.6105072463768115, "grad_norm": 0.12942696098688652, "learning_rate": 6.85689946686167e-05, "loss": 0.934, "step": 3556 }, { "epoch": 1.6109601449275361, "grad_norm": 0.12351936130846239, "learning_rate": 6.856014117979104e-05, "loss": 0.9588, "step": 3557 }, { "epoch": 1.6114130434782608, "grad_norm": 0.10481491441989975, "learning_rate": 6.85512848357628e-05, "loss": 0.9676, "step": 3558 }, { "epoch": 1.6118659420289854, "grad_norm": 0.09667489854861457, "learning_rate": 6.854242563741737e-05, "loss": 0.9677, "step": 3559 }, { "epoch": 1.6123188405797102, "grad_norm": 0.12179747632358483, "learning_rate": 6.853356358564043e-05, "loss": 0.9722, "step": 3560 }, { "epoch": 1.6127717391304348, "grad_norm": 0.1389506346996357, "learning_rate": 6.852469868131792e-05, "loss": 0.9484, "step": 3561 }, { "epoch": 1.6132246376811594, "grad_norm": 0.14323988588822908, "learning_rate": 6.851583092533608e-05, "loss": 0.9469, "step": 3562 }, { "epoch": 1.613677536231884, "grad_norm": 0.15740895368666608, "learning_rate": 6.850696031858143e-05, "loss": 0.9502, "step": 3563 }, { "epoch": 1.6141304347826086, "grad_norm": 0.1527473163988691, "learning_rate": 6.84980868619408e-05, "loss": 0.9557, "step": 3564 }, { "epoch": 1.6145833333333335, "grad_norm": 0.1123441148900947, "learning_rate": 6.848921055630125e-05, "loss": 0.9594, "step": 3565 }, { "epoch": 1.615036231884058, "grad_norm": 0.11199589824183662, "learning_rate": 6.848033140255019e-05, "loss": 0.9542, "step": 3566 }, { "epoch": 1.6154891304347827, "grad_norm": 0.14237694106971902, "learning_rate": 6.847144940157525e-05, "loss": 0.9424, "step": 3567 }, { "epoch": 1.6159420289855073, "grad_norm": 0.15509893978363343, "learning_rate": 6.84625645542644e-05, "loss": 0.938, "step": 3568 }, { "epoch": 1.616394927536232, "grad_norm": 0.15426751315985043, "learning_rate": 6.84536768615059e-05, "loss": 0.931, "step": 3569 }, { "epoch": 1.6168478260869565, "grad_norm": 0.1262043998476006, "learning_rate": 6.84447863241882e-05, "loss": 0.9512, "step": 3570 }, { "epoch": 1.6173007246376812, "grad_norm": 0.10512149826508273, "learning_rate": 6.843589294320014e-05, "loss": 0.9314, "step": 3571 }, { "epoch": 1.6177536231884058, "grad_norm": 0.13233257243378757, "learning_rate": 6.842699671943081e-05, "loss": 0.9595, "step": 3572 }, { "epoch": 1.6182065217391304, "grad_norm": 0.1385341942717211, "learning_rate": 6.841809765376957e-05, "loss": 0.9219, "step": 3573 }, { "epoch": 1.618659420289855, "grad_norm": 0.1267862629187995, "learning_rate": 6.840919574710608e-05, "loss": 0.9441, "step": 3574 }, { "epoch": 1.6191123188405796, "grad_norm": 0.09682852771748013, "learning_rate": 6.840029100033027e-05, "loss": 0.968, "step": 3575 }, { "epoch": 1.6195652173913042, "grad_norm": 0.10626617731962043, "learning_rate": 6.839138341433236e-05, "loss": 0.9394, "step": 3576 }, { "epoch": 1.6200181159420288, "grad_norm": 0.11636037834152799, "learning_rate": 6.838247299000287e-05, "loss": 0.9564, "step": 3577 }, { "epoch": 1.6204710144927537, "grad_norm": 0.10990584065510726, "learning_rate": 6.837355972823257e-05, "loss": 0.947, "step": 3578 }, { "epoch": 1.6209239130434783, "grad_norm": 0.11963360992849065, "learning_rate": 6.836464362991255e-05, "loss": 0.9581, "step": 3579 }, { "epoch": 1.6213768115942029, "grad_norm": 0.12916258773161843, "learning_rate": 6.835572469593415e-05, "loss": 0.9424, "step": 3580 }, { "epoch": 1.6218297101449275, "grad_norm": 0.11680977169724428, "learning_rate": 6.834680292718902e-05, "loss": 0.9446, "step": 3581 }, { "epoch": 1.6222826086956523, "grad_norm": 0.1277018872951057, "learning_rate": 6.833787832456909e-05, "loss": 0.9599, "step": 3582 }, { "epoch": 1.622735507246377, "grad_norm": 0.14533921102506542, "learning_rate": 6.832895088896655e-05, "loss": 0.9353, "step": 3583 }, { "epoch": 1.6231884057971016, "grad_norm": 0.1676212342246846, "learning_rate": 6.83200206212739e-05, "loss": 0.9532, "step": 3584 }, { "epoch": 1.6236413043478262, "grad_norm": 0.17640274055793798, "learning_rate": 6.831108752238392e-05, "loss": 0.9533, "step": 3585 }, { "epoch": 1.6240942028985508, "grad_norm": 0.18641712883382794, "learning_rate": 6.830215159318965e-05, "loss": 0.9359, "step": 3586 }, { "epoch": 1.6245471014492754, "grad_norm": 0.1671183496905194, "learning_rate": 6.829321283458445e-05, "loss": 0.9484, "step": 3587 }, { "epoch": 1.625, "grad_norm": 0.11705286152419761, "learning_rate": 6.828427124746191e-05, "loss": 0.9374, "step": 3588 }, { "epoch": 1.6254528985507246, "grad_norm": 0.11638131462269428, "learning_rate": 6.827532683271595e-05, "loss": 0.9347, "step": 3589 }, { "epoch": 1.6259057971014492, "grad_norm": 0.11876827282597939, "learning_rate": 6.826637959124078e-05, "loss": 0.9369, "step": 3590 }, { "epoch": 1.6263586956521738, "grad_norm": 0.1654742555074565, "learning_rate": 6.825742952393083e-05, "loss": 0.9603, "step": 3591 }, { "epoch": 1.6268115942028984, "grad_norm": 0.19099227096185564, "learning_rate": 6.824847663168089e-05, "loss": 0.9576, "step": 3592 }, { "epoch": 1.627264492753623, "grad_norm": 0.2219862977798668, "learning_rate": 6.823952091538598e-05, "loss": 0.9238, "step": 3593 }, { "epoch": 1.6277173913043477, "grad_norm": 0.2153454772856533, "learning_rate": 6.82305623759414e-05, "loss": 0.9412, "step": 3594 }, { "epoch": 1.6281702898550725, "grad_norm": 0.18971061054217822, "learning_rate": 6.822160101424276e-05, "loss": 0.9535, "step": 3595 }, { "epoch": 1.6286231884057971, "grad_norm": 0.15101737429041523, "learning_rate": 6.821263683118596e-05, "loss": 0.9282, "step": 3596 }, { "epoch": 1.6290760869565217, "grad_norm": 0.13170003855769236, "learning_rate": 6.820366982766715e-05, "loss": 0.9452, "step": 3597 }, { "epoch": 1.6295289855072463, "grad_norm": 0.13028328893856986, "learning_rate": 6.819470000458276e-05, "loss": 0.9325, "step": 3598 }, { "epoch": 1.6299818840579712, "grad_norm": 0.11702111270666944, "learning_rate": 6.818572736282955e-05, "loss": 0.9487, "step": 3599 }, { "epoch": 1.6304347826086958, "grad_norm": 0.10080667923811537, "learning_rate": 6.81767519033045e-05, "loss": 0.9379, "step": 3600 }, { "epoch": 1.6308876811594204, "grad_norm": 0.10038102406578268, "learning_rate": 6.816777362690491e-05, "loss": 0.963, "step": 3601 }, { "epoch": 1.631340579710145, "grad_norm": 0.13327069135877917, "learning_rate": 6.815879253452837e-05, "loss": 0.9548, "step": 3602 }, { "epoch": 1.6317934782608696, "grad_norm": 0.16563406509818498, "learning_rate": 6.81498086270727e-05, "loss": 0.941, "step": 3603 }, { "epoch": 1.6322463768115942, "grad_norm": 0.18832778070532039, "learning_rate": 6.814082190543607e-05, "loss": 0.956, "step": 3604 }, { "epoch": 1.6326992753623188, "grad_norm": 0.15275606854032686, "learning_rate": 6.813183237051687e-05, "loss": 0.9499, "step": 3605 }, { "epoch": 1.6331521739130435, "grad_norm": 0.12377635469471976, "learning_rate": 6.812284002321381e-05, "loss": 0.9651, "step": 3606 }, { "epoch": 1.633605072463768, "grad_norm": 0.1261071311445632, "learning_rate": 6.811384486442589e-05, "loss": 0.9473, "step": 3607 }, { "epoch": 1.6340579710144927, "grad_norm": 0.11824109461309668, "learning_rate": 6.810484689505232e-05, "loss": 0.943, "step": 3608 }, { "epoch": 1.6345108695652173, "grad_norm": 0.13156412675338514, "learning_rate": 6.809584611599268e-05, "loss": 0.9531, "step": 3609 }, { "epoch": 1.634963768115942, "grad_norm": 0.14606495576495832, "learning_rate": 6.808684252814678e-05, "loss": 0.9522, "step": 3610 }, { "epoch": 1.6354166666666665, "grad_norm": 0.11231448844401232, "learning_rate": 6.807783613241474e-05, "loss": 0.9423, "step": 3611 }, { "epoch": 1.6358695652173914, "grad_norm": 0.13840231644187986, "learning_rate": 6.80688269296969e-05, "loss": 0.9544, "step": 3612 }, { "epoch": 1.636322463768116, "grad_norm": 0.162886176045168, "learning_rate": 6.805981492089398e-05, "loss": 0.9387, "step": 3613 }, { "epoch": 1.6367753623188406, "grad_norm": 0.18626941489531504, "learning_rate": 6.805080010690688e-05, "loss": 0.9484, "step": 3614 }, { "epoch": 1.6372282608695652, "grad_norm": 0.1658303109805352, "learning_rate": 6.804178248863685e-05, "loss": 0.9469, "step": 3615 }, { "epoch": 1.6376811594202898, "grad_norm": 0.1313164242296602, "learning_rate": 6.803276206698539e-05, "loss": 0.9594, "step": 3616 }, { "epoch": 1.6381340579710146, "grad_norm": 0.1375430826071288, "learning_rate": 6.802373884285428e-05, "loss": 0.9773, "step": 3617 }, { "epoch": 1.6385869565217392, "grad_norm": 0.13562483371705875, "learning_rate": 6.80147128171456e-05, "loss": 0.9626, "step": 3618 }, { "epoch": 1.6390398550724639, "grad_norm": 0.1283872409696931, "learning_rate": 6.800568399076168e-05, "loss": 0.9525, "step": 3619 }, { "epoch": 1.6394927536231885, "grad_norm": 0.14404372981149607, "learning_rate": 6.799665236460514e-05, "loss": 0.941, "step": 3620 }, { "epoch": 1.639945652173913, "grad_norm": 0.1476157785108282, "learning_rate": 6.798761793957893e-05, "loss": 0.9297, "step": 3621 }, { "epoch": 1.6403985507246377, "grad_norm": 0.1440254815234608, "learning_rate": 6.797858071658618e-05, "loss": 0.9555, "step": 3622 }, { "epoch": 1.6408514492753623, "grad_norm": 0.12854115869166607, "learning_rate": 6.79695406965304e-05, "loss": 0.947, "step": 3623 }, { "epoch": 1.641304347826087, "grad_norm": 0.13260111851870812, "learning_rate": 6.796049788031528e-05, "loss": 0.9239, "step": 3624 }, { "epoch": 1.6417572463768115, "grad_norm": 0.1713670108481951, "learning_rate": 6.795145226884491e-05, "loss": 0.9489, "step": 3625 }, { "epoch": 1.6422101449275361, "grad_norm": 0.19222425040974062, "learning_rate": 6.794240386302356e-05, "loss": 0.9363, "step": 3626 }, { "epoch": 1.6426630434782608, "grad_norm": 0.2028013646620696, "learning_rate": 6.793335266375581e-05, "loss": 0.9359, "step": 3627 }, { "epoch": 1.6431159420289854, "grad_norm": 0.20244911133700844, "learning_rate": 6.792429867194653e-05, "loss": 0.9489, "step": 3628 }, { "epoch": 1.6435688405797102, "grad_norm": 0.17291494889077505, "learning_rate": 6.791524188850086e-05, "loss": 0.9395, "step": 3629 }, { "epoch": 1.6440217391304348, "grad_norm": 0.11859040339122479, "learning_rate": 6.790618231432422e-05, "loss": 0.9686, "step": 3630 }, { "epoch": 1.6444746376811594, "grad_norm": 0.15473510686344671, "learning_rate": 6.789711995032233e-05, "loss": 0.9418, "step": 3631 }, { "epoch": 1.644927536231884, "grad_norm": 0.18888019127147856, "learning_rate": 6.788805479740114e-05, "loss": 0.959, "step": 3632 }, { "epoch": 1.6453804347826086, "grad_norm": 0.16036797795520652, "learning_rate": 6.787898685646692e-05, "loss": 0.9672, "step": 3633 }, { "epoch": 1.6458333333333335, "grad_norm": 0.13154946319857216, "learning_rate": 6.786991612842621e-05, "loss": 0.9373, "step": 3634 }, { "epoch": 1.646286231884058, "grad_norm": 0.13701557471227016, "learning_rate": 6.786084261418581e-05, "loss": 0.9498, "step": 3635 }, { "epoch": 1.6467391304347827, "grad_norm": 0.14359273063165487, "learning_rate": 6.785176631465284e-05, "loss": 0.9543, "step": 3636 }, { "epoch": 1.6471920289855073, "grad_norm": 0.12131846854869453, "learning_rate": 6.784268723073463e-05, "loss": 0.9437, "step": 3637 }, { "epoch": 1.647644927536232, "grad_norm": 0.16423126812169828, "learning_rate": 6.78336053633389e-05, "loss": 0.9143, "step": 3638 }, { "epoch": 1.6480978260869565, "grad_norm": 0.15881022607362805, "learning_rate": 6.78245207133735e-05, "loss": 0.9482, "step": 3639 }, { "epoch": 1.6485507246376812, "grad_norm": 0.13454519313233249, "learning_rate": 6.781543328174669e-05, "loss": 0.9518, "step": 3640 }, { "epoch": 1.6490036231884058, "grad_norm": 0.14660029327553695, "learning_rate": 6.780634306936692e-05, "loss": 0.9435, "step": 3641 }, { "epoch": 1.6494565217391304, "grad_norm": 0.15527977248607558, "learning_rate": 6.7797250077143e-05, "loss": 0.9393, "step": 3642 }, { "epoch": 1.649909420289855, "grad_norm": 0.15716712087944928, "learning_rate": 6.77881543059839e-05, "loss": 0.965, "step": 3643 }, { "epoch": 1.6503623188405796, "grad_norm": 0.1369968241876398, "learning_rate": 6.777905575679902e-05, "loss": 0.9428, "step": 3644 }, { "epoch": 1.6508152173913042, "grad_norm": 0.12197914193428683, "learning_rate": 6.776995443049789e-05, "loss": 0.9512, "step": 3645 }, { "epoch": 1.6512681159420288, "grad_norm": 0.11637183438619983, "learning_rate": 6.776085032799042e-05, "loss": 0.9656, "step": 3646 }, { "epoch": 1.6517210144927537, "grad_norm": 0.14775153185798015, "learning_rate": 6.775174345018676e-05, "loss": 0.9818, "step": 3647 }, { "epoch": 1.6521739130434783, "grad_norm": 0.17899849161779144, "learning_rate": 6.774263379799732e-05, "loss": 0.9408, "step": 3648 }, { "epoch": 1.6526268115942029, "grad_norm": 0.19037024024400162, "learning_rate": 6.773352137233282e-05, "loss": 0.9396, "step": 3649 }, { "epoch": 1.6530797101449275, "grad_norm": 0.1371510400436091, "learning_rate": 6.772440617410424e-05, "loss": 0.9613, "step": 3650 }, { "epoch": 1.6535326086956523, "grad_norm": 0.1102934523049556, "learning_rate": 6.771528820422285e-05, "loss": 0.9403, "step": 3651 }, { "epoch": 1.653985507246377, "grad_norm": 0.1642972065098577, "learning_rate": 6.770616746360016e-05, "loss": 0.9408, "step": 3652 }, { "epoch": 1.6544384057971016, "grad_norm": 0.16433049801679422, "learning_rate": 6.769704395314803e-05, "loss": 0.9411, "step": 3653 }, { "epoch": 1.6548913043478262, "grad_norm": 0.13339479758652484, "learning_rate": 6.768791767377851e-05, "loss": 0.9467, "step": 3654 }, { "epoch": 1.6553442028985508, "grad_norm": 0.11437690961895679, "learning_rate": 6.767878862640398e-05, "loss": 0.9501, "step": 3655 }, { "epoch": 1.6557971014492754, "grad_norm": 0.13426796934590915, "learning_rate": 6.76696568119371e-05, "loss": 0.9476, "step": 3656 }, { "epoch": 1.65625, "grad_norm": 0.13952392246322465, "learning_rate": 6.766052223129079e-05, "loss": 0.9482, "step": 3657 }, { "epoch": 1.6567028985507246, "grad_norm": 0.11830967060295375, "learning_rate": 6.765138488537823e-05, "loss": 0.9404, "step": 3658 }, { "epoch": 1.6571557971014492, "grad_norm": 0.1218550215096922, "learning_rate": 6.76422447751129e-05, "loss": 0.9601, "step": 3659 }, { "epoch": 1.6576086956521738, "grad_norm": 0.11823657447812747, "learning_rate": 6.763310190140856e-05, "loss": 0.9609, "step": 3660 }, { "epoch": 1.6580615942028984, "grad_norm": 0.12131679064894531, "learning_rate": 6.762395626517924e-05, "loss": 0.9343, "step": 3661 }, { "epoch": 1.658514492753623, "grad_norm": 0.1090611805316783, "learning_rate": 6.761480786733924e-05, "loss": 0.9406, "step": 3662 }, { "epoch": 1.6589673913043477, "grad_norm": 0.09603201406338092, "learning_rate": 6.760565670880313e-05, "loss": 0.9304, "step": 3663 }, { "epoch": 1.6594202898550725, "grad_norm": 0.14214994575276926, "learning_rate": 6.759650279048577e-05, "loss": 0.9314, "step": 3664 }, { "epoch": 1.6598731884057971, "grad_norm": 0.1645994035392329, "learning_rate": 6.758734611330231e-05, "loss": 0.9668, "step": 3665 }, { "epoch": 1.6603260869565217, "grad_norm": 0.18002281086720295, "learning_rate": 6.757818667816815e-05, "loss": 0.9517, "step": 3666 }, { "epoch": 1.6607789855072463, "grad_norm": 0.15509634423955138, "learning_rate": 6.756902448599894e-05, "loss": 0.9223, "step": 3667 }, { "epoch": 1.6612318840579712, "grad_norm": 0.12313685701607294, "learning_rate": 6.755985953771068e-05, "loss": 0.9418, "step": 3668 }, { "epoch": 1.6616847826086958, "grad_norm": 0.09648709331368581, "learning_rate": 6.75506918342196e-05, "loss": 0.9488, "step": 3669 }, { "epoch": 1.6621376811594204, "grad_norm": 0.10990329840904729, "learning_rate": 6.754152137644218e-05, "loss": 0.9255, "step": 3670 }, { "epoch": 1.662590579710145, "grad_norm": 0.11567924128328448, "learning_rate": 6.753234816529522e-05, "loss": 0.9222, "step": 3671 }, { "epoch": 1.6630434782608696, "grad_norm": 0.10827540137610141, "learning_rate": 6.752317220169581e-05, "loss": 0.9507, "step": 3672 }, { "epoch": 1.6634963768115942, "grad_norm": 0.12653151626038245, "learning_rate": 6.751399348656124e-05, "loss": 0.9615, "step": 3673 }, { "epoch": 1.6639492753623188, "grad_norm": 0.1298500690918541, "learning_rate": 6.750481202080916e-05, "loss": 0.9477, "step": 3674 }, { "epoch": 1.6644021739130435, "grad_norm": 0.11033680924306373, "learning_rate": 6.749562780535743e-05, "loss": 0.9587, "step": 3675 }, { "epoch": 1.664855072463768, "grad_norm": 0.1065492069113908, "learning_rate": 6.748644084112421e-05, "loss": 0.9515, "step": 3676 }, { "epoch": 1.6653079710144927, "grad_norm": 0.10074336785039072, "learning_rate": 6.747725112902794e-05, "loss": 0.9653, "step": 3677 }, { "epoch": 1.6657608695652173, "grad_norm": 0.11171907008065338, "learning_rate": 6.746805866998735e-05, "loss": 0.9463, "step": 3678 }, { "epoch": 1.666213768115942, "grad_norm": 0.08909421467732682, "learning_rate": 6.74588634649214e-05, "loss": 0.9575, "step": 3679 }, { "epoch": 1.6666666666666665, "grad_norm": 0.12510190553077707, "learning_rate": 6.744966551474936e-05, "loss": 0.9492, "step": 3680 }, { "epoch": 1.6671195652173914, "grad_norm": 0.11718961880704074, "learning_rate": 6.744046482039075e-05, "loss": 0.9493, "step": 3681 }, { "epoch": 1.667572463768116, "grad_norm": 0.11602477002236707, "learning_rate": 6.743126138276539e-05, "loss": 0.9537, "step": 3682 }, { "epoch": 1.6680253623188406, "grad_norm": 0.0987079804498688, "learning_rate": 6.742205520279337e-05, "loss": 0.9386, "step": 3683 }, { "epoch": 1.6684782608695652, "grad_norm": 0.09580488512725951, "learning_rate": 6.741284628139504e-05, "loss": 0.9318, "step": 3684 }, { "epoch": 1.6689311594202898, "grad_norm": 0.12661500703685244, "learning_rate": 6.740363461949102e-05, "loss": 0.9385, "step": 3685 }, { "epoch": 1.6693840579710146, "grad_norm": 0.1282744301577829, "learning_rate": 6.739442021800223e-05, "loss": 0.9579, "step": 3686 }, { "epoch": 1.6698369565217392, "grad_norm": 0.16317437484641487, "learning_rate": 6.738520307784982e-05, "loss": 0.9353, "step": 3687 }, { "epoch": 1.6702898550724639, "grad_norm": 0.21358641614721172, "learning_rate": 6.737598319995527e-05, "loss": 0.9603, "step": 3688 }, { "epoch": 1.6707427536231885, "grad_norm": 0.19666202085690668, "learning_rate": 6.736676058524032e-05, "loss": 0.9482, "step": 3689 }, { "epoch": 1.671195652173913, "grad_norm": 0.14804332851232876, "learning_rate": 6.735753523462694e-05, "loss": 0.9535, "step": 3690 }, { "epoch": 1.6716485507246377, "grad_norm": 0.1272628731027746, "learning_rate": 6.73483071490374e-05, "loss": 0.945, "step": 3691 }, { "epoch": 1.6721014492753623, "grad_norm": 0.11258679217389952, "learning_rate": 6.733907632939426e-05, "loss": 0.9446, "step": 3692 }, { "epoch": 1.672554347826087, "grad_norm": 0.11838519765933175, "learning_rate": 6.732984277662035e-05, "loss": 0.9421, "step": 3693 }, { "epoch": 1.6730072463768115, "grad_norm": 0.11855369854359825, "learning_rate": 6.732060649163872e-05, "loss": 0.9371, "step": 3694 }, { "epoch": 1.6734601449275361, "grad_norm": 0.10038840385233876, "learning_rate": 6.73113674753728e-05, "loss": 0.9375, "step": 3695 }, { "epoch": 1.6739130434782608, "grad_norm": 0.12983025303419793, "learning_rate": 6.730212572874618e-05, "loss": 0.95, "step": 3696 }, { "epoch": 1.6743659420289854, "grad_norm": 0.1785851099261137, "learning_rate": 6.729288125268278e-05, "loss": 0.9387, "step": 3697 }, { "epoch": 1.6748188405797102, "grad_norm": 0.19514031601056314, "learning_rate": 6.72836340481068e-05, "loss": 0.9709, "step": 3698 }, { "epoch": 1.6752717391304348, "grad_norm": 0.1847494080231315, "learning_rate": 6.727438411594267e-05, "loss": 0.9652, "step": 3699 }, { "epoch": 1.6757246376811594, "grad_norm": 0.16491765730087443, "learning_rate": 6.726513145711517e-05, "loss": 0.955, "step": 3700 }, { "epoch": 1.676177536231884, "grad_norm": 0.17436783448024348, "learning_rate": 6.725587607254925e-05, "loss": 0.9697, "step": 3701 }, { "epoch": 1.6766304347826086, "grad_norm": 0.15172913135642746, "learning_rate": 6.724661796317023e-05, "loss": 0.9541, "step": 3702 }, { "epoch": 1.6770833333333335, "grad_norm": 0.1509107193283058, "learning_rate": 6.723735712990362e-05, "loss": 0.9503, "step": 3703 }, { "epoch": 1.677536231884058, "grad_norm": 0.1492549435109609, "learning_rate": 6.722809357367526e-05, "loss": 0.9471, "step": 3704 }, { "epoch": 1.6779891304347827, "grad_norm": 0.12555599010654853, "learning_rate": 6.721882729541125e-05, "loss": 0.9544, "step": 3705 }, { "epoch": 1.6784420289855073, "grad_norm": 0.13789973760485694, "learning_rate": 6.720955829603794e-05, "loss": 0.937, "step": 3706 }, { "epoch": 1.678894927536232, "grad_norm": 0.14340970069270023, "learning_rate": 6.720028657648195e-05, "loss": 0.9376, "step": 3707 }, { "epoch": 1.6793478260869565, "grad_norm": 0.1234058092856849, "learning_rate": 6.719101213767022e-05, "loss": 0.9512, "step": 3708 }, { "epoch": 1.6798007246376812, "grad_norm": 0.15533368146107907, "learning_rate": 6.718173498052993e-05, "loss": 0.9566, "step": 3709 }, { "epoch": 1.6802536231884058, "grad_norm": 0.19778988175502799, "learning_rate": 6.717245510598851e-05, "loss": 0.9737, "step": 3710 }, { "epoch": 1.6807065217391304, "grad_norm": 0.21807769518653075, "learning_rate": 6.71631725149737e-05, "loss": 0.9459, "step": 3711 }, { "epoch": 1.681159420289855, "grad_norm": 0.2040532631118406, "learning_rate": 6.715388720841349e-05, "loss": 0.9724, "step": 3712 }, { "epoch": 1.6816123188405796, "grad_norm": 0.1800359229120948, "learning_rate": 6.714459918723613e-05, "loss": 0.9554, "step": 3713 }, { "epoch": 1.6820652173913042, "grad_norm": 0.1813250428490553, "learning_rate": 6.713530845237019e-05, "loss": 0.9466, "step": 3714 }, { "epoch": 1.6825181159420288, "grad_norm": 0.14448502186240852, "learning_rate": 6.712601500474447e-05, "loss": 0.9227, "step": 3715 }, { "epoch": 1.6829710144927537, "grad_norm": 0.1202319666103279, "learning_rate": 6.711671884528804e-05, "loss": 0.9669, "step": 3716 }, { "epoch": 1.6834239130434783, "grad_norm": 0.14361765921643468, "learning_rate": 6.710741997493025e-05, "loss": 0.9559, "step": 3717 }, { "epoch": 1.6838768115942029, "grad_norm": 0.19330856339738772, "learning_rate": 6.709811839460074e-05, "loss": 0.9642, "step": 3718 }, { "epoch": 1.6843297101449275, "grad_norm": 0.23487536291060918, "learning_rate": 6.708881410522939e-05, "loss": 0.9541, "step": 3719 }, { "epoch": 1.6847826086956523, "grad_norm": 0.2216919064556467, "learning_rate": 6.707950710774636e-05, "loss": 0.9477, "step": 3720 }, { "epoch": 1.685235507246377, "grad_norm": 0.21832057043642106, "learning_rate": 6.70701974030821e-05, "loss": 0.9228, "step": 3721 }, { "epoch": 1.6856884057971016, "grad_norm": 0.20659077363985673, "learning_rate": 6.70608849921673e-05, "loss": 0.9496, "step": 3722 }, { "epoch": 1.6861413043478262, "grad_norm": 0.1599894007063981, "learning_rate": 6.705156987593295e-05, "loss": 0.9503, "step": 3723 }, { "epoch": 1.6865942028985508, "grad_norm": 0.14201116837195996, "learning_rate": 6.70422520553103e-05, "loss": 0.9218, "step": 3724 }, { "epoch": 1.6870471014492754, "grad_norm": 0.1355529988979336, "learning_rate": 6.703293153123087e-05, "loss": 0.9551, "step": 3725 }, { "epoch": 1.6875, "grad_norm": 0.1823284179862668, "learning_rate": 6.702360830462642e-05, "loss": 0.9664, "step": 3726 }, { "epoch": 1.6879528985507246, "grad_norm": 0.21459884805830542, "learning_rate": 6.701428237642904e-05, "loss": 0.9252, "step": 3727 }, { "epoch": 1.6884057971014492, "grad_norm": 0.2019437827804538, "learning_rate": 6.700495374757102e-05, "loss": 0.9338, "step": 3728 }, { "epoch": 1.6888586956521738, "grad_norm": 0.17216906208042093, "learning_rate": 6.699562241898498e-05, "loss": 0.9479, "step": 3729 }, { "epoch": 1.6893115942028984, "grad_norm": 0.15510536517705928, "learning_rate": 6.698628839160381e-05, "loss": 0.9447, "step": 3730 }, { "epoch": 1.689764492753623, "grad_norm": 0.11266256304380078, "learning_rate": 6.697695166636061e-05, "loss": 0.9473, "step": 3731 }, { "epoch": 1.6902173913043477, "grad_norm": 0.11233068855660998, "learning_rate": 6.696761224418882e-05, "loss": 0.9522, "step": 3732 }, { "epoch": 1.6906702898550725, "grad_norm": 0.15947561365336252, "learning_rate": 6.695827012602209e-05, "loss": 0.9504, "step": 3733 }, { "epoch": 1.6911231884057971, "grad_norm": 0.2034568187362632, "learning_rate": 6.694892531279439e-05, "loss": 0.9472, "step": 3734 }, { "epoch": 1.6915760869565217, "grad_norm": 0.1929919968904827, "learning_rate": 6.69395778054399e-05, "loss": 0.9152, "step": 3735 }, { "epoch": 1.6920289855072463, "grad_norm": 0.12474577504314115, "learning_rate": 6.693022760489314e-05, "loss": 0.9599, "step": 3736 }, { "epoch": 1.6924818840579712, "grad_norm": 0.13405806688014063, "learning_rate": 6.692087471208884e-05, "loss": 0.9285, "step": 3737 }, { "epoch": 1.6929347826086958, "grad_norm": 0.14661901454800172, "learning_rate": 6.691151912796205e-05, "loss": 0.9503, "step": 3738 }, { "epoch": 1.6933876811594204, "grad_norm": 0.14480952855385576, "learning_rate": 6.690216085344804e-05, "loss": 0.9511, "step": 3739 }, { "epoch": 1.693840579710145, "grad_norm": 0.10874985410450885, "learning_rate": 6.689279988948239e-05, "loss": 0.9518, "step": 3740 }, { "epoch": 1.6942934782608696, "grad_norm": 0.15077682033197023, "learning_rate": 6.68834362370009e-05, "loss": 0.9514, "step": 3741 }, { "epoch": 1.6947463768115942, "grad_norm": 0.17532498062680282, "learning_rate": 6.68740698969397e-05, "loss": 0.9434, "step": 3742 }, { "epoch": 1.6951992753623188, "grad_norm": 0.13264076492670818, "learning_rate": 6.686470087023514e-05, "loss": 0.9399, "step": 3743 }, { "epoch": 1.6956521739130435, "grad_norm": 0.15023417799339472, "learning_rate": 6.685532915782388e-05, "loss": 0.9398, "step": 3744 }, { "epoch": 1.696105072463768, "grad_norm": 0.17015348043590547, "learning_rate": 6.68459547606428e-05, "loss": 0.9765, "step": 3745 }, { "epoch": 1.6965579710144927, "grad_norm": 0.18439175507884847, "learning_rate": 6.683657767962908e-05, "loss": 0.9513, "step": 3746 }, { "epoch": 1.6970108695652173, "grad_norm": 0.17818528549300197, "learning_rate": 6.682719791572016e-05, "loss": 0.9474, "step": 3747 }, { "epoch": 1.697463768115942, "grad_norm": 0.14718106817053672, "learning_rate": 6.681781546985376e-05, "loss": 0.9644, "step": 3748 }, { "epoch": 1.6979166666666665, "grad_norm": 0.12461413876888147, "learning_rate": 6.680843034296785e-05, "loss": 0.9489, "step": 3749 }, { "epoch": 1.6983695652173914, "grad_norm": 0.09720663203356111, "learning_rate": 6.679904253600066e-05, "loss": 0.9488, "step": 3750 }, { "epoch": 1.698822463768116, "grad_norm": 0.10091720801544868, "learning_rate": 6.678965204989073e-05, "loss": 0.9563, "step": 3751 }, { "epoch": 1.6992753623188406, "grad_norm": 0.1309061793093035, "learning_rate": 6.678025888557684e-05, "loss": 0.9493, "step": 3752 }, { "epoch": 1.6997282608695652, "grad_norm": 0.15343205317654726, "learning_rate": 6.677086304399803e-05, "loss": 0.9432, "step": 3753 }, { "epoch": 1.7001811594202898, "grad_norm": 0.14208299242051906, "learning_rate": 6.676146452609364e-05, "loss": 0.9281, "step": 3754 }, { "epoch": 1.7006340579710146, "grad_norm": 0.18726885995109757, "learning_rate": 6.67520633328032e-05, "loss": 0.9623, "step": 3755 }, { "epoch": 1.7010869565217392, "grad_norm": 0.12341360855597935, "learning_rate": 6.67426594650666e-05, "loss": 0.9314, "step": 3756 }, { "epoch": 1.7015398550724639, "grad_norm": 0.12080208275456263, "learning_rate": 6.673325292382397e-05, "loss": 0.9484, "step": 3757 }, { "epoch": 1.7019927536231885, "grad_norm": 0.17909311516340712, "learning_rate": 6.672384371001569e-05, "loss": 0.946, "step": 3758 }, { "epoch": 1.702445652173913, "grad_norm": 0.1851895193233734, "learning_rate": 6.671443182458238e-05, "loss": 0.9533, "step": 3759 }, { "epoch": 1.7028985507246377, "grad_norm": 0.181111690628582, "learning_rate": 6.670501726846501e-05, "loss": 0.9204, "step": 3760 }, { "epoch": 1.7033514492753623, "grad_norm": 0.19100757785734987, "learning_rate": 6.669560004260474e-05, "loss": 0.9485, "step": 3761 }, { "epoch": 1.703804347826087, "grad_norm": 0.14471521334404525, "learning_rate": 6.668618014794302e-05, "loss": 0.9415, "step": 3762 }, { "epoch": 1.7042572463768115, "grad_norm": 0.1018790340922905, "learning_rate": 6.66767575854216e-05, "loss": 0.9377, "step": 3763 }, { "epoch": 1.7047101449275361, "grad_norm": 0.10635459980988908, "learning_rate": 6.666733235598245e-05, "loss": 0.9454, "step": 3764 }, { "epoch": 1.7051630434782608, "grad_norm": 0.14518111659759442, "learning_rate": 6.665790446056781e-05, "loss": 0.9571, "step": 3765 }, { "epoch": 1.7056159420289854, "grad_norm": 0.13514161985939058, "learning_rate": 6.664847390012023e-05, "loss": 0.9259, "step": 3766 }, { "epoch": 1.7060688405797102, "grad_norm": 0.15455601691360646, "learning_rate": 6.663904067558247e-05, "loss": 0.9647, "step": 3767 }, { "epoch": 1.7065217391304348, "grad_norm": 0.18775298185410358, "learning_rate": 6.662960478789762e-05, "loss": 0.9519, "step": 3768 }, { "epoch": 1.7069746376811594, "grad_norm": 0.22809609167900724, "learning_rate": 6.662016623800899e-05, "loss": 0.9543, "step": 3769 }, { "epoch": 1.707427536231884, "grad_norm": 0.20927867073684658, "learning_rate": 6.661072502686014e-05, "loss": 0.9334, "step": 3770 }, { "epoch": 1.7078804347826086, "grad_norm": 0.14799890493621704, "learning_rate": 6.660128115539494e-05, "loss": 0.959, "step": 3771 }, { "epoch": 1.7083333333333335, "grad_norm": 0.13007821616701357, "learning_rate": 6.659183462455751e-05, "loss": 0.9513, "step": 3772 }, { "epoch": 1.708786231884058, "grad_norm": 0.13736899045682446, "learning_rate": 6.658238543529225e-05, "loss": 0.9558, "step": 3773 }, { "epoch": 1.7092391304347827, "grad_norm": 0.14639164524947962, "learning_rate": 6.657293358854376e-05, "loss": 0.9546, "step": 3774 }, { "epoch": 1.7096920289855073, "grad_norm": 0.19864728850899732, "learning_rate": 6.656347908525704e-05, "loss": 0.9589, "step": 3775 }, { "epoch": 1.710144927536232, "grad_norm": 0.23290655243785702, "learning_rate": 6.655402192637719e-05, "loss": 0.959, "step": 3776 }, { "epoch": 1.7105978260869565, "grad_norm": 0.18045667672169124, "learning_rate": 6.654456211284971e-05, "loss": 0.9636, "step": 3777 }, { "epoch": 1.7110507246376812, "grad_norm": 0.14818524462572802, "learning_rate": 6.653509964562027e-05, "loss": 0.9276, "step": 3778 }, { "epoch": 1.7115036231884058, "grad_norm": 0.1563106509769745, "learning_rate": 6.652563452563487e-05, "loss": 0.9437, "step": 3779 }, { "epoch": 1.7119565217391304, "grad_norm": 0.14540613081529136, "learning_rate": 6.651616675383977e-05, "loss": 0.9631, "step": 3780 }, { "epoch": 1.712409420289855, "grad_norm": 0.15300434932725904, "learning_rate": 6.650669633118144e-05, "loss": 0.9467, "step": 3781 }, { "epoch": 1.7128623188405796, "grad_norm": 0.14162471336161067, "learning_rate": 6.64972232586067e-05, "loss": 0.9301, "step": 3782 }, { "epoch": 1.7133152173913042, "grad_norm": 0.08881265509680768, "learning_rate": 6.648774753706254e-05, "loss": 0.9482, "step": 3783 }, { "epoch": 1.7137681159420288, "grad_norm": 0.1024715653288292, "learning_rate": 6.64782691674963e-05, "loss": 0.9697, "step": 3784 }, { "epoch": 1.7142210144927537, "grad_norm": 0.14135361067125513, "learning_rate": 6.646878815085552e-05, "loss": 0.9561, "step": 3785 }, { "epoch": 1.7146739130434783, "grad_norm": 0.17817618733244608, "learning_rate": 6.645930448808806e-05, "loss": 0.9517, "step": 3786 }, { "epoch": 1.7151268115942029, "grad_norm": 0.19932788172351606, "learning_rate": 6.644981818014199e-05, "loss": 0.964, "step": 3787 }, { "epoch": 1.7155797101449275, "grad_norm": 0.19057546888310903, "learning_rate": 6.644032922796568e-05, "loss": 0.9489, "step": 3788 }, { "epoch": 1.7160326086956523, "grad_norm": 0.1570986846120772, "learning_rate": 6.643083763250776e-05, "loss": 0.9318, "step": 3789 }, { "epoch": 1.716485507246377, "grad_norm": 0.15866315216655802, "learning_rate": 6.642134339471713e-05, "loss": 0.9548, "step": 3790 }, { "epoch": 1.7169384057971016, "grad_norm": 0.14764936053362918, "learning_rate": 6.641184651554292e-05, "loss": 0.9482, "step": 3791 }, { "epoch": 1.7173913043478262, "grad_norm": 0.12800531310636307, "learning_rate": 6.640234699593457e-05, "loss": 0.9609, "step": 3792 }, { "epoch": 1.7178442028985508, "grad_norm": 0.11093035347630296, "learning_rate": 6.639284483684176e-05, "loss": 0.921, "step": 3793 }, { "epoch": 1.7182971014492754, "grad_norm": 0.12557686958564954, "learning_rate": 6.638334003921441e-05, "loss": 0.9417, "step": 3794 }, { "epoch": 1.71875, "grad_norm": 0.10342810873528681, "learning_rate": 6.637383260400276e-05, "loss": 0.9279, "step": 3795 }, { "epoch": 1.7192028985507246, "grad_norm": 0.10388427996929993, "learning_rate": 6.636432253215728e-05, "loss": 0.935, "step": 3796 }, { "epoch": 1.7196557971014492, "grad_norm": 0.11338288713601105, "learning_rate": 6.635480982462868e-05, "loss": 0.9608, "step": 3797 }, { "epoch": 1.7201086956521738, "grad_norm": 0.11534064889031455, "learning_rate": 6.634529448236798e-05, "loss": 0.9519, "step": 3798 }, { "epoch": 1.7205615942028984, "grad_norm": 0.1304606630534735, "learning_rate": 6.633577650632646e-05, "loss": 0.9487, "step": 3799 }, { "epoch": 1.721014492753623, "grad_norm": 0.1204693038007521, "learning_rate": 6.632625589745561e-05, "loss": 0.9361, "step": 3800 }, { "epoch": 1.7214673913043477, "grad_norm": 0.11466671601215593, "learning_rate": 6.631673265670726e-05, "loss": 0.9305, "step": 3801 }, { "epoch": 1.7219202898550725, "grad_norm": 0.11335435368075783, "learning_rate": 6.630720678503343e-05, "loss": 0.9177, "step": 3802 }, { "epoch": 1.7223731884057971, "grad_norm": 0.12160054271247091, "learning_rate": 6.629767828338645e-05, "loss": 0.953, "step": 3803 }, { "epoch": 1.7228260869565217, "grad_norm": 0.12046616024115529, "learning_rate": 6.628814715271891e-05, "loss": 0.9494, "step": 3804 }, { "epoch": 1.7232789855072463, "grad_norm": 0.12868034788824922, "learning_rate": 6.627861339398364e-05, "loss": 0.9508, "step": 3805 }, { "epoch": 1.7237318840579712, "grad_norm": 0.09796967154883401, "learning_rate": 6.626907700813373e-05, "loss": 0.9517, "step": 3806 }, { "epoch": 1.7241847826086958, "grad_norm": 0.09250511646536813, "learning_rate": 6.625953799612259e-05, "loss": 0.9336, "step": 3807 }, { "epoch": 1.7246376811594204, "grad_norm": 0.12111196263392161, "learning_rate": 6.62499963589038e-05, "loss": 0.9364, "step": 3808 }, { "epoch": 1.725090579710145, "grad_norm": 0.15747894376548838, "learning_rate": 6.624045209743129e-05, "loss": 0.9406, "step": 3809 }, { "epoch": 1.7255434782608696, "grad_norm": 0.15763036892330118, "learning_rate": 6.62309052126592e-05, "loss": 0.9278, "step": 3810 }, { "epoch": 1.7259963768115942, "grad_norm": 0.14930111421766484, "learning_rate": 6.622135570554195e-05, "loss": 0.9336, "step": 3811 }, { "epoch": 1.7264492753623188, "grad_norm": 0.1544625131101951, "learning_rate": 6.621180357703423e-05, "loss": 0.9411, "step": 3812 }, { "epoch": 1.7269021739130435, "grad_norm": 0.1627909893430614, "learning_rate": 6.620224882809096e-05, "loss": 0.9515, "step": 3813 }, { "epoch": 1.727355072463768, "grad_norm": 0.1293312605198681, "learning_rate": 6.619269145966735e-05, "loss": 0.9398, "step": 3814 }, { "epoch": 1.7278079710144927, "grad_norm": 0.11026882997713175, "learning_rate": 6.618313147271887e-05, "loss": 0.9429, "step": 3815 }, { "epoch": 1.7282608695652173, "grad_norm": 0.13410969777540305, "learning_rate": 6.617356886820125e-05, "loss": 0.9275, "step": 3816 }, { "epoch": 1.728713768115942, "grad_norm": 0.173536516307528, "learning_rate": 6.616400364707048e-05, "loss": 0.9736, "step": 3817 }, { "epoch": 1.7291666666666665, "grad_norm": 0.18538497590707392, "learning_rate": 6.615443581028279e-05, "loss": 0.9481, "step": 3818 }, { "epoch": 1.7296195652173914, "grad_norm": 0.17814853563243163, "learning_rate": 6.614486535879472e-05, "loss": 0.9598, "step": 3819 }, { "epoch": 1.730072463768116, "grad_norm": 0.14523449097655064, "learning_rate": 6.613529229356303e-05, "loss": 0.9309, "step": 3820 }, { "epoch": 1.7305253623188406, "grad_norm": 0.12825215413757474, "learning_rate": 6.612571661554474e-05, "loss": 0.9402, "step": 3821 }, { "epoch": 1.7309782608695652, "grad_norm": 0.11350363413501678, "learning_rate": 6.611613832569717e-05, "loss": 0.9546, "step": 3822 }, { "epoch": 1.7314311594202898, "grad_norm": 0.14579303553863032, "learning_rate": 6.610655742497787e-05, "loss": 0.9529, "step": 3823 }, { "epoch": 1.7318840579710146, "grad_norm": 0.18567299363995457, "learning_rate": 6.609697391434465e-05, "loss": 0.9448, "step": 3824 }, { "epoch": 1.7323369565217392, "grad_norm": 0.17301588957663158, "learning_rate": 6.60873877947556e-05, "loss": 0.9368, "step": 3825 }, { "epoch": 1.7327898550724639, "grad_norm": 0.13947824285168797, "learning_rate": 6.607779906716905e-05, "loss": 0.9295, "step": 3826 }, { "epoch": 1.7332427536231885, "grad_norm": 0.12593476448544455, "learning_rate": 6.60682077325436e-05, "loss": 0.9326, "step": 3827 }, { "epoch": 1.733695652173913, "grad_norm": 0.10307149653322797, "learning_rate": 6.605861379183811e-05, "loss": 0.9269, "step": 3828 }, { "epoch": 1.7341485507246377, "grad_norm": 0.14307183255267236, "learning_rate": 6.604901724601171e-05, "loss": 0.9489, "step": 3829 }, { "epoch": 1.7346014492753623, "grad_norm": 0.16179274236636232, "learning_rate": 6.60394180960238e-05, "loss": 0.9408, "step": 3830 }, { "epoch": 1.735054347826087, "grad_norm": 0.13937731765669992, "learning_rate": 6.602981634283396e-05, "loss": 0.9648, "step": 3831 }, { "epoch": 1.7355072463768115, "grad_norm": 0.10434404768381672, "learning_rate": 6.602021198740217e-05, "loss": 0.9396, "step": 3832 }, { "epoch": 1.7359601449275361, "grad_norm": 0.12792699843241295, "learning_rate": 6.601060503068853e-05, "loss": 0.9538, "step": 3833 }, { "epoch": 1.7364130434782608, "grad_norm": 0.16818295664370952, "learning_rate": 6.600099547365348e-05, "loss": 0.9464, "step": 3834 }, { "epoch": 1.7368659420289854, "grad_norm": 0.1752637833005123, "learning_rate": 6.599138331725774e-05, "loss": 0.9428, "step": 3835 }, { "epoch": 1.7373188405797102, "grad_norm": 0.1397908309905773, "learning_rate": 6.59817685624622e-05, "loss": 0.9489, "step": 3836 }, { "epoch": 1.7377717391304348, "grad_norm": 0.1338688574116277, "learning_rate": 6.597215121022809e-05, "loss": 0.959, "step": 3837 }, { "epoch": 1.7382246376811594, "grad_norm": 0.14935408393268385, "learning_rate": 6.596253126151688e-05, "loss": 0.9594, "step": 3838 }, { "epoch": 1.738677536231884, "grad_norm": 0.15279076330142333, "learning_rate": 6.595290871729027e-05, "loss": 0.9415, "step": 3839 }, { "epoch": 1.7391304347826086, "grad_norm": 0.11624803100462303, "learning_rate": 6.594328357851024e-05, "loss": 0.9449, "step": 3840 }, { "epoch": 1.7395833333333335, "grad_norm": 0.11388493872336992, "learning_rate": 6.593365584613906e-05, "loss": 0.9484, "step": 3841 }, { "epoch": 1.740036231884058, "grad_norm": 0.14677904654242066, "learning_rate": 6.59240255211392e-05, "loss": 0.9501, "step": 3842 }, { "epoch": 1.7404891304347827, "grad_norm": 0.11972339210394363, "learning_rate": 6.591439260447342e-05, "loss": 0.9526, "step": 3843 }, { "epoch": 1.7409420289855073, "grad_norm": 0.10350689031902795, "learning_rate": 6.590475709710476e-05, "loss": 0.9741, "step": 3844 }, { "epoch": 1.741394927536232, "grad_norm": 0.12701572813122605, "learning_rate": 6.589511899999647e-05, "loss": 0.9427, "step": 3845 }, { "epoch": 1.7418478260869565, "grad_norm": 0.14392117244634567, "learning_rate": 6.588547831411211e-05, "loss": 0.9501, "step": 3846 }, { "epoch": 1.7423007246376812, "grad_norm": 0.13482541084567162, "learning_rate": 6.587583504041545e-05, "loss": 0.9656, "step": 3847 }, { "epoch": 1.7427536231884058, "grad_norm": 0.1016565595687279, "learning_rate": 6.586618917987058e-05, "loss": 0.9244, "step": 3848 }, { "epoch": 1.7432065217391304, "grad_norm": 0.10748058755727241, "learning_rate": 6.585654073344177e-05, "loss": 0.9665, "step": 3849 }, { "epoch": 1.743659420289855, "grad_norm": 0.1089757802219838, "learning_rate": 6.584688970209361e-05, "loss": 0.9493, "step": 3850 }, { "epoch": 1.7441123188405796, "grad_norm": 0.11147719333586514, "learning_rate": 6.583723608679095e-05, "loss": 0.9383, "step": 3851 }, { "epoch": 1.7445652173913042, "grad_norm": 0.13488574339506354, "learning_rate": 6.582757988849884e-05, "loss": 0.9299, "step": 3852 }, { "epoch": 1.7450181159420288, "grad_norm": 0.15071386307437643, "learning_rate": 6.581792110818263e-05, "loss": 0.9595, "step": 3853 }, { "epoch": 1.7454710144927537, "grad_norm": 0.14546899099997615, "learning_rate": 6.580825974680794e-05, "loss": 0.9444, "step": 3854 }, { "epoch": 1.7459239130434783, "grad_norm": 0.11411406152750656, "learning_rate": 6.579859580534063e-05, "loss": 0.9431, "step": 3855 }, { "epoch": 1.7463768115942029, "grad_norm": 0.09897925351459609, "learning_rate": 6.578892928474682e-05, "loss": 0.9495, "step": 3856 }, { "epoch": 1.7468297101449275, "grad_norm": 0.10512306459007784, "learning_rate": 6.577926018599289e-05, "loss": 0.9367, "step": 3857 }, { "epoch": 1.7472826086956523, "grad_norm": 0.12228111742080537, "learning_rate": 6.576958851004546e-05, "loss": 0.9395, "step": 3858 }, { "epoch": 1.747735507246377, "grad_norm": 0.140636232606962, "learning_rate": 6.575991425787143e-05, "loss": 0.9478, "step": 3859 }, { "epoch": 1.7481884057971016, "grad_norm": 0.1352733065812045, "learning_rate": 6.575023743043795e-05, "loss": 0.9313, "step": 3860 }, { "epoch": 1.7486413043478262, "grad_norm": 0.14678818337545876, "learning_rate": 6.574055802871244e-05, "loss": 0.9373, "step": 3861 }, { "epoch": 1.7490942028985508, "grad_norm": 0.14422308950461046, "learning_rate": 6.573087605366255e-05, "loss": 0.9467, "step": 3862 }, { "epoch": 1.7495471014492754, "grad_norm": 0.12135842473318628, "learning_rate": 6.57211915062562e-05, "loss": 0.9506, "step": 3863 }, { "epoch": 1.75, "grad_norm": 0.1083533159287248, "learning_rate": 6.571150438746157e-05, "loss": 0.9751, "step": 3864 }, { "epoch": 1.7504528985507246, "grad_norm": 0.0868929845725774, "learning_rate": 6.570181469824712e-05, "loss": 0.9345, "step": 3865 }, { "epoch": 1.7509057971014492, "grad_norm": 0.10308009215405146, "learning_rate": 6.569212243958153e-05, "loss": 0.9329, "step": 3866 }, { "epoch": 1.7513586956521738, "grad_norm": 0.11677513782648286, "learning_rate": 6.568242761243373e-05, "loss": 0.9563, "step": 3867 }, { "epoch": 1.7518115942028984, "grad_norm": 0.1445640610759464, "learning_rate": 6.567273021777296e-05, "loss": 0.9607, "step": 3868 }, { "epoch": 1.752264492753623, "grad_norm": 0.11280489227827559, "learning_rate": 6.566303025656865e-05, "loss": 0.9618, "step": 3869 }, { "epoch": 1.7527173913043477, "grad_norm": 0.12058231829212956, "learning_rate": 6.565332772979055e-05, "loss": 0.9672, "step": 3870 }, { "epoch": 1.7531702898550725, "grad_norm": 0.12769092325303322, "learning_rate": 6.564362263840861e-05, "loss": 0.9482, "step": 3871 }, { "epoch": 1.7536231884057971, "grad_norm": 0.13628850576007265, "learning_rate": 6.563391498339312e-05, "loss": 0.961, "step": 3872 }, { "epoch": 1.7540760869565217, "grad_norm": 0.1374587336947709, "learning_rate": 6.56242047657145e-05, "loss": 0.9602, "step": 3873 }, { "epoch": 1.7545289855072463, "grad_norm": 0.12772869621199645, "learning_rate": 6.561449198634352e-05, "loss": 0.963, "step": 3874 }, { "epoch": 1.7549818840579712, "grad_norm": 0.1100386493515576, "learning_rate": 6.56047766462512e-05, "loss": 0.9359, "step": 3875 }, { "epoch": 1.7554347826086958, "grad_norm": 0.1008458034564621, "learning_rate": 6.559505874640875e-05, "loss": 0.9466, "step": 3876 }, { "epoch": 1.7558876811594204, "grad_norm": 0.11252914174713867, "learning_rate": 6.558533828778776e-05, "loss": 0.9497, "step": 3877 }, { "epoch": 1.756340579710145, "grad_norm": 0.1346063412553051, "learning_rate": 6.557561527135994e-05, "loss": 0.9429, "step": 3878 }, { "epoch": 1.7567934782608696, "grad_norm": 0.15714535645267413, "learning_rate": 6.556588969809733e-05, "loss": 0.9549, "step": 3879 }, { "epoch": 1.7572463768115942, "grad_norm": 0.16465888279740218, "learning_rate": 6.555616156897222e-05, "loss": 0.9505, "step": 3880 }, { "epoch": 1.7576992753623188, "grad_norm": 0.1501660460651466, "learning_rate": 6.554643088495715e-05, "loss": 0.9802, "step": 3881 }, { "epoch": 1.7581521739130435, "grad_norm": 0.10869901442377526, "learning_rate": 6.55366976470249e-05, "loss": 0.938, "step": 3882 }, { "epoch": 1.758605072463768, "grad_norm": 0.10172293639951324, "learning_rate": 6.552696185614852e-05, "loss": 0.9324, "step": 3883 }, { "epoch": 1.7590579710144927, "grad_norm": 0.12589575653763457, "learning_rate": 6.55172235133013e-05, "loss": 0.948, "step": 3884 }, { "epoch": 1.7595108695652173, "grad_norm": 0.12584416403273221, "learning_rate": 6.55074826194568e-05, "loss": 0.9654, "step": 3885 }, { "epoch": 1.759963768115942, "grad_norm": 0.09940957025073521, "learning_rate": 6.549773917558886e-05, "loss": 0.9556, "step": 3886 }, { "epoch": 1.7604166666666665, "grad_norm": 0.12520640648136291, "learning_rate": 6.548799318267154e-05, "loss": 0.9594, "step": 3887 }, { "epoch": 1.7608695652173914, "grad_norm": 0.12348176853327152, "learning_rate": 6.547824464167913e-05, "loss": 0.9645, "step": 3888 }, { "epoch": 1.761322463768116, "grad_norm": 0.113394935563682, "learning_rate": 6.546849355358624e-05, "loss": 0.9445, "step": 3889 }, { "epoch": 1.7617753623188406, "grad_norm": 0.13093283421497998, "learning_rate": 6.54587399193677e-05, "loss": 0.9644, "step": 3890 }, { "epoch": 1.7622282608695652, "grad_norm": 0.11038850139148236, "learning_rate": 6.544898373999857e-05, "loss": 0.946, "step": 3891 }, { "epoch": 1.7626811594202898, "grad_norm": 0.11902823427875782, "learning_rate": 6.543922501645424e-05, "loss": 0.9508, "step": 3892 }, { "epoch": 1.7631340579710146, "grad_norm": 0.113821198393128, "learning_rate": 6.542946374971024e-05, "loss": 0.9312, "step": 3893 }, { "epoch": 1.7635869565217392, "grad_norm": 0.09605486439002346, "learning_rate": 6.541969994074246e-05, "loss": 0.9329, "step": 3894 }, { "epoch": 1.7640398550724639, "grad_norm": 0.10092525114024542, "learning_rate": 6.5409933590527e-05, "loss": 0.9409, "step": 3895 }, { "epoch": 1.7644927536231885, "grad_norm": 0.09948541978575755, "learning_rate": 6.54001647000402e-05, "loss": 0.9451, "step": 3896 }, { "epoch": 1.764945652173913, "grad_norm": 0.10254989323918454, "learning_rate": 6.539039327025871e-05, "loss": 0.9331, "step": 3897 }, { "epoch": 1.7653985507246377, "grad_norm": 0.11500566674079017, "learning_rate": 6.538061930215936e-05, "loss": 0.9454, "step": 3898 }, { "epoch": 1.7658514492753623, "grad_norm": 0.15666944735772004, "learning_rate": 6.537084279671927e-05, "loss": 0.9421, "step": 3899 }, { "epoch": 1.766304347826087, "grad_norm": 0.1545084781204531, "learning_rate": 6.536106375491583e-05, "loss": 0.9487, "step": 3900 }, { "epoch": 1.7667572463768115, "grad_norm": 0.1055116893213569, "learning_rate": 6.535128217772667e-05, "loss": 0.9587, "step": 3901 }, { "epoch": 1.7672101449275361, "grad_norm": 0.09594801735373824, "learning_rate": 6.534149806612962e-05, "loss": 0.9371, "step": 3902 }, { "epoch": 1.7676630434782608, "grad_norm": 0.12920159348219348, "learning_rate": 6.53317114211029e-05, "loss": 0.9289, "step": 3903 }, { "epoch": 1.7681159420289854, "grad_norm": 0.14887389795156192, "learning_rate": 6.532192224362482e-05, "loss": 0.9269, "step": 3904 }, { "epoch": 1.7685688405797102, "grad_norm": 0.15809956832012623, "learning_rate": 6.531213053467404e-05, "loss": 0.9659, "step": 3905 }, { "epoch": 1.7690217391304348, "grad_norm": 0.18360120090024135, "learning_rate": 6.530233629522948e-05, "loss": 0.9489, "step": 3906 }, { "epoch": 1.7694746376811594, "grad_norm": 0.19217800364933857, "learning_rate": 6.529253952627026e-05, "loss": 0.9634, "step": 3907 }, { "epoch": 1.769927536231884, "grad_norm": 0.21836751966199866, "learning_rate": 6.528274022877578e-05, "loss": 0.9513, "step": 3908 }, { "epoch": 1.7703804347826086, "grad_norm": 0.18834751791178125, "learning_rate": 6.527293840372569e-05, "loss": 0.9651, "step": 3909 }, { "epoch": 1.7708333333333335, "grad_norm": 0.1825157040436153, "learning_rate": 6.526313405209991e-05, "loss": 0.9333, "step": 3910 }, { "epoch": 1.771286231884058, "grad_norm": 0.1439709969560925, "learning_rate": 6.525332717487859e-05, "loss": 0.9197, "step": 3911 }, { "epoch": 1.7717391304347827, "grad_norm": 0.15525884960480643, "learning_rate": 6.524351777304212e-05, "loss": 0.9225, "step": 3912 }, { "epoch": 1.7721920289855073, "grad_norm": 0.1787734285360064, "learning_rate": 6.523370584757118e-05, "loss": 0.9496, "step": 3913 }, { "epoch": 1.772644927536232, "grad_norm": 0.12939073763459888, "learning_rate": 6.522389139944668e-05, "loss": 0.952, "step": 3914 }, { "epoch": 1.7730978260869565, "grad_norm": 0.15982339122073388, "learning_rate": 6.52140744296498e-05, "loss": 0.9414, "step": 3915 }, { "epoch": 1.7735507246376812, "grad_norm": 0.1776105692165865, "learning_rate": 6.520425493916194e-05, "loss": 0.9746, "step": 3916 }, { "epoch": 1.7740036231884058, "grad_norm": 0.1664032253974723, "learning_rate": 6.519443292896476e-05, "loss": 0.9645, "step": 3917 }, { "epoch": 1.7744565217391304, "grad_norm": 0.14032135098345933, "learning_rate": 6.51846084000402e-05, "loss": 0.9499, "step": 3918 }, { "epoch": 1.774909420289855, "grad_norm": 0.1569377426730688, "learning_rate": 6.517478135337044e-05, "loss": 0.9236, "step": 3919 }, { "epoch": 1.7753623188405796, "grad_norm": 0.13047961031635688, "learning_rate": 6.516495178993788e-05, "loss": 0.9546, "step": 3920 }, { "epoch": 1.7758152173913042, "grad_norm": 0.11253698328579113, "learning_rate": 6.515511971072523e-05, "loss": 0.9502, "step": 3921 }, { "epoch": 1.7762681159420288, "grad_norm": 0.12013342206807644, "learning_rate": 6.51452851167154e-05, "loss": 0.9359, "step": 3922 }, { "epoch": 1.7767210144927537, "grad_norm": 0.1931765198491219, "learning_rate": 6.513544800889155e-05, "loss": 0.9421, "step": 3923 }, { "epoch": 1.7771739130434783, "grad_norm": 0.18182600716008657, "learning_rate": 6.512560838823715e-05, "loss": 0.9479, "step": 3924 }, { "epoch": 1.7776268115942029, "grad_norm": 0.139822787198771, "learning_rate": 6.511576625573585e-05, "loss": 0.9584, "step": 3925 }, { "epoch": 1.7780797101449275, "grad_norm": 0.1193771393346012, "learning_rate": 6.510592161237162e-05, "loss": 0.9203, "step": 3926 }, { "epoch": 1.7785326086956523, "grad_norm": 0.11246845017272399, "learning_rate": 6.50960744591286e-05, "loss": 0.9672, "step": 3927 }, { "epoch": 1.778985507246377, "grad_norm": 0.12491873481098376, "learning_rate": 6.508622479699126e-05, "loss": 0.947, "step": 3928 }, { "epoch": 1.7794384057971016, "grad_norm": 0.1306761980582313, "learning_rate": 6.507637262694427e-05, "loss": 0.9428, "step": 3929 }, { "epoch": 1.7798913043478262, "grad_norm": 0.1409528386588343, "learning_rate": 6.506651794997257e-05, "loss": 0.9451, "step": 3930 }, { "epoch": 1.7803442028985508, "grad_norm": 0.16035801229368152, "learning_rate": 6.505666076706136e-05, "loss": 0.9461, "step": 3931 }, { "epoch": 1.7807971014492754, "grad_norm": 0.16874889217806224, "learning_rate": 6.504680107919606e-05, "loss": 0.9362, "step": 3932 }, { "epoch": 1.78125, "grad_norm": 0.17073304903903572, "learning_rate": 6.503693888736238e-05, "loss": 0.9504, "step": 3933 }, { "epoch": 1.7817028985507246, "grad_norm": 0.1708488359421594, "learning_rate": 6.502707419254622e-05, "loss": 0.9255, "step": 3934 }, { "epoch": 1.7821557971014492, "grad_norm": 0.1456385494850413, "learning_rate": 6.50172069957338e-05, "loss": 0.9325, "step": 3935 }, { "epoch": 1.7826086956521738, "grad_norm": 0.13628066571840075, "learning_rate": 6.500733729791156e-05, "loss": 0.9295, "step": 3936 }, { "epoch": 1.7830615942028984, "grad_norm": 0.11895208955706355, "learning_rate": 6.499746510006618e-05, "loss": 0.953, "step": 3937 }, { "epoch": 1.783514492753623, "grad_norm": 0.15050315999301492, "learning_rate": 6.498759040318461e-05, "loss": 0.9479, "step": 3938 }, { "epoch": 1.7839673913043477, "grad_norm": 0.21371371498002137, "learning_rate": 6.497771320825402e-05, "loss": 0.9548, "step": 3939 }, { "epoch": 1.7844202898550725, "grad_norm": 0.26606369442498456, "learning_rate": 6.496783351626187e-05, "loss": 0.9467, "step": 3940 }, { "epoch": 1.7848731884057971, "grad_norm": 0.24583984658486238, "learning_rate": 6.495795132819585e-05, "loss": 0.9446, "step": 3941 }, { "epoch": 1.7853260869565217, "grad_norm": 0.16225146839838614, "learning_rate": 6.494806664504387e-05, "loss": 0.9467, "step": 3942 }, { "epoch": 1.7857789855072463, "grad_norm": 0.09609851787941938, "learning_rate": 6.493817946779412e-05, "loss": 0.9247, "step": 3943 }, { "epoch": 1.7862318840579712, "grad_norm": 0.12695409557659137, "learning_rate": 6.492828979743507e-05, "loss": 0.9476, "step": 3944 }, { "epoch": 1.7866847826086958, "grad_norm": 0.15199187253268978, "learning_rate": 6.491839763495538e-05, "loss": 0.927, "step": 3945 }, { "epoch": 1.7871376811594204, "grad_norm": 0.14463420916727485, "learning_rate": 6.490850298134399e-05, "loss": 0.9586, "step": 3946 }, { "epoch": 1.787590579710145, "grad_norm": 0.10413721854090073, "learning_rate": 6.48986058375901e-05, "loss": 0.9371, "step": 3947 }, { "epoch": 1.7880434782608696, "grad_norm": 0.11905421797885725, "learning_rate": 6.488870620468312e-05, "loss": 0.9295, "step": 3948 }, { "epoch": 1.7884963768115942, "grad_norm": 0.1688766775143016, "learning_rate": 6.487880408361273e-05, "loss": 0.9448, "step": 3949 }, { "epoch": 1.7889492753623188, "grad_norm": 0.13350362419587047, "learning_rate": 6.486889947536888e-05, "loss": 0.9704, "step": 3950 }, { "epoch": 1.7894021739130435, "grad_norm": 0.09321698644397824, "learning_rate": 6.485899238094175e-05, "loss": 0.9483, "step": 3951 }, { "epoch": 1.789855072463768, "grad_norm": 0.12693128890701263, "learning_rate": 6.484908280132175e-05, "loss": 0.9357, "step": 3952 }, { "epoch": 1.7903079710144927, "grad_norm": 0.15760337385846235, "learning_rate": 6.483917073749957e-05, "loss": 0.9116, "step": 3953 }, { "epoch": 1.7907608695652173, "grad_norm": 0.17269230044063097, "learning_rate": 6.482925619046614e-05, "loss": 0.923, "step": 3954 }, { "epoch": 1.791213768115942, "grad_norm": 0.15151970924921634, "learning_rate": 6.481933916121261e-05, "loss": 0.9354, "step": 3955 }, { "epoch": 1.7916666666666665, "grad_norm": 0.13815820911996762, "learning_rate": 6.480941965073041e-05, "loss": 0.9332, "step": 3956 }, { "epoch": 1.7921195652173914, "grad_norm": 0.22054388512185025, "learning_rate": 6.479949766001122e-05, "loss": 0.9883, "step": 3957 }, { "epoch": 1.792572463768116, "grad_norm": 0.21395467339773938, "learning_rate": 6.478957319004695e-05, "loss": 0.9556, "step": 3958 }, { "epoch": 1.7930253623188406, "grad_norm": 0.1521549782618116, "learning_rate": 6.477964624182975e-05, "loss": 0.9407, "step": 3959 }, { "epoch": 1.7934782608695652, "grad_norm": 0.12154208457748919, "learning_rate": 6.476971681635209e-05, "loss": 0.9568, "step": 3960 }, { "epoch": 1.7939311594202898, "grad_norm": 0.16050097620817125, "learning_rate": 6.475978491460656e-05, "loss": 0.9301, "step": 3961 }, { "epoch": 1.7943840579710146, "grad_norm": 0.16049598770564433, "learning_rate": 6.47498505375861e-05, "loss": 0.959, "step": 3962 }, { "epoch": 1.7948369565217392, "grad_norm": 0.11408329703492916, "learning_rate": 6.473991368628386e-05, "loss": 0.9694, "step": 3963 }, { "epoch": 1.7952898550724639, "grad_norm": 0.13979250209706834, "learning_rate": 6.472997436169324e-05, "loss": 0.954, "step": 3964 }, { "epoch": 1.7957427536231885, "grad_norm": 0.16855247770006532, "learning_rate": 6.472003256480791e-05, "loss": 0.9317, "step": 3965 }, { "epoch": 1.796195652173913, "grad_norm": 0.17621039918038084, "learning_rate": 6.471008829662175e-05, "loss": 0.9602, "step": 3966 }, { "epoch": 1.7966485507246377, "grad_norm": 0.18452049863010897, "learning_rate": 6.47001415581289e-05, "loss": 0.9331, "step": 3967 }, { "epoch": 1.7971014492753623, "grad_norm": 0.19742102134478817, "learning_rate": 6.469019235032376e-05, "loss": 0.9512, "step": 3968 }, { "epoch": 1.797554347826087, "grad_norm": 0.17863558289423334, "learning_rate": 6.468024067420097e-05, "loss": 0.9308, "step": 3969 }, { "epoch": 1.7980072463768115, "grad_norm": 0.12419386889445078, "learning_rate": 6.46702865307554e-05, "loss": 0.9362, "step": 3970 }, { "epoch": 1.7984601449275361, "grad_norm": 0.12198434083281089, "learning_rate": 6.46603299209822e-05, "loss": 0.927, "step": 3971 }, { "epoch": 1.7989130434782608, "grad_norm": 0.14229837057401823, "learning_rate": 6.465037084587673e-05, "loss": 0.9333, "step": 3972 }, { "epoch": 1.7993659420289854, "grad_norm": 0.16018367933683408, "learning_rate": 6.464040930643464e-05, "loss": 0.9208, "step": 3973 }, { "epoch": 1.7998188405797102, "grad_norm": 0.16227308030280393, "learning_rate": 6.463044530365178e-05, "loss": 0.9563, "step": 3974 }, { "epoch": 1.8002717391304348, "grad_norm": 0.15787217923005745, "learning_rate": 6.462047883852427e-05, "loss": 0.9533, "step": 3975 }, { "epoch": 1.8007246376811594, "grad_norm": 0.11030652436984106, "learning_rate": 6.461050991204846e-05, "loss": 0.9294, "step": 3976 }, { "epoch": 1.801177536231884, "grad_norm": 0.09774818129853503, "learning_rate": 6.4600538525221e-05, "loss": 0.955, "step": 3977 }, { "epoch": 1.8016304347826086, "grad_norm": 0.12580026063050612, "learning_rate": 6.45905646790387e-05, "loss": 0.9375, "step": 3978 }, { "epoch": 1.8020833333333335, "grad_norm": 0.11772513125382073, "learning_rate": 6.458058837449871e-05, "loss": 0.9215, "step": 3979 }, { "epoch": 1.802536231884058, "grad_norm": 0.11578505407542915, "learning_rate": 6.457060961259833e-05, "loss": 0.9399, "step": 3980 }, { "epoch": 1.8029891304347827, "grad_norm": 0.11669472275843862, "learning_rate": 6.456062839433518e-05, "loss": 0.9557, "step": 3981 }, { "epoch": 1.8034420289855073, "grad_norm": 0.1441046576758473, "learning_rate": 6.455064472070709e-05, "loss": 0.9423, "step": 3982 }, { "epoch": 1.803894927536232, "grad_norm": 0.12178343106836044, "learning_rate": 6.454065859271213e-05, "loss": 0.9539, "step": 3983 }, { "epoch": 1.8043478260869565, "grad_norm": 0.16881639165997012, "learning_rate": 6.453067001134866e-05, "loss": 0.9807, "step": 3984 }, { "epoch": 1.8048007246376812, "grad_norm": 0.1408295795915706, "learning_rate": 6.452067897761522e-05, "loss": 0.9344, "step": 3985 }, { "epoch": 1.8052536231884058, "grad_norm": 0.1297063701400138, "learning_rate": 6.451068549251066e-05, "loss": 0.9403, "step": 3986 }, { "epoch": 1.8057065217391304, "grad_norm": 0.13548343118689124, "learning_rate": 6.450068955703403e-05, "loss": 0.9456, "step": 3987 }, { "epoch": 1.806159420289855, "grad_norm": 0.17939498367649054, "learning_rate": 6.449069117218463e-05, "loss": 0.9262, "step": 3988 }, { "epoch": 1.8066123188405796, "grad_norm": 0.16687452437072747, "learning_rate": 6.448069033896203e-05, "loss": 0.9514, "step": 3989 }, { "epoch": 1.8070652173913042, "grad_norm": 0.12050525909900303, "learning_rate": 6.447068705836602e-05, "loss": 0.9355, "step": 3990 }, { "epoch": 1.8075181159420288, "grad_norm": 0.12385567163241783, "learning_rate": 6.446068133139666e-05, "loss": 0.9287, "step": 3991 }, { "epoch": 1.8079710144927537, "grad_norm": 0.13154381705550822, "learning_rate": 6.445067315905421e-05, "loss": 0.9129, "step": 3992 }, { "epoch": 1.8084239130434783, "grad_norm": 0.12459964259676329, "learning_rate": 6.444066254233923e-05, "loss": 0.9456, "step": 3993 }, { "epoch": 1.8088768115942029, "grad_norm": 0.14279472390603226, "learning_rate": 6.443064948225248e-05, "loss": 0.9595, "step": 3994 }, { "epoch": 1.8093297101449275, "grad_norm": 0.11620339111749176, "learning_rate": 6.4420633979795e-05, "loss": 0.9297, "step": 3995 }, { "epoch": 1.8097826086956523, "grad_norm": 0.09742335999322756, "learning_rate": 6.441061603596803e-05, "loss": 0.9607, "step": 3996 }, { "epoch": 1.810235507246377, "grad_norm": 0.10418657142704474, "learning_rate": 6.44005956517731e-05, "loss": 0.9462, "step": 3997 }, { "epoch": 1.8106884057971016, "grad_norm": 0.10192076995614667, "learning_rate": 6.439057282821195e-05, "loss": 0.9293, "step": 3998 }, { "epoch": 1.8111413043478262, "grad_norm": 0.12038255447068241, "learning_rate": 6.438054756628659e-05, "loss": 0.9441, "step": 3999 }, { "epoch": 1.8115942028985508, "grad_norm": 0.13103303876319383, "learning_rate": 6.437051986699924e-05, "loss": 0.9397, "step": 4000 }, { "epoch": 1.8120471014492754, "grad_norm": 0.15525416438404072, "learning_rate": 6.436048973135242e-05, "loss": 0.9729, "step": 4001 }, { "epoch": 1.8125, "grad_norm": 0.15099497003876242, "learning_rate": 6.435045716034883e-05, "loss": 0.9323, "step": 4002 }, { "epoch": 1.8129528985507246, "grad_norm": 0.09749244115309992, "learning_rate": 6.434042215499146e-05, "loss": 0.9443, "step": 4003 }, { "epoch": 1.8134057971014492, "grad_norm": 0.12986822750670307, "learning_rate": 6.43303847162835e-05, "loss": 0.9664, "step": 4004 }, { "epoch": 1.8138586956521738, "grad_norm": 0.1382286434063111, "learning_rate": 6.432034484522843e-05, "loss": 0.9479, "step": 4005 }, { "epoch": 1.8143115942028984, "grad_norm": 0.12226426743744354, "learning_rate": 6.431030254282995e-05, "loss": 0.9521, "step": 4006 }, { "epoch": 1.814764492753623, "grad_norm": 0.11490844565163492, "learning_rate": 6.430025781009202e-05, "loss": 0.959, "step": 4007 }, { "epoch": 1.8152173913043477, "grad_norm": 0.12772265732119534, "learning_rate": 6.429021064801879e-05, "loss": 0.9416, "step": 4008 }, { "epoch": 1.8156702898550725, "grad_norm": 0.14327050856704987, "learning_rate": 6.42801610576147e-05, "loss": 0.9363, "step": 4009 }, { "epoch": 1.8161231884057971, "grad_norm": 0.1301658074492359, "learning_rate": 6.427010903988445e-05, "loss": 0.945, "step": 4010 }, { "epoch": 1.8165760869565217, "grad_norm": 0.12237062877258542, "learning_rate": 6.426005459583295e-05, "loss": 0.9486, "step": 4011 }, { "epoch": 1.8170289855072463, "grad_norm": 0.12257498071377851, "learning_rate": 6.424999772646533e-05, "loss": 0.9444, "step": 4012 }, { "epoch": 1.8174818840579712, "grad_norm": 0.11723167377175212, "learning_rate": 6.423993843278702e-05, "loss": 0.9559, "step": 4013 }, { "epoch": 1.8179347826086958, "grad_norm": 0.09522857072077659, "learning_rate": 6.422987671580364e-05, "loss": 0.9363, "step": 4014 }, { "epoch": 1.8183876811594204, "grad_norm": 0.09683494314830977, "learning_rate": 6.42198125765211e-05, "loss": 0.9453, "step": 4015 }, { "epoch": 1.818840579710145, "grad_norm": 0.11037647155251776, "learning_rate": 6.420974601594552e-05, "loss": 0.9304, "step": 4016 }, { "epoch": 1.8192934782608696, "grad_norm": 0.13037298788035412, "learning_rate": 6.419967703508326e-05, "loss": 0.9209, "step": 4017 }, { "epoch": 1.8197463768115942, "grad_norm": 0.12745266925832408, "learning_rate": 6.418960563494095e-05, "loss": 0.9588, "step": 4018 }, { "epoch": 1.8201992753623188, "grad_norm": 0.12106046223641913, "learning_rate": 6.417953181652543e-05, "loss": 0.9345, "step": 4019 }, { "epoch": 1.8206521739130435, "grad_norm": 0.1382298860543519, "learning_rate": 6.416945558084379e-05, "loss": 0.9512, "step": 4020 }, { "epoch": 1.821105072463768, "grad_norm": 0.1562528259887903, "learning_rate": 6.415937692890337e-05, "loss": 0.9444, "step": 4021 }, { "epoch": 1.8215579710144927, "grad_norm": 0.1190687298653645, "learning_rate": 6.414929586171177e-05, "loss": 0.9624, "step": 4022 }, { "epoch": 1.8220108695652173, "grad_norm": 0.16296317631765098, "learning_rate": 6.41392123802768e-05, "loss": 0.9278, "step": 4023 }, { "epoch": 1.822463768115942, "grad_norm": 0.1727274270998275, "learning_rate": 6.41291264856065e-05, "loss": 0.9637, "step": 4024 }, { "epoch": 1.8229166666666665, "grad_norm": 0.10807196326667454, "learning_rate": 6.411903817870919e-05, "loss": 0.9423, "step": 4025 }, { "epoch": 1.8233695652173914, "grad_norm": 0.15076434128712243, "learning_rate": 6.410894746059342e-05, "loss": 0.935, "step": 4026 }, { "epoch": 1.823822463768116, "grad_norm": 0.18963125694625008, "learning_rate": 6.409885433226796e-05, "loss": 0.9329, "step": 4027 }, { "epoch": 1.8242753623188406, "grad_norm": 0.16085713932974116, "learning_rate": 6.408875879474186e-05, "loss": 0.94, "step": 4028 }, { "epoch": 1.8247282608695652, "grad_norm": 0.13303247830168796, "learning_rate": 6.407866084902437e-05, "loss": 0.9434, "step": 4029 }, { "epoch": 1.8251811594202898, "grad_norm": 0.1275025188091679, "learning_rate": 6.406856049612499e-05, "loss": 0.9218, "step": 4030 }, { "epoch": 1.8256340579710146, "grad_norm": 0.13179527451352366, "learning_rate": 6.405845773705346e-05, "loss": 0.9524, "step": 4031 }, { "epoch": 1.8260869565217392, "grad_norm": 0.1285196371412924, "learning_rate": 6.404835257281981e-05, "loss": 0.9412, "step": 4032 }, { "epoch": 1.8265398550724639, "grad_norm": 0.1419719168254877, "learning_rate": 6.403824500443425e-05, "loss": 0.9458, "step": 4033 }, { "epoch": 1.8269927536231885, "grad_norm": 0.11742960904708277, "learning_rate": 6.402813503290724e-05, "loss": 0.9582, "step": 4034 }, { "epoch": 1.827445652173913, "grad_norm": 0.0934801832338576, "learning_rate": 6.401802265924949e-05, "loss": 0.9342, "step": 4035 }, { "epoch": 1.8278985507246377, "grad_norm": 0.12645132116675184, "learning_rate": 6.400790788447195e-05, "loss": 0.9599, "step": 4036 }, { "epoch": 1.8283514492753623, "grad_norm": 0.1558040275419551, "learning_rate": 6.399779070958581e-05, "loss": 0.9369, "step": 4037 }, { "epoch": 1.828804347826087, "grad_norm": 0.1539165677031871, "learning_rate": 6.398767113560253e-05, "loss": 0.9376, "step": 4038 }, { "epoch": 1.8292572463768115, "grad_norm": 0.1615397766960858, "learning_rate": 6.397754916353373e-05, "loss": 0.9544, "step": 4039 }, { "epoch": 1.8297101449275361, "grad_norm": 0.1418857507623452, "learning_rate": 6.396742479439136e-05, "loss": 0.9602, "step": 4040 }, { "epoch": 1.8301630434782608, "grad_norm": 0.11985364777315469, "learning_rate": 6.395729802918754e-05, "loss": 0.943, "step": 4041 }, { "epoch": 1.8306159420289854, "grad_norm": 0.1280671975237335, "learning_rate": 6.394716886893468e-05, "loss": 0.9621, "step": 4042 }, { "epoch": 1.8310688405797102, "grad_norm": 0.1290704102840271, "learning_rate": 6.39370373146454e-05, "loss": 0.9387, "step": 4043 }, { "epoch": 1.8315217391304348, "grad_norm": 0.10659244562516643, "learning_rate": 6.392690336733257e-05, "loss": 0.9496, "step": 4044 }, { "epoch": 1.8319746376811594, "grad_norm": 0.11525835393318196, "learning_rate": 6.391676702800929e-05, "loss": 0.9528, "step": 4045 }, { "epoch": 1.832427536231884, "grad_norm": 0.14008332719849326, "learning_rate": 6.390662829768894e-05, "loss": 0.9482, "step": 4046 }, { "epoch": 1.8328804347826086, "grad_norm": 0.11827971876623142, "learning_rate": 6.389648717738504e-05, "loss": 0.9505, "step": 4047 }, { "epoch": 1.8333333333333335, "grad_norm": 0.1275838525478191, "learning_rate": 6.388634366811146e-05, "loss": 0.93, "step": 4048 }, { "epoch": 1.833786231884058, "grad_norm": 0.141412357587261, "learning_rate": 6.387619777088224e-05, "loss": 0.9419, "step": 4049 }, { "epoch": 1.8342391304347827, "grad_norm": 0.10136762629823436, "learning_rate": 6.386604948671171e-05, "loss": 0.9459, "step": 4050 }, { "epoch": 1.8346920289855073, "grad_norm": 0.10743281402225784, "learning_rate": 6.385589881661439e-05, "loss": 0.9257, "step": 4051 }, { "epoch": 1.835144927536232, "grad_norm": 0.1419444812203703, "learning_rate": 6.384574576160507e-05, "loss": 0.9427, "step": 4052 }, { "epoch": 1.8355978260869565, "grad_norm": 0.12815021431644344, "learning_rate": 6.383559032269874e-05, "loss": 0.945, "step": 4053 }, { "epoch": 1.8360507246376812, "grad_norm": 0.10899456621583094, "learning_rate": 6.38254325009107e-05, "loss": 0.955, "step": 4054 }, { "epoch": 1.8365036231884058, "grad_norm": 0.10805502337326538, "learning_rate": 6.38152722972564e-05, "loss": 0.9346, "step": 4055 }, { "epoch": 1.8369565217391304, "grad_norm": 0.10308780312627888, "learning_rate": 6.38051097127516e-05, "loss": 0.9564, "step": 4056 }, { "epoch": 1.837409420289855, "grad_norm": 0.09816089348874386, "learning_rate": 6.379494474841225e-05, "loss": 0.9638, "step": 4057 }, { "epoch": 1.8378623188405796, "grad_norm": 0.0988537198423467, "learning_rate": 6.378477740525456e-05, "loss": 0.952, "step": 4058 }, { "epoch": 1.8383152173913042, "grad_norm": 0.12038793668309762, "learning_rate": 6.377460768429502e-05, "loss": 0.9317, "step": 4059 }, { "epoch": 1.8387681159420288, "grad_norm": 0.11658169887541406, "learning_rate": 6.376443558655024e-05, "loss": 0.9413, "step": 4060 }, { "epoch": 1.8392210144927537, "grad_norm": 0.09795255717133146, "learning_rate": 6.375426111303718e-05, "loss": 0.9446, "step": 4061 }, { "epoch": 1.8396739130434783, "grad_norm": 0.11750244751868259, "learning_rate": 6.374408426477301e-05, "loss": 0.9512, "step": 4062 }, { "epoch": 1.8401268115942029, "grad_norm": 0.11499411574230434, "learning_rate": 6.373390504277509e-05, "loss": 0.9525, "step": 4063 }, { "epoch": 1.8405797101449275, "grad_norm": 0.11624559788743127, "learning_rate": 6.372372344806109e-05, "loss": 0.9339, "step": 4064 }, { "epoch": 1.8410326086956523, "grad_norm": 0.16795901665427415, "learning_rate": 6.371353948164885e-05, "loss": 0.948, "step": 4065 }, { "epoch": 1.841485507246377, "grad_norm": 0.17260526820863029, "learning_rate": 6.370335314455651e-05, "loss": 0.9413, "step": 4066 }, { "epoch": 1.8419384057971016, "grad_norm": 0.13885619738843183, "learning_rate": 6.369316443780238e-05, "loss": 0.9558, "step": 4067 }, { "epoch": 1.8423913043478262, "grad_norm": 0.10493305315633016, "learning_rate": 6.368297336240506e-05, "loss": 0.9433, "step": 4068 }, { "epoch": 1.8428442028985508, "grad_norm": 0.10693298067815402, "learning_rate": 6.367277991938337e-05, "loss": 0.9512, "step": 4069 }, { "epoch": 1.8432971014492754, "grad_norm": 0.1090499994581406, "learning_rate": 6.366258410975635e-05, "loss": 0.9426, "step": 4070 }, { "epoch": 1.84375, "grad_norm": 0.12401697955097243, "learning_rate": 6.365238593454331e-05, "loss": 0.9683, "step": 4071 }, { "epoch": 1.8442028985507246, "grad_norm": 0.13071968628261782, "learning_rate": 6.364218539476375e-05, "loss": 0.9452, "step": 4072 }, { "epoch": 1.8446557971014492, "grad_norm": 0.13617166425995886, "learning_rate": 6.363198249143746e-05, "loss": 0.9601, "step": 4073 }, { "epoch": 1.8451086956521738, "grad_norm": 0.12364443162618226, "learning_rate": 6.362177722558444e-05, "loss": 0.9369, "step": 4074 }, { "epoch": 1.8455615942028984, "grad_norm": 0.10514782565475743, "learning_rate": 6.361156959822492e-05, "loss": 0.941, "step": 4075 }, { "epoch": 1.846014492753623, "grad_norm": 0.11819835083520251, "learning_rate": 6.360135961037936e-05, "loss": 0.9517, "step": 4076 }, { "epoch": 1.8464673913043477, "grad_norm": 0.13305056910107288, "learning_rate": 6.359114726306847e-05, "loss": 0.9525, "step": 4077 }, { "epoch": 1.8469202898550725, "grad_norm": 0.10806806100845286, "learning_rate": 6.358093255731323e-05, "loss": 0.9437, "step": 4078 }, { "epoch": 1.8473731884057971, "grad_norm": 0.11193073303099221, "learning_rate": 6.357071549413478e-05, "loss": 0.9407, "step": 4079 }, { "epoch": 1.8478260869565217, "grad_norm": 0.12088733011201576, "learning_rate": 6.356049607455456e-05, "loss": 0.9504, "step": 4080 }, { "epoch": 1.8482789855072463, "grad_norm": 0.117639924539556, "learning_rate": 6.355027429959421e-05, "loss": 0.9314, "step": 4081 }, { "epoch": 1.8487318840579712, "grad_norm": 0.09762542972252565, "learning_rate": 6.354005017027563e-05, "loss": 0.9537, "step": 4082 }, { "epoch": 1.8491847826086958, "grad_norm": 0.1604138069240303, "learning_rate": 6.352982368762091e-05, "loss": 0.9485, "step": 4083 }, { "epoch": 1.8496376811594204, "grad_norm": 0.11947517014195407, "learning_rate": 6.351959485265245e-05, "loss": 0.9462, "step": 4084 }, { "epoch": 1.850090579710145, "grad_norm": 0.14745913639680963, "learning_rate": 6.350936366639281e-05, "loss": 0.9524, "step": 4085 }, { "epoch": 1.8505434782608696, "grad_norm": 0.1379831324801199, "learning_rate": 6.349913012986482e-05, "loss": 0.9264, "step": 4086 }, { "epoch": 1.8509963768115942, "grad_norm": 0.1187109924143669, "learning_rate": 6.348889424409158e-05, "loss": 0.9389, "step": 4087 }, { "epoch": 1.8514492753623188, "grad_norm": 0.08285713432773159, "learning_rate": 6.347865601009635e-05, "loss": 0.9505, "step": 4088 }, { "epoch": 1.8519021739130435, "grad_norm": 0.09381866037613257, "learning_rate": 6.346841542890266e-05, "loss": 0.9543, "step": 4089 }, { "epoch": 1.852355072463768, "grad_norm": 0.25899075256065995, "learning_rate": 6.34581725015343e-05, "loss": 0.9476, "step": 4090 }, { "epoch": 1.8528079710144927, "grad_norm": 0.12650386573927058, "learning_rate": 6.344792722901527e-05, "loss": 0.9499, "step": 4091 }, { "epoch": 1.8532608695652173, "grad_norm": 0.1388678599488625, "learning_rate": 6.34376796123698e-05, "loss": 0.933, "step": 4092 }, { "epoch": 1.853713768115942, "grad_norm": 0.1341518466084501, "learning_rate": 6.342742965262235e-05, "loss": 0.93, "step": 4093 }, { "epoch": 1.8541666666666665, "grad_norm": 0.10978658896761997, "learning_rate": 6.341717735079763e-05, "loss": 0.9342, "step": 4094 }, { "epoch": 1.8546195652173914, "grad_norm": 0.11170730428551981, "learning_rate": 6.340692270792059e-05, "loss": 0.9804, "step": 4095 }, { "epoch": 1.855072463768116, "grad_norm": 0.12272179974275638, "learning_rate": 6.339666572501639e-05, "loss": 0.9499, "step": 4096 }, { "epoch": 1.8555253623188406, "grad_norm": 0.09150901029974493, "learning_rate": 6.338640640311043e-05, "loss": 0.9301, "step": 4097 }, { "epoch": 1.8559782608695652, "grad_norm": 0.14607642510587932, "learning_rate": 6.33761447432284e-05, "loss": 0.9539, "step": 4098 }, { "epoch": 1.8564311594202898, "grad_norm": 0.12616464335954772, "learning_rate": 6.336588074639612e-05, "loss": 0.9519, "step": 4099 }, { "epoch": 1.8568840579710146, "grad_norm": 0.11846028094617676, "learning_rate": 6.33556144136397e-05, "loss": 0.9476, "step": 4100 }, { "epoch": 1.8573369565217392, "grad_norm": 0.11387323565933799, "learning_rate": 6.334534574598551e-05, "loss": 0.9454, "step": 4101 }, { "epoch": 1.8577898550724639, "grad_norm": 0.11726285459924407, "learning_rate": 6.333507474446011e-05, "loss": 0.936, "step": 4102 }, { "epoch": 1.8582427536231885, "grad_norm": 0.13951225446839582, "learning_rate": 6.332480141009031e-05, "loss": 0.9296, "step": 4103 }, { "epoch": 1.858695652173913, "grad_norm": 0.1181151145803519, "learning_rate": 6.331452574390314e-05, "loss": 0.9587, "step": 4104 }, { "epoch": 1.8591485507246377, "grad_norm": 0.15173451107650573, "learning_rate": 6.330424774692591e-05, "loss": 0.9524, "step": 4105 }, { "epoch": 1.8596014492753623, "grad_norm": 0.13320210996586787, "learning_rate": 6.329396742018609e-05, "loss": 0.9333, "step": 4106 }, { "epoch": 1.860054347826087, "grad_norm": 0.1157545181732515, "learning_rate": 6.328368476471143e-05, "loss": 0.9682, "step": 4107 }, { "epoch": 1.8605072463768115, "grad_norm": 0.11055361693781006, "learning_rate": 6.327339978152992e-05, "loss": 0.9127, "step": 4108 }, { "epoch": 1.8609601449275361, "grad_norm": 0.09452873333669652, "learning_rate": 6.326311247166975e-05, "loss": 0.9351, "step": 4109 }, { "epoch": 1.8614130434782608, "grad_norm": 0.09358568458236666, "learning_rate": 6.325282283615937e-05, "loss": 0.9361, "step": 4110 }, { "epoch": 1.8618659420289854, "grad_norm": 0.1075237601803835, "learning_rate": 6.324253087602743e-05, "loss": 0.9649, "step": 4111 }, { "epoch": 1.8623188405797102, "grad_norm": 0.10881820904972819, "learning_rate": 6.323223659230284e-05, "loss": 0.9613, "step": 4112 }, { "epoch": 1.8627717391304348, "grad_norm": 0.09757056980022494, "learning_rate": 6.322193998601476e-05, "loss": 0.9309, "step": 4113 }, { "epoch": 1.8632246376811594, "grad_norm": 0.09853661164511504, "learning_rate": 6.321164105819253e-05, "loss": 0.9549, "step": 4114 }, { "epoch": 1.863677536231884, "grad_norm": 0.10467228418863013, "learning_rate": 6.320133980986579e-05, "loss": 0.9174, "step": 4115 }, { "epoch": 1.8641304347826086, "grad_norm": 0.12885681053736187, "learning_rate": 6.319103624206431e-05, "loss": 0.9422, "step": 4116 }, { "epoch": 1.8645833333333335, "grad_norm": 0.1499325132661032, "learning_rate": 6.318073035581821e-05, "loss": 0.9332, "step": 4117 }, { "epoch": 1.865036231884058, "grad_norm": 0.19076904293153424, "learning_rate": 6.317042215215775e-05, "loss": 0.9158, "step": 4118 }, { "epoch": 1.8654891304347827, "grad_norm": 0.21998111045014185, "learning_rate": 6.316011163211346e-05, "loss": 0.9374, "step": 4119 }, { "epoch": 1.8659420289855073, "grad_norm": 0.21596338246593436, "learning_rate": 6.314979879671614e-05, "loss": 0.9391, "step": 4120 }, { "epoch": 1.866394927536232, "grad_norm": 0.17373911298836198, "learning_rate": 6.313948364699671e-05, "loss": 0.9277, "step": 4121 }, { "epoch": 1.8668478260869565, "grad_norm": 0.1117023254973443, "learning_rate": 6.312916618398646e-05, "loss": 0.9441, "step": 4122 }, { "epoch": 1.8673007246376812, "grad_norm": 0.10615287640090054, "learning_rate": 6.311884640871682e-05, "loss": 0.9466, "step": 4123 }, { "epoch": 1.8677536231884058, "grad_norm": 0.135339411565073, "learning_rate": 6.310852432221946e-05, "loss": 0.9509, "step": 4124 }, { "epoch": 1.8682065217391304, "grad_norm": 0.15905752553718527, "learning_rate": 6.30981999255263e-05, "loss": 0.9489, "step": 4125 }, { "epoch": 1.868659420289855, "grad_norm": 0.18525580382306367, "learning_rate": 6.30878732196695e-05, "loss": 0.937, "step": 4126 }, { "epoch": 1.8691123188405796, "grad_norm": 0.19083425574303836, "learning_rate": 6.307754420568143e-05, "loss": 0.957, "step": 4127 }, { "epoch": 1.8695652173913042, "grad_norm": 0.19169857443517413, "learning_rate": 6.30672128845947e-05, "loss": 0.9413, "step": 4128 }, { "epoch": 1.8700181159420288, "grad_norm": 0.16061912744836246, "learning_rate": 6.305687925744214e-05, "loss": 0.9644, "step": 4129 }, { "epoch": 1.8704710144927537, "grad_norm": 0.11956716741430623, "learning_rate": 6.304654332525681e-05, "loss": 0.9483, "step": 4130 }, { "epoch": 1.8709239130434783, "grad_norm": 0.10460155672333603, "learning_rate": 6.303620508907202e-05, "loss": 0.9371, "step": 4131 }, { "epoch": 1.8713768115942029, "grad_norm": 0.10646221551529451, "learning_rate": 6.302586454992133e-05, "loss": 0.9287, "step": 4132 }, { "epoch": 1.8718297101449275, "grad_norm": 0.10712537606382111, "learning_rate": 6.301552170883845e-05, "loss": 0.9319, "step": 4133 }, { "epoch": 1.8722826086956523, "grad_norm": 0.1348457742818261, "learning_rate": 6.30051765668574e-05, "loss": 0.9524, "step": 4134 }, { "epoch": 1.872735507246377, "grad_norm": 0.15150672177020566, "learning_rate": 6.299482912501241e-05, "loss": 0.9351, "step": 4135 }, { "epoch": 1.8731884057971016, "grad_norm": 0.12809198117903692, "learning_rate": 6.29844793843379e-05, "loss": 0.9527, "step": 4136 }, { "epoch": 1.8736413043478262, "grad_norm": 0.11800612750549928, "learning_rate": 6.297412734586856e-05, "loss": 0.9507, "step": 4137 }, { "epoch": 1.8740942028985508, "grad_norm": 0.13489146819084677, "learning_rate": 6.296377301063932e-05, "loss": 0.968, "step": 4138 }, { "epoch": 1.8745471014492754, "grad_norm": 0.12211594923753206, "learning_rate": 6.29534163796853e-05, "loss": 0.9518, "step": 4139 }, { "epoch": 1.875, "grad_norm": 0.09420955555381347, "learning_rate": 6.294305745404185e-05, "loss": 0.954, "step": 4140 }, { "epoch": 1.8754528985507246, "grad_norm": 0.10528496356664194, "learning_rate": 6.293269623474461e-05, "loss": 0.929, "step": 4141 }, { "epoch": 1.8759057971014492, "grad_norm": 0.12862111514912736, "learning_rate": 6.292233272282938e-05, "loss": 0.9271, "step": 4142 }, { "epoch": 1.8763586956521738, "grad_norm": 0.13786045851593684, "learning_rate": 6.291196691933222e-05, "loss": 0.9584, "step": 4143 }, { "epoch": 1.8768115942028984, "grad_norm": 0.15115131618497246, "learning_rate": 6.290159882528942e-05, "loss": 0.9207, "step": 4144 }, { "epoch": 1.877264492753623, "grad_norm": 0.15155368988620238, "learning_rate": 6.28912284417375e-05, "loss": 0.9476, "step": 4145 }, { "epoch": 1.8777173913043477, "grad_norm": 0.12790337011711297, "learning_rate": 6.28808557697132e-05, "loss": 0.9808, "step": 4146 }, { "epoch": 1.8781702898550725, "grad_norm": 0.10197235721646705, "learning_rate": 6.28704808102535e-05, "loss": 0.9673, "step": 4147 }, { "epoch": 1.8786231884057971, "grad_norm": 0.13105881477572776, "learning_rate": 6.286010356439558e-05, "loss": 0.9287, "step": 4148 }, { "epoch": 1.8790760869565217, "grad_norm": 0.14377339172482012, "learning_rate": 6.284972403317687e-05, "loss": 0.9397, "step": 4149 }, { "epoch": 1.8795289855072463, "grad_norm": 0.1283482657109791, "learning_rate": 6.283934221763506e-05, "loss": 0.9601, "step": 4150 }, { "epoch": 1.8799818840579712, "grad_norm": 0.12951237522810155, "learning_rate": 6.282895811880801e-05, "loss": 0.9645, "step": 4151 }, { "epoch": 1.8804347826086958, "grad_norm": 0.11035848720862944, "learning_rate": 6.281857173773384e-05, "loss": 0.953, "step": 4152 }, { "epoch": 1.8808876811594204, "grad_norm": 0.10988614381810385, "learning_rate": 6.280818307545089e-05, "loss": 0.9449, "step": 4153 }, { "epoch": 1.881340579710145, "grad_norm": 0.12308017885940245, "learning_rate": 6.279779213299775e-05, "loss": 0.9474, "step": 4154 }, { "epoch": 1.8817934782608696, "grad_norm": 0.12633094783209825, "learning_rate": 6.278739891141319e-05, "loss": 0.9692, "step": 4155 }, { "epoch": 1.8822463768115942, "grad_norm": 0.13541198911714789, "learning_rate": 6.277700341173625e-05, "loss": 0.9158, "step": 4156 }, { "epoch": 1.8826992753623188, "grad_norm": 0.13235124259357633, "learning_rate": 6.276660563500619e-05, "loss": 0.9429, "step": 4157 }, { "epoch": 1.8831521739130435, "grad_norm": 0.09311668946200338, "learning_rate": 6.27562055822625e-05, "loss": 0.9443, "step": 4158 }, { "epoch": 1.883605072463768, "grad_norm": 0.11870334397507876, "learning_rate": 6.274580325454485e-05, "loss": 0.9409, "step": 4159 }, { "epoch": 1.8840579710144927, "grad_norm": 0.16434271020589572, "learning_rate": 6.273539865289322e-05, "loss": 0.9439, "step": 4160 }, { "epoch": 1.8845108695652173, "grad_norm": 0.18780776114293274, "learning_rate": 6.272499177834777e-05, "loss": 0.9852, "step": 4161 }, { "epoch": 1.884963768115942, "grad_norm": 0.1697315084132526, "learning_rate": 6.271458263194888e-05, "loss": 0.9536, "step": 4162 }, { "epoch": 1.8854166666666665, "grad_norm": 0.15471838851089478, "learning_rate": 6.270417121473716e-05, "loss": 0.9518, "step": 4163 }, { "epoch": 1.8858695652173914, "grad_norm": 0.12041108205365167, "learning_rate": 6.269375752775346e-05, "loss": 0.9541, "step": 4164 }, { "epoch": 1.886322463768116, "grad_norm": 0.10976558358242487, "learning_rate": 6.268334157203886e-05, "loss": 0.9498, "step": 4165 }, { "epoch": 1.8867753623188406, "grad_norm": 0.1587224085949009, "learning_rate": 6.267292334863467e-05, "loss": 0.9319, "step": 4166 }, { "epoch": 1.8872282608695652, "grad_norm": 0.16384341923393247, "learning_rate": 6.26625028585824e-05, "loss": 0.936, "step": 4167 }, { "epoch": 1.8876811594202898, "grad_norm": 0.18247897333710206, "learning_rate": 6.26520801029238e-05, "loss": 0.9555, "step": 4168 }, { "epoch": 1.8881340579710146, "grad_norm": 0.19784603749607688, "learning_rate": 6.264165508270087e-05, "loss": 0.9502, "step": 4169 }, { "epoch": 1.8885869565217392, "grad_norm": 0.1737773542194102, "learning_rate": 6.26312277989558e-05, "loss": 0.9283, "step": 4170 }, { "epoch": 1.8890398550724639, "grad_norm": 0.11809873552461772, "learning_rate": 6.262079825273101e-05, "loss": 0.9472, "step": 4171 }, { "epoch": 1.8894927536231885, "grad_norm": 0.10905035896555332, "learning_rate": 6.261036644506918e-05, "loss": 0.9488, "step": 4172 }, { "epoch": 1.889945652173913, "grad_norm": 0.17453857520199817, "learning_rate": 6.25999323770132e-05, "loss": 0.9446, "step": 4173 }, { "epoch": 1.8903985507246377, "grad_norm": 0.1781574130925268, "learning_rate": 6.258949604960616e-05, "loss": 0.941, "step": 4174 }, { "epoch": 1.8908514492753623, "grad_norm": 0.16580097056209986, "learning_rate": 6.25790574638914e-05, "loss": 0.9272, "step": 4175 }, { "epoch": 1.891304347826087, "grad_norm": 0.12541298261789235, "learning_rate": 6.25686166209125e-05, "loss": 0.9335, "step": 4176 }, { "epoch": 1.8917572463768115, "grad_norm": 0.11242476905727805, "learning_rate": 6.25581735217132e-05, "loss": 0.9254, "step": 4177 }, { "epoch": 1.8922101449275361, "grad_norm": 0.10429584838108927, "learning_rate": 6.254772816733758e-05, "loss": 0.9475, "step": 4178 }, { "epoch": 1.8926630434782608, "grad_norm": 0.12219687166401605, "learning_rate": 6.253728055882983e-05, "loss": 0.9672, "step": 4179 }, { "epoch": 1.8931159420289854, "grad_norm": 0.15278975593531108, "learning_rate": 6.252683069723443e-05, "loss": 0.939, "step": 4180 }, { "epoch": 1.8935688405797102, "grad_norm": 0.14461241906581193, "learning_rate": 6.251637858359607e-05, "loss": 0.9512, "step": 4181 }, { "epoch": 1.8940217391304348, "grad_norm": 0.12522554086355517, "learning_rate": 6.250592421895966e-05, "loss": 0.9529, "step": 4182 }, { "epoch": 1.8944746376811594, "grad_norm": 0.10607251743739933, "learning_rate": 6.249546760437035e-05, "loss": 0.9433, "step": 4183 }, { "epoch": 1.894927536231884, "grad_norm": 0.09102860125224643, "learning_rate": 6.248500874087349e-05, "loss": 0.9405, "step": 4184 }, { "epoch": 1.8953804347826086, "grad_norm": 0.1081083702848434, "learning_rate": 6.247454762951468e-05, "loss": 0.9563, "step": 4185 }, { "epoch": 1.8958333333333335, "grad_norm": 0.13601938859211676, "learning_rate": 6.246408427133972e-05, "loss": 0.9458, "step": 4186 }, { "epoch": 1.896286231884058, "grad_norm": 0.14059254696719778, "learning_rate": 6.245361866739468e-05, "loss": 0.9379, "step": 4187 }, { "epoch": 1.8967391304347827, "grad_norm": 0.11724921425909048, "learning_rate": 6.244315081872577e-05, "loss": 0.9433, "step": 4188 }, { "epoch": 1.8971920289855073, "grad_norm": 0.11457911117863494, "learning_rate": 6.243268072637952e-05, "loss": 0.9724, "step": 4189 }, { "epoch": 1.897644927536232, "grad_norm": 0.10625659811502045, "learning_rate": 6.242220839140264e-05, "loss": 0.9439, "step": 4190 }, { "epoch": 1.8980978260869565, "grad_norm": 0.09313784136822542, "learning_rate": 6.241173381484204e-05, "loss": 0.9485, "step": 4191 }, { "epoch": 1.8985507246376812, "grad_norm": 0.12684117409394355, "learning_rate": 6.240125699774492e-05, "loss": 0.9557, "step": 4192 }, { "epoch": 1.8990036231884058, "grad_norm": 0.11615876693027931, "learning_rate": 6.239077794115863e-05, "loss": 0.9499, "step": 4193 }, { "epoch": 1.8994565217391304, "grad_norm": 0.10802641252284305, "learning_rate": 6.238029664613078e-05, "loss": 0.9648, "step": 4194 }, { "epoch": 1.899909420289855, "grad_norm": 0.1394874653238315, "learning_rate": 6.23698131137092e-05, "loss": 0.955, "step": 4195 }, { "epoch": 1.9003623188405796, "grad_norm": 0.1168387284239055, "learning_rate": 6.235932734494198e-05, "loss": 0.9434, "step": 4196 }, { "epoch": 1.9008152173913042, "grad_norm": 0.14423809543037322, "learning_rate": 6.234883934087737e-05, "loss": 0.9359, "step": 4197 }, { "epoch": 1.9012681159420288, "grad_norm": 0.16750513554741778, "learning_rate": 6.233834910256386e-05, "loss": 0.9332, "step": 4198 }, { "epoch": 1.9017210144927537, "grad_norm": 0.15211627309757947, "learning_rate": 6.23278566310502e-05, "loss": 0.9556, "step": 4199 }, { "epoch": 1.9021739130434783, "grad_norm": 0.14324208305260608, "learning_rate": 6.231736192738535e-05, "loss": 0.9461, "step": 4200 }, { "epoch": 1.9026268115942029, "grad_norm": 0.13224774469511688, "learning_rate": 6.230686499261844e-05, "loss": 0.9423, "step": 4201 }, { "epoch": 1.9030797101449275, "grad_norm": 0.14775026948854528, "learning_rate": 6.22963658277989e-05, "loss": 0.9508, "step": 4202 }, { "epoch": 1.9035326086956523, "grad_norm": 0.15791202621251815, "learning_rate": 6.228586443397632e-05, "loss": 0.9419, "step": 4203 }, { "epoch": 1.903985507246377, "grad_norm": 0.15424124139246428, "learning_rate": 6.227536081220057e-05, "loss": 0.9398, "step": 4204 }, { "epoch": 1.9044384057971016, "grad_norm": 0.1465282836461326, "learning_rate": 6.22648549635217e-05, "loss": 0.9394, "step": 4205 }, { "epoch": 1.9048913043478262, "grad_norm": 0.17459172303610557, "learning_rate": 6.225434688899e-05, "loss": 0.9329, "step": 4206 }, { "epoch": 1.9053442028985508, "grad_norm": 0.12698032968486694, "learning_rate": 6.224383658965597e-05, "loss": 0.9708, "step": 4207 }, { "epoch": 1.9057971014492754, "grad_norm": 0.10406467079548937, "learning_rate": 6.223332406657034e-05, "loss": 0.9552, "step": 4208 }, { "epoch": 1.90625, "grad_norm": 0.12006733761101542, "learning_rate": 6.22228093207841e-05, "loss": 0.9582, "step": 4209 }, { "epoch": 1.9067028985507246, "grad_norm": 0.09573365386547898, "learning_rate": 6.221229235334838e-05, "loss": 0.9282, "step": 4210 }, { "epoch": 1.9071557971014492, "grad_norm": 0.09722437350259248, "learning_rate": 6.220177316531459e-05, "loss": 0.9425, "step": 4211 }, { "epoch": 1.9076086956521738, "grad_norm": 0.09528513949209104, "learning_rate": 6.219125175773436e-05, "loss": 0.9383, "step": 4212 }, { "epoch": 1.9080615942028984, "grad_norm": 0.10961048709816057, "learning_rate": 6.218072813165952e-05, "loss": 0.9652, "step": 4213 }, { "epoch": 1.908514492753623, "grad_norm": 0.13073086790517044, "learning_rate": 6.217020228814215e-05, "loss": 0.9485, "step": 4214 }, { "epoch": 1.9089673913043477, "grad_norm": 0.136017405981404, "learning_rate": 6.215967422823452e-05, "loss": 0.9521, "step": 4215 }, { "epoch": 1.9094202898550725, "grad_norm": 0.1278783472278255, "learning_rate": 6.214914395298914e-05, "loss": 0.946, "step": 4216 }, { "epoch": 1.9098731884057971, "grad_norm": 0.14237267040450927, "learning_rate": 6.213861146345876e-05, "loss": 0.962, "step": 4217 }, { "epoch": 1.9103260869565217, "grad_norm": 0.10573151882892957, "learning_rate": 6.21280767606963e-05, "loss": 0.9394, "step": 4218 }, { "epoch": 1.9107789855072463, "grad_norm": 0.11424058558695353, "learning_rate": 6.211753984575493e-05, "loss": 0.9539, "step": 4219 }, { "epoch": 1.9112318840579712, "grad_norm": 0.13162013313070156, "learning_rate": 6.210700071968806e-05, "loss": 0.9629, "step": 4220 }, { "epoch": 1.9116847826086958, "grad_norm": 0.14737655119371648, "learning_rate": 6.20964593835493e-05, "loss": 0.9423, "step": 4221 }, { "epoch": 1.9121376811594204, "grad_norm": 0.14842597893734252, "learning_rate": 6.208591583839248e-05, "loss": 0.9662, "step": 4222 }, { "epoch": 1.912590579710145, "grad_norm": 0.12913053874856323, "learning_rate": 6.207537008527165e-05, "loss": 0.9335, "step": 4223 }, { "epoch": 1.9130434782608696, "grad_norm": 0.09262645479613332, "learning_rate": 6.206482212524109e-05, "loss": 0.9175, "step": 4224 }, { "epoch": 1.9134963768115942, "grad_norm": 0.11014919361479675, "learning_rate": 6.20542719593553e-05, "loss": 0.9387, "step": 4225 }, { "epoch": 1.9139492753623188, "grad_norm": 0.13639439581093768, "learning_rate": 6.204371958866901e-05, "loss": 0.9517, "step": 4226 }, { "epoch": 1.9144021739130435, "grad_norm": 0.13266710743919366, "learning_rate": 6.203316501423712e-05, "loss": 0.9434, "step": 4227 }, { "epoch": 1.914855072463768, "grad_norm": 0.14501749550802676, "learning_rate": 6.202260823711482e-05, "loss": 0.9211, "step": 4228 }, { "epoch": 1.9153079710144927, "grad_norm": 0.13155020107299678, "learning_rate": 6.201204925835747e-05, "loss": 0.9601, "step": 4229 }, { "epoch": 1.9157608695652173, "grad_norm": 0.17068830415735042, "learning_rate": 6.200148807902069e-05, "loss": 0.9366, "step": 4230 }, { "epoch": 1.916213768115942, "grad_norm": 0.15802310499492178, "learning_rate": 6.199092470016026e-05, "loss": 0.9492, "step": 4231 }, { "epoch": 1.9166666666666665, "grad_norm": 0.12001785280189115, "learning_rate": 6.198035912283225e-05, "loss": 0.9324, "step": 4232 }, { "epoch": 1.9171195652173914, "grad_norm": 0.1085151350006637, "learning_rate": 6.19697913480929e-05, "loss": 0.9458, "step": 4233 }, { "epoch": 1.917572463768116, "grad_norm": 0.1283166657975748, "learning_rate": 6.195922137699872e-05, "loss": 0.9299, "step": 4234 }, { "epoch": 1.9180253623188406, "grad_norm": 0.09290868452177051, "learning_rate": 6.194864921060636e-05, "loss": 0.9476, "step": 4235 }, { "epoch": 1.9184782608695652, "grad_norm": 0.11897038408622547, "learning_rate": 6.193807484997275e-05, "loss": 0.9413, "step": 4236 }, { "epoch": 1.9189311594202898, "grad_norm": 0.14159312898558676, "learning_rate": 6.192749829615505e-05, "loss": 0.9351, "step": 4237 }, { "epoch": 1.9193840579710146, "grad_norm": 0.10766719795514292, "learning_rate": 6.191691955021059e-05, "loss": 0.9414, "step": 4238 }, { "epoch": 1.9198369565217392, "grad_norm": 0.11132030901584729, "learning_rate": 6.190633861319696e-05, "loss": 0.9495, "step": 4239 }, { "epoch": 1.9202898550724639, "grad_norm": 0.12402031281866478, "learning_rate": 6.189575548617194e-05, "loss": 0.9347, "step": 4240 }, { "epoch": 1.9207427536231885, "grad_norm": 0.10775169826215776, "learning_rate": 6.188517017019355e-05, "loss": 0.9316, "step": 4241 }, { "epoch": 1.921195652173913, "grad_norm": 0.1010587047052906, "learning_rate": 6.187458266632002e-05, "loss": 0.9275, "step": 4242 }, { "epoch": 1.9216485507246377, "grad_norm": 0.13834958040925244, "learning_rate": 6.186399297560979e-05, "loss": 0.9407, "step": 4243 }, { "epoch": 1.9221014492753623, "grad_norm": 0.1387897901103358, "learning_rate": 6.185340109912155e-05, "loss": 0.943, "step": 4244 }, { "epoch": 1.922554347826087, "grad_norm": 0.11433824405895708, "learning_rate": 6.184280703791415e-05, "loss": 0.9629, "step": 4245 }, { "epoch": 1.9230072463768115, "grad_norm": 0.11505275928386029, "learning_rate": 6.183221079304673e-05, "loss": 0.9549, "step": 4246 }, { "epoch": 1.9234601449275361, "grad_norm": 0.12443458016175922, "learning_rate": 6.18216123655786e-05, "loss": 0.9499, "step": 4247 }, { "epoch": 1.9239130434782608, "grad_norm": 0.11882764110081333, "learning_rate": 6.181101175656932e-05, "loss": 0.9557, "step": 4248 }, { "epoch": 1.9243659420289854, "grad_norm": 0.09318533915334618, "learning_rate": 6.180040896707861e-05, "loss": 0.9473, "step": 4249 }, { "epoch": 1.9248188405797102, "grad_norm": 0.08474824274320224, "learning_rate": 6.178980399816647e-05, "loss": 0.9442, "step": 4250 }, { "epoch": 1.9252717391304348, "grad_norm": 0.10611268895377486, "learning_rate": 6.17791968508931e-05, "loss": 0.947, "step": 4251 }, { "epoch": 1.9257246376811594, "grad_norm": 0.1191271114230568, "learning_rate": 6.17685875263189e-05, "loss": 0.9389, "step": 4252 }, { "epoch": 1.926177536231884, "grad_norm": 0.13087108091997374, "learning_rate": 6.175797602550452e-05, "loss": 0.9429, "step": 4253 }, { "epoch": 1.9266304347826086, "grad_norm": 0.11798120871701308, "learning_rate": 6.17473623495108e-05, "loss": 0.9265, "step": 4254 }, { "epoch": 1.9270833333333335, "grad_norm": 0.11364591422879752, "learning_rate": 6.17367464993988e-05, "loss": 0.9483, "step": 4255 }, { "epoch": 1.927536231884058, "grad_norm": 0.1333317407523236, "learning_rate": 6.172612847622982e-05, "loss": 0.9366, "step": 4256 }, { "epoch": 1.9279891304347827, "grad_norm": 0.12377379511032134, "learning_rate": 6.171550828106534e-05, "loss": 0.9398, "step": 4257 }, { "epoch": 1.9284420289855073, "grad_norm": 0.12720303785030768, "learning_rate": 6.170488591496707e-05, "loss": 0.9452, "step": 4258 }, { "epoch": 1.928894927536232, "grad_norm": 0.13730456035195354, "learning_rate": 6.169426137899697e-05, "loss": 0.9557, "step": 4259 }, { "epoch": 1.9293478260869565, "grad_norm": 0.1093250284178659, "learning_rate": 6.168363467421719e-05, "loss": 0.9288, "step": 4260 }, { "epoch": 1.9298007246376812, "grad_norm": 0.10343016475191903, "learning_rate": 6.167300580169009e-05, "loss": 0.9382, "step": 4261 }, { "epoch": 1.9302536231884058, "grad_norm": 0.10144678593983521, "learning_rate": 6.166237476247826e-05, "loss": 0.9473, "step": 4262 }, { "epoch": 1.9307065217391304, "grad_norm": 0.11475076084570328, "learning_rate": 6.16517415576445e-05, "loss": 0.9454, "step": 4263 }, { "epoch": 1.931159420289855, "grad_norm": 0.1253989049635423, "learning_rate": 6.164110618825181e-05, "loss": 0.9425, "step": 4264 }, { "epoch": 1.9316123188405796, "grad_norm": 0.10839719471837414, "learning_rate": 6.163046865536345e-05, "loss": 0.9609, "step": 4265 }, { "epoch": 1.9320652173913042, "grad_norm": 0.10486167193327059, "learning_rate": 6.161982896004286e-05, "loss": 0.9533, "step": 4266 }, { "epoch": 1.9325181159420288, "grad_norm": 0.09687139501658097, "learning_rate": 6.16091871033537e-05, "loss": 0.9427, "step": 4267 }, { "epoch": 1.9329710144927537, "grad_norm": 0.07793737088826577, "learning_rate": 6.159854308635988e-05, "loss": 0.937, "step": 4268 }, { "epoch": 1.9334239130434783, "grad_norm": 0.08483531435121217, "learning_rate": 6.158789691012548e-05, "loss": 0.9316, "step": 4269 }, { "epoch": 1.9338768115942029, "grad_norm": 0.10414093808595984, "learning_rate": 6.15772485757148e-05, "loss": 0.9354, "step": 4270 }, { "epoch": 1.9343297101449275, "grad_norm": 0.08237520512293547, "learning_rate": 6.156659808419241e-05, "loss": 0.94, "step": 4271 }, { "epoch": 1.9347826086956523, "grad_norm": 0.10341928478426539, "learning_rate": 6.155594543662304e-05, "loss": 0.9373, "step": 4272 }, { "epoch": 1.935235507246377, "grad_norm": 0.09408834050496859, "learning_rate": 6.154529063407161e-05, "loss": 0.9344, "step": 4273 }, { "epoch": 1.9356884057971016, "grad_norm": 0.08247481297157601, "learning_rate": 6.153463367760336e-05, "loss": 0.9434, "step": 4274 }, { "epoch": 1.9361413043478262, "grad_norm": 0.10227801200249659, "learning_rate": 6.152397456828366e-05, "loss": 0.941, "step": 4275 }, { "epoch": 1.9365942028985508, "grad_norm": 0.11070066945152669, "learning_rate": 6.15133133071781e-05, "loss": 0.9227, "step": 4276 }, { "epoch": 1.9370471014492754, "grad_norm": 0.08799520279753548, "learning_rate": 6.150264989535252e-05, "loss": 0.9416, "step": 4277 }, { "epoch": 1.9375, "grad_norm": 0.09830024469336592, "learning_rate": 6.149198433387297e-05, "loss": 0.9214, "step": 4278 }, { "epoch": 1.9379528985507246, "grad_norm": 0.09309212818015218, "learning_rate": 6.148131662380567e-05, "loss": 0.9561, "step": 4279 }, { "epoch": 1.9384057971014492, "grad_norm": 0.10695641323686811, "learning_rate": 6.147064676621713e-05, "loss": 0.9416, "step": 4280 }, { "epoch": 1.9388586956521738, "grad_norm": 0.0968591028780841, "learning_rate": 6.145997476217398e-05, "loss": 0.962, "step": 4281 }, { "epoch": 1.9393115942028984, "grad_norm": 0.09229032684330288, "learning_rate": 6.144930061274317e-05, "loss": 0.9299, "step": 4282 }, { "epoch": 1.939764492753623, "grad_norm": 0.10295991659667342, "learning_rate": 6.143862431899177e-05, "loss": 0.9654, "step": 4283 }, { "epoch": 1.9402173913043477, "grad_norm": 0.14329452659110814, "learning_rate": 6.142794588198715e-05, "loss": 0.9676, "step": 4284 }, { "epoch": 1.9406702898550725, "grad_norm": 0.15067120611688384, "learning_rate": 6.14172653027968e-05, "loss": 0.9692, "step": 4285 }, { "epoch": 1.9411231884057971, "grad_norm": 0.13191903307611663, "learning_rate": 6.14065825824885e-05, "loss": 0.9657, "step": 4286 }, { "epoch": 1.9415760869565217, "grad_norm": 0.10773691594279258, "learning_rate": 6.139589772213024e-05, "loss": 0.9375, "step": 4287 }, { "epoch": 1.9420289855072463, "grad_norm": 0.09740208669853133, "learning_rate": 6.138521072279017e-05, "loss": 0.9616, "step": 4288 }, { "epoch": 1.9424818840579712, "grad_norm": 0.11975149410259922, "learning_rate": 6.137452158553668e-05, "loss": 0.9394, "step": 4289 }, { "epoch": 1.9429347826086958, "grad_norm": 0.17694671018738087, "learning_rate": 6.136383031143842e-05, "loss": 0.9548, "step": 4290 }, { "epoch": 1.9433876811594204, "grad_norm": 0.17522732504646318, "learning_rate": 6.135313690156418e-05, "loss": 0.9348, "step": 4291 }, { "epoch": 1.943840579710145, "grad_norm": 0.14382670301164982, "learning_rate": 6.1342441356983e-05, "loss": 0.9476, "step": 4292 }, { "epoch": 1.9442934782608696, "grad_norm": 0.1364438657825465, "learning_rate": 6.133174367876415e-05, "loss": 0.9422, "step": 4293 }, { "epoch": 1.9447463768115942, "grad_norm": 0.12278779346565491, "learning_rate": 6.132104386797707e-05, "loss": 0.9397, "step": 4294 }, { "epoch": 1.9451992753623188, "grad_norm": 0.1309478346364202, "learning_rate": 6.131034192569145e-05, "loss": 0.9487, "step": 4295 }, { "epoch": 1.9456521739130435, "grad_norm": 0.12354438447821532, "learning_rate": 6.12996378529772e-05, "loss": 0.9349, "step": 4296 }, { "epoch": 1.946105072463768, "grad_norm": 0.10164647368950337, "learning_rate": 6.128893165090437e-05, "loss": 0.9314, "step": 4297 }, { "epoch": 1.9465579710144927, "grad_norm": 0.1273803690168168, "learning_rate": 6.127822332054332e-05, "loss": 0.9555, "step": 4298 }, { "epoch": 1.9470108695652173, "grad_norm": 0.10577649281459353, "learning_rate": 6.126751286296456e-05, "loss": 0.9361, "step": 4299 }, { "epoch": 1.947463768115942, "grad_norm": 0.12115094418209156, "learning_rate": 6.125680027923884e-05, "loss": 0.9514, "step": 4300 }, { "epoch": 1.9479166666666665, "grad_norm": 0.11319752286573773, "learning_rate": 6.124608557043713e-05, "loss": 0.925, "step": 4301 }, { "epoch": 1.9483695652173914, "grad_norm": 0.15740305606697907, "learning_rate": 6.123536873763056e-05, "loss": 0.9556, "step": 4302 }, { "epoch": 1.948822463768116, "grad_norm": 0.1977249504897113, "learning_rate": 6.122464978189053e-05, "loss": 0.9403, "step": 4303 }, { "epoch": 1.9492753623188406, "grad_norm": 0.17258034390981583, "learning_rate": 6.121392870428864e-05, "loss": 0.9338, "step": 4304 }, { "epoch": 1.9497282608695652, "grad_norm": 0.11303754028606944, "learning_rate": 6.120320550589667e-05, "loss": 0.9359, "step": 4305 }, { "epoch": 1.9501811594202898, "grad_norm": 0.15970609064098926, "learning_rate": 6.119248018778665e-05, "loss": 0.9414, "step": 4306 }, { "epoch": 1.9506340579710146, "grad_norm": 0.17786851274892546, "learning_rate": 6.118175275103079e-05, "loss": 0.9648, "step": 4307 }, { "epoch": 1.9510869565217392, "grad_norm": 0.2112755207922738, "learning_rate": 6.117102319670156e-05, "loss": 0.9452, "step": 4308 }, { "epoch": 1.9515398550724639, "grad_norm": 0.21765973439220498, "learning_rate": 6.116029152587159e-05, "loss": 0.951, "step": 4309 }, { "epoch": 1.9519927536231885, "grad_norm": 0.1907815760313227, "learning_rate": 6.114955773961374e-05, "loss": 0.9468, "step": 4310 }, { "epoch": 1.952445652173913, "grad_norm": 0.1643566208739979, "learning_rate": 6.113882183900108e-05, "loss": 0.9168, "step": 4311 }, { "epoch": 1.9528985507246377, "grad_norm": 0.14235123595643084, "learning_rate": 6.112808382510692e-05, "loss": 0.9301, "step": 4312 }, { "epoch": 1.9533514492753623, "grad_norm": 0.10836811128762791, "learning_rate": 6.111734369900474e-05, "loss": 0.9507, "step": 4313 }, { "epoch": 1.953804347826087, "grad_norm": 0.11892057065986879, "learning_rate": 6.110660146176823e-05, "loss": 0.9322, "step": 4314 }, { "epoch": 1.9542572463768115, "grad_norm": 0.11573550027907013, "learning_rate": 6.109585711447134e-05, "loss": 0.9519, "step": 4315 }, { "epoch": 1.9547101449275361, "grad_norm": 0.13774490484377738, "learning_rate": 6.108511065818818e-05, "loss": 0.9174, "step": 4316 }, { "epoch": 1.9551630434782608, "grad_norm": 0.13019459616241585, "learning_rate": 6.107436209399309e-05, "loss": 0.9355, "step": 4317 }, { "epoch": 1.9556159420289854, "grad_norm": 0.10904770967055201, "learning_rate": 6.106361142296064e-05, "loss": 0.9341, "step": 4318 }, { "epoch": 1.9560688405797102, "grad_norm": 0.11487945395969278, "learning_rate": 6.105285864616556e-05, "loss": 0.9548, "step": 4319 }, { "epoch": 1.9565217391304348, "grad_norm": 0.1253822417152419, "learning_rate": 6.104210376468284e-05, "loss": 0.9437, "step": 4320 }, { "epoch": 1.9569746376811594, "grad_norm": 0.11322952573909845, "learning_rate": 6.1031346779587677e-05, "loss": 0.9336, "step": 4321 }, { "epoch": 1.957427536231884, "grad_norm": 0.12400446160678642, "learning_rate": 6.1020587691955446e-05, "loss": 0.9468, "step": 4322 }, { "epoch": 1.9578804347826086, "grad_norm": 0.11851033828263689, "learning_rate": 6.100982650286176e-05, "loss": 0.9473, "step": 4323 }, { "epoch": 1.9583333333333335, "grad_norm": 0.11404816086807211, "learning_rate": 6.099906321338241e-05, "loss": 0.9391, "step": 4324 }, { "epoch": 1.958786231884058, "grad_norm": 0.11439885868423183, "learning_rate": 6.0988297824593444e-05, "loss": 0.9307, "step": 4325 }, { "epoch": 1.9592391304347827, "grad_norm": 0.14508562380297121, "learning_rate": 6.097753033757109e-05, "loss": 0.9561, "step": 4326 }, { "epoch": 1.9596920289855073, "grad_norm": 0.13206963456808873, "learning_rate": 6.096676075339178e-05, "loss": 0.9554, "step": 4327 }, { "epoch": 1.960144927536232, "grad_norm": 0.10247494770747796, "learning_rate": 6.095598907313217e-05, "loss": 0.9623, "step": 4328 }, { "epoch": 1.9605978260869565, "grad_norm": 0.12990069410929636, "learning_rate": 6.094521529786912e-05, "loss": 0.9335, "step": 4329 }, { "epoch": 1.9610507246376812, "grad_norm": 0.11805483333756323, "learning_rate": 6.0934439428679726e-05, "loss": 0.9415, "step": 4330 }, { "epoch": 1.9615036231884058, "grad_norm": 0.11730620530020128, "learning_rate": 6.092366146664123e-05, "loss": 0.9423, "step": 4331 }, { "epoch": 1.9619565217391304, "grad_norm": 0.11145729302556368, "learning_rate": 6.0912881412831145e-05, "loss": 0.9568, "step": 4332 }, { "epoch": 1.962409420289855, "grad_norm": 0.12522178778933854, "learning_rate": 6.0902099268327174e-05, "loss": 0.9569, "step": 4333 }, { "epoch": 1.9628623188405796, "grad_norm": 0.10464682302723069, "learning_rate": 6.089131503420721e-05, "loss": 0.9592, "step": 4334 }, { "epoch": 1.9633152173913042, "grad_norm": 0.09859065240105219, "learning_rate": 6.088052871154938e-05, "loss": 0.9453, "step": 4335 }, { "epoch": 1.9637681159420288, "grad_norm": 0.11781149151643494, "learning_rate": 6.0869740301432e-05, "loss": 0.9439, "step": 4336 }, { "epoch": 1.9642210144927537, "grad_norm": 0.15562739326494382, "learning_rate": 6.0858949804933616e-05, "loss": 0.9364, "step": 4337 }, { "epoch": 1.9646739130434783, "grad_norm": 0.15813606442470535, "learning_rate": 6.084815722313297e-05, "loss": 0.9635, "step": 4338 }, { "epoch": 1.9651268115942029, "grad_norm": 0.14395384006264564, "learning_rate": 6.0837362557109e-05, "loss": 0.9594, "step": 4339 }, { "epoch": 1.9655797101449275, "grad_norm": 0.14619235827585494, "learning_rate": 6.082656580794089e-05, "loss": 0.9504, "step": 4340 }, { "epoch": 1.9660326086956523, "grad_norm": 0.1297614071288853, "learning_rate": 6.081576697670798e-05, "loss": 0.9488, "step": 4341 }, { "epoch": 1.966485507246377, "grad_norm": 0.1115444175933301, "learning_rate": 6.080496606448986e-05, "loss": 0.939, "step": 4342 }, { "epoch": 1.9669384057971016, "grad_norm": 0.10677764185302316, "learning_rate": 6.079416307236632e-05, "loss": 0.9325, "step": 4343 }, { "epoch": 1.9673913043478262, "grad_norm": 0.11579769160795475, "learning_rate": 6.078335800141735e-05, "loss": 0.9578, "step": 4344 }, { "epoch": 1.9678442028985508, "grad_norm": 0.1341039047674676, "learning_rate": 6.0772550852723136e-05, "loss": 0.9261, "step": 4345 }, { "epoch": 1.9682971014492754, "grad_norm": 0.14745429136408056, "learning_rate": 6.0761741627364104e-05, "loss": 0.9445, "step": 4346 }, { "epoch": 1.96875, "grad_norm": 0.10243836113288515, "learning_rate": 6.075093032642087e-05, "loss": 0.9513, "step": 4347 }, { "epoch": 1.9692028985507246, "grad_norm": 0.11001001361697477, "learning_rate": 6.074011695097423e-05, "loss": 0.9701, "step": 4348 }, { "epoch": 1.9696557971014492, "grad_norm": 0.13435164187378723, "learning_rate": 6.072930150210525e-05, "loss": 0.951, "step": 4349 }, { "epoch": 1.9701086956521738, "grad_norm": 0.13769976812216655, "learning_rate": 6.071848398089513e-05, "loss": 0.9083, "step": 4350 }, { "epoch": 1.9705615942028984, "grad_norm": 0.13661672976003894, "learning_rate": 6.070766438842535e-05, "loss": 0.9608, "step": 4351 }, { "epoch": 1.971014492753623, "grad_norm": 0.1547473677222076, "learning_rate": 6.069684272577755e-05, "loss": 0.9259, "step": 4352 }, { "epoch": 1.9714673913043477, "grad_norm": 0.1370589244119211, "learning_rate": 6.0686018994033574e-05, "loss": 0.9617, "step": 4353 }, { "epoch": 1.9719202898550725, "grad_norm": 0.11669729422355797, "learning_rate": 6.06751931942755e-05, "loss": 0.96, "step": 4354 }, { "epoch": 1.9723731884057971, "grad_norm": 0.12489727039976521, "learning_rate": 6.0664365327585614e-05, "loss": 0.957, "step": 4355 }, { "epoch": 1.9728260869565217, "grad_norm": 0.11546233969722898, "learning_rate": 6.065353539504637e-05, "loss": 0.9603, "step": 4356 }, { "epoch": 1.9732789855072463, "grad_norm": 0.1223436010087934, "learning_rate": 6.0642703397740464e-05, "loss": 0.9337, "step": 4357 }, { "epoch": 1.9737318840579712, "grad_norm": 0.1151004800477813, "learning_rate": 6.063186933675079e-05, "loss": 0.9592, "step": 4358 }, { "epoch": 1.9741847826086958, "grad_norm": 0.10974815614985697, "learning_rate": 6.062103321316045e-05, "loss": 0.9501, "step": 4359 }, { "epoch": 1.9746376811594204, "grad_norm": 0.12371873537898216, "learning_rate": 6.061019502805273e-05, "loss": 0.9396, "step": 4360 }, { "epoch": 1.975090579710145, "grad_norm": 0.12831921768054563, "learning_rate": 6.059935478251117e-05, "loss": 0.9358, "step": 4361 }, { "epoch": 1.9755434782608696, "grad_norm": 0.1269051711471379, "learning_rate": 6.0588512477619453e-05, "loss": 0.9513, "step": 4362 }, { "epoch": 1.9759963768115942, "grad_norm": 0.10135078206011564, "learning_rate": 6.0577668114461534e-05, "loss": 0.9497, "step": 4363 }, { "epoch": 1.9764492753623188, "grad_norm": 0.09583755249799344, "learning_rate": 6.0566821694121506e-05, "loss": 0.9478, "step": 4364 }, { "epoch": 1.9769021739130435, "grad_norm": 0.12886998111287307, "learning_rate": 6.055597321768375e-05, "loss": 0.9293, "step": 4365 }, { "epoch": 1.977355072463768, "grad_norm": 0.11683437499033848, "learning_rate": 6.054512268623275e-05, "loss": 0.9443, "step": 4366 }, { "epoch": 1.9778079710144927, "grad_norm": 0.12115207194285516, "learning_rate": 6.0534270100853284e-05, "loss": 0.9401, "step": 4367 }, { "epoch": 1.9782608695652173, "grad_norm": 0.12641697919750483, "learning_rate": 6.052341546263032e-05, "loss": 0.9505, "step": 4368 }, { "epoch": 1.978713768115942, "grad_norm": 0.10780527624254367, "learning_rate": 6.051255877264897e-05, "loss": 0.9226, "step": 4369 }, { "epoch": 1.9791666666666665, "grad_norm": 0.09505163908508166, "learning_rate": 6.0501700031994613e-05, "loss": 0.9438, "step": 4370 }, { "epoch": 1.9796195652173914, "grad_norm": 0.10872878076150452, "learning_rate": 6.049083924175282e-05, "loss": 0.9314, "step": 4371 }, { "epoch": 1.980072463768116, "grad_norm": 0.13110398647835603, "learning_rate": 6.047997640300936e-05, "loss": 0.9525, "step": 4372 }, { "epoch": 1.9805253623188406, "grad_norm": 0.1638253606798469, "learning_rate": 6.0469111516850204e-05, "loss": 0.9482, "step": 4373 }, { "epoch": 1.9809782608695652, "grad_norm": 0.1750810489340822, "learning_rate": 6.045824458436153e-05, "loss": 0.9342, "step": 4374 }, { "epoch": 1.9814311594202898, "grad_norm": 0.1748762295503762, "learning_rate": 6.044737560662972e-05, "loss": 0.9356, "step": 4375 }, { "epoch": 1.9818840579710146, "grad_norm": 0.16622923951879867, "learning_rate": 6.043650458474139e-05, "loss": 0.9471, "step": 4376 }, { "epoch": 1.9823369565217392, "grad_norm": 0.12298239580325475, "learning_rate": 6.0425631519783304e-05, "loss": 0.9658, "step": 4377 }, { "epoch": 1.9827898550724639, "grad_norm": 0.08224685725786833, "learning_rate": 6.041475641284246e-05, "loss": 0.9473, "step": 4378 }, { "epoch": 1.9832427536231885, "grad_norm": 0.11214900684207207, "learning_rate": 6.0403879265006064e-05, "loss": 0.9695, "step": 4379 }, { "epoch": 1.983695652173913, "grad_norm": 0.17645789988190558, "learning_rate": 6.039300007736154e-05, "loss": 0.9231, "step": 4380 }, { "epoch": 1.9841485507246377, "grad_norm": 0.19860459357159466, "learning_rate": 6.038211885099648e-05, "loss": 0.9064, "step": 4381 }, { "epoch": 1.9846014492753623, "grad_norm": 0.17034976556587178, "learning_rate": 6.037123558699868e-05, "loss": 0.9336, "step": 4382 }, { "epoch": 1.985054347826087, "grad_norm": 0.12126388031775791, "learning_rate": 6.03603502864562e-05, "loss": 0.9276, "step": 4383 }, { "epoch": 1.9855072463768115, "grad_norm": 0.08909131867852355, "learning_rate": 6.0349462950457234e-05, "loss": 0.9443, "step": 4384 }, { "epoch": 1.9859601449275361, "grad_norm": 0.11134154448723237, "learning_rate": 6.0338573580090214e-05, "loss": 0.9573, "step": 4385 }, { "epoch": 1.9864130434782608, "grad_norm": 0.14132033473769287, "learning_rate": 6.032768217644375e-05, "loss": 0.9528, "step": 4386 }, { "epoch": 1.9868659420289854, "grad_norm": 0.13991635535676616, "learning_rate": 6.031678874060669e-05, "loss": 0.9205, "step": 4387 }, { "epoch": 1.9873188405797102, "grad_norm": 0.14161975472523458, "learning_rate": 6.030589327366807e-05, "loss": 0.9221, "step": 4388 }, { "epoch": 1.9877717391304348, "grad_norm": 0.1331565918258822, "learning_rate": 6.029499577671713e-05, "loss": 0.9492, "step": 4389 }, { "epoch": 1.9882246376811594, "grad_norm": 0.11678005369131858, "learning_rate": 6.0284096250843295e-05, "loss": 0.9584, "step": 4390 }, { "epoch": 1.988677536231884, "grad_norm": 0.12875967101492042, "learning_rate": 6.027319469713621e-05, "loss": 0.9288, "step": 4391 }, { "epoch": 1.9891304347826086, "grad_norm": 0.12974696818946047, "learning_rate": 6.026229111668572e-05, "loss": 0.9589, "step": 4392 }, { "epoch": 1.9895833333333335, "grad_norm": 0.11572387392638347, "learning_rate": 6.02513855105819e-05, "loss": 0.9589, "step": 4393 }, { "epoch": 1.990036231884058, "grad_norm": 0.09941873906320954, "learning_rate": 6.024047787991496e-05, "loss": 0.9282, "step": 4394 }, { "epoch": 1.9904891304347827, "grad_norm": 0.0914283262603787, "learning_rate": 6.022956822577538e-05, "loss": 0.9279, "step": 4395 }, { "epoch": 1.9909420289855073, "grad_norm": 0.14009878581035445, "learning_rate": 6.021865654925381e-05, "loss": 0.9605, "step": 4396 }, { "epoch": 1.991394927536232, "grad_norm": 0.1477798773605948, "learning_rate": 6.020774285144111e-05, "loss": 0.948, "step": 4397 }, { "epoch": 1.9918478260869565, "grad_norm": 0.13959034591749983, "learning_rate": 6.019682713342833e-05, "loss": 0.9765, "step": 4398 }, { "epoch": 1.9923007246376812, "grad_norm": 0.11614172881835273, "learning_rate": 6.0185909396306727e-05, "loss": 0.9349, "step": 4399 }, { "epoch": 1.9927536231884058, "grad_norm": 0.10517241803611133, "learning_rate": 6.017498964116778e-05, "loss": 0.9508, "step": 4400 }, { "epoch": 1.9932065217391304, "grad_norm": 0.10800651354630997, "learning_rate": 6.016406786910315e-05, "loss": 0.9707, "step": 4401 }, { "epoch": 1.993659420289855, "grad_norm": 0.08210135797071089, "learning_rate": 6.0153144081204715e-05, "loss": 0.9635, "step": 4402 }, { "epoch": 1.9941123188405796, "grad_norm": 0.11868097729474561, "learning_rate": 6.0142218278564516e-05, "loss": 0.9346, "step": 4403 }, { "epoch": 1.9945652173913042, "grad_norm": 0.11356855923162981, "learning_rate": 6.0131290462274835e-05, "loss": 0.9533, "step": 4404 }, { "epoch": 1.9950181159420288, "grad_norm": 0.10843373126426634, "learning_rate": 6.012036063342815e-05, "loss": 0.9685, "step": 4405 }, { "epoch": 1.9954710144927537, "grad_norm": 0.1006117160640939, "learning_rate": 6.010942879311714e-05, "loss": 0.9508, "step": 4406 }, { "epoch": 1.9959239130434783, "grad_norm": 0.12569510314560822, "learning_rate": 6.009849494243465e-05, "loss": 0.9436, "step": 4407 }, { "epoch": 1.9963768115942029, "grad_norm": 0.10771055867015195, "learning_rate": 6.008755908247378e-05, "loss": 0.9599, "step": 4408 }, { "epoch": 1.9968297101449275, "grad_norm": 0.08941865584323024, "learning_rate": 6.00766212143278e-05, "loss": 0.9477, "step": 4409 }, { "epoch": 1.9972826086956523, "grad_norm": 0.11754270356913349, "learning_rate": 6.006568133909018e-05, "loss": 0.9432, "step": 4410 }, { "epoch": 1.997735507246377, "grad_norm": 0.0937843472059306, "learning_rate": 6.00547394578546e-05, "loss": 0.9646, "step": 4411 }, { "epoch": 1.9981884057971016, "grad_norm": 0.12656314864879153, "learning_rate": 6.004379557171493e-05, "loss": 0.944, "step": 4412 }, { "epoch": 1.9986413043478262, "grad_norm": 0.15771770296986792, "learning_rate": 6.0032849681765265e-05, "loss": 0.9464, "step": 4413 }, { "epoch": 1.9990942028985508, "grad_norm": 0.16508827940819987, "learning_rate": 6.002190178909986e-05, "loss": 0.9398, "step": 4414 }, { "epoch": 1.9995471014492754, "grad_norm": 0.15666649975344046, "learning_rate": 6.0010951894813227e-05, "loss": 0.9574, "step": 4415 }, { "epoch": 2.0, "grad_norm": 0.1116591349805633, "learning_rate": 6.000000000000001e-05, "loss": 0.9231, "step": 4416 }, { "epoch": 2.0004528985507246, "grad_norm": 0.12266509730539149, "learning_rate": 5.99890461057551e-05, "loss": 0.9397, "step": 4417 }, { "epoch": 2.0009057971014492, "grad_norm": 0.12108615651780603, "learning_rate": 5.9978090213173575e-05, "loss": 0.9348, "step": 4418 }, { "epoch": 2.001358695652174, "grad_norm": 0.10258995463984416, "learning_rate": 5.996713232335071e-05, "loss": 0.935, "step": 4419 }, { "epoch": 2.0018115942028984, "grad_norm": 0.10049192804080329, "learning_rate": 5.9956172437381993e-05, "loss": 0.9348, "step": 4420 }, { "epoch": 2.002264492753623, "grad_norm": 0.08949421629531683, "learning_rate": 5.994521055636309e-05, "loss": 0.9119, "step": 4421 }, { "epoch": 2.0027173913043477, "grad_norm": 0.09265959497490532, "learning_rate": 5.993424668138988e-05, "loss": 0.9512, "step": 4422 }, { "epoch": 2.0031702898550723, "grad_norm": 0.10689492099548947, "learning_rate": 5.992328081355846e-05, "loss": 0.964, "step": 4423 }, { "epoch": 2.003623188405797, "grad_norm": 0.12668812206917193, "learning_rate": 5.991231295396507e-05, "loss": 0.9169, "step": 4424 }, { "epoch": 2.004076086956522, "grad_norm": 0.12098269131035136, "learning_rate": 5.990134310370619e-05, "loss": 0.9384, "step": 4425 }, { "epoch": 2.0045289855072466, "grad_norm": 0.1219767016805419, "learning_rate": 5.989037126387852e-05, "loss": 0.9549, "step": 4426 }, { "epoch": 2.004981884057971, "grad_norm": 0.12436804316481061, "learning_rate": 5.9879397435578914e-05, "loss": 0.9384, "step": 4427 }, { "epoch": 2.005434782608696, "grad_norm": 0.11485630245905239, "learning_rate": 5.986842161990443e-05, "loss": 0.9273, "step": 4428 }, { "epoch": 2.0058876811594204, "grad_norm": 0.09691170977912711, "learning_rate": 5.9857443817952366e-05, "loss": 0.9449, "step": 4429 }, { "epoch": 2.006340579710145, "grad_norm": 0.10129406425737425, "learning_rate": 5.9846464030820166e-05, "loss": 0.9195, "step": 4430 }, { "epoch": 2.0067934782608696, "grad_norm": 0.1248750822053866, "learning_rate": 5.9835482259605526e-05, "loss": 0.9294, "step": 4431 }, { "epoch": 2.0072463768115942, "grad_norm": 0.12448356507626841, "learning_rate": 5.9824498505406275e-05, "loss": 0.9324, "step": 4432 }, { "epoch": 2.007699275362319, "grad_norm": 0.10317573032898837, "learning_rate": 5.9813512769320505e-05, "loss": 0.9247, "step": 4433 }, { "epoch": 2.0081521739130435, "grad_norm": 0.09441002518875373, "learning_rate": 5.980252505244646e-05, "loss": 0.9448, "step": 4434 }, { "epoch": 2.008605072463768, "grad_norm": 0.09938806298105216, "learning_rate": 5.979153535588261e-05, "loss": 0.9534, "step": 4435 }, { "epoch": 2.0090579710144927, "grad_norm": 0.10803022259445451, "learning_rate": 5.97805436807276e-05, "loss": 0.9375, "step": 4436 }, { "epoch": 2.0095108695652173, "grad_norm": 0.13633276467875116, "learning_rate": 5.97695500280803e-05, "loss": 0.9414, "step": 4437 }, { "epoch": 2.009963768115942, "grad_norm": 0.13692466495851197, "learning_rate": 5.975855439903975e-05, "loss": 0.955, "step": 4438 }, { "epoch": 2.0104166666666665, "grad_norm": 0.10703088838227483, "learning_rate": 5.9747556794705213e-05, "loss": 0.9392, "step": 4439 }, { "epoch": 2.010869565217391, "grad_norm": 0.11681788022170622, "learning_rate": 5.973655721617613e-05, "loss": 0.9445, "step": 4440 }, { "epoch": 2.0113224637681157, "grad_norm": 0.1124038972638146, "learning_rate": 5.9725555664552147e-05, "loss": 0.9559, "step": 4441 }, { "epoch": 2.011775362318841, "grad_norm": 0.08990073008617752, "learning_rate": 5.9714552140933104e-05, "loss": 0.9317, "step": 4442 }, { "epoch": 2.0122282608695654, "grad_norm": 0.08443197786236607, "learning_rate": 5.970354664641904e-05, "loss": 0.9455, "step": 4443 }, { "epoch": 2.01268115942029, "grad_norm": 0.10168311246195699, "learning_rate": 5.9692539182110205e-05, "loss": 0.9368, "step": 4444 }, { "epoch": 2.0131340579710146, "grad_norm": 0.10274951042876186, "learning_rate": 5.968152974910702e-05, "loss": 0.9571, "step": 4445 }, { "epoch": 2.0135869565217392, "grad_norm": 0.09379793577194968, "learning_rate": 5.9670518348510114e-05, "loss": 0.9555, "step": 4446 }, { "epoch": 2.014039855072464, "grad_norm": 0.11004125888468631, "learning_rate": 5.965950498142031e-05, "loss": 0.9263, "step": 4447 }, { "epoch": 2.0144927536231885, "grad_norm": 0.12367648176385374, "learning_rate": 5.964848964893866e-05, "loss": 0.9291, "step": 4448 }, { "epoch": 2.014945652173913, "grad_norm": 0.13910736250278902, "learning_rate": 5.9637472352166345e-05, "loss": 0.9303, "step": 4449 }, { "epoch": 2.0153985507246377, "grad_norm": 0.10520329883673606, "learning_rate": 5.962645309220481e-05, "loss": 0.9374, "step": 4450 }, { "epoch": 2.0158514492753623, "grad_norm": 0.10607883090131989, "learning_rate": 5.961543187015565e-05, "loss": 0.9133, "step": 4451 }, { "epoch": 2.016304347826087, "grad_norm": 0.09240964443338823, "learning_rate": 5.96044086871207e-05, "loss": 0.9414, "step": 4452 }, { "epoch": 2.0167572463768115, "grad_norm": 0.11542874967236384, "learning_rate": 5.959338354420194e-05, "loss": 0.9411, "step": 4453 }, { "epoch": 2.017210144927536, "grad_norm": 0.12055456900055904, "learning_rate": 5.958235644250157e-05, "loss": 0.9379, "step": 4454 }, { "epoch": 2.0176630434782608, "grad_norm": 0.1046119653392342, "learning_rate": 5.9571327383122e-05, "loss": 0.9605, "step": 4455 }, { "epoch": 2.0181159420289854, "grad_norm": 0.10537143073418796, "learning_rate": 5.956029636716581e-05, "loss": 0.9385, "step": 4456 }, { "epoch": 2.01856884057971, "grad_norm": 0.09407710413116005, "learning_rate": 5.9549263395735814e-05, "loss": 0.9528, "step": 4457 }, { "epoch": 2.0190217391304346, "grad_norm": 0.0941015871020405, "learning_rate": 5.953822846993497e-05, "loss": 0.9446, "step": 4458 }, { "epoch": 2.019474637681159, "grad_norm": 0.10552571052998952, "learning_rate": 5.9527191590866464e-05, "loss": 0.9456, "step": 4459 }, { "epoch": 2.0199275362318843, "grad_norm": 0.09985022508623649, "learning_rate": 5.951615275963367e-05, "loss": 0.9386, "step": 4460 }, { "epoch": 2.020380434782609, "grad_norm": 0.09340298572132698, "learning_rate": 5.950511197734017e-05, "loss": 0.9163, "step": 4461 }, { "epoch": 2.0208333333333335, "grad_norm": 0.12327789487618332, "learning_rate": 5.94940692450897e-05, "loss": 0.9386, "step": 4462 }, { "epoch": 2.021286231884058, "grad_norm": 0.11798478404129856, "learning_rate": 5.948302456398624e-05, "loss": 0.943, "step": 4463 }, { "epoch": 2.0217391304347827, "grad_norm": 0.11002922239555599, "learning_rate": 5.947197793513394e-05, "loss": 0.9329, "step": 4464 }, { "epoch": 2.0221920289855073, "grad_norm": 0.12538357772621941, "learning_rate": 5.946092935963716e-05, "loss": 0.9457, "step": 4465 }, { "epoch": 2.022644927536232, "grad_norm": 0.10946844906193483, "learning_rate": 5.944987883860043e-05, "loss": 0.9561, "step": 4466 }, { "epoch": 2.0230978260869565, "grad_norm": 0.10138023621406626, "learning_rate": 5.9438826373128485e-05, "loss": 0.9348, "step": 4467 }, { "epoch": 2.023550724637681, "grad_norm": 0.10258314097507026, "learning_rate": 5.942777196432627e-05, "loss": 0.9296, "step": 4468 }, { "epoch": 2.0240036231884058, "grad_norm": 0.08923956800415622, "learning_rate": 5.941671561329891e-05, "loss": 0.94, "step": 4469 }, { "epoch": 2.0244565217391304, "grad_norm": 0.12164511418725556, "learning_rate": 5.940565732115172e-05, "loss": 0.9543, "step": 4470 }, { "epoch": 2.024909420289855, "grad_norm": 0.12532540030969333, "learning_rate": 5.939459708899021e-05, "loss": 0.944, "step": 4471 }, { "epoch": 2.0253623188405796, "grad_norm": 0.11496173374009978, "learning_rate": 5.9383534917920096e-05, "loss": 0.9492, "step": 4472 }, { "epoch": 2.025815217391304, "grad_norm": 0.13291401144810197, "learning_rate": 5.9372470809047286e-05, "loss": 0.9159, "step": 4473 }, { "epoch": 2.026268115942029, "grad_norm": 0.10714296620084174, "learning_rate": 5.9361404763477865e-05, "loss": 0.9458, "step": 4474 }, { "epoch": 2.0267210144927534, "grad_norm": 0.09711709486130442, "learning_rate": 5.935033678231813e-05, "loss": 0.9204, "step": 4475 }, { "epoch": 2.027173913043478, "grad_norm": 0.1433800793651384, "learning_rate": 5.933926686667456e-05, "loss": 0.9297, "step": 4476 }, { "epoch": 2.027626811594203, "grad_norm": 0.17158881828901307, "learning_rate": 5.932819501765384e-05, "loss": 0.9307, "step": 4477 }, { "epoch": 2.0280797101449277, "grad_norm": 0.15251110159679251, "learning_rate": 5.931712123636285e-05, "loss": 0.9452, "step": 4478 }, { "epoch": 2.0285326086956523, "grad_norm": 0.11430802786400343, "learning_rate": 5.930604552390862e-05, "loss": 0.9337, "step": 4479 }, { "epoch": 2.028985507246377, "grad_norm": 0.09232841576912734, "learning_rate": 5.929496788139843e-05, "loss": 0.9483, "step": 4480 }, { "epoch": 2.0294384057971016, "grad_norm": 0.11857829311210948, "learning_rate": 5.928388830993973e-05, "loss": 0.9576, "step": 4481 }, { "epoch": 2.029891304347826, "grad_norm": 0.21678252797526568, "learning_rate": 5.927280681064016e-05, "loss": 0.961, "step": 4482 }, { "epoch": 2.0303442028985508, "grad_norm": 0.13632374296918856, "learning_rate": 5.9261723384607546e-05, "loss": 0.9289, "step": 4483 }, { "epoch": 2.0307971014492754, "grad_norm": 0.13988149861076313, "learning_rate": 5.9250638032949916e-05, "loss": 0.9354, "step": 4484 }, { "epoch": 2.03125, "grad_norm": 0.12406964183817244, "learning_rate": 5.923955075677551e-05, "loss": 0.9313, "step": 4485 }, { "epoch": 2.0317028985507246, "grad_norm": 0.12338936427206333, "learning_rate": 5.9228461557192724e-05, "loss": 0.9503, "step": 4486 }, { "epoch": 2.0321557971014492, "grad_norm": 0.10582443093861214, "learning_rate": 5.921737043531016e-05, "loss": 0.9351, "step": 4487 }, { "epoch": 2.032608695652174, "grad_norm": 0.12174074685815649, "learning_rate": 5.920627739223663e-05, "loss": 0.9366, "step": 4488 }, { "epoch": 2.0330615942028984, "grad_norm": 0.11632403307652577, "learning_rate": 5.9195182429081103e-05, "loss": 0.9124, "step": 4489 }, { "epoch": 2.033514492753623, "grad_norm": 0.11950089040723404, "learning_rate": 5.918408554695278e-05, "loss": 0.957, "step": 4490 }, { "epoch": 2.0339673913043477, "grad_norm": 0.1206823060860949, "learning_rate": 5.9172986746961025e-05, "loss": 0.9286, "step": 4491 }, { "epoch": 2.0344202898550723, "grad_norm": 0.09494231492481395, "learning_rate": 5.91618860302154e-05, "loss": 0.9352, "step": 4492 }, { "epoch": 2.034873188405797, "grad_norm": 0.10706463354994118, "learning_rate": 5.915078339782566e-05, "loss": 0.9477, "step": 4493 }, { "epoch": 2.035326086956522, "grad_norm": 0.1316453636757553, "learning_rate": 5.913967885090177e-05, "loss": 0.9643, "step": 4494 }, { "epoch": 2.0357789855072466, "grad_norm": 0.14366082894340026, "learning_rate": 5.9128572390553846e-05, "loss": 0.9346, "step": 4495 }, { "epoch": 2.036231884057971, "grad_norm": 0.1497055702827042, "learning_rate": 5.911746401789223e-05, "loss": 0.9225, "step": 4496 }, { "epoch": 2.036684782608696, "grad_norm": 0.1434107642309668, "learning_rate": 5.910635373402743e-05, "loss": 0.9623, "step": 4497 }, { "epoch": 2.0371376811594204, "grad_norm": 0.13080730748568137, "learning_rate": 5.909524154007019e-05, "loss": 0.9359, "step": 4498 }, { "epoch": 2.037590579710145, "grad_norm": 0.12899756241958113, "learning_rate": 5.908412743713139e-05, "loss": 0.9401, "step": 4499 }, { "epoch": 2.0380434782608696, "grad_norm": 0.10401389295520402, "learning_rate": 5.9073011426322116e-05, "loss": 0.9311, "step": 4500 }, { "epoch": 2.0384963768115942, "grad_norm": 0.09819171651365273, "learning_rate": 5.9061893508753676e-05, "loss": 0.9471, "step": 4501 }, { "epoch": 2.038949275362319, "grad_norm": 0.11948996958739215, "learning_rate": 5.905077368553754e-05, "loss": 0.9348, "step": 4502 }, { "epoch": 2.0394021739130435, "grad_norm": 0.12259817928755458, "learning_rate": 5.903965195778537e-05, "loss": 0.9545, "step": 4503 }, { "epoch": 2.039855072463768, "grad_norm": 0.0981572786601327, "learning_rate": 5.902852832660901e-05, "loss": 0.9297, "step": 4504 }, { "epoch": 2.0403079710144927, "grad_norm": 0.10056810638928575, "learning_rate": 5.901740279312053e-05, "loss": 0.9434, "step": 4505 }, { "epoch": 2.0407608695652173, "grad_norm": 0.11002348828955168, "learning_rate": 5.9006275358432145e-05, "loss": 0.9367, "step": 4506 }, { "epoch": 2.041213768115942, "grad_norm": 0.12764064617196144, "learning_rate": 5.8995146023656315e-05, "loss": 0.9429, "step": 4507 }, { "epoch": 2.0416666666666665, "grad_norm": 0.12770407752373256, "learning_rate": 5.8984014789905625e-05, "loss": 0.9253, "step": 4508 }, { "epoch": 2.042119565217391, "grad_norm": 0.11821688377844275, "learning_rate": 5.897288165829289e-05, "loss": 0.9332, "step": 4509 }, { "epoch": 2.0425724637681157, "grad_norm": 0.13250546626868587, "learning_rate": 5.896174662993111e-05, "loss": 0.9509, "step": 4510 }, { "epoch": 2.043025362318841, "grad_norm": 0.12597213019308134, "learning_rate": 5.895060970593348e-05, "loss": 0.937, "step": 4511 }, { "epoch": 2.0434782608695654, "grad_norm": 0.11110165255560138, "learning_rate": 5.8939470887413364e-05, "loss": 0.9335, "step": 4512 }, { "epoch": 2.04393115942029, "grad_norm": 0.10611962330914905, "learning_rate": 5.892833017548433e-05, "loss": 0.9317, "step": 4513 }, { "epoch": 2.0443840579710146, "grad_norm": 0.10361226046988962, "learning_rate": 5.891718757126013e-05, "loss": 0.9521, "step": 4514 }, { "epoch": 2.0448369565217392, "grad_norm": 0.10733028930695884, "learning_rate": 5.890604307585473e-05, "loss": 0.9318, "step": 4515 }, { "epoch": 2.045289855072464, "grad_norm": 0.10408774810752534, "learning_rate": 5.889489669038223e-05, "loss": 0.9374, "step": 4516 }, { "epoch": 2.0457427536231885, "grad_norm": 0.1267203254216454, "learning_rate": 5.888374841595697e-05, "loss": 0.9348, "step": 4517 }, { "epoch": 2.046195652173913, "grad_norm": 0.11730464484654994, "learning_rate": 5.887259825369345e-05, "loss": 0.938, "step": 4518 }, { "epoch": 2.0466485507246377, "grad_norm": 0.12347821715436168, "learning_rate": 5.886144620470638e-05, "loss": 0.9322, "step": 4519 }, { "epoch": 2.0471014492753623, "grad_norm": 0.12972421187959912, "learning_rate": 5.885029227011066e-05, "loss": 0.9265, "step": 4520 }, { "epoch": 2.047554347826087, "grad_norm": 0.12524417047385306, "learning_rate": 5.8839136451021337e-05, "loss": 0.9281, "step": 4521 }, { "epoch": 2.0480072463768115, "grad_norm": 0.13715730584418864, "learning_rate": 5.8827978748553685e-05, "loss": 0.9464, "step": 4522 }, { "epoch": 2.048460144927536, "grad_norm": 0.1414830147506915, "learning_rate": 5.8816819163823166e-05, "loss": 0.9146, "step": 4523 }, { "epoch": 2.0489130434782608, "grad_norm": 0.10477487196094931, "learning_rate": 5.880565769794541e-05, "loss": 0.9218, "step": 4524 }, { "epoch": 2.0493659420289854, "grad_norm": 0.10870407968875526, "learning_rate": 5.879449435203627e-05, "loss": 0.938, "step": 4525 }, { "epoch": 2.04981884057971, "grad_norm": 0.10662393897167818, "learning_rate": 5.878332912721174e-05, "loss": 0.9511, "step": 4526 }, { "epoch": 2.0502717391304346, "grad_norm": 0.09042676771187184, "learning_rate": 5.877216202458803e-05, "loss": 0.9274, "step": 4527 }, { "epoch": 2.050724637681159, "grad_norm": 0.10191254125541077, "learning_rate": 5.876099304528153e-05, "loss": 0.9435, "step": 4528 }, { "epoch": 2.0511775362318843, "grad_norm": 0.109596309906628, "learning_rate": 5.874982219040884e-05, "loss": 0.9303, "step": 4529 }, { "epoch": 2.051630434782609, "grad_norm": 0.13256700899475785, "learning_rate": 5.87386494610867e-05, "loss": 0.9354, "step": 4530 }, { "epoch": 2.0520833333333335, "grad_norm": 0.1275006837458429, "learning_rate": 5.8727474858432085e-05, "loss": 0.94, "step": 4531 }, { "epoch": 2.052536231884058, "grad_norm": 0.13519799924643866, "learning_rate": 5.871629838356214e-05, "loss": 0.9274, "step": 4532 }, { "epoch": 2.0529891304347827, "grad_norm": 0.1393371649232877, "learning_rate": 5.870512003759418e-05, "loss": 0.9401, "step": 4533 }, { "epoch": 2.0534420289855073, "grad_norm": 0.13052555405997016, "learning_rate": 5.869393982164573e-05, "loss": 0.9441, "step": 4534 }, { "epoch": 2.053894927536232, "grad_norm": 0.10184839499544826, "learning_rate": 5.868275773683449e-05, "loss": 0.9239, "step": 4535 }, { "epoch": 2.0543478260869565, "grad_norm": 0.11367832508666217, "learning_rate": 5.867157378427835e-05, "loss": 0.9429, "step": 4536 }, { "epoch": 2.054800724637681, "grad_norm": 0.09638699658245187, "learning_rate": 5.866038796509541e-05, "loss": 0.9297, "step": 4537 }, { "epoch": 2.0552536231884058, "grad_norm": 0.09377911518009634, "learning_rate": 5.8649200280403895e-05, "loss": 0.9286, "step": 4538 }, { "epoch": 2.0557065217391304, "grad_norm": 0.10475573028104743, "learning_rate": 5.863801073132228e-05, "loss": 0.9396, "step": 4539 }, { "epoch": 2.056159420289855, "grad_norm": 0.10273083945618307, "learning_rate": 5.86268193189692e-05, "loss": 0.9461, "step": 4540 }, { "epoch": 2.0566123188405796, "grad_norm": 0.10477817439135541, "learning_rate": 5.861562604446349e-05, "loss": 0.933, "step": 4541 }, { "epoch": 2.057065217391304, "grad_norm": 0.08309024721086626, "learning_rate": 5.8604430908924126e-05, "loss": 0.9439, "step": 4542 }, { "epoch": 2.057518115942029, "grad_norm": 0.10384274946542296, "learning_rate": 5.8593233913470326e-05, "loss": 0.9163, "step": 4543 }, { "epoch": 2.0579710144927534, "grad_norm": 0.0922153077181473, "learning_rate": 5.858203505922147e-05, "loss": 0.9422, "step": 4544 }, { "epoch": 2.058423913043478, "grad_norm": 0.08373381856644237, "learning_rate": 5.857083434729711e-05, "loss": 0.9062, "step": 4545 }, { "epoch": 2.058876811594203, "grad_norm": 0.09953277354637236, "learning_rate": 5.8559631778817046e-05, "loss": 0.9103, "step": 4546 }, { "epoch": 2.0593297101449277, "grad_norm": 0.10427567598211279, "learning_rate": 5.854842735490116e-05, "loss": 0.9089, "step": 4547 }, { "epoch": 2.0597826086956523, "grad_norm": 0.10286920482816612, "learning_rate": 5.85372210766696e-05, "loss": 0.9412, "step": 4548 }, { "epoch": 2.060235507246377, "grad_norm": 0.11521807121454253, "learning_rate": 5.852601294524268e-05, "loss": 0.9245, "step": 4549 }, { "epoch": 2.0606884057971016, "grad_norm": 0.12161601601472398, "learning_rate": 5.851480296174089e-05, "loss": 0.9652, "step": 4550 }, { "epoch": 2.061141304347826, "grad_norm": 0.13433020851544447, "learning_rate": 5.850359112728489e-05, "loss": 0.9467, "step": 4551 }, { "epoch": 2.0615942028985508, "grad_norm": 0.11945686250260534, "learning_rate": 5.849237744299558e-05, "loss": 0.9439, "step": 4552 }, { "epoch": 2.0620471014492754, "grad_norm": 0.09993701812200904, "learning_rate": 5.848116190999399e-05, "loss": 0.9427, "step": 4553 }, { "epoch": 2.0625, "grad_norm": 0.09782238315850388, "learning_rate": 5.846994452940137e-05, "loss": 0.9447, "step": 4554 }, { "epoch": 2.0629528985507246, "grad_norm": 0.10499012750252652, "learning_rate": 5.845872530233911e-05, "loss": 0.9413, "step": 4555 }, { "epoch": 2.0634057971014492, "grad_norm": 0.07634601292189544, "learning_rate": 5.844750422992883e-05, "loss": 0.9313, "step": 4556 }, { "epoch": 2.063858695652174, "grad_norm": 0.09765296139110277, "learning_rate": 5.843628131329233e-05, "loss": 0.9373, "step": 4557 }, { "epoch": 2.0643115942028984, "grad_norm": 0.1035925447570422, "learning_rate": 5.842505655355158e-05, "loss": 0.9721, "step": 4558 }, { "epoch": 2.064764492753623, "grad_norm": 0.1029212413714837, "learning_rate": 5.841382995182871e-05, "loss": 0.9469, "step": 4559 }, { "epoch": 2.0652173913043477, "grad_norm": 0.09347714806678668, "learning_rate": 5.840260150924609e-05, "loss": 0.9368, "step": 4560 }, { "epoch": 2.0656702898550723, "grad_norm": 0.0853728932517277, "learning_rate": 5.839137122692624e-05, "loss": 0.9097, "step": 4561 }, { "epoch": 2.066123188405797, "grad_norm": 0.10074804886630767, "learning_rate": 5.838013910599186e-05, "loss": 0.936, "step": 4562 }, { "epoch": 2.066576086956522, "grad_norm": 0.0845706091537935, "learning_rate": 5.836890514756585e-05, "loss": 0.9512, "step": 4563 }, { "epoch": 2.0670289855072466, "grad_norm": 0.08415841977092155, "learning_rate": 5.835766935277128e-05, "loss": 0.9203, "step": 4564 }, { "epoch": 2.067481884057971, "grad_norm": 0.09476004683984812, "learning_rate": 5.834643172273141e-05, "loss": 0.945, "step": 4565 }, { "epoch": 2.067934782608696, "grad_norm": 0.10345604063508784, "learning_rate": 5.833519225856969e-05, "loss": 0.9464, "step": 4566 }, { "epoch": 2.0683876811594204, "grad_norm": 0.10667601750910823, "learning_rate": 5.8323950961409753e-05, "loss": 0.9571, "step": 4567 }, { "epoch": 2.068840579710145, "grad_norm": 0.09201108583533156, "learning_rate": 5.83127078323754e-05, "loss": 0.9256, "step": 4568 }, { "epoch": 2.0692934782608696, "grad_norm": 0.10534958584981957, "learning_rate": 5.830146287259062e-05, "loss": 0.9584, "step": 4569 }, { "epoch": 2.0697463768115942, "grad_norm": 0.10593321867587127, "learning_rate": 5.82902160831796e-05, "loss": 0.961, "step": 4570 }, { "epoch": 2.070199275362319, "grad_norm": 0.0900354483775419, "learning_rate": 5.8278967465266706e-05, "loss": 0.9347, "step": 4571 }, { "epoch": 2.0706521739130435, "grad_norm": 0.1331293574883339, "learning_rate": 5.826771701997645e-05, "loss": 0.955, "step": 4572 }, { "epoch": 2.071105072463768, "grad_norm": 0.1636627900828472, "learning_rate": 5.825646474843359e-05, "loss": 0.9394, "step": 4573 }, { "epoch": 2.0715579710144927, "grad_norm": 0.1759653602656374, "learning_rate": 5.824521065176301e-05, "loss": 0.9477, "step": 4574 }, { "epoch": 2.0720108695652173, "grad_norm": 0.15612393353075182, "learning_rate": 5.8233954731089816e-05, "loss": 0.931, "step": 4575 }, { "epoch": 2.072463768115942, "grad_norm": 0.11722151780362815, "learning_rate": 5.8222696987539264e-05, "loss": 0.963, "step": 4576 }, { "epoch": 2.0729166666666665, "grad_norm": 0.0853436171473364, "learning_rate": 5.821143742223682e-05, "loss": 0.9503, "step": 4577 }, { "epoch": 2.073369565217391, "grad_norm": 0.1324940882997077, "learning_rate": 5.820017603630812e-05, "loss": 0.9428, "step": 4578 }, { "epoch": 2.0738224637681157, "grad_norm": 0.13699818597203542, "learning_rate": 5.818891283087899e-05, "loss": 0.9518, "step": 4579 }, { "epoch": 2.074275362318841, "grad_norm": 0.16961086127870253, "learning_rate": 5.8177647807075413e-05, "loss": 0.9393, "step": 4580 }, { "epoch": 2.0747282608695654, "grad_norm": 0.20174786169209652, "learning_rate": 5.816638096602358e-05, "loss": 0.9304, "step": 4581 }, { "epoch": 2.07518115942029, "grad_norm": 0.20165457719745625, "learning_rate": 5.815511230884985e-05, "loss": 0.9381, "step": 4582 }, { "epoch": 2.0756340579710146, "grad_norm": 0.1840432561898969, "learning_rate": 5.814384183668079e-05, "loss": 0.9355, "step": 4583 }, { "epoch": 2.0760869565217392, "grad_norm": 0.1701087502221928, "learning_rate": 5.81325695506431e-05, "loss": 0.9565, "step": 4584 }, { "epoch": 2.076539855072464, "grad_norm": 0.1300619119081219, "learning_rate": 5.812129545186369e-05, "loss": 0.9459, "step": 4585 }, { "epoch": 2.0769927536231885, "grad_norm": 0.1231087188402432, "learning_rate": 5.8110019541469663e-05, "loss": 0.9618, "step": 4586 }, { "epoch": 2.077445652173913, "grad_norm": 0.15227051811842038, "learning_rate": 5.8098741820588286e-05, "loss": 0.9413, "step": 4587 }, { "epoch": 2.0778985507246377, "grad_norm": 0.18623025829160222, "learning_rate": 5.808746229034702e-05, "loss": 0.9347, "step": 4588 }, { "epoch": 2.0783514492753623, "grad_norm": 0.2147665544651519, "learning_rate": 5.807618095187347e-05, "loss": 0.9408, "step": 4589 }, { "epoch": 2.078804347826087, "grad_norm": 0.17729912178992116, "learning_rate": 5.806489780629547e-05, "loss": 0.9222, "step": 4590 }, { "epoch": 2.0792572463768115, "grad_norm": 0.16887319308854906, "learning_rate": 5.805361285474101e-05, "loss": 0.946, "step": 4591 }, { "epoch": 2.079710144927536, "grad_norm": 0.137387846634061, "learning_rate": 5.8042326098338285e-05, "loss": 0.9458, "step": 4592 }, { "epoch": 2.0801630434782608, "grad_norm": 0.17318412373861342, "learning_rate": 5.8031037538215606e-05, "loss": 0.9378, "step": 4593 }, { "epoch": 2.0806159420289854, "grad_norm": 0.18794132679453063, "learning_rate": 5.801974717550154e-05, "loss": 0.9438, "step": 4594 }, { "epoch": 2.08106884057971, "grad_norm": 0.1772727636436015, "learning_rate": 5.800845501132479e-05, "loss": 0.9365, "step": 4595 }, { "epoch": 2.0815217391304346, "grad_norm": 0.16836870985691826, "learning_rate": 5.799716104681426e-05, "loss": 0.9268, "step": 4596 }, { "epoch": 2.081974637681159, "grad_norm": 0.1559313006450975, "learning_rate": 5.798586528309901e-05, "loss": 0.9056, "step": 4597 }, { "epoch": 2.0824275362318843, "grad_norm": 0.13817374498894294, "learning_rate": 5.797456772130832e-05, "loss": 0.9143, "step": 4598 }, { "epoch": 2.082880434782609, "grad_norm": 0.127261584082018, "learning_rate": 5.7963268362571606e-05, "loss": 0.9276, "step": 4599 }, { "epoch": 2.0833333333333335, "grad_norm": 0.1435463229463979, "learning_rate": 5.79519672080185e-05, "loss": 0.95, "step": 4600 }, { "epoch": 2.083786231884058, "grad_norm": 0.14143457284263455, "learning_rate": 5.7940664258778764e-05, "loss": 0.9657, "step": 4601 }, { "epoch": 2.0842391304347827, "grad_norm": 0.11615025836823577, "learning_rate": 5.792935951598241e-05, "loss": 0.9529, "step": 4602 }, { "epoch": 2.0846920289855073, "grad_norm": 0.14917801935247946, "learning_rate": 5.7918052980759554e-05, "loss": 0.9431, "step": 4603 }, { "epoch": 2.085144927536232, "grad_norm": 0.13011450554419926, "learning_rate": 5.790674465424057e-05, "loss": 0.9425, "step": 4604 }, { "epoch": 2.0855978260869565, "grad_norm": 0.19818536463032427, "learning_rate": 5.7895434537555936e-05, "loss": 0.9486, "step": 4605 }, { "epoch": 2.086050724637681, "grad_norm": 0.10620882498800206, "learning_rate": 5.7884122631836364e-05, "loss": 0.9392, "step": 4606 }, { "epoch": 2.0865036231884058, "grad_norm": 0.11922938897896415, "learning_rate": 5.78728089382127e-05, "loss": 0.9352, "step": 4607 }, { "epoch": 2.0869565217391304, "grad_norm": 0.13033920966059484, "learning_rate": 5.7861493457816005e-05, "loss": 0.9204, "step": 4608 }, { "epoch": 2.087409420289855, "grad_norm": 0.10349276594763165, "learning_rate": 5.7850176191777514e-05, "loss": 0.935, "step": 4609 }, { "epoch": 2.0878623188405796, "grad_norm": 0.12972971104776632, "learning_rate": 5.783885714122862e-05, "loss": 0.941, "step": 4610 }, { "epoch": 2.088315217391304, "grad_norm": 0.10610173513554758, "learning_rate": 5.7827536307300904e-05, "loss": 0.9366, "step": 4611 }, { "epoch": 2.088768115942029, "grad_norm": 0.15538594812695442, "learning_rate": 5.781621369112613e-05, "loss": 0.9435, "step": 4612 }, { "epoch": 2.0892210144927534, "grad_norm": 0.19090280704139556, "learning_rate": 5.7804889293836254e-05, "loss": 0.9488, "step": 4613 }, { "epoch": 2.089673913043478, "grad_norm": 0.17041874726886846, "learning_rate": 5.7793563116563376e-05, "loss": 0.9294, "step": 4614 }, { "epoch": 2.090126811594203, "grad_norm": 0.14279214318943986, "learning_rate": 5.778223516043979e-05, "loss": 0.9347, "step": 4615 }, { "epoch": 2.0905797101449277, "grad_norm": 0.13580954387943112, "learning_rate": 5.7770905426597986e-05, "loss": 0.9527, "step": 4616 }, { "epoch": 2.0910326086956523, "grad_norm": 0.09460369973545248, "learning_rate": 5.775957391617061e-05, "loss": 0.9263, "step": 4617 }, { "epoch": 2.091485507246377, "grad_norm": 0.09831574728123794, "learning_rate": 5.774824063029048e-05, "loss": 0.9423, "step": 4618 }, { "epoch": 2.0919384057971016, "grad_norm": 0.16345714896343425, "learning_rate": 5.77369055700906e-05, "loss": 0.9389, "step": 4619 }, { "epoch": 2.092391304347826, "grad_norm": 0.17907951299612454, "learning_rate": 5.772556873670417e-05, "loss": 0.9397, "step": 4620 }, { "epoch": 2.0928442028985508, "grad_norm": 0.16640293602862777, "learning_rate": 5.771423013126455e-05, "loss": 0.9386, "step": 4621 }, { "epoch": 2.0932971014492754, "grad_norm": 0.12736324616876502, "learning_rate": 5.7702889754905276e-05, "loss": 0.9506, "step": 4622 }, { "epoch": 2.09375, "grad_norm": 0.10002772701705706, "learning_rate": 5.7691547608760055e-05, "loss": 0.958, "step": 4623 }, { "epoch": 2.0942028985507246, "grad_norm": 0.15096906305022814, "learning_rate": 5.768020369396278e-05, "loss": 0.9248, "step": 4624 }, { "epoch": 2.0946557971014492, "grad_norm": 0.1710821671560023, "learning_rate": 5.766885801164754e-05, "loss": 0.9389, "step": 4625 }, { "epoch": 2.095108695652174, "grad_norm": 0.14713604388989307, "learning_rate": 5.7657510562948555e-05, "loss": 0.9476, "step": 4626 }, { "epoch": 2.0955615942028984, "grad_norm": 0.1276922501487786, "learning_rate": 5.764616134900026e-05, "loss": 0.934, "step": 4627 }, { "epoch": 2.096014492753623, "grad_norm": 0.10861715622250316, "learning_rate": 5.763481037093725e-05, "loss": 0.904, "step": 4628 }, { "epoch": 2.0964673913043477, "grad_norm": 0.1603839224248179, "learning_rate": 5.7623457629894314e-05, "loss": 0.9311, "step": 4629 }, { "epoch": 2.0969202898550723, "grad_norm": 0.20964771682968258, "learning_rate": 5.7612103127006375e-05, "loss": 0.9215, "step": 4630 }, { "epoch": 2.097373188405797, "grad_norm": 0.2028437002917894, "learning_rate": 5.760074686340858e-05, "loss": 0.9153, "step": 4631 }, { "epoch": 2.097826086956522, "grad_norm": 0.1431786209585607, "learning_rate": 5.7589388840236235e-05, "loss": 0.9392, "step": 4632 }, { "epoch": 2.0982789855072466, "grad_norm": 0.09502146116778029, "learning_rate": 5.757802905862481e-05, "loss": 0.9507, "step": 4633 }, { "epoch": 2.098731884057971, "grad_norm": 0.08117972838875244, "learning_rate": 5.756666751970998e-05, "loss": 0.9286, "step": 4634 }, { "epoch": 2.099184782608696, "grad_norm": 0.11849080894223944, "learning_rate": 5.755530422462755e-05, "loss": 0.9575, "step": 4635 }, { "epoch": 2.0996376811594204, "grad_norm": 0.16363795936256453, "learning_rate": 5.754393917451354e-05, "loss": 0.9263, "step": 4636 }, { "epoch": 2.100090579710145, "grad_norm": 0.16740711474329453, "learning_rate": 5.7532572370504126e-05, "loss": 0.939, "step": 4637 }, { "epoch": 2.1005434782608696, "grad_norm": 0.11173703153828746, "learning_rate": 5.752120381373568e-05, "loss": 0.9085, "step": 4638 }, { "epoch": 2.1009963768115942, "grad_norm": 0.08603399963250292, "learning_rate": 5.750983350534471e-05, "loss": 0.9491, "step": 4639 }, { "epoch": 2.101449275362319, "grad_norm": 0.09249880408849016, "learning_rate": 5.749846144646794e-05, "loss": 0.9328, "step": 4640 }, { "epoch": 2.1019021739130435, "grad_norm": 0.13841633270632464, "learning_rate": 5.7487087638242265e-05, "loss": 0.9426, "step": 4641 }, { "epoch": 2.102355072463768, "grad_norm": 0.13169595945650223, "learning_rate": 5.747571208180472e-05, "loss": 0.9447, "step": 4642 }, { "epoch": 2.1028079710144927, "grad_norm": 0.11695108008697003, "learning_rate": 5.7464334778292535e-05, "loss": 0.9302, "step": 4643 }, { "epoch": 2.1032608695652173, "grad_norm": 0.10142513613040546, "learning_rate": 5.745295572884315e-05, "loss": 0.9394, "step": 4644 }, { "epoch": 2.103713768115942, "grad_norm": 0.08525468186807952, "learning_rate": 5.74415749345941e-05, "loss": 0.9389, "step": 4645 }, { "epoch": 2.1041666666666665, "grad_norm": 0.09613578965189279, "learning_rate": 5.743019239668318e-05, "loss": 0.9199, "step": 4646 }, { "epoch": 2.104619565217391, "grad_norm": 0.11677350339691234, "learning_rate": 5.741880811624831e-05, "loss": 0.9438, "step": 4647 }, { "epoch": 2.1050724637681157, "grad_norm": 0.1193590907509134, "learning_rate": 5.740742209442758e-05, "loss": 0.9425, "step": 4648 }, { "epoch": 2.105525362318841, "grad_norm": 0.09416857288481487, "learning_rate": 5.739603433235927e-05, "loss": 0.908, "step": 4649 }, { "epoch": 2.1059782608695654, "grad_norm": 0.07790890126901329, "learning_rate": 5.738464483118186e-05, "loss": 0.9393, "step": 4650 }, { "epoch": 2.10643115942029, "grad_norm": 0.07602560296224645, "learning_rate": 5.737325359203395e-05, "loss": 0.9407, "step": 4651 }, { "epoch": 2.1068840579710146, "grad_norm": 0.0901587774685276, "learning_rate": 5.736186061605435e-05, "loss": 0.9276, "step": 4652 }, { "epoch": 2.1073369565217392, "grad_norm": 0.08994307536086972, "learning_rate": 5.735046590438202e-05, "loss": 0.9379, "step": 4653 }, { "epoch": 2.107789855072464, "grad_norm": 0.10048979405831011, "learning_rate": 5.733906945815613e-05, "loss": 0.9317, "step": 4654 }, { "epoch": 2.1082427536231885, "grad_norm": 0.10485310376305004, "learning_rate": 5.732767127851599e-05, "loss": 0.917, "step": 4655 }, { "epoch": 2.108695652173913, "grad_norm": 0.11424282190059168, "learning_rate": 5.73162713666011e-05, "loss": 0.9332, "step": 4656 }, { "epoch": 2.1091485507246377, "grad_norm": 0.10184377427538341, "learning_rate": 5.730486972355112e-05, "loss": 0.9188, "step": 4657 }, { "epoch": 2.1096014492753623, "grad_norm": 0.07729331001814782, "learning_rate": 5.729346635050588e-05, "loss": 0.9303, "step": 4658 }, { "epoch": 2.110054347826087, "grad_norm": 0.08750641346741053, "learning_rate": 5.728206124860541e-05, "loss": 0.9695, "step": 4659 }, { "epoch": 2.1105072463768115, "grad_norm": 0.09900585614463268, "learning_rate": 5.7270654418989904e-05, "loss": 0.9636, "step": 4660 }, { "epoch": 2.110960144927536, "grad_norm": 0.12503049155710447, "learning_rate": 5.725924586279969e-05, "loss": 0.9426, "step": 4661 }, { "epoch": 2.1114130434782608, "grad_norm": 0.09334709288775289, "learning_rate": 5.7247835581175326e-05, "loss": 0.9508, "step": 4662 }, { "epoch": 2.1118659420289854, "grad_norm": 0.0927069514799653, "learning_rate": 5.7236423575257496e-05, "loss": 0.9413, "step": 4663 }, { "epoch": 2.11231884057971, "grad_norm": 0.1248928612656325, "learning_rate": 5.722500984618711e-05, "loss": 0.9353, "step": 4664 }, { "epoch": 2.1127717391304346, "grad_norm": 0.10545576015629651, "learning_rate": 5.721359439510518e-05, "loss": 0.9148, "step": 4665 }, { "epoch": 2.113224637681159, "grad_norm": 0.10724235332214924, "learning_rate": 5.720217722315294e-05, "loss": 0.9547, "step": 4666 }, { "epoch": 2.1136775362318843, "grad_norm": 0.12124851684047593, "learning_rate": 5.719075833147179e-05, "loss": 0.9319, "step": 4667 }, { "epoch": 2.114130434782609, "grad_norm": 0.10797769938024901, "learning_rate": 5.717933772120329e-05, "loss": 0.9481, "step": 4668 }, { "epoch": 2.1145833333333335, "grad_norm": 0.10063537120758996, "learning_rate": 5.716791539348917e-05, "loss": 0.9303, "step": 4669 }, { "epoch": 2.115036231884058, "grad_norm": 0.10004615683490797, "learning_rate": 5.7156491349471336e-05, "loss": 0.9389, "step": 4670 }, { "epoch": 2.1154891304347827, "grad_norm": 0.1307714791839524, "learning_rate": 5.714506559029188e-05, "loss": 0.9532, "step": 4671 }, { "epoch": 2.1159420289855073, "grad_norm": 0.14077559382038424, "learning_rate": 5.713363811709306e-05, "loss": 0.9377, "step": 4672 }, { "epoch": 2.116394927536232, "grad_norm": 0.1501050309453569, "learning_rate": 5.7122208931017276e-05, "loss": 0.9283, "step": 4673 }, { "epoch": 2.1168478260869565, "grad_norm": 0.12490881147728441, "learning_rate": 5.711077803320713e-05, "loss": 0.9385, "step": 4674 }, { "epoch": 2.117300724637681, "grad_norm": 0.09577310191320826, "learning_rate": 5.7099345424805394e-05, "loss": 0.9189, "step": 4675 }, { "epoch": 2.1177536231884058, "grad_norm": 0.11177392755961389, "learning_rate": 5.7087911106955e-05, "loss": 0.9175, "step": 4676 }, { "epoch": 2.1182065217391304, "grad_norm": 0.13325999312553305, "learning_rate": 5.7076475080799055e-05, "loss": 0.9491, "step": 4677 }, { "epoch": 2.118659420289855, "grad_norm": 0.16639606689237207, "learning_rate": 5.706503734748084e-05, "loss": 0.9424, "step": 4678 }, { "epoch": 2.1191123188405796, "grad_norm": 0.18610661200694872, "learning_rate": 5.7053597908143797e-05, "loss": 0.9593, "step": 4679 }, { "epoch": 2.119565217391304, "grad_norm": 0.13855011674628545, "learning_rate": 5.704215676393154e-05, "loss": 0.9509, "step": 4680 }, { "epoch": 2.120018115942029, "grad_norm": 0.094146285680405, "learning_rate": 5.703071391598787e-05, "loss": 0.941, "step": 4681 }, { "epoch": 2.1204710144927534, "grad_norm": 0.10711221404856289, "learning_rate": 5.701926936545675e-05, "loss": 0.9235, "step": 4682 }, { "epoch": 2.120923913043478, "grad_norm": 0.13961816508213518, "learning_rate": 5.7007823113482286e-05, "loss": 0.9057, "step": 4683 }, { "epoch": 2.121376811594203, "grad_norm": 0.12895788563883728, "learning_rate": 5.699637516120881e-05, "loss": 0.9195, "step": 4684 }, { "epoch": 2.1218297101449277, "grad_norm": 0.11872768537426777, "learning_rate": 5.6984925509780774e-05, "loss": 0.9538, "step": 4685 }, { "epoch": 2.1222826086956523, "grad_norm": 0.11695369779037995, "learning_rate": 5.697347416034281e-05, "loss": 0.9278, "step": 4686 }, { "epoch": 2.122735507246377, "grad_norm": 0.09051303735435834, "learning_rate": 5.696202111403974e-05, "loss": 0.9467, "step": 4687 }, { "epoch": 2.1231884057971016, "grad_norm": 0.1151763232756116, "learning_rate": 5.695056637201655e-05, "loss": 0.9586, "step": 4688 }, { "epoch": 2.123641304347826, "grad_norm": 0.12254463729187938, "learning_rate": 5.693910993541837e-05, "loss": 0.9152, "step": 4689 }, { "epoch": 2.1240942028985508, "grad_norm": 0.11350327125124864, "learning_rate": 5.692765180539053e-05, "loss": 0.9348, "step": 4690 }, { "epoch": 2.1245471014492754, "grad_norm": 0.10360988792823095, "learning_rate": 5.691619198307852e-05, "loss": 0.9502, "step": 4691 }, { "epoch": 2.125, "grad_norm": 0.11213623000442136, "learning_rate": 5.6904730469627985e-05, "loss": 0.934, "step": 4692 }, { "epoch": 2.1254528985507246, "grad_norm": 0.119589529681403, "learning_rate": 5.689326726618476e-05, "loss": 0.9321, "step": 4693 }, { "epoch": 2.1259057971014492, "grad_norm": 0.11542296273501512, "learning_rate": 5.688180237389484e-05, "loss": 0.9511, "step": 4694 }, { "epoch": 2.126358695652174, "grad_norm": 0.08014121541032122, "learning_rate": 5.687033579390438e-05, "loss": 0.9422, "step": 4695 }, { "epoch": 2.1268115942028984, "grad_norm": 0.08387836681052506, "learning_rate": 5.6858867527359706e-05, "loss": 0.9442, "step": 4696 }, { "epoch": 2.127264492753623, "grad_norm": 0.09511248887070774, "learning_rate": 5.684739757540735e-05, "loss": 0.9518, "step": 4697 }, { "epoch": 2.1277173913043477, "grad_norm": 0.0981447129212835, "learning_rate": 5.683592593919395e-05, "loss": 0.9481, "step": 4698 }, { "epoch": 2.1281702898550723, "grad_norm": 0.0943954378025377, "learning_rate": 5.682445261986636e-05, "loss": 0.9293, "step": 4699 }, { "epoch": 2.128623188405797, "grad_norm": 0.09398643267443164, "learning_rate": 5.681297761857158e-05, "loss": 0.9222, "step": 4700 }, { "epoch": 2.1290760869565215, "grad_norm": 0.08874928650355472, "learning_rate": 5.680150093645678e-05, "loss": 0.9385, "step": 4701 }, { "epoch": 2.1295289855072466, "grad_norm": 0.10639825715389377, "learning_rate": 5.6790022574669324e-05, "loss": 0.9418, "step": 4702 }, { "epoch": 2.129981884057971, "grad_norm": 0.09548426824924135, "learning_rate": 5.677854253435669e-05, "loss": 0.9388, "step": 4703 }, { "epoch": 2.130434782608696, "grad_norm": 0.11530250601812973, "learning_rate": 5.676706081666657e-05, "loss": 0.9552, "step": 4704 }, { "epoch": 2.1308876811594204, "grad_norm": 0.11531445505286757, "learning_rate": 5.6755577422746805e-05, "loss": 0.9482, "step": 4705 }, { "epoch": 2.131340579710145, "grad_norm": 0.0908857379588881, "learning_rate": 5.6744092353745444e-05, "loss": 0.9408, "step": 4706 }, { "epoch": 2.1317934782608696, "grad_norm": 0.09094690039418102, "learning_rate": 5.67326056108106e-05, "loss": 0.9349, "step": 4707 }, { "epoch": 2.1322463768115942, "grad_norm": 0.10066916839409713, "learning_rate": 5.672111719509068e-05, "loss": 0.9366, "step": 4708 }, { "epoch": 2.132699275362319, "grad_norm": 0.08896561941455693, "learning_rate": 5.670962710773417e-05, "loss": 0.9467, "step": 4709 }, { "epoch": 2.1331521739130435, "grad_norm": 0.11260573632038973, "learning_rate": 5.669813534988977e-05, "loss": 0.9461, "step": 4710 }, { "epoch": 2.133605072463768, "grad_norm": 0.09566032850590026, "learning_rate": 5.66866419227063e-05, "loss": 0.9157, "step": 4711 }, { "epoch": 2.1340579710144927, "grad_norm": 0.11468559769095087, "learning_rate": 5.6675146827332805e-05, "loss": 0.9339, "step": 4712 }, { "epoch": 2.1345108695652173, "grad_norm": 0.13113738270608663, "learning_rate": 5.6663650064918456e-05, "loss": 0.9358, "step": 4713 }, { "epoch": 2.134963768115942, "grad_norm": 0.1567157014214192, "learning_rate": 5.665215163661261e-05, "loss": 0.9616, "step": 4714 }, { "epoch": 2.1354166666666665, "grad_norm": 0.13419151581390443, "learning_rate": 5.664065154356477e-05, "loss": 0.9503, "step": 4715 }, { "epoch": 2.135869565217391, "grad_norm": 0.0920990383270525, "learning_rate": 5.662914978692463e-05, "loss": 0.9575, "step": 4716 }, { "epoch": 2.1363224637681157, "grad_norm": 0.1207710832436002, "learning_rate": 5.6617646367842025e-05, "loss": 0.9257, "step": 4717 }, { "epoch": 2.136775362318841, "grad_norm": 0.15645788490953294, "learning_rate": 5.6606141287466996e-05, "loss": 0.9439, "step": 4718 }, { "epoch": 2.1372282608695654, "grad_norm": 0.18560238951311825, "learning_rate": 5.659463454694969e-05, "loss": 0.9629, "step": 4719 }, { "epoch": 2.13768115942029, "grad_norm": 0.15963630715154362, "learning_rate": 5.6583126147440486e-05, "loss": 0.9471, "step": 4720 }, { "epoch": 2.1381340579710146, "grad_norm": 0.11741000712819216, "learning_rate": 5.657161609008987e-05, "loss": 0.9198, "step": 4721 }, { "epoch": 2.1385869565217392, "grad_norm": 0.11706245959422668, "learning_rate": 5.656010437604854e-05, "loss": 0.9268, "step": 4722 }, { "epoch": 2.139039855072464, "grad_norm": 0.09550748783848471, "learning_rate": 5.654859100646734e-05, "loss": 0.9541, "step": 4723 }, { "epoch": 2.1394927536231885, "grad_norm": 0.10940854184681072, "learning_rate": 5.653707598249726e-05, "loss": 0.9364, "step": 4724 }, { "epoch": 2.139945652173913, "grad_norm": 0.08981156444470702, "learning_rate": 5.6525559305289506e-05, "loss": 0.954, "step": 4725 }, { "epoch": 2.1403985507246377, "grad_norm": 0.10496634989993894, "learning_rate": 5.651404097599539e-05, "loss": 0.9396, "step": 4726 }, { "epoch": 2.1408514492753623, "grad_norm": 0.11256029659613195, "learning_rate": 5.650252099576644e-05, "loss": 0.9361, "step": 4727 }, { "epoch": 2.141304347826087, "grad_norm": 0.10670693443316222, "learning_rate": 5.6490999365754304e-05, "loss": 0.9369, "step": 4728 }, { "epoch": 2.1417572463768115, "grad_norm": 0.1035068684561167, "learning_rate": 5.6479476087110835e-05, "loss": 0.9392, "step": 4729 }, { "epoch": 2.142210144927536, "grad_norm": 0.07264030081208961, "learning_rate": 5.646795116098801e-05, "loss": 0.9575, "step": 4730 }, { "epoch": 2.1426630434782608, "grad_norm": 0.1061811616316879, "learning_rate": 5.645642458853804e-05, "loss": 0.9322, "step": 4731 }, { "epoch": 2.1431159420289854, "grad_norm": 0.13861424675955902, "learning_rate": 5.644489637091321e-05, "loss": 0.9486, "step": 4732 }, { "epoch": 2.14356884057971, "grad_norm": 0.10859800695245776, "learning_rate": 5.643336650926603e-05, "loss": 0.9509, "step": 4733 }, { "epoch": 2.1440217391304346, "grad_norm": 0.10667027184106868, "learning_rate": 5.642183500474916e-05, "loss": 0.9443, "step": 4734 }, { "epoch": 2.144474637681159, "grad_norm": 0.12542214523632866, "learning_rate": 5.6410301858515435e-05, "loss": 0.9419, "step": 4735 }, { "epoch": 2.1449275362318843, "grad_norm": 0.09621979392224474, "learning_rate": 5.639876707171782e-05, "loss": 0.952, "step": 4736 }, { "epoch": 2.145380434782609, "grad_norm": 0.10558391728760519, "learning_rate": 5.638723064550947e-05, "loss": 0.9077, "step": 4737 }, { "epoch": 2.1458333333333335, "grad_norm": 0.11471362583693878, "learning_rate": 5.6375692581043705e-05, "loss": 0.9262, "step": 4738 }, { "epoch": 2.146286231884058, "grad_norm": 0.09271104652781723, "learning_rate": 5.636415287947401e-05, "loss": 0.925, "step": 4739 }, { "epoch": 2.1467391304347827, "grad_norm": 0.09421644420958379, "learning_rate": 5.635261154195402e-05, "loss": 0.943, "step": 4740 }, { "epoch": 2.1471920289855073, "grad_norm": 0.10648556952447275, "learning_rate": 5.634106856963753e-05, "loss": 0.9232, "step": 4741 }, { "epoch": 2.147644927536232, "grad_norm": 0.09242921654585115, "learning_rate": 5.6329523963678525e-05, "loss": 0.9283, "step": 4742 }, { "epoch": 2.1480978260869565, "grad_norm": 0.11611056213796482, "learning_rate": 5.631797772523113e-05, "loss": 0.9378, "step": 4743 }, { "epoch": 2.148550724637681, "grad_norm": 0.10301201659513863, "learning_rate": 5.630642985544967e-05, "loss": 0.9228, "step": 4744 }, { "epoch": 2.1490036231884058, "grad_norm": 0.10636726863593034, "learning_rate": 5.6294880355488555e-05, "loss": 0.9178, "step": 4745 }, { "epoch": 2.1494565217391304, "grad_norm": 0.11428718030778487, "learning_rate": 5.628332922650242e-05, "loss": 0.9175, "step": 4746 }, { "epoch": 2.149909420289855, "grad_norm": 0.13293939716243836, "learning_rate": 5.627177646964607e-05, "loss": 0.9246, "step": 4747 }, { "epoch": 2.1503623188405796, "grad_norm": 0.14392793279249635, "learning_rate": 5.626022208607445e-05, "loss": 0.9427, "step": 4748 }, { "epoch": 2.150815217391304, "grad_norm": 0.12453554067801753, "learning_rate": 5.624866607694266e-05, "loss": 0.9212, "step": 4749 }, { "epoch": 2.151268115942029, "grad_norm": 0.1054181760667429, "learning_rate": 5.623710844340596e-05, "loss": 0.9407, "step": 4750 }, { "epoch": 2.1517210144927534, "grad_norm": 0.1318007301491443, "learning_rate": 5.622554918661981e-05, "loss": 0.9466, "step": 4751 }, { "epoch": 2.1521739130434785, "grad_norm": 0.13403652013978892, "learning_rate": 5.621398830773979e-05, "loss": 0.9105, "step": 4752 }, { "epoch": 2.152626811594203, "grad_norm": 0.1113927667582332, "learning_rate": 5.620242580792167e-05, "loss": 0.9415, "step": 4753 }, { "epoch": 2.1530797101449277, "grad_norm": 0.12294484446009209, "learning_rate": 5.6190861688321366e-05, "loss": 0.9491, "step": 4754 }, { "epoch": 2.1535326086956523, "grad_norm": 0.13016266546719543, "learning_rate": 5.617929595009497e-05, "loss": 0.9045, "step": 4755 }, { "epoch": 2.153985507246377, "grad_norm": 0.10907678427339676, "learning_rate": 5.616772859439871e-05, "loss": 0.9599, "step": 4756 }, { "epoch": 2.1544384057971016, "grad_norm": 0.08526811651712446, "learning_rate": 5.6156159622389006e-05, "loss": 0.9347, "step": 4757 }, { "epoch": 2.154891304347826, "grad_norm": 0.11223106444546793, "learning_rate": 5.6144589035222416e-05, "loss": 0.8961, "step": 4758 }, { "epoch": 2.1553442028985508, "grad_norm": 0.09012043460125833, "learning_rate": 5.613301683405568e-05, "loss": 0.9439, "step": 4759 }, { "epoch": 2.1557971014492754, "grad_norm": 0.12750722454537708, "learning_rate": 5.612144302004569e-05, "loss": 0.9433, "step": 4760 }, { "epoch": 2.15625, "grad_norm": 0.14871861211233173, "learning_rate": 5.61098675943495e-05, "loss": 0.9429, "step": 4761 }, { "epoch": 2.1567028985507246, "grad_norm": 0.11786483605772595, "learning_rate": 5.6098290558124304e-05, "loss": 0.9279, "step": 4762 }, { "epoch": 2.1571557971014492, "grad_norm": 0.10760328131361467, "learning_rate": 5.6086711912527496e-05, "loss": 0.9484, "step": 4763 }, { "epoch": 2.157608695652174, "grad_norm": 0.09830180766957115, "learning_rate": 5.607513165871661e-05, "loss": 0.9287, "step": 4764 }, { "epoch": 2.1580615942028984, "grad_norm": 0.10995579082846806, "learning_rate": 5.606354979784934e-05, "loss": 0.9388, "step": 4765 }, { "epoch": 2.158514492753623, "grad_norm": 0.08648016165769455, "learning_rate": 5.605196633108354e-05, "loss": 0.9324, "step": 4766 }, { "epoch": 2.1589673913043477, "grad_norm": 0.10975420391975588, "learning_rate": 5.604038125957722e-05, "loss": 0.9483, "step": 4767 }, { "epoch": 2.1594202898550723, "grad_norm": 0.10274379021805717, "learning_rate": 5.6028794584488585e-05, "loss": 0.9513, "step": 4768 }, { "epoch": 2.159873188405797, "grad_norm": 0.11348854005753203, "learning_rate": 5.601720630697595e-05, "loss": 0.9299, "step": 4769 }, { "epoch": 2.1603260869565215, "grad_norm": 0.1051092946139274, "learning_rate": 5.6005616428197816e-05, "loss": 0.9367, "step": 4770 }, { "epoch": 2.1607789855072466, "grad_norm": 0.10661841511337572, "learning_rate": 5.5994024949312845e-05, "loss": 0.9593, "step": 4771 }, { "epoch": 2.161231884057971, "grad_norm": 0.10266684721659951, "learning_rate": 5.5982431871479856e-05, "loss": 0.9259, "step": 4772 }, { "epoch": 2.161684782608696, "grad_norm": 0.08889684408908079, "learning_rate": 5.597083719585782e-05, "loss": 0.9554, "step": 4773 }, { "epoch": 2.1621376811594204, "grad_norm": 0.08207832529826163, "learning_rate": 5.595924092360589e-05, "loss": 0.9243, "step": 4774 }, { "epoch": 2.162590579710145, "grad_norm": 0.10333055220609458, "learning_rate": 5.5947643055883355e-05, "loss": 0.9209, "step": 4775 }, { "epoch": 2.1630434782608696, "grad_norm": 0.11164456403802202, "learning_rate": 5.593604359384967e-05, "loss": 0.9184, "step": 4776 }, { "epoch": 2.1634963768115942, "grad_norm": 0.1152110915241035, "learning_rate": 5.5924442538664454e-05, "loss": 0.9522, "step": 4777 }, { "epoch": 2.163949275362319, "grad_norm": 0.08869115960544846, "learning_rate": 5.5912839891487484e-05, "loss": 0.937, "step": 4778 }, { "epoch": 2.1644021739130435, "grad_norm": 0.08827620181853364, "learning_rate": 5.590123565347869e-05, "loss": 0.9252, "step": 4779 }, { "epoch": 2.164855072463768, "grad_norm": 0.1130207586399869, "learning_rate": 5.5889629825798164e-05, "loss": 0.9364, "step": 4780 }, { "epoch": 2.1653079710144927, "grad_norm": 0.09678420995894055, "learning_rate": 5.587802240960617e-05, "loss": 0.9405, "step": 4781 }, { "epoch": 2.1657608695652173, "grad_norm": 0.12401851702201695, "learning_rate": 5.5866413406063136e-05, "loss": 0.9531, "step": 4782 }, { "epoch": 2.166213768115942, "grad_norm": 0.12583598016203196, "learning_rate": 5.5854802816329587e-05, "loss": 0.9524, "step": 4783 }, { "epoch": 2.1666666666666665, "grad_norm": 0.12330634715717911, "learning_rate": 5.584319064156628e-05, "loss": 0.9528, "step": 4784 }, { "epoch": 2.167119565217391, "grad_norm": 0.10831538226788137, "learning_rate": 5.5831576882934104e-05, "loss": 0.9176, "step": 4785 }, { "epoch": 2.1675724637681157, "grad_norm": 0.09578322171656062, "learning_rate": 5.5819961541594104e-05, "loss": 0.921, "step": 4786 }, { "epoch": 2.168025362318841, "grad_norm": 0.0944593493706682, "learning_rate": 5.580834461870747e-05, "loss": 0.924, "step": 4787 }, { "epoch": 2.1684782608695654, "grad_norm": 0.09830721816893416, "learning_rate": 5.579672611543557e-05, "loss": 0.9515, "step": 4788 }, { "epoch": 2.16893115942029, "grad_norm": 0.08950840244122604, "learning_rate": 5.578510603293994e-05, "loss": 0.9264, "step": 4789 }, { "epoch": 2.1693840579710146, "grad_norm": 0.08680708793505783, "learning_rate": 5.577348437238223e-05, "loss": 0.9265, "step": 4790 }, { "epoch": 2.1698369565217392, "grad_norm": 0.10376686438742547, "learning_rate": 5.576186113492432e-05, "loss": 0.9517, "step": 4791 }, { "epoch": 2.170289855072464, "grad_norm": 0.09708350456480132, "learning_rate": 5.575023632172816e-05, "loss": 0.9342, "step": 4792 }, { "epoch": 2.1707427536231885, "grad_norm": 0.08630933143486756, "learning_rate": 5.573860993395591e-05, "loss": 0.9051, "step": 4793 }, { "epoch": 2.171195652173913, "grad_norm": 0.10121541678014807, "learning_rate": 5.572698197276989e-05, "loss": 0.945, "step": 4794 }, { "epoch": 2.1716485507246377, "grad_norm": 0.19916480761365965, "learning_rate": 5.571535243933257e-05, "loss": 0.9448, "step": 4795 }, { "epoch": 2.1721014492753623, "grad_norm": 0.0970926125838818, "learning_rate": 5.5703721334806555e-05, "loss": 0.9735, "step": 4796 }, { "epoch": 2.172554347826087, "grad_norm": 0.10239728330813626, "learning_rate": 5.569208866035462e-05, "loss": 0.9497, "step": 4797 }, { "epoch": 2.1730072463768115, "grad_norm": 0.10006238762230939, "learning_rate": 5.568045441713973e-05, "loss": 0.9384, "step": 4798 }, { "epoch": 2.173460144927536, "grad_norm": 0.09824575315221931, "learning_rate": 5.5668818606324975e-05, "loss": 0.9414, "step": 4799 }, { "epoch": 2.1739130434782608, "grad_norm": 0.13664627651630123, "learning_rate": 5.5657181229073584e-05, "loss": 0.9176, "step": 4800 }, { "epoch": 2.1743659420289854, "grad_norm": 0.14614809094986927, "learning_rate": 5.5645542286548965e-05, "loss": 0.9439, "step": 4801 }, { "epoch": 2.17481884057971, "grad_norm": 0.08902064275483354, "learning_rate": 5.563390177991469e-05, "loss": 0.9549, "step": 4802 }, { "epoch": 2.1752717391304346, "grad_norm": 0.10776811439034821, "learning_rate": 5.56222597103345e-05, "loss": 0.9413, "step": 4803 }, { "epoch": 2.175724637681159, "grad_norm": 0.10978362019184497, "learning_rate": 5.561061607897224e-05, "loss": 0.9232, "step": 4804 }, { "epoch": 2.1761775362318843, "grad_norm": 0.10526805639204803, "learning_rate": 5.5598970886991956e-05, "loss": 0.9423, "step": 4805 }, { "epoch": 2.176630434782609, "grad_norm": 0.09525236761710806, "learning_rate": 5.558732413555783e-05, "loss": 0.9245, "step": 4806 }, { "epoch": 2.1770833333333335, "grad_norm": 0.10405022892109637, "learning_rate": 5.5575675825834215e-05, "loss": 0.9233, "step": 4807 }, { "epoch": 2.177536231884058, "grad_norm": 0.11425022967408305, "learning_rate": 5.55640259589856e-05, "loss": 0.9252, "step": 4808 }, { "epoch": 2.1779891304347827, "grad_norm": 0.12611969016536798, "learning_rate": 5.555237453617665e-05, "loss": 0.9449, "step": 4809 }, { "epoch": 2.1784420289855073, "grad_norm": 0.09102439641075608, "learning_rate": 5.554072155857217e-05, "loss": 0.9257, "step": 4810 }, { "epoch": 2.178894927536232, "grad_norm": 0.11319596954543429, "learning_rate": 5.552906702733713e-05, "loss": 0.9367, "step": 4811 }, { "epoch": 2.1793478260869565, "grad_norm": 0.11444104677918275, "learning_rate": 5.5517410943636654e-05, "loss": 0.9343, "step": 4812 }, { "epoch": 2.179800724637681, "grad_norm": 0.1067585300073556, "learning_rate": 5.5505753308636015e-05, "loss": 0.9625, "step": 4813 }, { "epoch": 2.1802536231884058, "grad_norm": 0.12435182216795924, "learning_rate": 5.549409412350065e-05, "loss": 0.9246, "step": 4814 }, { "epoch": 2.1807065217391304, "grad_norm": 0.1230008943636331, "learning_rate": 5.548243338939613e-05, "loss": 0.9359, "step": 4815 }, { "epoch": 2.181159420289855, "grad_norm": 0.14704448346247637, "learning_rate": 5.547077110748823e-05, "loss": 0.9432, "step": 4816 }, { "epoch": 2.1816123188405796, "grad_norm": 0.08356700782260001, "learning_rate": 5.5459107278942815e-05, "loss": 0.944, "step": 4817 }, { "epoch": 2.182065217391304, "grad_norm": 0.11773106156431214, "learning_rate": 5.544744190492594e-05, "loss": 0.9576, "step": 4818 }, { "epoch": 2.182518115942029, "grad_norm": 0.11836689597241078, "learning_rate": 5.5435774986603835e-05, "loss": 0.9411, "step": 4819 }, { "epoch": 2.1829710144927534, "grad_norm": 0.1071813769500214, "learning_rate": 5.542410652514284e-05, "loss": 0.9185, "step": 4820 }, { "epoch": 2.1834239130434785, "grad_norm": 0.11033086028347627, "learning_rate": 5.5412436521709476e-05, "loss": 0.9393, "step": 4821 }, { "epoch": 2.183876811594203, "grad_norm": 0.11535808212768492, "learning_rate": 5.5400764977470405e-05, "loss": 0.9426, "step": 4822 }, { "epoch": 2.1843297101449277, "grad_norm": 0.07977707995913344, "learning_rate": 5.5389091893592454e-05, "loss": 0.9103, "step": 4823 }, { "epoch": 2.1847826086956523, "grad_norm": 0.08326870719874745, "learning_rate": 5.5377417271242605e-05, "loss": 0.9198, "step": 4824 }, { "epoch": 2.185235507246377, "grad_norm": 0.08994293648518592, "learning_rate": 5.536574111158798e-05, "loss": 0.9247, "step": 4825 }, { "epoch": 2.1856884057971016, "grad_norm": 0.10100512743314527, "learning_rate": 5.535406341579587e-05, "loss": 0.9378, "step": 4826 }, { "epoch": 2.186141304347826, "grad_norm": 0.09033130054466645, "learning_rate": 5.534238418503371e-05, "loss": 0.9372, "step": 4827 }, { "epoch": 2.1865942028985508, "grad_norm": 0.11377159536918882, "learning_rate": 5.533070342046909e-05, "loss": 0.9426, "step": 4828 }, { "epoch": 2.1870471014492754, "grad_norm": 0.09526478973482327, "learning_rate": 5.5319021123269756e-05, "loss": 0.9467, "step": 4829 }, { "epoch": 2.1875, "grad_norm": 0.08979852846462592, "learning_rate": 5.5307337294603595e-05, "loss": 0.9478, "step": 4830 }, { "epoch": 2.1879528985507246, "grad_norm": 0.09187112369239062, "learning_rate": 5.529565193563868e-05, "loss": 0.9507, "step": 4831 }, { "epoch": 2.1884057971014492, "grad_norm": 0.12582267252964757, "learning_rate": 5.528396504754319e-05, "loss": 0.9536, "step": 4832 }, { "epoch": 2.188858695652174, "grad_norm": 0.12277120193454989, "learning_rate": 5.5272276631485516e-05, "loss": 0.9354, "step": 4833 }, { "epoch": 2.1893115942028984, "grad_norm": 0.09933909925448027, "learning_rate": 5.526058668863413e-05, "loss": 0.9361, "step": 4834 }, { "epoch": 2.189764492753623, "grad_norm": 0.11312173495718257, "learning_rate": 5.524889522015771e-05, "loss": 0.9448, "step": 4835 }, { "epoch": 2.1902173913043477, "grad_norm": 0.10914721261928374, "learning_rate": 5.523720222722508e-05, "loss": 0.9389, "step": 4836 }, { "epoch": 2.1906702898550723, "grad_norm": 0.0997953383530955, "learning_rate": 5.5225507711005196e-05, "loss": 0.9352, "step": 4837 }, { "epoch": 2.191123188405797, "grad_norm": 0.12679727310576253, "learning_rate": 5.521381167266718e-05, "loss": 0.9382, "step": 4838 }, { "epoch": 2.1915760869565215, "grad_norm": 0.11930176443435592, "learning_rate": 5.5202114113380296e-05, "loss": 0.926, "step": 4839 }, { "epoch": 2.1920289855072466, "grad_norm": 0.12442075525459038, "learning_rate": 5.519041503431398e-05, "loss": 0.9341, "step": 4840 }, { "epoch": 2.192481884057971, "grad_norm": 0.0984127523566489, "learning_rate": 5.5178714436637806e-05, "loss": 0.9366, "step": 4841 }, { "epoch": 2.192934782608696, "grad_norm": 0.0952489102499137, "learning_rate": 5.516701232152148e-05, "loss": 0.9355, "step": 4842 }, { "epoch": 2.1933876811594204, "grad_norm": 0.11487646911225442, "learning_rate": 5.515530869013491e-05, "loss": 0.9367, "step": 4843 }, { "epoch": 2.193840579710145, "grad_norm": 0.11217936242937437, "learning_rate": 5.5143603543648114e-05, "loss": 0.938, "step": 4844 }, { "epoch": 2.1942934782608696, "grad_norm": 0.11810693415572046, "learning_rate": 5.5131896883231275e-05, "loss": 0.9323, "step": 4845 }, { "epoch": 2.1947463768115942, "grad_norm": 0.08935121690503466, "learning_rate": 5.512018871005472e-05, "loss": 0.9604, "step": 4846 }, { "epoch": 2.195199275362319, "grad_norm": 0.10547813928389188, "learning_rate": 5.5108479025288946e-05, "loss": 0.9604, "step": 4847 }, { "epoch": 2.1956521739130435, "grad_norm": 0.09691062089683765, "learning_rate": 5.5096767830104574e-05, "loss": 0.9415, "step": 4848 }, { "epoch": 2.196105072463768, "grad_norm": 0.0824845652671712, "learning_rate": 5.508505512567242e-05, "loss": 0.9322, "step": 4849 }, { "epoch": 2.1965579710144927, "grad_norm": 0.09355315411061739, "learning_rate": 5.507334091316338e-05, "loss": 0.9403, "step": 4850 }, { "epoch": 2.1970108695652173, "grad_norm": 0.10363185723757055, "learning_rate": 5.506162519374857e-05, "loss": 0.9276, "step": 4851 }, { "epoch": 2.197463768115942, "grad_norm": 0.0911459892024758, "learning_rate": 5.504990796859922e-05, "loss": 0.9233, "step": 4852 }, { "epoch": 2.1979166666666665, "grad_norm": 0.10381699003110596, "learning_rate": 5.5038189238886724e-05, "loss": 0.9732, "step": 4853 }, { "epoch": 2.198369565217391, "grad_norm": 0.1080206369809082, "learning_rate": 5.502646900578263e-05, "loss": 0.9144, "step": 4854 }, { "epoch": 2.1988224637681157, "grad_norm": 0.10571705243833371, "learning_rate": 5.5014747270458605e-05, "loss": 0.945, "step": 4855 }, { "epoch": 2.199275362318841, "grad_norm": 0.10983294048522846, "learning_rate": 5.5003024034086504e-05, "loss": 0.9469, "step": 4856 }, { "epoch": 2.1997282608695654, "grad_norm": 0.10162120514981865, "learning_rate": 5.4991299297838306e-05, "loss": 0.9149, "step": 4857 }, { "epoch": 2.20018115942029, "grad_norm": 0.12293380962613647, "learning_rate": 5.497957306288618e-05, "loss": 0.9685, "step": 4858 }, { "epoch": 2.2006340579710146, "grad_norm": 0.10624101732182281, "learning_rate": 5.49678453304024e-05, "loss": 0.9309, "step": 4859 }, { "epoch": 2.2010869565217392, "grad_norm": 0.10150469405635927, "learning_rate": 5.495611610155939e-05, "loss": 0.9251, "step": 4860 }, { "epoch": 2.201539855072464, "grad_norm": 0.09260245899236552, "learning_rate": 5.4944385377529755e-05, "loss": 0.9401, "step": 4861 }, { "epoch": 2.2019927536231885, "grad_norm": 0.09761717299547236, "learning_rate": 5.4932653159486236e-05, "loss": 0.9575, "step": 4862 }, { "epoch": 2.202445652173913, "grad_norm": 0.09336200522227728, "learning_rate": 5.492091944860171e-05, "loss": 0.9371, "step": 4863 }, { "epoch": 2.2028985507246377, "grad_norm": 0.09541863343390229, "learning_rate": 5.490918424604923e-05, "loss": 0.9553, "step": 4864 }, { "epoch": 2.2033514492753623, "grad_norm": 0.10653830789380038, "learning_rate": 5.4897447553001976e-05, "loss": 0.9307, "step": 4865 }, { "epoch": 2.203804347826087, "grad_norm": 0.08262005653420537, "learning_rate": 5.4885709370633286e-05, "loss": 0.9473, "step": 4866 }, { "epoch": 2.2042572463768115, "grad_norm": 0.07741594960717908, "learning_rate": 5.487396970011664e-05, "loss": 0.9408, "step": 4867 }, { "epoch": 2.204710144927536, "grad_norm": 0.07776410295403918, "learning_rate": 5.4862228542625665e-05, "loss": 0.9423, "step": 4868 }, { "epoch": 2.2051630434782608, "grad_norm": 0.08947776715879617, "learning_rate": 5.4850485899334154e-05, "loss": 0.9366, "step": 4869 }, { "epoch": 2.2056159420289854, "grad_norm": 0.08884989220030874, "learning_rate": 5.483874177141605e-05, "loss": 0.9267, "step": 4870 }, { "epoch": 2.20606884057971, "grad_norm": 0.10917598366108397, "learning_rate": 5.4826996160045396e-05, "loss": 0.9516, "step": 4871 }, { "epoch": 2.2065217391304346, "grad_norm": 0.10033628177408546, "learning_rate": 5.481524906639645e-05, "loss": 0.9355, "step": 4872 }, { "epoch": 2.206974637681159, "grad_norm": 0.09275668445129949, "learning_rate": 5.480350049164358e-05, "loss": 0.9391, "step": 4873 }, { "epoch": 2.2074275362318843, "grad_norm": 0.08287862443854196, "learning_rate": 5.4791750436961295e-05, "loss": 0.9476, "step": 4874 }, { "epoch": 2.207880434782609, "grad_norm": 0.09167244197066393, "learning_rate": 5.47799989035243e-05, "loss": 0.9275, "step": 4875 }, { "epoch": 2.2083333333333335, "grad_norm": 0.10489589482281503, "learning_rate": 5.476824589250738e-05, "loss": 0.9615, "step": 4876 }, { "epoch": 2.208786231884058, "grad_norm": 0.10902462815800876, "learning_rate": 5.4756491405085516e-05, "loss": 0.9339, "step": 4877 }, { "epoch": 2.2092391304347827, "grad_norm": 0.13591311817833032, "learning_rate": 5.474473544243383e-05, "loss": 0.9364, "step": 4878 }, { "epoch": 2.2096920289855073, "grad_norm": 0.13190410729174082, "learning_rate": 5.4732978005727575e-05, "loss": 0.9366, "step": 4879 }, { "epoch": 2.210144927536232, "grad_norm": 0.09859443134295796, "learning_rate": 5.472121909614216e-05, "loss": 0.9503, "step": 4880 }, { "epoch": 2.2105978260869565, "grad_norm": 0.09764035873964855, "learning_rate": 5.470945871485314e-05, "loss": 0.9398, "step": 4881 }, { "epoch": 2.211050724637681, "grad_norm": 0.10433110065179746, "learning_rate": 5.4697696863036224e-05, "loss": 0.937, "step": 4882 }, { "epoch": 2.2115036231884058, "grad_norm": 0.104373132682754, "learning_rate": 5.4685933541867266e-05, "loss": 0.9309, "step": 4883 }, { "epoch": 2.2119565217391304, "grad_norm": 0.08644919194047491, "learning_rate": 5.467416875252227e-05, "loss": 0.9448, "step": 4884 }, { "epoch": 2.212409420289855, "grad_norm": 0.10896874330639948, "learning_rate": 5.466240249617736e-05, "loss": 0.9504, "step": 4885 }, { "epoch": 2.2128623188405796, "grad_norm": 0.1183192133738814, "learning_rate": 5.465063477400883e-05, "loss": 0.9347, "step": 4886 }, { "epoch": 2.213315217391304, "grad_norm": 0.10979186504501799, "learning_rate": 5.4638865587193136e-05, "loss": 0.9304, "step": 4887 }, { "epoch": 2.213768115942029, "grad_norm": 0.08890345924796247, "learning_rate": 5.462709493690686e-05, "loss": 0.9239, "step": 4888 }, { "epoch": 2.2142210144927534, "grad_norm": 0.10213197291716992, "learning_rate": 5.4615322824326697e-05, "loss": 0.9491, "step": 4889 }, { "epoch": 2.2146739130434785, "grad_norm": 0.10698798911476168, "learning_rate": 5.460354925062957e-05, "loss": 0.9428, "step": 4890 }, { "epoch": 2.215126811594203, "grad_norm": 0.11717530509248654, "learning_rate": 5.459177421699249e-05, "loss": 0.9473, "step": 4891 }, { "epoch": 2.2155797101449277, "grad_norm": 0.12831720732196447, "learning_rate": 5.457999772459261e-05, "loss": 0.937, "step": 4892 }, { "epoch": 2.2160326086956523, "grad_norm": 0.08638640172533704, "learning_rate": 5.4568219774607255e-05, "loss": 0.9463, "step": 4893 }, { "epoch": 2.216485507246377, "grad_norm": 0.10339856699114659, "learning_rate": 5.455644036821388e-05, "loss": 0.9498, "step": 4894 }, { "epoch": 2.2169384057971016, "grad_norm": 0.11590179046375025, "learning_rate": 5.45446595065901e-05, "loss": 0.9376, "step": 4895 }, { "epoch": 2.217391304347826, "grad_norm": 0.11455019830739593, "learning_rate": 5.453287719091366e-05, "loss": 0.9371, "step": 4896 }, { "epoch": 2.2178442028985508, "grad_norm": 0.13247297310402398, "learning_rate": 5.452109342236245e-05, "loss": 0.9176, "step": 4897 }, { "epoch": 2.2182971014492754, "grad_norm": 0.10137652964910802, "learning_rate": 5.4509308202114535e-05, "loss": 0.9597, "step": 4898 }, { "epoch": 2.21875, "grad_norm": 0.09814038839799152, "learning_rate": 5.4497521531348066e-05, "loss": 0.9348, "step": 4899 }, { "epoch": 2.2192028985507246, "grad_norm": 0.12843408619269156, "learning_rate": 5.448573341124142e-05, "loss": 0.9411, "step": 4900 }, { "epoch": 2.2196557971014492, "grad_norm": 0.12157517215169965, "learning_rate": 5.4473943842973035e-05, "loss": 0.9418, "step": 4901 }, { "epoch": 2.220108695652174, "grad_norm": 0.12365513775759507, "learning_rate": 5.446215282772155e-05, "loss": 0.933, "step": 4902 }, { "epoch": 2.2205615942028984, "grad_norm": 0.13500432013284203, "learning_rate": 5.445036036666573e-05, "loss": 0.9301, "step": 4903 }, { "epoch": 2.221014492753623, "grad_norm": 0.12185758451460009, "learning_rate": 5.4438566460984486e-05, "loss": 0.933, "step": 4904 }, { "epoch": 2.2214673913043477, "grad_norm": 0.10832667427635446, "learning_rate": 5.442677111185689e-05, "loss": 0.9583, "step": 4905 }, { "epoch": 2.2219202898550723, "grad_norm": 0.09656327867437628, "learning_rate": 5.44149743204621e-05, "loss": 0.9269, "step": 4906 }, { "epoch": 2.222373188405797, "grad_norm": 0.08764132634142546, "learning_rate": 5.4403176087979485e-05, "loss": 0.9403, "step": 4907 }, { "epoch": 2.2228260869565215, "grad_norm": 0.1393684098397121, "learning_rate": 5.4391376415588554e-05, "loss": 0.9518, "step": 4908 }, { "epoch": 2.2232789855072466, "grad_norm": 0.16789114114711334, "learning_rate": 5.43795753044689e-05, "loss": 0.9506, "step": 4909 }, { "epoch": 2.223731884057971, "grad_norm": 0.1334923752365794, "learning_rate": 5.4367772755800324e-05, "loss": 0.9303, "step": 4910 }, { "epoch": 2.224184782608696, "grad_norm": 0.10760468834504769, "learning_rate": 5.435596877076273e-05, "loss": 0.9221, "step": 4911 }, { "epoch": 2.2246376811594204, "grad_norm": 0.09188768632500076, "learning_rate": 5.4344163350536196e-05, "loss": 0.944, "step": 4912 }, { "epoch": 2.225090579710145, "grad_norm": 0.11906976672264986, "learning_rate": 5.4332356496300925e-05, "loss": 0.9391, "step": 4913 }, { "epoch": 2.2255434782608696, "grad_norm": 0.13031376158796784, "learning_rate": 5.432054820923725e-05, "loss": 0.9319, "step": 4914 }, { "epoch": 2.2259963768115942, "grad_norm": 0.09614363178706195, "learning_rate": 5.430873849052569e-05, "loss": 0.9208, "step": 4915 }, { "epoch": 2.226449275362319, "grad_norm": 0.1083217618027032, "learning_rate": 5.4296927341346864e-05, "loss": 0.9158, "step": 4916 }, { "epoch": 2.2269021739130435, "grad_norm": 0.12251425699804834, "learning_rate": 5.4285114762881564e-05, "loss": 0.95, "step": 4917 }, { "epoch": 2.227355072463768, "grad_norm": 0.08840568171623736, "learning_rate": 5.427330075631069e-05, "loss": 0.9466, "step": 4918 }, { "epoch": 2.2278079710144927, "grad_norm": 0.08684091482612913, "learning_rate": 5.426148532281532e-05, "loss": 0.9265, "step": 4919 }, { "epoch": 2.2282608695652173, "grad_norm": 0.1027104715940045, "learning_rate": 5.424966846357668e-05, "loss": 0.9288, "step": 4920 }, { "epoch": 2.228713768115942, "grad_norm": 0.11320977361352913, "learning_rate": 5.42378501797761e-05, "loss": 0.9284, "step": 4921 }, { "epoch": 2.2291666666666665, "grad_norm": 0.11380949200676682, "learning_rate": 5.4226030472595075e-05, "loss": 0.9658, "step": 4922 }, { "epoch": 2.229619565217391, "grad_norm": 0.08202160779128088, "learning_rate": 5.4214209343215234e-05, "loss": 0.9285, "step": 4923 }, { "epoch": 2.2300724637681157, "grad_norm": 0.10589571832600168, "learning_rate": 5.420238679281836e-05, "loss": 0.9281, "step": 4924 }, { "epoch": 2.230525362318841, "grad_norm": 0.09837663180261634, "learning_rate": 5.419056282258637e-05, "loss": 0.9434, "step": 4925 }, { "epoch": 2.2309782608695654, "grad_norm": 0.07588405700024388, "learning_rate": 5.417873743370134e-05, "loss": 0.9441, "step": 4926 }, { "epoch": 2.23143115942029, "grad_norm": 0.11823632350131313, "learning_rate": 5.416691062734545e-05, "loss": 0.9278, "step": 4927 }, { "epoch": 2.2318840579710146, "grad_norm": 0.13604268934351124, "learning_rate": 5.4155082404701056e-05, "loss": 0.9381, "step": 4928 }, { "epoch": 2.2323369565217392, "grad_norm": 0.09787714466882717, "learning_rate": 5.4143252766950653e-05, "loss": 0.9316, "step": 4929 }, { "epoch": 2.232789855072464, "grad_norm": 0.1126809877521636, "learning_rate": 5.4131421715276856e-05, "loss": 0.9429, "step": 4930 }, { "epoch": 2.2332427536231885, "grad_norm": 0.1200587062289803, "learning_rate": 5.411958925086243e-05, "loss": 0.948, "step": 4931 }, { "epoch": 2.233695652173913, "grad_norm": 0.09719965430856066, "learning_rate": 5.410775537489029e-05, "loss": 0.9133, "step": 4932 }, { "epoch": 2.2341485507246377, "grad_norm": 0.09630030260028787, "learning_rate": 5.40959200885435e-05, "loss": 0.9288, "step": 4933 }, { "epoch": 2.2346014492753623, "grad_norm": 0.10791482913671491, "learning_rate": 5.4084083393005235e-05, "loss": 0.9543, "step": 4934 }, { "epoch": 2.235054347826087, "grad_norm": 0.0903115086549773, "learning_rate": 5.407224528945884e-05, "loss": 0.9305, "step": 4935 }, { "epoch": 2.2355072463768115, "grad_norm": 0.11625699500288046, "learning_rate": 5.406040577908778e-05, "loss": 0.9353, "step": 4936 }, { "epoch": 2.235960144927536, "grad_norm": 0.11054130276370172, "learning_rate": 5.4048564863075674e-05, "loss": 0.9398, "step": 4937 }, { "epoch": 2.2364130434782608, "grad_norm": 0.12112986511821587, "learning_rate": 5.403672254260628e-05, "loss": 0.9274, "step": 4938 }, { "epoch": 2.2368659420289854, "grad_norm": 0.10938055091289776, "learning_rate": 5.402487881886349e-05, "loss": 0.9424, "step": 4939 }, { "epoch": 2.23731884057971, "grad_norm": 0.09799445570278188, "learning_rate": 5.4013033693031336e-05, "loss": 0.9455, "step": 4940 }, { "epoch": 2.2377717391304346, "grad_norm": 0.10759328637564751, "learning_rate": 5.4001187166294e-05, "loss": 0.9582, "step": 4941 }, { "epoch": 2.238224637681159, "grad_norm": 0.12913404300253647, "learning_rate": 5.3989339239835795e-05, "loss": 0.926, "step": 4942 }, { "epoch": 2.2386775362318843, "grad_norm": 0.11419791428853518, "learning_rate": 5.397748991484119e-05, "loss": 0.9399, "step": 4943 }, { "epoch": 2.239130434782609, "grad_norm": 0.09284359603916971, "learning_rate": 5.3965639192494746e-05, "loss": 0.9105, "step": 4944 }, { "epoch": 2.2395833333333335, "grad_norm": 0.11283486199540096, "learning_rate": 5.3953787073981236e-05, "loss": 0.9313, "step": 4945 }, { "epoch": 2.240036231884058, "grad_norm": 0.11749158432251107, "learning_rate": 5.394193356048551e-05, "loss": 0.9373, "step": 4946 }, { "epoch": 2.2404891304347827, "grad_norm": 0.11278171143430353, "learning_rate": 5.393007865319262e-05, "loss": 0.9246, "step": 4947 }, { "epoch": 2.2409420289855073, "grad_norm": 0.12464559236357237, "learning_rate": 5.391822235328767e-05, "loss": 0.945, "step": 4948 }, { "epoch": 2.241394927536232, "grad_norm": 0.11909101183771093, "learning_rate": 5.390636466195598e-05, "loss": 0.9232, "step": 4949 }, { "epoch": 2.2418478260869565, "grad_norm": 0.10842253675701877, "learning_rate": 5.389450558038297e-05, "loss": 0.9508, "step": 4950 }, { "epoch": 2.242300724637681, "grad_norm": 0.10706443027996436, "learning_rate": 5.3882645109754234e-05, "loss": 0.914, "step": 4951 }, { "epoch": 2.2427536231884058, "grad_norm": 0.10555198216970368, "learning_rate": 5.3870783251255464e-05, "loss": 0.9072, "step": 4952 }, { "epoch": 2.2432065217391304, "grad_norm": 0.10596163498696036, "learning_rate": 5.385892000607251e-05, "loss": 0.9242, "step": 4953 }, { "epoch": 2.243659420289855, "grad_norm": 0.11323532905926748, "learning_rate": 5.384705537539136e-05, "loss": 0.9456, "step": 4954 }, { "epoch": 2.2441123188405796, "grad_norm": 0.1378002962422579, "learning_rate": 5.3835189360398146e-05, "loss": 0.9326, "step": 4955 }, { "epoch": 2.244565217391304, "grad_norm": 0.11621574061724783, "learning_rate": 5.3823321962279126e-05, "loss": 0.9234, "step": 4956 }, { "epoch": 2.245018115942029, "grad_norm": 0.10666335096594709, "learning_rate": 5.3811453182220696e-05, "loss": 0.9271, "step": 4957 }, { "epoch": 2.2454710144927534, "grad_norm": 0.1041411775752192, "learning_rate": 5.3799583021409414e-05, "loss": 0.9452, "step": 4958 }, { "epoch": 2.2459239130434785, "grad_norm": 0.10379592686679263, "learning_rate": 5.378771148103195e-05, "loss": 0.9292, "step": 4959 }, { "epoch": 2.246376811594203, "grad_norm": 0.09988959533005061, "learning_rate": 5.377583856227512e-05, "loss": 0.9302, "step": 4960 }, { "epoch": 2.2468297101449277, "grad_norm": 0.09760304036448324, "learning_rate": 5.376396426632588e-05, "loss": 0.9289, "step": 4961 }, { "epoch": 2.2472826086956523, "grad_norm": 0.11566203086189539, "learning_rate": 5.3752088594371324e-05, "loss": 0.9501, "step": 4962 }, { "epoch": 2.247735507246377, "grad_norm": 0.10281934134758795, "learning_rate": 5.374021154759866e-05, "loss": 0.9513, "step": 4963 }, { "epoch": 2.2481884057971016, "grad_norm": 0.0935504406298784, "learning_rate": 5.372833312719531e-05, "loss": 0.9351, "step": 4964 }, { "epoch": 2.248641304347826, "grad_norm": 0.10160958920444031, "learning_rate": 5.3716453334348724e-05, "loss": 0.938, "step": 4965 }, { "epoch": 2.2490942028985508, "grad_norm": 0.10150979761521187, "learning_rate": 5.3704572170246545e-05, "loss": 0.9334, "step": 4966 }, { "epoch": 2.2495471014492754, "grad_norm": 0.1247362315609664, "learning_rate": 5.369268963607659e-05, "loss": 0.915, "step": 4967 }, { "epoch": 2.25, "grad_norm": 0.10534465466846242, "learning_rate": 5.368080573302676e-05, "loss": 0.9531, "step": 4968 }, { "epoch": 2.2504528985507246, "grad_norm": 0.09157021985292242, "learning_rate": 5.366892046228509e-05, "loss": 0.9322, "step": 4969 }, { "epoch": 2.2509057971014492, "grad_norm": 0.13426890849521897, "learning_rate": 5.365703382503977e-05, "loss": 0.9212, "step": 4970 }, { "epoch": 2.251358695652174, "grad_norm": 0.14771525726918733, "learning_rate": 5.364514582247916e-05, "loss": 0.9474, "step": 4971 }, { "epoch": 2.2518115942028984, "grad_norm": 0.09454650655285944, "learning_rate": 5.363325645579169e-05, "loss": 0.9209, "step": 4972 }, { "epoch": 2.252264492753623, "grad_norm": 0.0956162763334495, "learning_rate": 5.3621365726165966e-05, "loss": 0.9399, "step": 4973 }, { "epoch": 2.2527173913043477, "grad_norm": 0.0974685368379608, "learning_rate": 5.3609473634790724e-05, "loss": 0.9386, "step": 4974 }, { "epoch": 2.2531702898550723, "grad_norm": 0.09466866930929459, "learning_rate": 5.359758018285484e-05, "loss": 0.9335, "step": 4975 }, { "epoch": 2.253623188405797, "grad_norm": 0.13466696846275075, "learning_rate": 5.3585685371547326e-05, "loss": 0.9366, "step": 4976 }, { "epoch": 2.2540760869565215, "grad_norm": 0.14849233772226142, "learning_rate": 5.3573789202057306e-05, "loss": 0.9405, "step": 4977 }, { "epoch": 2.2545289855072466, "grad_norm": 0.10854079129136546, "learning_rate": 5.356189167557407e-05, "loss": 0.9433, "step": 4978 }, { "epoch": 2.254981884057971, "grad_norm": 0.09830649706375137, "learning_rate": 5.354999279328703e-05, "loss": 0.9465, "step": 4979 }, { "epoch": 2.255434782608696, "grad_norm": 0.09425399939109005, "learning_rate": 5.3538092556385754e-05, "loss": 0.9422, "step": 4980 }, { "epoch": 2.2558876811594204, "grad_norm": 0.16611985001715152, "learning_rate": 5.35261909660599e-05, "loss": 0.9477, "step": 4981 }, { "epoch": 2.256340579710145, "grad_norm": 0.1090384860652094, "learning_rate": 5.35142880234993e-05, "loss": 0.9334, "step": 4982 }, { "epoch": 2.2567934782608696, "grad_norm": 0.11441820836573785, "learning_rate": 5.350238372989392e-05, "loss": 0.9491, "step": 4983 }, { "epoch": 2.2572463768115942, "grad_norm": 0.09945401914135107, "learning_rate": 5.3490478086433836e-05, "loss": 0.934, "step": 4984 }, { "epoch": 2.257699275362319, "grad_norm": 0.09180814458110843, "learning_rate": 5.347857109430928e-05, "loss": 0.9467, "step": 4985 }, { "epoch": 2.2581521739130435, "grad_norm": 0.09481363419067299, "learning_rate": 5.346666275471061e-05, "loss": 0.9487, "step": 4986 }, { "epoch": 2.258605072463768, "grad_norm": 0.1010862253063532, "learning_rate": 5.345475306882833e-05, "loss": 0.9319, "step": 4987 }, { "epoch": 2.2590579710144927, "grad_norm": 0.10806071164391712, "learning_rate": 5.344284203785305e-05, "loss": 0.9143, "step": 4988 }, { "epoch": 2.2595108695652173, "grad_norm": 0.10990416557785794, "learning_rate": 5.343092966297558e-05, "loss": 0.953, "step": 4989 }, { "epoch": 2.259963768115942, "grad_norm": 0.11977521058677812, "learning_rate": 5.3419015945386766e-05, "loss": 0.9147, "step": 4990 }, { "epoch": 2.2604166666666665, "grad_norm": 0.11452061115173784, "learning_rate": 5.340710088627766e-05, "loss": 0.9319, "step": 4991 }, { "epoch": 2.260869565217391, "grad_norm": 0.11711939641622723, "learning_rate": 5.339518448683945e-05, "loss": 0.9297, "step": 4992 }, { "epoch": 2.261322463768116, "grad_norm": 0.11882791333448205, "learning_rate": 5.338326674826341e-05, "loss": 0.9257, "step": 4993 }, { "epoch": 2.261775362318841, "grad_norm": 0.11728921726867747, "learning_rate": 5.3371347671740994e-05, "loss": 0.9348, "step": 4994 }, { "epoch": 2.2622282608695654, "grad_norm": 0.09935862832789094, "learning_rate": 5.3359427258463745e-05, "loss": 0.9381, "step": 4995 }, { "epoch": 2.26268115942029, "grad_norm": 0.10709574430538642, "learning_rate": 5.33475055096234e-05, "loss": 0.9374, "step": 4996 }, { "epoch": 2.2631340579710146, "grad_norm": 0.09833789055499409, "learning_rate": 5.333558242641178e-05, "loss": 0.9552, "step": 4997 }, { "epoch": 2.2635869565217392, "grad_norm": 0.10848056764421979, "learning_rate": 5.332365801002084e-05, "loss": 0.9155, "step": 4998 }, { "epoch": 2.264039855072464, "grad_norm": 0.11869743066835958, "learning_rate": 5.33117322616427e-05, "loss": 0.9421, "step": 4999 }, { "epoch": 2.2644927536231885, "grad_norm": 0.10472437786840036, "learning_rate": 5.3299805182469594e-05, "loss": 0.9321, "step": 5000 }, { "epoch": 2.264945652173913, "grad_norm": 0.12283245091939565, "learning_rate": 5.328787677369389e-05, "loss": 0.9231, "step": 5001 }, { "epoch": 2.2653985507246377, "grad_norm": 0.14079160234277532, "learning_rate": 5.327594703650809e-05, "loss": 0.9529, "step": 5002 }, { "epoch": 2.2658514492753623, "grad_norm": 0.11720675347556218, "learning_rate": 5.3264015972104814e-05, "loss": 0.933, "step": 5003 }, { "epoch": 2.266304347826087, "grad_norm": 0.10236892435785058, "learning_rate": 5.325208358167685e-05, "loss": 0.9468, "step": 5004 }, { "epoch": 2.2667572463768115, "grad_norm": 0.09880359829660687, "learning_rate": 5.32401498664171e-05, "loss": 0.9397, "step": 5005 }, { "epoch": 2.267210144927536, "grad_norm": 0.12493884704091704, "learning_rate": 5.322821482751858e-05, "loss": 0.923, "step": 5006 }, { "epoch": 2.2676630434782608, "grad_norm": 0.14112986012420764, "learning_rate": 5.3216278466174456e-05, "loss": 0.9308, "step": 5007 }, { "epoch": 2.2681159420289854, "grad_norm": 0.12540209809414943, "learning_rate": 5.320434078357803e-05, "loss": 0.935, "step": 5008 }, { "epoch": 2.26856884057971, "grad_norm": 0.08590427806017321, "learning_rate": 5.3192401780922736e-05, "loss": 0.9158, "step": 5009 }, { "epoch": 2.2690217391304346, "grad_norm": 0.12370850535298758, "learning_rate": 5.318046145940214e-05, "loss": 0.9594, "step": 5010 }, { "epoch": 2.269474637681159, "grad_norm": 0.09758111593556168, "learning_rate": 5.316851982020992e-05, "loss": 0.9323, "step": 5011 }, { "epoch": 2.269927536231884, "grad_norm": 0.09261890870353583, "learning_rate": 5.31565768645399e-05, "loss": 0.9429, "step": 5012 }, { "epoch": 2.270380434782609, "grad_norm": 0.11510456885952856, "learning_rate": 5.3144632593586046e-05, "loss": 0.9477, "step": 5013 }, { "epoch": 2.2708333333333335, "grad_norm": 0.10878693140635153, "learning_rate": 5.3132687008542454e-05, "loss": 0.9526, "step": 5014 }, { "epoch": 2.271286231884058, "grad_norm": 0.09116619370171462, "learning_rate": 5.312074011060331e-05, "loss": 0.9268, "step": 5015 }, { "epoch": 2.2717391304347827, "grad_norm": 0.08633302929933723, "learning_rate": 5.3108791900963e-05, "loss": 0.9231, "step": 5016 }, { "epoch": 2.2721920289855073, "grad_norm": 0.1017486155423251, "learning_rate": 5.309684238081599e-05, "loss": 0.94, "step": 5017 }, { "epoch": 2.272644927536232, "grad_norm": 0.12776470479033747, "learning_rate": 5.3084891551356886e-05, "loss": 0.9328, "step": 5018 }, { "epoch": 2.2730978260869565, "grad_norm": 0.11534601622151748, "learning_rate": 5.307293941378045e-05, "loss": 0.9534, "step": 5019 }, { "epoch": 2.273550724637681, "grad_norm": 0.12030957351637062, "learning_rate": 5.306098596928154e-05, "loss": 0.936, "step": 5020 }, { "epoch": 2.2740036231884058, "grad_norm": 0.12290693277913295, "learning_rate": 5.304903121905516e-05, "loss": 0.9266, "step": 5021 }, { "epoch": 2.2744565217391304, "grad_norm": 0.10557771088214679, "learning_rate": 5.303707516429645e-05, "loss": 0.9367, "step": 5022 }, { "epoch": 2.274909420289855, "grad_norm": 0.10711006874472057, "learning_rate": 5.302511780620069e-05, "loss": 0.9449, "step": 5023 }, { "epoch": 2.2753623188405796, "grad_norm": 0.18242144256000783, "learning_rate": 5.301315914596325e-05, "loss": 0.9375, "step": 5024 }, { "epoch": 2.275815217391304, "grad_norm": 0.2056933402091454, "learning_rate": 5.3001199184779676e-05, "loss": 0.9366, "step": 5025 }, { "epoch": 2.276268115942029, "grad_norm": 0.18097916173088846, "learning_rate": 5.298923792384561e-05, "loss": 0.9394, "step": 5026 }, { "epoch": 2.2767210144927534, "grad_norm": 0.11223466751409124, "learning_rate": 5.297727536435686e-05, "loss": 0.9327, "step": 5027 }, { "epoch": 2.2771739130434785, "grad_norm": 0.08974655719609945, "learning_rate": 5.2965311507509314e-05, "loss": 0.9434, "step": 5028 }, { "epoch": 2.277626811594203, "grad_norm": 0.19310094643899617, "learning_rate": 5.2953346354499036e-05, "loss": 0.9583, "step": 5029 }, { "epoch": 2.2780797101449277, "grad_norm": 0.22651268539295633, "learning_rate": 5.294137990652219e-05, "loss": 0.9452, "step": 5030 }, { "epoch": 2.2785326086956523, "grad_norm": 0.16297229297658786, "learning_rate": 5.292941216477509e-05, "loss": 0.9267, "step": 5031 }, { "epoch": 2.278985507246377, "grad_norm": 0.09229960116374294, "learning_rate": 5.2917443130454174e-05, "loss": 0.9265, "step": 5032 }, { "epoch": 2.2794384057971016, "grad_norm": 0.12145057470343394, "learning_rate": 5.2905472804755995e-05, "loss": 0.9414, "step": 5033 }, { "epoch": 2.279891304347826, "grad_norm": 0.13105141621972244, "learning_rate": 5.289350118887725e-05, "loss": 0.9379, "step": 5034 }, { "epoch": 2.2803442028985508, "grad_norm": 0.10287770954727782, "learning_rate": 5.2881528284014764e-05, "loss": 0.9364, "step": 5035 }, { "epoch": 2.2807971014492754, "grad_norm": 0.09751351975763571, "learning_rate": 5.286955409136548e-05, "loss": 0.9541, "step": 5036 }, { "epoch": 2.28125, "grad_norm": 0.12124309578960159, "learning_rate": 5.2857578612126466e-05, "loss": 0.9445, "step": 5037 }, { "epoch": 2.2817028985507246, "grad_norm": 0.143411435652336, "learning_rate": 5.2845601847494963e-05, "loss": 0.9247, "step": 5038 }, { "epoch": 2.2821557971014492, "grad_norm": 0.1120627667178068, "learning_rate": 5.283362379866828e-05, "loss": 0.9445, "step": 5039 }, { "epoch": 2.282608695652174, "grad_norm": 0.10814860279047485, "learning_rate": 5.282164446684389e-05, "loss": 0.9361, "step": 5040 }, { "epoch": 2.2830615942028984, "grad_norm": 0.1177210367833744, "learning_rate": 5.280966385321939e-05, "loss": 0.9447, "step": 5041 }, { "epoch": 2.283514492753623, "grad_norm": 0.09957163730842036, "learning_rate": 5.279768195899248e-05, "loss": 0.9438, "step": 5042 }, { "epoch": 2.2839673913043477, "grad_norm": 0.08210633385597746, "learning_rate": 5.278569878536104e-05, "loss": 0.9314, "step": 5043 }, { "epoch": 2.2844202898550723, "grad_norm": 0.09922197925436474, "learning_rate": 5.277371433352302e-05, "loss": 0.9346, "step": 5044 }, { "epoch": 2.284873188405797, "grad_norm": 0.10648753026890616, "learning_rate": 5.2761728604676557e-05, "loss": 0.9112, "step": 5045 }, { "epoch": 2.2853260869565215, "grad_norm": 0.09414682341130046, "learning_rate": 5.274974160001984e-05, "loss": 0.9378, "step": 5046 }, { "epoch": 2.2857789855072466, "grad_norm": 0.08740716583982033, "learning_rate": 5.273775332075126e-05, "loss": 0.9305, "step": 5047 }, { "epoch": 2.286231884057971, "grad_norm": 0.09855531772306135, "learning_rate": 5.27257637680693e-05, "loss": 0.9517, "step": 5048 }, { "epoch": 2.286684782608696, "grad_norm": 0.12418456726748019, "learning_rate": 5.271377294317256e-05, "loss": 0.9149, "step": 5049 }, { "epoch": 2.2871376811594204, "grad_norm": 0.1294198753113423, "learning_rate": 5.2701780847259806e-05, "loss": 0.9313, "step": 5050 }, { "epoch": 2.287590579710145, "grad_norm": 0.10698153534083443, "learning_rate": 5.2689787481529885e-05, "loss": 0.9265, "step": 5051 }, { "epoch": 2.2880434782608696, "grad_norm": 0.11482924066693567, "learning_rate": 5.267779284718181e-05, "loss": 0.9647, "step": 5052 }, { "epoch": 2.2884963768115942, "grad_norm": 0.10493408083312368, "learning_rate": 5.266579694541468e-05, "loss": 0.9439, "step": 5053 }, { "epoch": 2.288949275362319, "grad_norm": 0.1064899635724097, "learning_rate": 5.265379977742776e-05, "loss": 0.9287, "step": 5054 }, { "epoch": 2.2894021739130435, "grad_norm": 0.09426817011560655, "learning_rate": 5.264180134442043e-05, "loss": 0.9186, "step": 5055 }, { "epoch": 2.289855072463768, "grad_norm": 0.08946051005514558, "learning_rate": 5.262980164759219e-05, "loss": 0.9379, "step": 5056 }, { "epoch": 2.2903079710144927, "grad_norm": 0.09575408209700255, "learning_rate": 5.261780068814267e-05, "loss": 0.9608, "step": 5057 }, { "epoch": 2.2907608695652173, "grad_norm": 0.08243064768232013, "learning_rate": 5.2605798467271605e-05, "loss": 0.936, "step": 5058 }, { "epoch": 2.291213768115942, "grad_norm": 0.09051993833720373, "learning_rate": 5.25937949861789e-05, "loss": 0.9309, "step": 5059 }, { "epoch": 2.2916666666666665, "grad_norm": 0.10499794863171709, "learning_rate": 5.258179024606455e-05, "loss": 0.9248, "step": 5060 }, { "epoch": 2.292119565217391, "grad_norm": 0.09084376647486446, "learning_rate": 5.25697842481287e-05, "loss": 0.9292, "step": 5061 }, { "epoch": 2.292572463768116, "grad_norm": 0.09161971807050427, "learning_rate": 5.255777699357159e-05, "loss": 0.9352, "step": 5062 }, { "epoch": 2.293025362318841, "grad_norm": 0.3353432786608795, "learning_rate": 5.254576848359362e-05, "loss": 0.9481, "step": 5063 }, { "epoch": 2.2934782608695654, "grad_norm": 0.09892486873360845, "learning_rate": 5.2533758719395294e-05, "loss": 0.9484, "step": 5064 }, { "epoch": 2.29393115942029, "grad_norm": 0.12546145794484576, "learning_rate": 5.2521747702177246e-05, "loss": 0.9683, "step": 5065 }, { "epoch": 2.2943840579710146, "grad_norm": 0.11679688044245591, "learning_rate": 5.2509735433140244e-05, "loss": 0.9389, "step": 5066 }, { "epoch": 2.2948369565217392, "grad_norm": 0.10330926151873175, "learning_rate": 5.2497721913485163e-05, "loss": 0.9352, "step": 5067 }, { "epoch": 2.295289855072464, "grad_norm": 0.11085110945210401, "learning_rate": 5.248570714441301e-05, "loss": 0.9307, "step": 5068 }, { "epoch": 2.2957427536231885, "grad_norm": 0.1194622972591039, "learning_rate": 5.247369112712495e-05, "loss": 0.9441, "step": 5069 }, { "epoch": 2.296195652173913, "grad_norm": 0.11398078673148641, "learning_rate": 5.2461673862822196e-05, "loss": 0.9315, "step": 5070 }, { "epoch": 2.2966485507246377, "grad_norm": 0.1005174171920614, "learning_rate": 5.244965535270617e-05, "loss": 0.9592, "step": 5071 }, { "epoch": 2.2971014492753623, "grad_norm": 0.07873726601063032, "learning_rate": 5.243763559797836e-05, "loss": 0.9498, "step": 5072 }, { "epoch": 2.297554347826087, "grad_norm": 0.10542465908526912, "learning_rate": 5.2425614599840436e-05, "loss": 0.9448, "step": 5073 }, { "epoch": 2.2980072463768115, "grad_norm": 0.10583890041959265, "learning_rate": 5.2413592359494105e-05, "loss": 0.9344, "step": 5074 }, { "epoch": 2.298460144927536, "grad_norm": 0.07869305136696869, "learning_rate": 5.240156887814129e-05, "loss": 0.9421, "step": 5075 }, { "epoch": 2.2989130434782608, "grad_norm": 0.10019465929173346, "learning_rate": 5.2389544156983975e-05, "loss": 0.9342, "step": 5076 }, { "epoch": 2.2993659420289854, "grad_norm": 0.10599589300148628, "learning_rate": 5.2377518197224304e-05, "loss": 0.8966, "step": 5077 }, { "epoch": 2.29981884057971, "grad_norm": 0.07493987691543531, "learning_rate": 5.236549100006452e-05, "loss": 0.9278, "step": 5078 }, { "epoch": 2.3002717391304346, "grad_norm": 0.09675712353874813, "learning_rate": 5.235346256670701e-05, "loss": 0.9308, "step": 5079 }, { "epoch": 2.300724637681159, "grad_norm": 0.10522670131505854, "learning_rate": 5.234143289835426e-05, "loss": 0.9171, "step": 5080 }, { "epoch": 2.301177536231884, "grad_norm": 0.09507372009177975, "learning_rate": 5.232940199620891e-05, "loss": 0.9329, "step": 5081 }, { "epoch": 2.301630434782609, "grad_norm": 0.12892702264554556, "learning_rate": 5.231736986147372e-05, "loss": 0.9581, "step": 5082 }, { "epoch": 2.3020833333333335, "grad_norm": 0.13236746992365528, "learning_rate": 5.2305336495351536e-05, "loss": 0.929, "step": 5083 }, { "epoch": 2.302536231884058, "grad_norm": 0.10306606832062826, "learning_rate": 5.2293301899045363e-05, "loss": 0.9588, "step": 5084 }, { "epoch": 2.3029891304347827, "grad_norm": 0.09549724626706142, "learning_rate": 5.228126607375831e-05, "loss": 0.95, "step": 5085 }, { "epoch": 2.3034420289855073, "grad_norm": 0.12788972663060644, "learning_rate": 5.2269229020693635e-05, "loss": 0.9328, "step": 5086 }, { "epoch": 2.303894927536232, "grad_norm": 0.15794048563129628, "learning_rate": 5.2257190741054697e-05, "loss": 0.9675, "step": 5087 }, { "epoch": 2.3043478260869565, "grad_norm": 0.15530806969072944, "learning_rate": 5.2245151236044974e-05, "loss": 0.9288, "step": 5088 }, { "epoch": 2.304800724637681, "grad_norm": 0.13537966860706827, "learning_rate": 5.223311050686807e-05, "loss": 0.9341, "step": 5089 }, { "epoch": 2.3052536231884058, "grad_norm": 0.14365936301077803, "learning_rate": 5.222106855472774e-05, "loss": 0.9442, "step": 5090 }, { "epoch": 2.3057065217391304, "grad_norm": 0.12425176569253402, "learning_rate": 5.220902538082781e-05, "loss": 0.9324, "step": 5091 }, { "epoch": 2.306159420289855, "grad_norm": 0.2031442029243552, "learning_rate": 5.219698098637226e-05, "loss": 0.9366, "step": 5092 }, { "epoch": 2.3066123188405796, "grad_norm": 0.19314374740057233, "learning_rate": 5.2184935372565204e-05, "loss": 0.9026, "step": 5093 }, { "epoch": 2.307065217391304, "grad_norm": 0.1495091415700564, "learning_rate": 5.217288854061085e-05, "loss": 0.9464, "step": 5094 }, { "epoch": 2.307518115942029, "grad_norm": 0.1361976561842333, "learning_rate": 5.216084049171353e-05, "loss": 0.9383, "step": 5095 }, { "epoch": 2.3079710144927534, "grad_norm": 0.13800591253025934, "learning_rate": 5.2148791227077715e-05, "loss": 0.9032, "step": 5096 }, { "epoch": 2.3084239130434785, "grad_norm": 0.18967823691200128, "learning_rate": 5.2136740747907996e-05, "loss": 0.9217, "step": 5097 }, { "epoch": 2.308876811594203, "grad_norm": 0.17311470313145733, "learning_rate": 5.212468905540907e-05, "loss": 0.9304, "step": 5098 }, { "epoch": 2.3093297101449277, "grad_norm": 0.12303649039363646, "learning_rate": 5.2112636150785774e-05, "loss": 0.9306, "step": 5099 }, { "epoch": 2.3097826086956523, "grad_norm": 0.10576513689110101, "learning_rate": 5.210058203524304e-05, "loss": 0.9317, "step": 5100 }, { "epoch": 2.310235507246377, "grad_norm": 0.12621113032985876, "learning_rate": 5.208852670998594e-05, "loss": 0.9193, "step": 5101 }, { "epoch": 2.3106884057971016, "grad_norm": 0.12893534541967247, "learning_rate": 5.207647017621967e-05, "loss": 0.9128, "step": 5102 }, { "epoch": 2.311141304347826, "grad_norm": 0.10330168368488471, "learning_rate": 5.206441243514955e-05, "loss": 0.9456, "step": 5103 }, { "epoch": 2.3115942028985508, "grad_norm": 0.1176325387641481, "learning_rate": 5.2052353487981e-05, "loss": 0.9442, "step": 5104 }, { "epoch": 2.3120471014492754, "grad_norm": 0.09354321066001661, "learning_rate": 5.204029333591956e-05, "loss": 0.9383, "step": 5105 }, { "epoch": 2.3125, "grad_norm": 0.10039395643304619, "learning_rate": 5.202823198017092e-05, "loss": 0.9279, "step": 5106 }, { "epoch": 2.3129528985507246, "grad_norm": 0.12017179330323234, "learning_rate": 5.201616942194088e-05, "loss": 0.9287, "step": 5107 }, { "epoch": 2.3134057971014492, "grad_norm": 0.12478230153155155, "learning_rate": 5.200410566243534e-05, "loss": 0.9345, "step": 5108 }, { "epoch": 2.313858695652174, "grad_norm": 0.09319563585351259, "learning_rate": 5.1992040702860325e-05, "loss": 0.9142, "step": 5109 }, { "epoch": 2.3143115942028984, "grad_norm": 0.12218519250667433, "learning_rate": 5.1979974544422e-05, "loss": 0.9498, "step": 5110 }, { "epoch": 2.314764492753623, "grad_norm": 0.10394187454825694, "learning_rate": 5.196790718832665e-05, "loss": 0.9301, "step": 5111 }, { "epoch": 2.3152173913043477, "grad_norm": 0.12768841060307973, "learning_rate": 5.195583863578064e-05, "loss": 0.9479, "step": 5112 }, { "epoch": 2.3156702898550723, "grad_norm": 0.09835481004346204, "learning_rate": 5.194376888799049e-05, "loss": 0.9142, "step": 5113 }, { "epoch": 2.316123188405797, "grad_norm": 0.10127085987360819, "learning_rate": 5.193169794616285e-05, "loss": 0.9219, "step": 5114 }, { "epoch": 2.3165760869565215, "grad_norm": 0.1124579652293973, "learning_rate": 5.191962581150446e-05, "loss": 0.9446, "step": 5115 }, { "epoch": 2.3170289855072466, "grad_norm": 0.10461946322486815, "learning_rate": 5.190755248522218e-05, "loss": 0.9527, "step": 5116 }, { "epoch": 2.317481884057971, "grad_norm": 0.08462041235167836, "learning_rate": 5.1895477968523013e-05, "loss": 0.9181, "step": 5117 }, { "epoch": 2.317934782608696, "grad_norm": 0.08303341238451942, "learning_rate": 5.188340226261406e-05, "loss": 0.9251, "step": 5118 }, { "epoch": 2.3183876811594204, "grad_norm": 0.093128464185023, "learning_rate": 5.1871325368702553e-05, "loss": 0.9254, "step": 5119 }, { "epoch": 2.318840579710145, "grad_norm": 0.07807855948347181, "learning_rate": 5.1859247287995846e-05, "loss": 0.9483, "step": 5120 }, { "epoch": 2.3192934782608696, "grad_norm": 0.09320903045555098, "learning_rate": 5.1847168021701384e-05, "loss": 0.9415, "step": 5121 }, { "epoch": 2.3197463768115942, "grad_norm": 0.09030252116016088, "learning_rate": 5.183508757102676e-05, "loss": 0.9546, "step": 5122 }, { "epoch": 2.320199275362319, "grad_norm": 0.10078170675464328, "learning_rate": 5.1823005937179684e-05, "loss": 0.9471, "step": 5123 }, { "epoch": 2.3206521739130435, "grad_norm": 0.09363435237796822, "learning_rate": 5.1810923121367975e-05, "loss": 0.9219, "step": 5124 }, { "epoch": 2.321105072463768, "grad_norm": 0.10693712497524205, "learning_rate": 5.1798839124799554e-05, "loss": 0.9376, "step": 5125 }, { "epoch": 2.3215579710144927, "grad_norm": 0.0829003205121708, "learning_rate": 5.1786753948682486e-05, "loss": 0.9552, "step": 5126 }, { "epoch": 2.3220108695652173, "grad_norm": 0.08672689622746958, "learning_rate": 5.177466759422496e-05, "loss": 0.9269, "step": 5127 }, { "epoch": 2.322463768115942, "grad_norm": 0.08171024773765057, "learning_rate": 5.176258006263526e-05, "loss": 0.927, "step": 5128 }, { "epoch": 2.3229166666666665, "grad_norm": 0.08969618819093395, "learning_rate": 5.1750491355121776e-05, "loss": 0.9378, "step": 5129 }, { "epoch": 2.323369565217391, "grad_norm": 0.0968614761444213, "learning_rate": 5.1738401472893055e-05, "loss": 0.9371, "step": 5130 }, { "epoch": 2.323822463768116, "grad_norm": 0.0875454960847928, "learning_rate": 5.1726310417157744e-05, "loss": 0.929, "step": 5131 }, { "epoch": 2.324275362318841, "grad_norm": 0.09889495238505377, "learning_rate": 5.1714218189124605e-05, "loss": 0.9451, "step": 5132 }, { "epoch": 2.3247282608695654, "grad_norm": 0.09734121902987881, "learning_rate": 5.1702124790002497e-05, "loss": 0.9466, "step": 5133 }, { "epoch": 2.32518115942029, "grad_norm": 0.08291636269618535, "learning_rate": 5.1690030221000434e-05, "loss": 0.9528, "step": 5134 }, { "epoch": 2.3256340579710146, "grad_norm": 0.1020227137410204, "learning_rate": 5.1677934483327525e-05, "loss": 0.9398, "step": 5135 }, { "epoch": 2.3260869565217392, "grad_norm": 0.08318287986273634, "learning_rate": 5.1665837578193e-05, "loss": 0.946, "step": 5136 }, { "epoch": 2.326539855072464, "grad_norm": 0.3152726979736329, "learning_rate": 5.1653739506806216e-05, "loss": 0.9255, "step": 5137 }, { "epoch": 2.3269927536231885, "grad_norm": 0.0990094469179097, "learning_rate": 5.164164027037661e-05, "loss": 0.9464, "step": 5138 }, { "epoch": 2.327445652173913, "grad_norm": 0.0846087236977147, "learning_rate": 5.1629539870113786e-05, "loss": 0.9436, "step": 5139 }, { "epoch": 2.3278985507246377, "grad_norm": 0.09531096120182181, "learning_rate": 5.161743830722743e-05, "loss": 0.9259, "step": 5140 }, { "epoch": 2.3283514492753623, "grad_norm": 0.0857722207359222, "learning_rate": 5.160533558292738e-05, "loss": 0.944, "step": 5141 }, { "epoch": 2.328804347826087, "grad_norm": 0.07657165852879588, "learning_rate": 5.159323169842352e-05, "loss": 0.9266, "step": 5142 }, { "epoch": 2.3292572463768115, "grad_norm": 0.08126130686673413, "learning_rate": 5.158112665492591e-05, "loss": 0.9463, "step": 5143 }, { "epoch": 2.329710144927536, "grad_norm": 0.09193341700589731, "learning_rate": 5.1569020453644726e-05, "loss": 0.958, "step": 5144 }, { "epoch": 2.3301630434782608, "grad_norm": 0.11673411180720875, "learning_rate": 5.1556913095790234e-05, "loss": 0.9373, "step": 5145 }, { "epoch": 2.3306159420289854, "grad_norm": 0.10557406450483965, "learning_rate": 5.154480458257283e-05, "loss": 0.9479, "step": 5146 }, { "epoch": 2.33106884057971, "grad_norm": 0.1007594912986818, "learning_rate": 5.1532694915203005e-05, "loss": 0.9406, "step": 5147 }, { "epoch": 2.3315217391304346, "grad_norm": 0.08932278984319178, "learning_rate": 5.15205840948914e-05, "loss": 0.9225, "step": 5148 }, { "epoch": 2.331974637681159, "grad_norm": 0.090781317415309, "learning_rate": 5.150847212284875e-05, "loss": 0.9364, "step": 5149 }, { "epoch": 2.332427536231884, "grad_norm": 0.09724101860861478, "learning_rate": 5.149635900028591e-05, "loss": 0.9507, "step": 5150 }, { "epoch": 2.332880434782609, "grad_norm": 0.12057972989566072, "learning_rate": 5.1484244728413834e-05, "loss": 0.9223, "step": 5151 }, { "epoch": 2.3333333333333335, "grad_norm": 0.0861640624290816, "learning_rate": 5.1472129308443616e-05, "loss": 0.9543, "step": 5152 }, { "epoch": 2.333786231884058, "grad_norm": 0.08321968821187756, "learning_rate": 5.146001274158646e-05, "loss": 0.9341, "step": 5153 }, { "epoch": 2.3342391304347827, "grad_norm": 0.11022090712083477, "learning_rate": 5.144789502905366e-05, "loss": 0.9268, "step": 5154 }, { "epoch": 2.3346920289855073, "grad_norm": 0.13404583664716324, "learning_rate": 5.1435776172056665e-05, "loss": 0.9374, "step": 5155 }, { "epoch": 2.335144927536232, "grad_norm": 0.10504114326912627, "learning_rate": 5.1423656171806986e-05, "loss": 0.9355, "step": 5156 }, { "epoch": 2.3355978260869565, "grad_norm": 0.11176755575826551, "learning_rate": 5.141153502951631e-05, "loss": 0.9156, "step": 5157 }, { "epoch": 2.336050724637681, "grad_norm": 0.09711506075668871, "learning_rate": 5.1399412746396394e-05, "loss": 0.9395, "step": 5158 }, { "epoch": 2.3365036231884058, "grad_norm": 0.08551878564542086, "learning_rate": 5.1387289323659114e-05, "loss": 0.9197, "step": 5159 }, { "epoch": 2.3369565217391304, "grad_norm": 0.09797671415081267, "learning_rate": 5.137516476251648e-05, "loss": 0.938, "step": 5160 }, { "epoch": 2.337409420289855, "grad_norm": 0.09433358232390009, "learning_rate": 5.13630390641806e-05, "loss": 0.9279, "step": 5161 }, { "epoch": 2.3378623188405796, "grad_norm": 0.1003991292955644, "learning_rate": 5.1350912229863704e-05, "loss": 0.9291, "step": 5162 }, { "epoch": 2.338315217391304, "grad_norm": 0.08518249936749672, "learning_rate": 5.133878426077812e-05, "loss": 0.9514, "step": 5163 }, { "epoch": 2.338768115942029, "grad_norm": 0.08161679646527505, "learning_rate": 5.132665515813631e-05, "loss": 0.9458, "step": 5164 }, { "epoch": 2.3392210144927534, "grad_norm": 0.08772014066312535, "learning_rate": 5.131452492315083e-05, "loss": 0.9174, "step": 5165 }, { "epoch": 2.3396739130434785, "grad_norm": 0.08406930562130327, "learning_rate": 5.130239355703436e-05, "loss": 0.9362, "step": 5166 }, { "epoch": 2.340126811594203, "grad_norm": 0.09712021556400094, "learning_rate": 5.129026106099971e-05, "loss": 0.9438, "step": 5167 }, { "epoch": 2.3405797101449277, "grad_norm": 0.11664156272930136, "learning_rate": 5.127812743625976e-05, "loss": 0.912, "step": 5168 }, { "epoch": 2.3410326086956523, "grad_norm": 0.0868994565121893, "learning_rate": 5.126599268402753e-05, "loss": 0.9362, "step": 5169 }, { "epoch": 2.341485507246377, "grad_norm": 0.08821446515131232, "learning_rate": 5.125385680551617e-05, "loss": 0.9603, "step": 5170 }, { "epoch": 2.3419384057971016, "grad_norm": 0.09148958245461868, "learning_rate": 5.124171980193891e-05, "loss": 0.9303, "step": 5171 }, { "epoch": 2.342391304347826, "grad_norm": 0.09246505445838693, "learning_rate": 5.122958167450911e-05, "loss": 0.9583, "step": 5172 }, { "epoch": 2.3428442028985508, "grad_norm": 0.08868407733417544, "learning_rate": 5.121744242444023e-05, "loss": 0.933, "step": 5173 }, { "epoch": 2.3432971014492754, "grad_norm": 0.08763796957902718, "learning_rate": 5.1205302052945856e-05, "loss": 0.9477, "step": 5174 }, { "epoch": 2.34375, "grad_norm": 0.10495536296984467, "learning_rate": 5.1193160561239694e-05, "loss": 0.9234, "step": 5175 }, { "epoch": 2.3442028985507246, "grad_norm": 0.1028851209099174, "learning_rate": 5.118101795053552e-05, "loss": 0.9447, "step": 5176 }, { "epoch": 2.3446557971014492, "grad_norm": 0.07767500968900685, "learning_rate": 5.116887422204726e-05, "loss": 0.9265, "step": 5177 }, { "epoch": 2.345108695652174, "grad_norm": 0.08925960911010924, "learning_rate": 5.115672937698895e-05, "loss": 0.9349, "step": 5178 }, { "epoch": 2.3455615942028984, "grad_norm": 0.09066018387641686, "learning_rate": 5.114458341657473e-05, "loss": 0.9201, "step": 5179 }, { "epoch": 2.346014492753623, "grad_norm": 0.08342806835021352, "learning_rate": 5.113243634201884e-05, "loss": 0.916, "step": 5180 }, { "epoch": 2.3464673913043477, "grad_norm": 0.08636363009492726, "learning_rate": 5.112028815453565e-05, "loss": 0.9447, "step": 5181 }, { "epoch": 2.3469202898550723, "grad_norm": 0.09526931144625993, "learning_rate": 5.1108138855339633e-05, "loss": 0.9337, "step": 5182 }, { "epoch": 2.347373188405797, "grad_norm": 0.12363026015635202, "learning_rate": 5.109598844564537e-05, "loss": 0.9465, "step": 5183 }, { "epoch": 2.3478260869565215, "grad_norm": 0.15037257804689075, "learning_rate": 5.108383692666756e-05, "loss": 0.9148, "step": 5184 }, { "epoch": 2.3482789855072466, "grad_norm": 0.162824380514373, "learning_rate": 5.107168429962101e-05, "loss": 0.9507, "step": 5185 }, { "epoch": 2.348731884057971, "grad_norm": 0.08800931928083269, "learning_rate": 5.1059530565720624e-05, "loss": 0.928, "step": 5186 }, { "epoch": 2.349184782608696, "grad_norm": 0.1248222535999134, "learning_rate": 5.104737572618145e-05, "loss": 0.9554, "step": 5187 }, { "epoch": 2.3496376811594204, "grad_norm": 0.16293555831183748, "learning_rate": 5.103521978221863e-05, "loss": 0.9347, "step": 5188 }, { "epoch": 2.350090579710145, "grad_norm": 0.13649674927900365, "learning_rate": 5.102306273504739e-05, "loss": 0.915, "step": 5189 }, { "epoch": 2.3505434782608696, "grad_norm": 0.12411300572069972, "learning_rate": 5.10109045858831e-05, "loss": 0.9321, "step": 5190 }, { "epoch": 2.3509963768115942, "grad_norm": 0.13479607619938377, "learning_rate": 5.099874533594123e-05, "loss": 0.9475, "step": 5191 }, { "epoch": 2.351449275362319, "grad_norm": 0.1312660186114406, "learning_rate": 5.0986584986437366e-05, "loss": 0.9591, "step": 5192 }, { "epoch": 2.3519021739130435, "grad_norm": 0.09910905324547754, "learning_rate": 5.0974423538587194e-05, "loss": 0.9297, "step": 5193 }, { "epoch": 2.352355072463768, "grad_norm": 0.11674814424326843, "learning_rate": 5.09622609936065e-05, "loss": 0.9209, "step": 5194 }, { "epoch": 2.3528079710144927, "grad_norm": 0.11983124067575615, "learning_rate": 5.09500973527112e-05, "loss": 0.9418, "step": 5195 }, { "epoch": 2.3532608695652173, "grad_norm": 0.11865756957172523, "learning_rate": 5.093793261711733e-05, "loss": 0.9251, "step": 5196 }, { "epoch": 2.353713768115942, "grad_norm": 0.13744833436281706, "learning_rate": 5.0925766788041e-05, "loss": 0.9421, "step": 5197 }, { "epoch": 2.3541666666666665, "grad_norm": 0.12887026787380668, "learning_rate": 5.091359986669845e-05, "loss": 0.9409, "step": 5198 }, { "epoch": 2.354619565217391, "grad_norm": 0.09236418211500479, "learning_rate": 5.0901431854306024e-05, "loss": 0.9376, "step": 5199 }, { "epoch": 2.355072463768116, "grad_norm": 0.09366542142348876, "learning_rate": 5.0889262752080185e-05, "loss": 0.943, "step": 5200 }, { "epoch": 2.355525362318841, "grad_norm": 0.11909468413474654, "learning_rate": 5.0877092561237484e-05, "loss": 0.9467, "step": 5201 }, { "epoch": 2.3559782608695654, "grad_norm": 0.11567508547372365, "learning_rate": 5.086492128299461e-05, "loss": 0.9327, "step": 5202 }, { "epoch": 2.35643115942029, "grad_norm": 0.11858041487841924, "learning_rate": 5.0852748918568335e-05, "loss": 0.9254, "step": 5203 }, { "epoch": 2.3568840579710146, "grad_norm": 0.08403971943394505, "learning_rate": 5.084057546917557e-05, "loss": 0.9356, "step": 5204 }, { "epoch": 2.3573369565217392, "grad_norm": 0.09836853285508276, "learning_rate": 5.082840093603327e-05, "loss": 0.9186, "step": 5205 }, { "epoch": 2.357789855072464, "grad_norm": 0.09944511658271928, "learning_rate": 5.0816225320358586e-05, "loss": 0.9398, "step": 5206 }, { "epoch": 2.3582427536231885, "grad_norm": 0.1130272205942115, "learning_rate": 5.08040486233687e-05, "loss": 0.947, "step": 5207 }, { "epoch": 2.358695652173913, "grad_norm": 0.09677230355624747, "learning_rate": 5.0791870846280974e-05, "loss": 0.9388, "step": 5208 }, { "epoch": 2.3591485507246377, "grad_norm": 0.08005199377241408, "learning_rate": 5.077969199031282e-05, "loss": 0.9438, "step": 5209 }, { "epoch": 2.3596014492753623, "grad_norm": 0.10011102433546198, "learning_rate": 5.076751205668176e-05, "loss": 0.9432, "step": 5210 }, { "epoch": 2.360054347826087, "grad_norm": 0.13301264824960496, "learning_rate": 5.075533104660546e-05, "loss": 0.9348, "step": 5211 }, { "epoch": 2.3605072463768115, "grad_norm": 0.10406877295970669, "learning_rate": 5.074314896130167e-05, "loss": 0.9407, "step": 5212 }, { "epoch": 2.360960144927536, "grad_norm": 0.09316285238888433, "learning_rate": 5.073096580198827e-05, "loss": 0.913, "step": 5213 }, { "epoch": 2.3614130434782608, "grad_norm": 0.11393492930131069, "learning_rate": 5.071878156988319e-05, "loss": 0.9396, "step": 5214 }, { "epoch": 2.3618659420289854, "grad_norm": 0.15219919602330867, "learning_rate": 5.0706596266204555e-05, "loss": 0.9419, "step": 5215 }, { "epoch": 2.36231884057971, "grad_norm": 0.1505440514347863, "learning_rate": 5.069440989217051e-05, "loss": 0.9819, "step": 5216 }, { "epoch": 2.3627717391304346, "grad_norm": 0.11520135980790498, "learning_rate": 5.068222244899938e-05, "loss": 0.9371, "step": 5217 }, { "epoch": 2.363224637681159, "grad_norm": 0.09274507175786163, "learning_rate": 5.067003393790952e-05, "loss": 0.9163, "step": 5218 }, { "epoch": 2.363677536231884, "grad_norm": 0.09697094499114504, "learning_rate": 5.065784436011947e-05, "loss": 0.9299, "step": 5219 }, { "epoch": 2.364130434782609, "grad_norm": 0.09974999915625317, "learning_rate": 5.064565371684783e-05, "loss": 0.943, "step": 5220 }, { "epoch": 2.3645833333333335, "grad_norm": 0.1013690643576167, "learning_rate": 5.0633462009313315e-05, "loss": 0.9489, "step": 5221 }, { "epoch": 2.365036231884058, "grad_norm": 0.1185053035170641, "learning_rate": 5.0621269238734764e-05, "loss": 0.9032, "step": 5222 }, { "epoch": 2.3654891304347827, "grad_norm": 0.14113022179163404, "learning_rate": 5.0609075406331084e-05, "loss": 0.9591, "step": 5223 }, { "epoch": 2.3659420289855073, "grad_norm": 0.10295308488912502, "learning_rate": 5.059688051332132e-05, "loss": 0.9402, "step": 5224 }, { "epoch": 2.366394927536232, "grad_norm": 0.08646998197573062, "learning_rate": 5.058468456092464e-05, "loss": 0.9456, "step": 5225 }, { "epoch": 2.3668478260869565, "grad_norm": 0.09442503372168458, "learning_rate": 5.0572487550360256e-05, "loss": 0.9353, "step": 5226 }, { "epoch": 2.367300724637681, "grad_norm": 0.1274883967431796, "learning_rate": 5.0560289482847534e-05, "loss": 0.9267, "step": 5227 }, { "epoch": 2.3677536231884058, "grad_norm": 0.16345202741505316, "learning_rate": 5.054809035960595e-05, "loss": 0.9366, "step": 5228 }, { "epoch": 2.3682065217391304, "grad_norm": 0.1420872025310281, "learning_rate": 5.0535890181855044e-05, "loss": 0.9404, "step": 5229 }, { "epoch": 2.368659420289855, "grad_norm": 0.13321525865271672, "learning_rate": 5.052368895081453e-05, "loss": 0.9411, "step": 5230 }, { "epoch": 2.3691123188405796, "grad_norm": 0.12135122921314072, "learning_rate": 5.051148666770413e-05, "loss": 0.9566, "step": 5231 }, { "epoch": 2.369565217391304, "grad_norm": 0.10649319100498683, "learning_rate": 5.0499283333743754e-05, "loss": 0.9342, "step": 5232 }, { "epoch": 2.370018115942029, "grad_norm": 0.14695422834813218, "learning_rate": 5.048707895015338e-05, "loss": 0.9194, "step": 5233 }, { "epoch": 2.3704710144927534, "grad_norm": 0.13548692657136394, "learning_rate": 5.047487351815312e-05, "loss": 0.9355, "step": 5234 }, { "epoch": 2.3709239130434785, "grad_norm": 0.10799384422614502, "learning_rate": 5.046266703896315e-05, "loss": 0.9381, "step": 5235 }, { "epoch": 2.371376811594203, "grad_norm": 0.08786690897515945, "learning_rate": 5.045045951380378e-05, "loss": 0.9404, "step": 5236 }, { "epoch": 2.3718297101449277, "grad_norm": 0.12608814878175392, "learning_rate": 5.0438250943895404e-05, "loss": 0.9509, "step": 5237 }, { "epoch": 2.3722826086956523, "grad_norm": 0.11568128242246933, "learning_rate": 5.042604133045855e-05, "loss": 0.9632, "step": 5238 }, { "epoch": 2.372735507246377, "grad_norm": 0.1000229318461058, "learning_rate": 5.0413830674713815e-05, "loss": 0.9639, "step": 5239 }, { "epoch": 2.3731884057971016, "grad_norm": 0.07810837616158466, "learning_rate": 5.0401618977881924e-05, "loss": 0.9187, "step": 5240 }, { "epoch": 2.373641304347826, "grad_norm": 0.08324269821852116, "learning_rate": 5.03894062411837e-05, "loss": 0.921, "step": 5241 }, { "epoch": 2.3740942028985508, "grad_norm": 0.08826629024421066, "learning_rate": 5.037719246584008e-05, "loss": 0.948, "step": 5242 }, { "epoch": 2.3745471014492754, "grad_norm": 0.10908532000197295, "learning_rate": 5.036497765307207e-05, "loss": 0.9071, "step": 5243 }, { "epoch": 2.375, "grad_norm": 0.08076850660374287, "learning_rate": 5.0352761804100835e-05, "loss": 0.9523, "step": 5244 }, { "epoch": 2.3754528985507246, "grad_norm": 0.10320676198418403, "learning_rate": 5.034054492014758e-05, "loss": 0.9324, "step": 5245 }, { "epoch": 2.3759057971014492, "grad_norm": 0.10211420648288254, "learning_rate": 5.0328327002433686e-05, "loss": 0.9283, "step": 5246 }, { "epoch": 2.376358695652174, "grad_norm": 0.11667199811958025, "learning_rate": 5.031610805218057e-05, "loss": 0.9367, "step": 5247 }, { "epoch": 2.3768115942028984, "grad_norm": 0.0976406650367475, "learning_rate": 5.030388807060977e-05, "loss": 0.9537, "step": 5248 }, { "epoch": 2.377264492753623, "grad_norm": 0.09592394412204945, "learning_rate": 5.0291667058942964e-05, "loss": 0.9418, "step": 5249 }, { "epoch": 2.3777173913043477, "grad_norm": 0.09484582400814935, "learning_rate": 5.0279445018401894e-05, "loss": 0.9447, "step": 5250 }, { "epoch": 2.3781702898550723, "grad_norm": 0.10459933891550095, "learning_rate": 5.02672219502084e-05, "loss": 0.9476, "step": 5251 }, { "epoch": 2.378623188405797, "grad_norm": 0.09661530696380646, "learning_rate": 5.0254997855584474e-05, "loss": 0.921, "step": 5252 }, { "epoch": 2.3790760869565215, "grad_norm": 0.0964679072824957, "learning_rate": 5.024277273575216e-05, "loss": 0.9335, "step": 5253 }, { "epoch": 2.3795289855072466, "grad_norm": 0.12154800296282019, "learning_rate": 5.023054659193362e-05, "loss": 0.9371, "step": 5254 }, { "epoch": 2.379981884057971, "grad_norm": 0.11705777006578726, "learning_rate": 5.021831942535113e-05, "loss": 0.9376, "step": 5255 }, { "epoch": 2.380434782608696, "grad_norm": 0.11787059793382407, "learning_rate": 5.020609123722705e-05, "loss": 0.9327, "step": 5256 }, { "epoch": 2.3808876811594204, "grad_norm": 0.12473925651622703, "learning_rate": 5.019386202878386e-05, "loss": 0.9218, "step": 5257 }, { "epoch": 2.381340579710145, "grad_norm": 0.13272244064962066, "learning_rate": 5.018163180124412e-05, "loss": 0.938, "step": 5258 }, { "epoch": 2.3817934782608696, "grad_norm": 0.13005848975033618, "learning_rate": 5.016940055583053e-05, "loss": 0.9572, "step": 5259 }, { "epoch": 2.3822463768115942, "grad_norm": 0.10604313815875852, "learning_rate": 5.015716829376584e-05, "loss": 0.9224, "step": 5260 }, { "epoch": 2.382699275362319, "grad_norm": 0.11580214259982392, "learning_rate": 5.0144935016272944e-05, "loss": 0.9351, "step": 5261 }, { "epoch": 2.3831521739130435, "grad_norm": 0.1090906048103441, "learning_rate": 5.0132700724574814e-05, "loss": 0.934, "step": 5262 }, { "epoch": 2.383605072463768, "grad_norm": 0.08925678116131515, "learning_rate": 5.0120465419894554e-05, "loss": 0.9375, "step": 5263 }, { "epoch": 2.3840579710144927, "grad_norm": 0.1282651565986984, "learning_rate": 5.010822910345531e-05, "loss": 0.9369, "step": 5264 }, { "epoch": 2.3845108695652173, "grad_norm": 0.11036666954671107, "learning_rate": 5.009599177648039e-05, "loss": 0.9324, "step": 5265 }, { "epoch": 2.384963768115942, "grad_norm": 0.10220965665039557, "learning_rate": 5.008375344019318e-05, "loss": 0.9641, "step": 5266 }, { "epoch": 2.3854166666666665, "grad_norm": 0.09930132993444712, "learning_rate": 5.007151409581715e-05, "loss": 0.9491, "step": 5267 }, { "epoch": 2.385869565217391, "grad_norm": 0.09122922850430715, "learning_rate": 5.0059273744575906e-05, "loss": 0.9425, "step": 5268 }, { "epoch": 2.386322463768116, "grad_norm": 0.07396610794963396, "learning_rate": 5.004703238769312e-05, "loss": 0.9153, "step": 5269 }, { "epoch": 2.386775362318841, "grad_norm": 0.08457013989902347, "learning_rate": 5.0034790026392574e-05, "loss": 0.9357, "step": 5270 }, { "epoch": 2.3872282608695654, "grad_norm": 0.09514653017232004, "learning_rate": 5.002254666189818e-05, "loss": 0.9542, "step": 5271 }, { "epoch": 2.38768115942029, "grad_norm": 0.08849112577317074, "learning_rate": 5.001030229543392e-05, "loss": 0.9216, "step": 5272 }, { "epoch": 2.3881340579710146, "grad_norm": 0.08466969631375428, "learning_rate": 4.9998056928223865e-05, "loss": 0.9439, "step": 5273 }, { "epoch": 2.3885869565217392, "grad_norm": 0.10905213176235326, "learning_rate": 4.998581056149222e-05, "loss": 0.9517, "step": 5274 }, { "epoch": 2.389039855072464, "grad_norm": 0.10759327444537226, "learning_rate": 4.9973563196463266e-05, "loss": 0.9219, "step": 5275 }, { "epoch": 2.3894927536231885, "grad_norm": 0.12448508182161346, "learning_rate": 4.99613148343614e-05, "loss": 0.9172, "step": 5276 }, { "epoch": 2.389945652173913, "grad_norm": 0.09612768312231677, "learning_rate": 4.99490654764111e-05, "loss": 0.9537, "step": 5277 }, { "epoch": 2.3903985507246377, "grad_norm": 0.08669190713310118, "learning_rate": 4.9936815123836956e-05, "loss": 0.9411, "step": 5278 }, { "epoch": 2.3908514492753623, "grad_norm": 0.10680596536138583, "learning_rate": 4.9924563777863666e-05, "loss": 0.9275, "step": 5279 }, { "epoch": 2.391304347826087, "grad_norm": 0.10609194620744103, "learning_rate": 4.9912311439716e-05, "loss": 0.9225, "step": 5280 }, { "epoch": 2.3917572463768115, "grad_norm": 0.08704563884571684, "learning_rate": 4.990005811061886e-05, "loss": 0.9483, "step": 5281 }, { "epoch": 2.392210144927536, "grad_norm": 0.10395834189602027, "learning_rate": 4.988780379179722e-05, "loss": 0.9675, "step": 5282 }, { "epoch": 2.3926630434782608, "grad_norm": 0.11337623229661561, "learning_rate": 4.9875548484476166e-05, "loss": 0.9417, "step": 5283 }, { "epoch": 2.3931159420289854, "grad_norm": 0.10110136191878984, "learning_rate": 4.986329218988088e-05, "loss": 0.9298, "step": 5284 }, { "epoch": 2.39356884057971, "grad_norm": 0.09119150733351032, "learning_rate": 4.9851034909236655e-05, "loss": 0.9285, "step": 5285 }, { "epoch": 2.3940217391304346, "grad_norm": 0.10949859064295309, "learning_rate": 4.983877664376884e-05, "loss": 0.9299, "step": 5286 }, { "epoch": 2.394474637681159, "grad_norm": 0.10545262712793683, "learning_rate": 4.982651739470294e-05, "loss": 0.9419, "step": 5287 }, { "epoch": 2.394927536231884, "grad_norm": 0.0988298495124862, "learning_rate": 4.981425716326453e-05, "loss": 0.9268, "step": 5288 }, { "epoch": 2.395380434782609, "grad_norm": 0.12632884804137245, "learning_rate": 4.980199595067928e-05, "loss": 0.9498, "step": 5289 }, { "epoch": 2.3958333333333335, "grad_norm": 0.1307196253157213, "learning_rate": 4.9789733758172956e-05, "loss": 0.9277, "step": 5290 }, { "epoch": 2.396286231884058, "grad_norm": 0.09531720446285157, "learning_rate": 4.9777470586971444e-05, "loss": 0.9472, "step": 5291 }, { "epoch": 2.3967391304347827, "grad_norm": 0.10312905670909889, "learning_rate": 4.9765206438300694e-05, "loss": 0.9419, "step": 5292 }, { "epoch": 2.3971920289855073, "grad_norm": 0.1177421503097571, "learning_rate": 4.975294131338679e-05, "loss": 0.9334, "step": 5293 }, { "epoch": 2.397644927536232, "grad_norm": 0.10396708893980533, "learning_rate": 4.974067521345588e-05, "loss": 0.9091, "step": 5294 }, { "epoch": 2.3980978260869565, "grad_norm": 0.09711358480185528, "learning_rate": 4.972840813973424e-05, "loss": 0.9207, "step": 5295 }, { "epoch": 2.398550724637681, "grad_norm": 0.090028034087384, "learning_rate": 4.971614009344822e-05, "loss": 0.9351, "step": 5296 }, { "epoch": 2.3990036231884058, "grad_norm": 0.09990709244058835, "learning_rate": 4.970387107582427e-05, "loss": 0.928, "step": 5297 }, { "epoch": 2.3994565217391304, "grad_norm": 0.10829674864246633, "learning_rate": 4.9691601088088967e-05, "loss": 0.9551, "step": 5298 }, { "epoch": 2.399909420289855, "grad_norm": 0.09878645219049169, "learning_rate": 4.967933013146894e-05, "loss": 0.9339, "step": 5299 }, { "epoch": 2.4003623188405796, "grad_norm": 0.10240393141420759, "learning_rate": 4.966705820719093e-05, "loss": 0.9646, "step": 5300 }, { "epoch": 2.400815217391304, "grad_norm": 0.12043508234045353, "learning_rate": 4.96547853164818e-05, "loss": 0.9285, "step": 5301 }, { "epoch": 2.401268115942029, "grad_norm": 0.09107776706310945, "learning_rate": 4.9642511460568495e-05, "loss": 0.9401, "step": 5302 }, { "epoch": 2.4017210144927534, "grad_norm": 0.10431560365565842, "learning_rate": 4.9630236640678025e-05, "loss": 0.9347, "step": 5303 }, { "epoch": 2.4021739130434785, "grad_norm": 0.11393913389535731, "learning_rate": 4.961796085803754e-05, "loss": 0.9294, "step": 5304 }, { "epoch": 2.402626811594203, "grad_norm": 0.08145053489473617, "learning_rate": 4.960568411387427e-05, "loss": 0.9284, "step": 5305 }, { "epoch": 2.4030797101449277, "grad_norm": 0.09523721453706675, "learning_rate": 4.959340640941554e-05, "loss": 0.9575, "step": 5306 }, { "epoch": 2.4035326086956523, "grad_norm": 0.08474369165635809, "learning_rate": 4.958112774588876e-05, "loss": 0.9211, "step": 5307 }, { "epoch": 2.403985507246377, "grad_norm": 0.08336102267134328, "learning_rate": 4.956884812452146e-05, "loss": 0.9394, "step": 5308 }, { "epoch": 2.4044384057971016, "grad_norm": 0.1093203987550815, "learning_rate": 4.9556567546541256e-05, "loss": 0.9246, "step": 5309 }, { "epoch": 2.404891304347826, "grad_norm": 0.08784095502439528, "learning_rate": 4.954428601317585e-05, "loss": 0.9252, "step": 5310 }, { "epoch": 2.4053442028985508, "grad_norm": 0.0934333355603288, "learning_rate": 4.953200352565304e-05, "loss": 0.95, "step": 5311 }, { "epoch": 2.4057971014492754, "grad_norm": 0.10580696059792415, "learning_rate": 4.9519720085200734e-05, "loss": 0.9487, "step": 5312 }, { "epoch": 2.40625, "grad_norm": 0.08468160523147027, "learning_rate": 4.950743569304693e-05, "loss": 0.9197, "step": 5313 }, { "epoch": 2.4067028985507246, "grad_norm": 0.08461982855441379, "learning_rate": 4.9495150350419716e-05, "loss": 0.9435, "step": 5314 }, { "epoch": 2.4071557971014492, "grad_norm": 0.12013816514670948, "learning_rate": 4.9482864058547265e-05, "loss": 0.9557, "step": 5315 }, { "epoch": 2.407608695652174, "grad_norm": 0.12291667590215756, "learning_rate": 4.9470576818657873e-05, "loss": 0.931, "step": 5316 }, { "epoch": 2.4080615942028984, "grad_norm": 0.11064094610695993, "learning_rate": 4.9458288631979916e-05, "loss": 0.9125, "step": 5317 }, { "epoch": 2.408514492753623, "grad_norm": 0.08889342127499172, "learning_rate": 4.9445999499741854e-05, "loss": 0.9351, "step": 5318 }, { "epoch": 2.4089673913043477, "grad_norm": 0.14222574054746123, "learning_rate": 4.9433709423172245e-05, "loss": 0.9495, "step": 5319 }, { "epoch": 2.4094202898550723, "grad_norm": 0.17517921928369587, "learning_rate": 4.942141840349976e-05, "loss": 0.9562, "step": 5320 }, { "epoch": 2.409873188405797, "grad_norm": 0.12984084683872918, "learning_rate": 4.940912644195314e-05, "loss": 0.9246, "step": 5321 }, { "epoch": 2.4103260869565215, "grad_norm": 0.09215304228497855, "learning_rate": 4.939683353976125e-05, "loss": 0.9378, "step": 5322 }, { "epoch": 2.4107789855072466, "grad_norm": 0.09743034583212408, "learning_rate": 4.9384539698153023e-05, "loss": 0.9033, "step": 5323 }, { "epoch": 2.411231884057971, "grad_norm": 0.12205277660481617, "learning_rate": 4.937224491835748e-05, "loss": 0.9109, "step": 5324 }, { "epoch": 2.411684782608696, "grad_norm": 0.1090550480512961, "learning_rate": 4.9359949201603764e-05, "loss": 0.9171, "step": 5325 }, { "epoch": 2.4121376811594204, "grad_norm": 0.11090921438109505, "learning_rate": 4.93476525491211e-05, "loss": 0.92, "step": 5326 }, { "epoch": 2.412590579710145, "grad_norm": 0.14147586660946349, "learning_rate": 4.93353549621388e-05, "loss": 0.9458, "step": 5327 }, { "epoch": 2.4130434782608696, "grad_norm": 0.10010606659912523, "learning_rate": 4.932305644188627e-05, "loss": 0.9417, "step": 5328 }, { "epoch": 2.4134963768115942, "grad_norm": 0.10381376714362592, "learning_rate": 4.9310756989593006e-05, "loss": 0.9457, "step": 5329 }, { "epoch": 2.413949275362319, "grad_norm": 0.09041242091592305, "learning_rate": 4.9298456606488616e-05, "loss": 0.9312, "step": 5330 }, { "epoch": 2.4144021739130435, "grad_norm": 0.0779002065207505, "learning_rate": 4.9286155293802796e-05, "loss": 0.9298, "step": 5331 }, { "epoch": 2.414855072463768, "grad_norm": 0.09661953160905459, "learning_rate": 4.92738530527653e-05, "loss": 0.9242, "step": 5332 }, { "epoch": 2.4153079710144927, "grad_norm": 0.08643958825458181, "learning_rate": 4.9261549884606024e-05, "loss": 0.9432, "step": 5333 }, { "epoch": 2.4157608695652173, "grad_norm": 0.1014069153320367, "learning_rate": 4.9249245790554924e-05, "loss": 0.9592, "step": 5334 }, { "epoch": 2.416213768115942, "grad_norm": 0.09434217899870724, "learning_rate": 4.9236940771842075e-05, "loss": 0.9377, "step": 5335 }, { "epoch": 2.4166666666666665, "grad_norm": 0.08596611405595915, "learning_rate": 4.922463482969761e-05, "loss": 0.9098, "step": 5336 }, { "epoch": 2.417119565217391, "grad_norm": 0.10089561945940297, "learning_rate": 4.921232796535179e-05, "loss": 0.9521, "step": 5337 }, { "epoch": 2.417572463768116, "grad_norm": 0.09493995440851458, "learning_rate": 4.920002018003495e-05, "loss": 0.9329, "step": 5338 }, { "epoch": 2.418025362318841, "grad_norm": 0.08151194465759713, "learning_rate": 4.9187711474977506e-05, "loss": 0.9502, "step": 5339 }, { "epoch": 2.4184782608695654, "grad_norm": 0.11257862888520488, "learning_rate": 4.9175401851409986e-05, "loss": 0.9528, "step": 5340 }, { "epoch": 2.41893115942029, "grad_norm": 0.08633211132690607, "learning_rate": 4.9163091310563e-05, "loss": 0.9525, "step": 5341 }, { "epoch": 2.4193840579710146, "grad_norm": 0.09843340666406342, "learning_rate": 4.915077985366727e-05, "loss": 0.9287, "step": 5342 }, { "epoch": 2.4198369565217392, "grad_norm": 0.11548834239517355, "learning_rate": 4.913846748195356e-05, "loss": 0.9247, "step": 5343 }, { "epoch": 2.420289855072464, "grad_norm": 0.09257331054092383, "learning_rate": 4.91261541966528e-05, "loss": 0.9396, "step": 5344 }, { "epoch": 2.4207427536231885, "grad_norm": 0.09137408153212057, "learning_rate": 4.911383999899592e-05, "loss": 0.9237, "step": 5345 }, { "epoch": 2.421195652173913, "grad_norm": 0.09995726013686289, "learning_rate": 4.910152489021402e-05, "loss": 0.9361, "step": 5346 }, { "epoch": 2.4216485507246377, "grad_norm": 0.07376903087749509, "learning_rate": 4.908920887153825e-05, "loss": 0.927, "step": 5347 }, { "epoch": 2.4221014492753623, "grad_norm": 0.09417625882151907, "learning_rate": 4.907689194419988e-05, "loss": 0.9387, "step": 5348 }, { "epoch": 2.422554347826087, "grad_norm": 0.11522817730344334, "learning_rate": 4.906457410943022e-05, "loss": 0.9418, "step": 5349 }, { "epoch": 2.4230072463768115, "grad_norm": 0.11703261093580508, "learning_rate": 4.9052255368460736e-05, "loss": 0.9313, "step": 5350 }, { "epoch": 2.423460144927536, "grad_norm": 0.13848915401878628, "learning_rate": 4.9039935722522926e-05, "loss": 0.9423, "step": 5351 }, { "epoch": 2.4239130434782608, "grad_norm": 0.11030091487369476, "learning_rate": 4.902761517284843e-05, "loss": 0.9359, "step": 5352 }, { "epoch": 2.4243659420289854, "grad_norm": 0.07602545582411588, "learning_rate": 4.901529372066893e-05, "loss": 0.9257, "step": 5353 }, { "epoch": 2.42481884057971, "grad_norm": 0.11304770246440791, "learning_rate": 4.900297136721624e-05, "loss": 0.9359, "step": 5354 }, { "epoch": 2.4252717391304346, "grad_norm": 0.10348439849108604, "learning_rate": 4.8990648113722226e-05, "loss": 0.917, "step": 5355 }, { "epoch": 2.425724637681159, "grad_norm": 0.08767087435683502, "learning_rate": 4.8978323961418887e-05, "loss": 0.9203, "step": 5356 }, { "epoch": 2.426177536231884, "grad_norm": 0.11080423450739801, "learning_rate": 4.896599891153826e-05, "loss": 0.9306, "step": 5357 }, { "epoch": 2.426630434782609, "grad_norm": 0.09065870475329889, "learning_rate": 4.8953672965312515e-05, "loss": 0.9448, "step": 5358 }, { "epoch": 2.4270833333333335, "grad_norm": 0.09193349276160569, "learning_rate": 4.89413461239739e-05, "loss": 0.9535, "step": 5359 }, { "epoch": 2.427536231884058, "grad_norm": 0.10170409819836307, "learning_rate": 4.892901838875476e-05, "loss": 0.9298, "step": 5360 }, { "epoch": 2.4279891304347827, "grad_norm": 0.1021541789466702, "learning_rate": 4.891668976088747e-05, "loss": 0.929, "step": 5361 }, { "epoch": 2.4284420289855073, "grad_norm": 0.09764039956499684, "learning_rate": 4.890436024160459e-05, "loss": 0.9488, "step": 5362 }, { "epoch": 2.428894927536232, "grad_norm": 0.09447374991785538, "learning_rate": 4.8892029832138704e-05, "loss": 0.9414, "step": 5363 }, { "epoch": 2.4293478260869565, "grad_norm": 0.11531062300792408, "learning_rate": 4.8879698533722514e-05, "loss": 0.9231, "step": 5364 }, { "epoch": 2.429800724637681, "grad_norm": 0.12189359331196643, "learning_rate": 4.88673663475888e-05, "loss": 0.9559, "step": 5365 }, { "epoch": 2.4302536231884058, "grad_norm": 0.12493705743071107, "learning_rate": 4.88550332749704e-05, "loss": 0.9399, "step": 5366 }, { "epoch": 2.4307065217391304, "grad_norm": 0.09320308270722515, "learning_rate": 4.884269931710029e-05, "loss": 0.9107, "step": 5367 }, { "epoch": 2.431159420289855, "grad_norm": 0.09880178748582427, "learning_rate": 4.883036447521152e-05, "loss": 0.931, "step": 5368 }, { "epoch": 2.4316123188405796, "grad_norm": 0.15726148128887726, "learning_rate": 4.8818028750537235e-05, "loss": 0.9421, "step": 5369 }, { "epoch": 2.432065217391304, "grad_norm": 0.15895878996980634, "learning_rate": 4.880569214431063e-05, "loss": 0.9219, "step": 5370 }, { "epoch": 2.432518115942029, "grad_norm": 0.13453099074300223, "learning_rate": 4.879335465776502e-05, "loss": 0.9326, "step": 5371 }, { "epoch": 2.4329710144927534, "grad_norm": 0.1030324323555312, "learning_rate": 4.878101629213382e-05, "loss": 0.9333, "step": 5372 }, { "epoch": 2.4334239130434785, "grad_norm": 0.1413731159110437, "learning_rate": 4.876867704865052e-05, "loss": 0.9309, "step": 5373 }, { "epoch": 2.433876811594203, "grad_norm": 0.1871198133747382, "learning_rate": 4.875633692854866e-05, "loss": 0.9522, "step": 5374 }, { "epoch": 2.4343297101449277, "grad_norm": 0.17518117989893012, "learning_rate": 4.874399593306192e-05, "loss": 0.9493, "step": 5375 }, { "epoch": 2.4347826086956523, "grad_norm": 0.10933178973624703, "learning_rate": 4.8731654063424066e-05, "loss": 0.9183, "step": 5376 }, { "epoch": 2.435235507246377, "grad_norm": 0.08565995270257848, "learning_rate": 4.871931132086892e-05, "loss": 0.9538, "step": 5377 }, { "epoch": 2.4356884057971016, "grad_norm": 0.10792981981637947, "learning_rate": 4.8706967706630395e-05, "loss": 0.9234, "step": 5378 }, { "epoch": 2.436141304347826, "grad_norm": 0.1038828265616549, "learning_rate": 4.869462322194251e-05, "loss": 0.952, "step": 5379 }, { "epoch": 2.4365942028985508, "grad_norm": 0.11080452203224886, "learning_rate": 4.868227786803936e-05, "loss": 0.9413, "step": 5380 }, { "epoch": 2.4370471014492754, "grad_norm": 0.10514366077087382, "learning_rate": 4.866993164615515e-05, "loss": 0.9237, "step": 5381 }, { "epoch": 2.4375, "grad_norm": 0.0880094201388878, "learning_rate": 4.8657584557524116e-05, "loss": 0.924, "step": 5382 }, { "epoch": 2.4379528985507246, "grad_norm": 0.09795394062763788, "learning_rate": 4.8645236603380645e-05, "loss": 0.9186, "step": 5383 }, { "epoch": 2.4384057971014492, "grad_norm": 0.1100608237664423, "learning_rate": 4.8632887784959165e-05, "loss": 0.9501, "step": 5384 }, { "epoch": 2.438858695652174, "grad_norm": 0.07741252647359853, "learning_rate": 4.862053810349421e-05, "loss": 0.9281, "step": 5385 }, { "epoch": 2.4393115942028984, "grad_norm": 0.09203267100120345, "learning_rate": 4.860818756022042e-05, "loss": 0.9107, "step": 5386 }, { "epoch": 2.439764492753623, "grad_norm": 0.08815113409998926, "learning_rate": 4.8595836156372465e-05, "loss": 0.9406, "step": 5387 }, { "epoch": 2.4402173913043477, "grad_norm": 0.09086142140145305, "learning_rate": 4.858348389318515e-05, "loss": 0.9451, "step": 5388 }, { "epoch": 2.4406702898550723, "grad_norm": 0.09288450150576512, "learning_rate": 4.8571130771893354e-05, "loss": 0.9494, "step": 5389 }, { "epoch": 2.441123188405797, "grad_norm": 0.10937162793765791, "learning_rate": 4.855877679373202e-05, "loss": 0.9342, "step": 5390 }, { "epoch": 2.4415760869565215, "grad_norm": 0.1126404571968321, "learning_rate": 4.8546421959936216e-05, "loss": 0.9205, "step": 5391 }, { "epoch": 2.4420289855072466, "grad_norm": 0.08940562151994115, "learning_rate": 4.8534066271741065e-05, "loss": 0.9229, "step": 5392 }, { "epoch": 2.442481884057971, "grad_norm": 0.079890189104895, "learning_rate": 4.8521709730381783e-05, "loss": 0.9263, "step": 5393 }, { "epoch": 2.442934782608696, "grad_norm": 0.08762117496387115, "learning_rate": 4.850935233709368e-05, "loss": 0.9167, "step": 5394 }, { "epoch": 2.4433876811594204, "grad_norm": 0.09871164664928819, "learning_rate": 4.849699409311214e-05, "loss": 0.9309, "step": 5395 }, { "epoch": 2.443840579710145, "grad_norm": 0.08597493741952612, "learning_rate": 4.848463499967262e-05, "loss": 0.9071, "step": 5396 }, { "epoch": 2.4442934782608696, "grad_norm": 0.1056645904004699, "learning_rate": 4.847227505801071e-05, "loss": 0.9388, "step": 5397 }, { "epoch": 2.4447463768115942, "grad_norm": 0.11354499202349336, "learning_rate": 4.845991426936203e-05, "loss": 0.9388, "step": 5398 }, { "epoch": 2.445199275362319, "grad_norm": 0.0955539568467073, "learning_rate": 4.844755263496233e-05, "loss": 0.9441, "step": 5399 }, { "epoch": 2.4456521739130435, "grad_norm": 0.07986082980200211, "learning_rate": 4.843519015604739e-05, "loss": 0.9455, "step": 5400 }, { "epoch": 2.446105072463768, "grad_norm": 0.10081876037231065, "learning_rate": 4.8422826833853126e-05, "loss": 0.9456, "step": 5401 }, { "epoch": 2.4465579710144927, "grad_norm": 0.10483582364146672, "learning_rate": 4.8410462669615516e-05, "loss": 0.9634, "step": 5402 }, { "epoch": 2.4470108695652173, "grad_norm": 0.09865665193421394, "learning_rate": 4.839809766457063e-05, "loss": 0.9506, "step": 5403 }, { "epoch": 2.447463768115942, "grad_norm": 0.08956548384114262, "learning_rate": 4.838573181995461e-05, "loss": 0.9292, "step": 5404 }, { "epoch": 2.4479166666666665, "grad_norm": 0.09058351619797811, "learning_rate": 4.837336513700369e-05, "loss": 0.9219, "step": 5405 }, { "epoch": 2.448369565217391, "grad_norm": 0.12431483267188397, "learning_rate": 4.836099761695419e-05, "loss": 0.9565, "step": 5406 }, { "epoch": 2.448822463768116, "grad_norm": 0.12659931833631763, "learning_rate": 4.8348629261042516e-05, "loss": 0.9215, "step": 5407 }, { "epoch": 2.449275362318841, "grad_norm": 0.10209260227415695, "learning_rate": 4.833626007050513e-05, "loss": 0.9411, "step": 5408 }, { "epoch": 2.4497282608695654, "grad_norm": 0.09352971738384655, "learning_rate": 4.832389004657862e-05, "loss": 0.9221, "step": 5409 }, { "epoch": 2.45018115942029, "grad_norm": 0.09800304261450116, "learning_rate": 4.831151919049964e-05, "loss": 0.9346, "step": 5410 }, { "epoch": 2.4506340579710146, "grad_norm": 0.10060971306873083, "learning_rate": 4.82991475035049e-05, "loss": 0.9094, "step": 5411 }, { "epoch": 2.4510869565217392, "grad_norm": 0.11395945306472531, "learning_rate": 4.828677498683125e-05, "loss": 0.931, "step": 5412 }, { "epoch": 2.451539855072464, "grad_norm": 0.11088078593325754, "learning_rate": 4.827440164171557e-05, "loss": 0.9373, "step": 5413 }, { "epoch": 2.4519927536231885, "grad_norm": 0.09175708063635925, "learning_rate": 4.826202746939484e-05, "loss": 0.934, "step": 5414 }, { "epoch": 2.452445652173913, "grad_norm": 0.10208217609245963, "learning_rate": 4.824965247110614e-05, "loss": 0.9422, "step": 5415 }, { "epoch": 2.4528985507246377, "grad_norm": 0.10814023758034341, "learning_rate": 4.82372766480866e-05, "loss": 0.9527, "step": 5416 }, { "epoch": 2.4533514492753623, "grad_norm": 0.1237013731732114, "learning_rate": 4.822490000157347e-05, "loss": 0.9371, "step": 5417 }, { "epoch": 2.453804347826087, "grad_norm": 0.10324210553690986, "learning_rate": 4.8212522532804046e-05, "loss": 0.9284, "step": 5418 }, { "epoch": 2.4542572463768115, "grad_norm": 0.09488199961246782, "learning_rate": 4.8200144243015724e-05, "loss": 0.9337, "step": 5419 }, { "epoch": 2.454710144927536, "grad_norm": 0.0838239096966423, "learning_rate": 4.8187765133446006e-05, "loss": 0.9565, "step": 5420 }, { "epoch": 2.4551630434782608, "grad_norm": 0.08793305213394999, "learning_rate": 4.817538520533242e-05, "loss": 0.9222, "step": 5421 }, { "epoch": 2.4556159420289854, "grad_norm": 0.09307634237134861, "learning_rate": 4.816300445991262e-05, "loss": 0.9384, "step": 5422 }, { "epoch": 2.45606884057971, "grad_norm": 0.10029620130833561, "learning_rate": 4.815062289842433e-05, "loss": 0.9114, "step": 5423 }, { "epoch": 2.4565217391304346, "grad_norm": 0.1104004724085343, "learning_rate": 4.8138240522105365e-05, "loss": 0.9392, "step": 5424 }, { "epoch": 2.456974637681159, "grad_norm": 0.08983157703737285, "learning_rate": 4.812585733219359e-05, "loss": 0.9467, "step": 5425 }, { "epoch": 2.457427536231884, "grad_norm": 0.09304414457617174, "learning_rate": 4.8113473329926975e-05, "loss": 0.9349, "step": 5426 }, { "epoch": 2.457880434782609, "grad_norm": 0.19632547611716492, "learning_rate": 4.8101088516543574e-05, "loss": 0.9289, "step": 5427 }, { "epoch": 2.4583333333333335, "grad_norm": 0.07404709516830113, "learning_rate": 4.808870289328153e-05, "loss": 0.9564, "step": 5428 }, { "epoch": 2.458786231884058, "grad_norm": 0.09114233872226335, "learning_rate": 4.807631646137903e-05, "loss": 0.9107, "step": 5429 }, { "epoch": 2.4592391304347827, "grad_norm": 0.09511295705628957, "learning_rate": 4.806392922207438e-05, "loss": 0.9382, "step": 5430 }, { "epoch": 2.4596920289855073, "grad_norm": 0.09218354270870821, "learning_rate": 4.8051541176605934e-05, "loss": 0.9454, "step": 5431 }, { "epoch": 2.460144927536232, "grad_norm": 0.0819042036232475, "learning_rate": 4.803915232621217e-05, "loss": 0.9314, "step": 5432 }, { "epoch": 2.4605978260869565, "grad_norm": 0.14997236877725847, "learning_rate": 4.802676267213161e-05, "loss": 0.9373, "step": 5433 }, { "epoch": 2.461050724637681, "grad_norm": 0.11705361666469187, "learning_rate": 4.801437221560286e-05, "loss": 0.9728, "step": 5434 }, { "epoch": 2.4615036231884058, "grad_norm": 0.08451214549128777, "learning_rate": 4.8001980957864624e-05, "loss": 0.9375, "step": 5435 }, { "epoch": 2.4619565217391304, "grad_norm": 0.10920129171647466, "learning_rate": 4.7989588900155675e-05, "loss": 0.9173, "step": 5436 }, { "epoch": 2.462409420289855, "grad_norm": 0.11788424798454061, "learning_rate": 4.7977196043714856e-05, "loss": 0.9337, "step": 5437 }, { "epoch": 2.4628623188405796, "grad_norm": 0.08986344236847069, "learning_rate": 4.7964802389781114e-05, "loss": 0.9208, "step": 5438 }, { "epoch": 2.463315217391304, "grad_norm": 0.09299356937133108, "learning_rate": 4.7952407939593446e-05, "loss": 0.9346, "step": 5439 }, { "epoch": 2.463768115942029, "grad_norm": 0.09759139944439259, "learning_rate": 4.794001269439096e-05, "loss": 0.9384, "step": 5440 }, { "epoch": 2.4642210144927534, "grad_norm": 0.08852690999432654, "learning_rate": 4.792761665541283e-05, "loss": 0.9419, "step": 5441 }, { "epoch": 2.4646739130434785, "grad_norm": 0.0916683870092829, "learning_rate": 4.791521982389829e-05, "loss": 0.9345, "step": 5442 }, { "epoch": 2.465126811594203, "grad_norm": 0.07572567200164235, "learning_rate": 4.790282220108669e-05, "loss": 0.96, "step": 5443 }, { "epoch": 2.4655797101449277, "grad_norm": 0.09742652041152326, "learning_rate": 4.789042378821742e-05, "loss": 0.9213, "step": 5444 }, { "epoch": 2.4660326086956523, "grad_norm": 0.07804554726295591, "learning_rate": 4.7878024586529994e-05, "loss": 0.913, "step": 5445 }, { "epoch": 2.466485507246377, "grad_norm": 0.07308114993077507, "learning_rate": 4.786562459726396e-05, "loss": 0.9376, "step": 5446 }, { "epoch": 2.4669384057971016, "grad_norm": 0.07275626031641542, "learning_rate": 4.7853223821658965e-05, "loss": 0.9327, "step": 5447 }, { "epoch": 2.467391304347826, "grad_norm": 0.33505826965725266, "learning_rate": 4.784082226095474e-05, "loss": 0.9514, "step": 5448 }, { "epoch": 2.4678442028985508, "grad_norm": 0.08367310126390629, "learning_rate": 4.7828419916391105e-05, "loss": 0.9404, "step": 5449 }, { "epoch": 2.4682971014492754, "grad_norm": 0.07757468198488662, "learning_rate": 4.7816016789207913e-05, "loss": 0.9178, "step": 5450 }, { "epoch": 2.46875, "grad_norm": 0.0792088581705604, "learning_rate": 4.780361288064514e-05, "loss": 0.9616, "step": 5451 }, { "epoch": 2.4692028985507246, "grad_norm": 0.07751953693527658, "learning_rate": 4.7791208191942815e-05, "loss": 0.9388, "step": 5452 }, { "epoch": 2.4696557971014492, "grad_norm": 0.09162207803435647, "learning_rate": 4.777880272434107e-05, "loss": 0.9308, "step": 5453 }, { "epoch": 2.470108695652174, "grad_norm": 0.1145377223304378, "learning_rate": 4.7766396479080086e-05, "loss": 0.9376, "step": 5454 }, { "epoch": 2.4705615942028984, "grad_norm": 0.12166587621371894, "learning_rate": 4.7753989457400137e-05, "loss": 0.9452, "step": 5455 }, { "epoch": 2.471014492753623, "grad_norm": 0.13030981260648766, "learning_rate": 4.774158166054157e-05, "loss": 0.9555, "step": 5456 }, { "epoch": 2.4714673913043477, "grad_norm": 0.10524618395561737, "learning_rate": 4.772917308974483e-05, "loss": 0.9349, "step": 5457 }, { "epoch": 2.4719202898550723, "grad_norm": 0.09919754514655796, "learning_rate": 4.771676374625041e-05, "loss": 0.9353, "step": 5458 }, { "epoch": 2.472373188405797, "grad_norm": 0.11294814794883414, "learning_rate": 4.7704353631298875e-05, "loss": 0.9454, "step": 5459 }, { "epoch": 2.4728260869565215, "grad_norm": 0.10679151226261333, "learning_rate": 4.7691942746130904e-05, "loss": 0.9282, "step": 5460 }, { "epoch": 2.4732789855072466, "grad_norm": 0.12899062263006114, "learning_rate": 4.767953109198723e-05, "loss": 0.9399, "step": 5461 }, { "epoch": 2.473731884057971, "grad_norm": 0.10265936501732106, "learning_rate": 4.766711867010867e-05, "loss": 0.9511, "step": 5462 }, { "epoch": 2.474184782608696, "grad_norm": 0.09463793884458417, "learning_rate": 4.7654705481736094e-05, "loss": 0.9157, "step": 5463 }, { "epoch": 2.4746376811594204, "grad_norm": 0.16783129860347443, "learning_rate": 4.764229152811049e-05, "loss": 0.9498, "step": 5464 }, { "epoch": 2.475090579710145, "grad_norm": 0.10003905441055365, "learning_rate": 4.762987681047289e-05, "loss": 0.9229, "step": 5465 }, { "epoch": 2.4755434782608696, "grad_norm": 0.09407981890392773, "learning_rate": 4.761746133006442e-05, "loss": 0.941, "step": 5466 }, { "epoch": 2.4759963768115942, "grad_norm": 0.0945158492533157, "learning_rate": 4.760504508812626e-05, "loss": 0.9189, "step": 5467 }, { "epoch": 2.476449275362319, "grad_norm": 0.07353827541831807, "learning_rate": 4.759262808589969e-05, "loss": 0.9393, "step": 5468 }, { "epoch": 2.4769021739130435, "grad_norm": 0.09696564685553691, "learning_rate": 4.7580210324626064e-05, "loss": 0.9454, "step": 5469 }, { "epoch": 2.477355072463768, "grad_norm": 0.08993072395586085, "learning_rate": 4.756779180554681e-05, "loss": 0.9422, "step": 5470 }, { "epoch": 2.4778079710144927, "grad_norm": 0.10482854235249252, "learning_rate": 4.7555372529903404e-05, "loss": 0.934, "step": 5471 }, { "epoch": 2.4782608695652173, "grad_norm": 0.07250485363932077, "learning_rate": 4.7542952498937437e-05, "loss": 0.9248, "step": 5472 }, { "epoch": 2.478713768115942, "grad_norm": 0.12546248914670285, "learning_rate": 4.753053171389055e-05, "loss": 0.918, "step": 5473 }, { "epoch": 2.4791666666666665, "grad_norm": 0.11726253758640873, "learning_rate": 4.751811017600448e-05, "loss": 0.9332, "step": 5474 }, { "epoch": 2.479619565217391, "grad_norm": 0.11566782029037682, "learning_rate": 4.7505687886521026e-05, "loss": 0.9312, "step": 5475 }, { "epoch": 2.480072463768116, "grad_norm": 0.09069738411718725, "learning_rate": 4.7493264846682055e-05, "loss": 0.9451, "step": 5476 }, { "epoch": 2.480525362318841, "grad_norm": 0.12151467401211073, "learning_rate": 4.7480841057729524e-05, "loss": 0.9591, "step": 5477 }, { "epoch": 2.4809782608695654, "grad_norm": 0.1250642606833255, "learning_rate": 4.7468416520905454e-05, "loss": 0.9561, "step": 5478 }, { "epoch": 2.48143115942029, "grad_norm": 0.1314491465287764, "learning_rate": 4.745599123745196e-05, "loss": 0.9295, "step": 5479 }, { "epoch": 2.4818840579710146, "grad_norm": 0.11129786824618208, "learning_rate": 4.7443565208611196e-05, "loss": 0.9379, "step": 5480 }, { "epoch": 2.4823369565217392, "grad_norm": 0.09732993745511305, "learning_rate": 4.743113843562543e-05, "loss": 0.9453, "step": 5481 }, { "epoch": 2.482789855072464, "grad_norm": 0.09404028349604822, "learning_rate": 4.741871091973697e-05, "loss": 0.9381, "step": 5482 }, { "epoch": 2.4832427536231885, "grad_norm": 0.09319503036837705, "learning_rate": 4.740628266218823e-05, "loss": 0.9119, "step": 5483 }, { "epoch": 2.483695652173913, "grad_norm": 0.13068711382872802, "learning_rate": 4.739385366422168e-05, "loss": 0.9367, "step": 5484 }, { "epoch": 2.4841485507246377, "grad_norm": 0.08816432298388878, "learning_rate": 4.7381423927079854e-05, "loss": 0.9358, "step": 5485 }, { "epoch": 2.4846014492753623, "grad_norm": 0.08321749739857315, "learning_rate": 4.736899345200539e-05, "loss": 0.9432, "step": 5486 }, { "epoch": 2.485054347826087, "grad_norm": 0.08704561942333465, "learning_rate": 4.735656224024097e-05, "loss": 0.9503, "step": 5487 }, { "epoch": 2.4855072463768115, "grad_norm": 0.10984297299298779, "learning_rate": 4.734413029302936e-05, "loss": 0.9333, "step": 5488 }, { "epoch": 2.485960144927536, "grad_norm": 0.12206267641837643, "learning_rate": 4.733169761161342e-05, "loss": 0.929, "step": 5489 }, { "epoch": 2.4864130434782608, "grad_norm": 0.12086184999610809, "learning_rate": 4.7319264197236034e-05, "loss": 0.956, "step": 5490 }, { "epoch": 2.4868659420289854, "grad_norm": 0.08548911631307611, "learning_rate": 4.7306830051140225e-05, "loss": 0.936, "step": 5491 }, { "epoch": 2.48731884057971, "grad_norm": 0.1431795516368664, "learning_rate": 4.729439517456903e-05, "loss": 0.9372, "step": 5492 }, { "epoch": 2.4877717391304346, "grad_norm": 0.1497651206065526, "learning_rate": 4.728195956876558e-05, "loss": 0.9334, "step": 5493 }, { "epoch": 2.488224637681159, "grad_norm": 0.1122613903520887, "learning_rate": 4.726952323497311e-05, "loss": 0.9324, "step": 5494 }, { "epoch": 2.488677536231884, "grad_norm": 0.09930364272660065, "learning_rate": 4.725708617443487e-05, "loss": 0.9466, "step": 5495 }, { "epoch": 2.489130434782609, "grad_norm": 0.14461494598805366, "learning_rate": 4.7244648388394235e-05, "loss": 0.9422, "step": 5496 }, { "epoch": 2.4895833333333335, "grad_norm": 0.19778838701906745, "learning_rate": 4.723220987809462e-05, "loss": 0.9396, "step": 5497 }, { "epoch": 2.490036231884058, "grad_norm": 0.15690220276926267, "learning_rate": 4.7219770644779514e-05, "loss": 0.9484, "step": 5498 }, { "epoch": 2.4904891304347827, "grad_norm": 0.0930894344725153, "learning_rate": 4.720733068969249e-05, "loss": 0.9351, "step": 5499 }, { "epoch": 2.4909420289855073, "grad_norm": 0.11359108741534321, "learning_rate": 4.719489001407722e-05, "loss": 0.9319, "step": 5500 }, { "epoch": 2.491394927536232, "grad_norm": 0.1405898542179796, "learning_rate": 4.718244861917738e-05, "loss": 0.9422, "step": 5501 }, { "epoch": 2.4918478260869565, "grad_norm": 0.12924725269485982, "learning_rate": 4.717000650623677e-05, "loss": 0.9058, "step": 5502 }, { "epoch": 2.492300724637681, "grad_norm": 0.0932438513150993, "learning_rate": 4.715756367649924e-05, "loss": 0.9531, "step": 5503 }, { "epoch": 2.4927536231884058, "grad_norm": 0.086160533492623, "learning_rate": 4.714512013120874e-05, "loss": 0.9528, "step": 5504 }, { "epoch": 2.4932065217391304, "grad_norm": 0.10656868899507473, "learning_rate": 4.713267587160926e-05, "loss": 0.9249, "step": 5505 }, { "epoch": 2.493659420289855, "grad_norm": 0.13944521694547737, "learning_rate": 4.712023089894486e-05, "loss": 0.9378, "step": 5506 }, { "epoch": 2.4941123188405796, "grad_norm": 0.10261472684472742, "learning_rate": 4.710778521445969e-05, "loss": 0.9415, "step": 5507 }, { "epoch": 2.494565217391304, "grad_norm": 0.10293495689207116, "learning_rate": 4.709533881939798e-05, "loss": 0.9379, "step": 5508 }, { "epoch": 2.495018115942029, "grad_norm": 0.09741475592317561, "learning_rate": 4.708289171500401e-05, "loss": 0.9268, "step": 5509 }, { "epoch": 2.4954710144927534, "grad_norm": 0.10023725972804191, "learning_rate": 4.707044390252212e-05, "loss": 0.9509, "step": 5510 }, { "epoch": 2.4959239130434785, "grad_norm": 0.089214983096439, "learning_rate": 4.7057995383196745e-05, "loss": 0.9267, "step": 5511 }, { "epoch": 2.496376811594203, "grad_norm": 0.08118463585681877, "learning_rate": 4.7045546158272404e-05, "loss": 0.9524, "step": 5512 }, { "epoch": 2.4968297101449277, "grad_norm": 0.10020737277690318, "learning_rate": 4.7033096228993634e-05, "loss": 0.9445, "step": 5513 }, { "epoch": 2.4972826086956523, "grad_norm": 0.0864494252772091, "learning_rate": 4.702064559660511e-05, "loss": 0.9407, "step": 5514 }, { "epoch": 2.497735507246377, "grad_norm": 0.08142502119340232, "learning_rate": 4.70081942623515e-05, "loss": 0.9291, "step": 5515 }, { "epoch": 2.4981884057971016, "grad_norm": 0.07918102171606567, "learning_rate": 4.6995742227477606e-05, "loss": 0.9366, "step": 5516 }, { "epoch": 2.498641304347826, "grad_norm": 0.0847437016200416, "learning_rate": 4.698328949322828e-05, "loss": 0.9389, "step": 5517 }, { "epoch": 2.4990942028985508, "grad_norm": 0.0817756446568333, "learning_rate": 4.6970836060848434e-05, "loss": 0.9409, "step": 5518 }, { "epoch": 2.4995471014492754, "grad_norm": 0.07817062462824037, "learning_rate": 4.695838193158306e-05, "loss": 0.9288, "step": 5519 }, { "epoch": 2.5, "grad_norm": 0.09027302732167523, "learning_rate": 4.694592710667723e-05, "loss": 0.9339, "step": 5520 }, { "epoch": 2.5004528985507246, "grad_norm": 0.09099582523019893, "learning_rate": 4.693347158737604e-05, "loss": 0.9405, "step": 5521 }, { "epoch": 2.5009057971014492, "grad_norm": 0.08769424896321833, "learning_rate": 4.692101537492472e-05, "loss": 0.9236, "step": 5522 }, { "epoch": 2.501358695652174, "grad_norm": 0.10295405649657416, "learning_rate": 4.690855847056853e-05, "loss": 0.9452, "step": 5523 }, { "epoch": 2.5018115942028984, "grad_norm": 0.11115457190125218, "learning_rate": 4.689610087555279e-05, "loss": 0.9545, "step": 5524 }, { "epoch": 2.502264492753623, "grad_norm": 0.12767749298549894, "learning_rate": 4.688364259112293e-05, "loss": 0.9484, "step": 5525 }, { "epoch": 2.5027173913043477, "grad_norm": 0.10033578196074248, "learning_rate": 4.687118361852441e-05, "loss": 0.9602, "step": 5526 }, { "epoch": 2.5031702898550723, "grad_norm": 0.09214697344792969, "learning_rate": 4.685872395900277e-05, "loss": 0.9386, "step": 5527 }, { "epoch": 2.503623188405797, "grad_norm": 0.110295305727354, "learning_rate": 4.684626361380363e-05, "loss": 0.9296, "step": 5528 }, { "epoch": 2.5040760869565215, "grad_norm": 0.08432717833105656, "learning_rate": 4.683380258417266e-05, "loss": 0.9235, "step": 5529 }, { "epoch": 2.504528985507246, "grad_norm": 0.08836625325025586, "learning_rate": 4.6821340871355635e-05, "loss": 0.9338, "step": 5530 }, { "epoch": 2.504981884057971, "grad_norm": 0.09362825425098066, "learning_rate": 4.680887847659834e-05, "loss": 0.9022, "step": 5531 }, { "epoch": 2.505434782608696, "grad_norm": 0.09704044890394692, "learning_rate": 4.679641540114667e-05, "loss": 0.9452, "step": 5532 }, { "epoch": 2.5058876811594204, "grad_norm": 0.10309984826657405, "learning_rate": 4.678395164624659e-05, "loss": 0.9325, "step": 5533 }, { "epoch": 2.506340579710145, "grad_norm": 0.10315200558307228, "learning_rate": 4.677148721314412e-05, "loss": 0.9413, "step": 5534 }, { "epoch": 2.5067934782608696, "grad_norm": 0.07991074727561473, "learning_rate": 4.675902210308534e-05, "loss": 0.9316, "step": 5535 }, { "epoch": 2.5072463768115942, "grad_norm": 0.12215169289476875, "learning_rate": 4.6746556317316394e-05, "loss": 0.936, "step": 5536 }, { "epoch": 2.507699275362319, "grad_norm": 0.1253418744514549, "learning_rate": 4.6734089857083536e-05, "loss": 0.9486, "step": 5537 }, { "epoch": 2.5081521739130435, "grad_norm": 0.10682469113457163, "learning_rate": 4.672162272363305e-05, "loss": 0.959, "step": 5538 }, { "epoch": 2.508605072463768, "grad_norm": 0.1048978461207663, "learning_rate": 4.670915491821127e-05, "loss": 0.9397, "step": 5539 }, { "epoch": 2.5090579710144927, "grad_norm": 0.0902102294700386, "learning_rate": 4.6696686442064664e-05, "loss": 0.9509, "step": 5540 }, { "epoch": 2.5095108695652173, "grad_norm": 0.07592225118429284, "learning_rate": 4.668421729643968e-05, "loss": 0.9425, "step": 5541 }, { "epoch": 2.509963768115942, "grad_norm": 0.10597501850643666, "learning_rate": 4.667174748258291e-05, "loss": 0.9372, "step": 5542 }, { "epoch": 2.5104166666666665, "grad_norm": 0.10185656672924238, "learning_rate": 4.6659277001740984e-05, "loss": 0.9355, "step": 5543 }, { "epoch": 2.5108695652173916, "grad_norm": 0.11333255455455435, "learning_rate": 4.664680585516057e-05, "loss": 0.9185, "step": 5544 }, { "epoch": 2.511322463768116, "grad_norm": 0.08207698212880819, "learning_rate": 4.663433404408844e-05, "loss": 0.9401, "step": 5545 }, { "epoch": 2.511775362318841, "grad_norm": 0.09609928213398836, "learning_rate": 4.662186156977143e-05, "loss": 0.9548, "step": 5546 }, { "epoch": 2.5122282608695654, "grad_norm": 0.09440493392975695, "learning_rate": 4.660938843345642e-05, "loss": 0.962, "step": 5547 }, { "epoch": 2.51268115942029, "grad_norm": 0.08786920867072733, "learning_rate": 4.659691463639038e-05, "loss": 0.9522, "step": 5548 }, { "epoch": 2.5131340579710146, "grad_norm": 0.08866701604711179, "learning_rate": 4.658444017982033e-05, "loss": 0.959, "step": 5549 }, { "epoch": 2.5135869565217392, "grad_norm": 0.0801779140623412, "learning_rate": 4.657196506499335e-05, "loss": 0.9166, "step": 5550 }, { "epoch": 2.514039855072464, "grad_norm": 0.11534608992121302, "learning_rate": 4.655948929315662e-05, "loss": 0.9326, "step": 5551 }, { "epoch": 2.5144927536231885, "grad_norm": 0.11618760702739017, "learning_rate": 4.6547012865557354e-05, "loss": 0.9422, "step": 5552 }, { "epoch": 2.514945652173913, "grad_norm": 0.09250631575537942, "learning_rate": 4.653453578344283e-05, "loss": 0.91, "step": 5553 }, { "epoch": 2.5153985507246377, "grad_norm": 0.07846209993390153, "learning_rate": 4.652205804806041e-05, "loss": 0.9289, "step": 5554 }, { "epoch": 2.5158514492753623, "grad_norm": 0.10134692176020343, "learning_rate": 4.6509579660657525e-05, "loss": 0.9305, "step": 5555 }, { "epoch": 2.516304347826087, "grad_norm": 0.10711075211318828, "learning_rate": 4.6497100622481625e-05, "loss": 0.9272, "step": 5556 }, { "epoch": 2.5167572463768115, "grad_norm": 0.09666608079866737, "learning_rate": 4.648462093478029e-05, "loss": 0.9051, "step": 5557 }, { "epoch": 2.517210144927536, "grad_norm": 0.07578055264094186, "learning_rate": 4.647214059880112e-05, "loss": 0.9502, "step": 5558 }, { "epoch": 2.5176630434782608, "grad_norm": 0.08685228583435553, "learning_rate": 4.645965961579181e-05, "loss": 0.9406, "step": 5559 }, { "epoch": 2.5181159420289854, "grad_norm": 0.11276491628443774, "learning_rate": 4.6447177987000075e-05, "loss": 0.9414, "step": 5560 }, { "epoch": 2.51856884057971, "grad_norm": 0.12048162840846552, "learning_rate": 4.6434695713673746e-05, "loss": 0.9395, "step": 5561 }, { "epoch": 2.5190217391304346, "grad_norm": 0.1161690145192835, "learning_rate": 4.64222127970607e-05, "loss": 0.9347, "step": 5562 }, { "epoch": 2.519474637681159, "grad_norm": 0.08910287337395964, "learning_rate": 4.6409729238408844e-05, "loss": 0.936, "step": 5563 }, { "epoch": 2.519927536231884, "grad_norm": 0.07703169000463957, "learning_rate": 4.639724503896622e-05, "loss": 0.9509, "step": 5564 }, { "epoch": 2.5203804347826084, "grad_norm": 0.09735245742466268, "learning_rate": 4.638476019998086e-05, "loss": 0.9337, "step": 5565 }, { "epoch": 2.5208333333333335, "grad_norm": 0.10989374655901951, "learning_rate": 4.637227472270091e-05, "loss": 0.947, "step": 5566 }, { "epoch": 2.521286231884058, "grad_norm": 0.1179733473676211, "learning_rate": 4.6359788608374553e-05, "loss": 0.9498, "step": 5567 }, { "epoch": 2.5217391304347827, "grad_norm": 0.07212176189548164, "learning_rate": 4.634730185825006e-05, "loss": 0.9511, "step": 5568 }, { "epoch": 2.5221920289855073, "grad_norm": 0.11688395867057283, "learning_rate": 4.6334814473575735e-05, "loss": 0.9277, "step": 5569 }, { "epoch": 2.522644927536232, "grad_norm": 0.14701761862118623, "learning_rate": 4.6322326455599974e-05, "loss": 0.9385, "step": 5570 }, { "epoch": 2.5230978260869565, "grad_norm": 0.1389117084439135, "learning_rate": 4.630983780557121e-05, "loss": 0.9232, "step": 5571 }, { "epoch": 2.523550724637681, "grad_norm": 0.09115096233444915, "learning_rate": 4.629734852473797e-05, "loss": 0.9196, "step": 5572 }, { "epoch": 2.5240036231884058, "grad_norm": 0.07781022047915231, "learning_rate": 4.6284858614348815e-05, "loss": 0.9174, "step": 5573 }, { "epoch": 2.5244565217391304, "grad_norm": 0.10936958626957077, "learning_rate": 4.627236807565238e-05, "loss": 0.9297, "step": 5574 }, { "epoch": 2.524909420289855, "grad_norm": 0.11448238245210242, "learning_rate": 4.625987690989738e-05, "loss": 0.9402, "step": 5575 }, { "epoch": 2.5253623188405796, "grad_norm": 0.08042559607733306, "learning_rate": 4.624738511833257e-05, "loss": 0.9267, "step": 5576 }, { "epoch": 2.525815217391304, "grad_norm": 0.0834681143689755, "learning_rate": 4.623489270220676e-05, "loss": 0.9248, "step": 5577 }, { "epoch": 2.526268115942029, "grad_norm": 0.09505805395655273, "learning_rate": 4.622239966276884e-05, "loss": 0.9296, "step": 5578 }, { "epoch": 2.526721014492754, "grad_norm": 0.09894161194816173, "learning_rate": 4.620990600126777e-05, "loss": 0.929, "step": 5579 }, { "epoch": 2.5271739130434785, "grad_norm": 0.09633598649280029, "learning_rate": 4.619741171895257e-05, "loss": 0.9339, "step": 5580 }, { "epoch": 2.527626811594203, "grad_norm": 0.09665302342787872, "learning_rate": 4.618491681707228e-05, "loss": 0.947, "step": 5581 }, { "epoch": 2.5280797101449277, "grad_norm": 0.0909245468547898, "learning_rate": 4.617242129687606e-05, "loss": 0.9349, "step": 5582 }, { "epoch": 2.5285326086956523, "grad_norm": 0.10899144020208841, "learning_rate": 4.6159925159613106e-05, "loss": 0.93, "step": 5583 }, { "epoch": 2.528985507246377, "grad_norm": 0.0970461367007598, "learning_rate": 4.614742840653267e-05, "loss": 0.9356, "step": 5584 }, { "epoch": 2.5294384057971016, "grad_norm": 0.0921374408697039, "learning_rate": 4.613493103888409e-05, "loss": 0.9363, "step": 5585 }, { "epoch": 2.529891304347826, "grad_norm": 0.07591107587876429, "learning_rate": 4.612243305791672e-05, "loss": 0.9402, "step": 5586 }, { "epoch": 2.5303442028985508, "grad_norm": 0.08097609782045834, "learning_rate": 4.610993446488e-05, "loss": 0.9257, "step": 5587 }, { "epoch": 2.5307971014492754, "grad_norm": 0.08479157634259829, "learning_rate": 4.609743526102347e-05, "loss": 0.9341, "step": 5588 }, { "epoch": 2.53125, "grad_norm": 0.07999853927156908, "learning_rate": 4.608493544759667e-05, "loss": 0.9399, "step": 5589 }, { "epoch": 2.5317028985507246, "grad_norm": 0.11468058601270265, "learning_rate": 4.6072435025849235e-05, "loss": 0.9152, "step": 5590 }, { "epoch": 2.5321557971014492, "grad_norm": 0.12317914970560755, "learning_rate": 4.605993399703084e-05, "loss": 0.9315, "step": 5591 }, { "epoch": 2.532608695652174, "grad_norm": 0.08319430489543027, "learning_rate": 4.604743236239124e-05, "loss": 0.9342, "step": 5592 }, { "epoch": 2.5330615942028984, "grad_norm": 0.07044718252592001, "learning_rate": 4.603493012318025e-05, "loss": 0.9485, "step": 5593 }, { "epoch": 2.533514492753623, "grad_norm": 0.12109010135362905, "learning_rate": 4.6022427280647734e-05, "loss": 0.9426, "step": 5594 }, { "epoch": 2.5339673913043477, "grad_norm": 0.12454547655397614, "learning_rate": 4.600992383604361e-05, "loss": 0.9404, "step": 5595 }, { "epoch": 2.5344202898550723, "grad_norm": 0.0733829080543733, "learning_rate": 4.599741979061788e-05, "loss": 0.9228, "step": 5596 }, { "epoch": 2.534873188405797, "grad_norm": 0.0980097413091997, "learning_rate": 4.59849151456206e-05, "loss": 0.9297, "step": 5597 }, { "epoch": 2.5353260869565215, "grad_norm": 0.11690839223549891, "learning_rate": 4.597240990230185e-05, "loss": 0.9509, "step": 5598 }, { "epoch": 2.535778985507246, "grad_norm": 0.10637066391946233, "learning_rate": 4.595990406191184e-05, "loss": 0.9526, "step": 5599 }, { "epoch": 2.536231884057971, "grad_norm": 0.0858482838177871, "learning_rate": 4.5947397625700756e-05, "loss": 0.9418, "step": 5600 }, { "epoch": 2.536684782608696, "grad_norm": 0.0869468288444715, "learning_rate": 4.593489059491893e-05, "loss": 0.9248, "step": 5601 }, { "epoch": 2.5371376811594204, "grad_norm": 0.10612740048738145, "learning_rate": 4.5922382970816665e-05, "loss": 0.9092, "step": 5602 }, { "epoch": 2.537590579710145, "grad_norm": 0.09728395285906777, "learning_rate": 4.59098747546444e-05, "loss": 0.9488, "step": 5603 }, { "epoch": 2.5380434782608696, "grad_norm": 0.07559662686059808, "learning_rate": 4.5897365947652584e-05, "loss": 0.9228, "step": 5604 }, { "epoch": 2.5384963768115942, "grad_norm": 0.07735759635526701, "learning_rate": 4.5884856551091756e-05, "loss": 0.9342, "step": 5605 }, { "epoch": 2.538949275362319, "grad_norm": 0.07779693159500145, "learning_rate": 4.5872346566212486e-05, "loss": 0.9206, "step": 5606 }, { "epoch": 2.5394021739130435, "grad_norm": 0.07278944605931066, "learning_rate": 4.5859835994265434e-05, "loss": 0.9248, "step": 5607 }, { "epoch": 2.539855072463768, "grad_norm": 0.07539450360044339, "learning_rate": 4.584732483650128e-05, "loss": 0.9279, "step": 5608 }, { "epoch": 2.5403079710144927, "grad_norm": 0.07871187541192197, "learning_rate": 4.58348130941708e-05, "loss": 0.9318, "step": 5609 }, { "epoch": 2.5407608695652173, "grad_norm": 0.07489043738580166, "learning_rate": 4.582230076852482e-05, "loss": 0.9506, "step": 5610 }, { "epoch": 2.541213768115942, "grad_norm": 0.0739450242310112, "learning_rate": 4.58097878608142e-05, "loss": 0.9363, "step": 5611 }, { "epoch": 2.5416666666666665, "grad_norm": 0.07598001889537759, "learning_rate": 4.579727437228987e-05, "loss": 0.9218, "step": 5612 }, { "epoch": 2.5421195652173916, "grad_norm": 0.06927820833211398, "learning_rate": 4.578476030420284e-05, "loss": 0.9106, "step": 5613 }, { "epoch": 2.542572463768116, "grad_norm": 0.08014594060296716, "learning_rate": 4.5772245657804166e-05, "loss": 0.9356, "step": 5614 }, { "epoch": 2.543025362318841, "grad_norm": 0.10060482162492236, "learning_rate": 4.575973043434494e-05, "loss": 0.9372, "step": 5615 }, { "epoch": 2.5434782608695654, "grad_norm": 0.10154245961027847, "learning_rate": 4.574721463507633e-05, "loss": 0.9377, "step": 5616 }, { "epoch": 2.54393115942029, "grad_norm": 0.08868679896297366, "learning_rate": 4.5734698261249575e-05, "loss": 0.9307, "step": 5617 }, { "epoch": 2.5443840579710146, "grad_norm": 0.07341156765447938, "learning_rate": 4.572218131411594e-05, "loss": 0.9195, "step": 5618 }, { "epoch": 2.5448369565217392, "grad_norm": 0.10920404476648043, "learning_rate": 4.570966379492678e-05, "loss": 0.9482, "step": 5619 }, { "epoch": 2.545289855072464, "grad_norm": 0.1267041652601431, "learning_rate": 4.5697145704933477e-05, "loss": 0.9545, "step": 5620 }, { "epoch": 2.5457427536231885, "grad_norm": 0.10774157122703994, "learning_rate": 4.568462704538748e-05, "loss": 0.9187, "step": 5621 }, { "epoch": 2.546195652173913, "grad_norm": 0.07461204946651093, "learning_rate": 4.567210781754033e-05, "loss": 0.9313, "step": 5622 }, { "epoch": 2.5466485507246377, "grad_norm": 0.07688786443810455, "learning_rate": 4.565958802264356e-05, "loss": 0.9391, "step": 5623 }, { "epoch": 2.5471014492753623, "grad_norm": 0.09092192392641207, "learning_rate": 4.56470676619488e-05, "loss": 0.9417, "step": 5624 }, { "epoch": 2.547554347826087, "grad_norm": 0.0937917826432324, "learning_rate": 4.563454673670776e-05, "loss": 0.9392, "step": 5625 }, { "epoch": 2.5480072463768115, "grad_norm": 0.08784679314960321, "learning_rate": 4.562202524817214e-05, "loss": 0.9623, "step": 5626 }, { "epoch": 2.548460144927536, "grad_norm": 0.07978467025053206, "learning_rate": 4.560950319759375e-05, "loss": 0.9083, "step": 5627 }, { "epoch": 2.5489130434782608, "grad_norm": 0.07857424025726852, "learning_rate": 4.559698058622443e-05, "loss": 0.9395, "step": 5628 }, { "epoch": 2.5493659420289854, "grad_norm": 0.09506356444668278, "learning_rate": 4.558445741531609e-05, "loss": 0.9318, "step": 5629 }, { "epoch": 2.54981884057971, "grad_norm": 0.11392962871254193, "learning_rate": 4.55719336861207e-05, "loss": 0.9283, "step": 5630 }, { "epoch": 2.5502717391304346, "grad_norm": 0.10933535137315334, "learning_rate": 4.555940939989027e-05, "loss": 0.9074, "step": 5631 }, { "epoch": 2.550724637681159, "grad_norm": 0.08381140514584122, "learning_rate": 4.554688455787688e-05, "loss": 0.9413, "step": 5632 }, { "epoch": 2.551177536231884, "grad_norm": 0.0954797016354905, "learning_rate": 4.5534359161332634e-05, "loss": 0.9114, "step": 5633 }, { "epoch": 2.5516304347826084, "grad_norm": 0.14241050729992066, "learning_rate": 4.552183321150974e-05, "loss": 0.9297, "step": 5634 }, { "epoch": 2.5520833333333335, "grad_norm": 0.11068245921372583, "learning_rate": 4.550930670966043e-05, "loss": 0.9406, "step": 5635 }, { "epoch": 2.552536231884058, "grad_norm": 0.07349985956904091, "learning_rate": 4.5496779657037e-05, "loss": 0.9502, "step": 5636 }, { "epoch": 2.5529891304347827, "grad_norm": 0.09440695469917869, "learning_rate": 4.54842520548918e-05, "loss": 0.9422, "step": 5637 }, { "epoch": 2.5534420289855073, "grad_norm": 0.11066838941312224, "learning_rate": 4.5471723904477234e-05, "loss": 0.9271, "step": 5638 }, { "epoch": 2.553894927536232, "grad_norm": 0.10064056513725508, "learning_rate": 4.5459195207045745e-05, "loss": 0.9185, "step": 5639 }, { "epoch": 2.5543478260869565, "grad_norm": 0.10071409837375922, "learning_rate": 4.5446665963849874e-05, "loss": 0.9342, "step": 5640 }, { "epoch": 2.554800724637681, "grad_norm": 0.07777010211397585, "learning_rate": 4.543413617614217e-05, "loss": 0.9344, "step": 5641 }, { "epoch": 2.5552536231884058, "grad_norm": 0.10189149434483322, "learning_rate": 4.5421605845175254e-05, "loss": 0.9202, "step": 5642 }, { "epoch": 2.5557065217391304, "grad_norm": 0.09639269262881112, "learning_rate": 4.540907497220182e-05, "loss": 0.9211, "step": 5643 }, { "epoch": 2.556159420289855, "grad_norm": 0.10169665070818741, "learning_rate": 4.539654355847459e-05, "loss": 0.9253, "step": 5644 }, { "epoch": 2.5566123188405796, "grad_norm": 0.12456625883506213, "learning_rate": 4.5384011605246343e-05, "loss": 0.9149, "step": 5645 }, { "epoch": 2.557065217391304, "grad_norm": 0.09526364647050205, "learning_rate": 4.537147911376993e-05, "loss": 0.936, "step": 5646 }, { "epoch": 2.557518115942029, "grad_norm": 0.09441012194898683, "learning_rate": 4.535894608529823e-05, "loss": 0.9297, "step": 5647 }, { "epoch": 2.557971014492754, "grad_norm": 0.10521796784203459, "learning_rate": 4.53464125210842e-05, "loss": 0.9286, "step": 5648 }, { "epoch": 2.5584239130434785, "grad_norm": 0.1440430339081469, "learning_rate": 4.533387842238084e-05, "loss": 0.9221, "step": 5649 }, { "epoch": 2.558876811594203, "grad_norm": 0.11979673741642058, "learning_rate": 4.53213437904412e-05, "loss": 0.9142, "step": 5650 }, { "epoch": 2.5593297101449277, "grad_norm": 0.08673220090414228, "learning_rate": 4.53088086265184e-05, "loss": 0.9427, "step": 5651 }, { "epoch": 2.5597826086956523, "grad_norm": 0.09534063600200567, "learning_rate": 4.529627293186558e-05, "loss": 0.9443, "step": 5652 }, { "epoch": 2.560235507246377, "grad_norm": 0.1252848158993433, "learning_rate": 4.528373670773597e-05, "loss": 0.9307, "step": 5653 }, { "epoch": 2.5606884057971016, "grad_norm": 0.10780547096327352, "learning_rate": 4.527119995538281e-05, "loss": 0.9193, "step": 5654 }, { "epoch": 2.561141304347826, "grad_norm": 0.08678137374939797, "learning_rate": 4.5258662676059445e-05, "loss": 0.9434, "step": 5655 }, { "epoch": 2.5615942028985508, "grad_norm": 0.08975677706696256, "learning_rate": 4.5246124871019246e-05, "loss": 0.9449, "step": 5656 }, { "epoch": 2.5620471014492754, "grad_norm": 0.106277378101624, "learning_rate": 4.523358654151563e-05, "loss": 0.9294, "step": 5657 }, { "epoch": 2.5625, "grad_norm": 0.10894980271388463, "learning_rate": 4.522104768880208e-05, "loss": 0.9187, "step": 5658 }, { "epoch": 2.5629528985507246, "grad_norm": 0.0787988172552057, "learning_rate": 4.52085083141321e-05, "loss": 0.9184, "step": 5659 }, { "epoch": 2.5634057971014492, "grad_norm": 0.09733270171613237, "learning_rate": 4.5195968418759304e-05, "loss": 0.9239, "step": 5660 }, { "epoch": 2.563858695652174, "grad_norm": 0.09589243966135316, "learning_rate": 4.518342800393731e-05, "loss": 0.915, "step": 5661 }, { "epoch": 2.5643115942028984, "grad_norm": 0.08255168538705955, "learning_rate": 4.51708870709198e-05, "loss": 0.9347, "step": 5662 }, { "epoch": 2.564764492753623, "grad_norm": 0.09110859872619416, "learning_rate": 4.515834562096052e-05, "loss": 0.9027, "step": 5663 }, { "epoch": 2.5652173913043477, "grad_norm": 0.08055907948005368, "learning_rate": 4.514580365531325e-05, "loss": 0.9198, "step": 5664 }, { "epoch": 2.5656702898550723, "grad_norm": 0.07415755591611507, "learning_rate": 4.513326117523185e-05, "loss": 0.938, "step": 5665 }, { "epoch": 2.566123188405797, "grad_norm": 0.0765229307405415, "learning_rate": 4.512071818197017e-05, "loss": 0.9272, "step": 5666 }, { "epoch": 2.5665760869565215, "grad_norm": 0.09493868830670761, "learning_rate": 4.5108174676782196e-05, "loss": 0.9172, "step": 5667 }, { "epoch": 2.567028985507246, "grad_norm": 0.10333871842730105, "learning_rate": 4.50956306609219e-05, "loss": 0.9294, "step": 5668 }, { "epoch": 2.567481884057971, "grad_norm": 0.09818065642073504, "learning_rate": 4.508308613564333e-05, "loss": 0.9415, "step": 5669 }, { "epoch": 2.567934782608696, "grad_norm": 0.0820402007439246, "learning_rate": 4.507054110220059e-05, "loss": 0.9239, "step": 5670 }, { "epoch": 2.5683876811594204, "grad_norm": 0.07410255601334877, "learning_rate": 4.505799556184782e-05, "loss": 0.9263, "step": 5671 }, { "epoch": 2.568840579710145, "grad_norm": 0.09126930102648273, "learning_rate": 4.504544951583922e-05, "loss": 0.934, "step": 5672 }, { "epoch": 2.5692934782608696, "grad_norm": 0.108361937223778, "learning_rate": 4.503290296542903e-05, "loss": 0.9383, "step": 5673 }, { "epoch": 2.5697463768115942, "grad_norm": 0.08736026984036176, "learning_rate": 4.502035591187156e-05, "loss": 0.9172, "step": 5674 }, { "epoch": 2.570199275362319, "grad_norm": 0.07909505636649736, "learning_rate": 4.5007808356421145e-05, "loss": 0.9365, "step": 5675 }, { "epoch": 2.5706521739130435, "grad_norm": 0.11130270908665844, "learning_rate": 4.4995260300332204e-05, "loss": 0.9483, "step": 5676 }, { "epoch": 2.571105072463768, "grad_norm": 0.1021469757810746, "learning_rate": 4.4982711744859167e-05, "loss": 0.9354, "step": 5677 }, { "epoch": 2.5715579710144927, "grad_norm": 0.09188291336910533, "learning_rate": 4.4970162691256556e-05, "loss": 0.9241, "step": 5678 }, { "epoch": 2.5720108695652173, "grad_norm": 0.09949548531721437, "learning_rate": 4.4957613140778895e-05, "loss": 0.9668, "step": 5679 }, { "epoch": 2.572463768115942, "grad_norm": 0.10360476086968358, "learning_rate": 4.4945063094680784e-05, "loss": 0.934, "step": 5680 }, { "epoch": 2.5729166666666665, "grad_norm": 0.11937085185077272, "learning_rate": 4.4932512554216886e-05, "loss": 0.9396, "step": 5681 }, { "epoch": 2.5733695652173916, "grad_norm": 0.09969745259528197, "learning_rate": 4.49199615206419e-05, "loss": 0.9427, "step": 5682 }, { "epoch": 2.573822463768116, "grad_norm": 0.09976591342527634, "learning_rate": 4.4907409995210554e-05, "loss": 0.9192, "step": 5683 }, { "epoch": 2.574275362318841, "grad_norm": 0.10999766985990726, "learning_rate": 4.489485797917766e-05, "loss": 0.9427, "step": 5684 }, { "epoch": 2.5747282608695654, "grad_norm": 0.09207708685142862, "learning_rate": 4.488230547379805e-05, "loss": 0.9262, "step": 5685 }, { "epoch": 2.57518115942029, "grad_norm": 0.0960397825711681, "learning_rate": 4.4869752480326645e-05, "loss": 0.9309, "step": 5686 }, { "epoch": 2.5756340579710146, "grad_norm": 0.10554066745056453, "learning_rate": 4.485719900001835e-05, "loss": 0.9392, "step": 5687 }, { "epoch": 2.5760869565217392, "grad_norm": 0.09330284123374172, "learning_rate": 4.484464503412817e-05, "loss": 0.9342, "step": 5688 }, { "epoch": 2.576539855072464, "grad_norm": 0.105509876763356, "learning_rate": 4.483209058391115e-05, "loss": 0.9394, "step": 5689 }, { "epoch": 2.5769927536231885, "grad_norm": 0.0945727000107292, "learning_rate": 4.481953565062239e-05, "loss": 0.9306, "step": 5690 }, { "epoch": 2.577445652173913, "grad_norm": 0.0916676060735365, "learning_rate": 4.480698023551701e-05, "loss": 0.942, "step": 5691 }, { "epoch": 2.5778985507246377, "grad_norm": 0.12097607717626925, "learning_rate": 4.4794424339850196e-05, "loss": 0.9414, "step": 5692 }, { "epoch": 2.5783514492753623, "grad_norm": 0.11555016946983225, "learning_rate": 4.478186796487718e-05, "loss": 0.9458, "step": 5693 }, { "epoch": 2.578804347826087, "grad_norm": 0.10793866840345871, "learning_rate": 4.476931111185325e-05, "loss": 0.9256, "step": 5694 }, { "epoch": 2.5792572463768115, "grad_norm": 0.11486541872399629, "learning_rate": 4.475675378203373e-05, "loss": 0.9155, "step": 5695 }, { "epoch": 2.579710144927536, "grad_norm": 0.08365712922590073, "learning_rate": 4.474419597667399e-05, "loss": 0.9423, "step": 5696 }, { "epoch": 2.5801630434782608, "grad_norm": 0.11904732699906939, "learning_rate": 4.473163769702947e-05, "loss": 0.9366, "step": 5697 }, { "epoch": 2.5806159420289854, "grad_norm": 0.09855602261690095, "learning_rate": 4.471907894435562e-05, "loss": 0.9267, "step": 5698 }, { "epoch": 2.58106884057971, "grad_norm": 0.10915932124036744, "learning_rate": 4.470651971990799e-05, "loss": 0.9334, "step": 5699 }, { "epoch": 2.5815217391304346, "grad_norm": 0.10386166588790287, "learning_rate": 4.469396002494211e-05, "loss": 0.9365, "step": 5700 }, { "epoch": 2.581974637681159, "grad_norm": 0.10720004900136769, "learning_rate": 4.4681399860713606e-05, "loss": 0.949, "step": 5701 }, { "epoch": 2.582427536231884, "grad_norm": 0.09748533445682987, "learning_rate": 4.466883922847815e-05, "loss": 0.9575, "step": 5702 }, { "epoch": 2.5828804347826084, "grad_norm": 0.08262062533789866, "learning_rate": 4.465627812949144e-05, "loss": 0.9338, "step": 5703 }, { "epoch": 2.5833333333333335, "grad_norm": 0.09601317894116156, "learning_rate": 4.464371656500921e-05, "loss": 0.9251, "step": 5704 }, { "epoch": 2.583786231884058, "grad_norm": 0.08121227921073955, "learning_rate": 4.463115453628729e-05, "loss": 0.9137, "step": 5705 }, { "epoch": 2.5842391304347827, "grad_norm": 0.0917423249763145, "learning_rate": 4.461859204458151e-05, "loss": 0.9186, "step": 5706 }, { "epoch": 2.5846920289855073, "grad_norm": 0.08602679928558242, "learning_rate": 4.4606029091147764e-05, "loss": 0.9434, "step": 5707 }, { "epoch": 2.585144927536232, "grad_norm": 0.08066439156554399, "learning_rate": 4.4593465677241985e-05, "loss": 0.9198, "step": 5708 }, { "epoch": 2.5855978260869565, "grad_norm": 0.09332700493463472, "learning_rate": 4.458090180412017e-05, "loss": 0.9361, "step": 5709 }, { "epoch": 2.586050724637681, "grad_norm": 0.07237139926802448, "learning_rate": 4.4568337473038336e-05, "loss": 0.913, "step": 5710 }, { "epoch": 2.5865036231884058, "grad_norm": 0.08151916641999923, "learning_rate": 4.455577268525258e-05, "loss": 0.9317, "step": 5711 }, { "epoch": 2.5869565217391304, "grad_norm": 0.08270574739606079, "learning_rate": 4.4543207442019e-05, "loss": 0.9508, "step": 5712 }, { "epoch": 2.587409420289855, "grad_norm": 0.1011377979896286, "learning_rate": 4.4530641744593774e-05, "loss": 0.917, "step": 5713 }, { "epoch": 2.5878623188405796, "grad_norm": 0.07869999958895794, "learning_rate": 4.4518075594233106e-05, "loss": 0.9025, "step": 5714 }, { "epoch": 2.588315217391304, "grad_norm": 0.07253702087736213, "learning_rate": 4.450550899219328e-05, "loss": 0.9134, "step": 5715 }, { "epoch": 2.588768115942029, "grad_norm": 0.08406009095746234, "learning_rate": 4.449294193973056e-05, "loss": 0.9276, "step": 5716 }, { "epoch": 2.589221014492754, "grad_norm": 0.07929135435966932, "learning_rate": 4.448037443810132e-05, "loss": 0.9397, "step": 5717 }, { "epoch": 2.5896739130434785, "grad_norm": 0.08391186412342799, "learning_rate": 4.4467806488561954e-05, "loss": 0.916, "step": 5718 }, { "epoch": 2.590126811594203, "grad_norm": 0.08688943657065125, "learning_rate": 4.445523809236888e-05, "loss": 0.9474, "step": 5719 }, { "epoch": 2.5905797101449277, "grad_norm": 0.08923461170450192, "learning_rate": 4.4442669250778614e-05, "loss": 0.9443, "step": 5720 }, { "epoch": 2.5910326086956523, "grad_norm": 0.09695209801122956, "learning_rate": 4.443009996504765e-05, "loss": 0.9418, "step": 5721 }, { "epoch": 2.591485507246377, "grad_norm": 0.0811592475417188, "learning_rate": 4.4417530236432566e-05, "loss": 0.9348, "step": 5722 }, { "epoch": 2.5919384057971016, "grad_norm": 0.08215754018875426, "learning_rate": 4.440496006619e-05, "loss": 0.9344, "step": 5723 }, { "epoch": 2.592391304347826, "grad_norm": 0.09448634852540072, "learning_rate": 4.43923894555766e-05, "loss": 0.9057, "step": 5724 }, { "epoch": 2.5928442028985508, "grad_norm": 0.09389477356265738, "learning_rate": 4.437981840584905e-05, "loss": 0.9406, "step": 5725 }, { "epoch": 2.5932971014492754, "grad_norm": 0.09332975773137135, "learning_rate": 4.436724691826411e-05, "loss": 0.9338, "step": 5726 }, { "epoch": 2.59375, "grad_norm": 0.07742925533608992, "learning_rate": 4.4354674994078585e-05, "loss": 0.9416, "step": 5727 }, { "epoch": 2.5942028985507246, "grad_norm": 0.07385056095056187, "learning_rate": 4.434210263454931e-05, "loss": 0.9246, "step": 5728 }, { "epoch": 2.5946557971014492, "grad_norm": 0.07569341414711045, "learning_rate": 4.432952984093313e-05, "loss": 0.9316, "step": 5729 }, { "epoch": 2.595108695652174, "grad_norm": 0.07910195026548525, "learning_rate": 4.431695661448701e-05, "loss": 0.9469, "step": 5730 }, { "epoch": 2.5955615942028984, "grad_norm": 0.0864750180829277, "learning_rate": 4.430438295646789e-05, "loss": 0.9448, "step": 5731 }, { "epoch": 2.596014492753623, "grad_norm": 0.07669825468339594, "learning_rate": 4.4291808868132786e-05, "loss": 0.9349, "step": 5732 }, { "epoch": 2.5964673913043477, "grad_norm": 0.08217508873974555, "learning_rate": 4.4279234350738746e-05, "loss": 0.9234, "step": 5733 }, { "epoch": 2.5969202898550723, "grad_norm": 0.09718056242440776, "learning_rate": 4.426665940554287e-05, "loss": 0.9267, "step": 5734 }, { "epoch": 2.597373188405797, "grad_norm": 0.107320957263872, "learning_rate": 4.425408403380228e-05, "loss": 0.9265, "step": 5735 }, { "epoch": 2.5978260869565215, "grad_norm": 0.09375398086872107, "learning_rate": 4.424150823677419e-05, "loss": 0.9082, "step": 5736 }, { "epoch": 2.598278985507246, "grad_norm": 0.07979733978074365, "learning_rate": 4.422893201571579e-05, "loss": 0.9215, "step": 5737 }, { "epoch": 2.598731884057971, "grad_norm": 0.08129377677728791, "learning_rate": 4.421635537188435e-05, "loss": 0.9403, "step": 5738 }, { "epoch": 2.599184782608696, "grad_norm": 0.08017545445705507, "learning_rate": 4.420377830653719e-05, "loss": 0.9467, "step": 5739 }, { "epoch": 2.5996376811594204, "grad_norm": 0.0907854627450386, "learning_rate": 4.419120082093164e-05, "loss": 0.9465, "step": 5740 }, { "epoch": 2.600090579710145, "grad_norm": 0.09201312945664368, "learning_rate": 4.417862291632512e-05, "loss": 0.9351, "step": 5741 }, { "epoch": 2.6005434782608696, "grad_norm": 0.08644886057916377, "learning_rate": 4.416604459397503e-05, "loss": 0.9243, "step": 5742 }, { "epoch": 2.6009963768115942, "grad_norm": 0.07623102101357007, "learning_rate": 4.4153465855138864e-05, "loss": 0.9265, "step": 5743 }, { "epoch": 2.601449275362319, "grad_norm": 0.12554617562039824, "learning_rate": 4.414088670107413e-05, "loss": 0.9537, "step": 5744 }, { "epoch": 2.6019021739130435, "grad_norm": 0.11666072747846759, "learning_rate": 4.412830713303841e-05, "loss": 0.93, "step": 5745 }, { "epoch": 2.602355072463768, "grad_norm": 0.0931891729036086, "learning_rate": 4.411572715228927e-05, "loss": 0.9254, "step": 5746 }, { "epoch": 2.6028079710144927, "grad_norm": 0.08973149448406749, "learning_rate": 4.410314676008436e-05, "loss": 0.9472, "step": 5747 }, { "epoch": 2.6032608695652173, "grad_norm": 0.08979082132694209, "learning_rate": 4.409056595768137e-05, "loss": 0.9258, "step": 5748 }, { "epoch": 2.603713768115942, "grad_norm": 0.08702768409814075, "learning_rate": 4.4077984746338026e-05, "loss": 0.9415, "step": 5749 }, { "epoch": 2.6041666666666665, "grad_norm": 0.10820861726655215, "learning_rate": 4.406540312731208e-05, "loss": 0.9447, "step": 5750 }, { "epoch": 2.6046195652173916, "grad_norm": 0.0930124794476522, "learning_rate": 4.405282110186134e-05, "loss": 0.9406, "step": 5751 }, { "epoch": 2.605072463768116, "grad_norm": 0.08403909316529205, "learning_rate": 4.404023867124365e-05, "loss": 0.9467, "step": 5752 }, { "epoch": 2.605525362318841, "grad_norm": 0.10514719269584993, "learning_rate": 4.402765583671691e-05, "loss": 0.937, "step": 5753 }, { "epoch": 2.6059782608695654, "grad_norm": 0.07578838742238701, "learning_rate": 4.401507259953902e-05, "loss": 0.9433, "step": 5754 }, { "epoch": 2.60643115942029, "grad_norm": 0.11028954241267701, "learning_rate": 4.4002488960967964e-05, "loss": 0.9338, "step": 5755 }, { "epoch": 2.6068840579710146, "grad_norm": 0.0985891741992498, "learning_rate": 4.3989904922261746e-05, "loss": 0.9216, "step": 5756 }, { "epoch": 2.6073369565217392, "grad_norm": 0.11633550412879959, "learning_rate": 4.397732048467841e-05, "loss": 0.9354, "step": 5757 }, { "epoch": 2.607789855072464, "grad_norm": 0.11239420868145447, "learning_rate": 4.396473564947604e-05, "loss": 0.958, "step": 5758 }, { "epoch": 2.6082427536231885, "grad_norm": 0.08588571685204535, "learning_rate": 4.395215041791277e-05, "loss": 0.9237, "step": 5759 }, { "epoch": 2.608695652173913, "grad_norm": 0.09060195097771404, "learning_rate": 4.3939564791246744e-05, "loss": 0.9322, "step": 5760 }, { "epoch": 2.6091485507246377, "grad_norm": 0.08079703100947194, "learning_rate": 4.39269787707362e-05, "loss": 0.9398, "step": 5761 }, { "epoch": 2.6096014492753623, "grad_norm": 0.09318803128018743, "learning_rate": 4.391439235763938e-05, "loss": 0.9225, "step": 5762 }, { "epoch": 2.610054347826087, "grad_norm": 0.08510852988297173, "learning_rate": 4.3901805553214525e-05, "loss": 0.9449, "step": 5763 }, { "epoch": 2.6105072463768115, "grad_norm": 0.08090910860464971, "learning_rate": 4.3889218358719994e-05, "loss": 0.9302, "step": 5764 }, { "epoch": 2.610960144927536, "grad_norm": 0.08596534920403771, "learning_rate": 4.387663077541414e-05, "loss": 0.9508, "step": 5765 }, { "epoch": 2.6114130434782608, "grad_norm": 0.08164086065764499, "learning_rate": 4.3864042804555363e-05, "loss": 0.9251, "step": 5766 }, { "epoch": 2.6118659420289854, "grad_norm": 0.08443259011530467, "learning_rate": 4.385145444740211e-05, "loss": 0.9321, "step": 5767 }, { "epoch": 2.61231884057971, "grad_norm": 0.07633616778691682, "learning_rate": 4.3838865705212854e-05, "loss": 0.9672, "step": 5768 }, { "epoch": 2.6127717391304346, "grad_norm": 0.09104918974060985, "learning_rate": 4.3826276579246095e-05, "loss": 0.9198, "step": 5769 }, { "epoch": 2.613224637681159, "grad_norm": 0.11455409459382242, "learning_rate": 4.3813687070760416e-05, "loss": 0.9738, "step": 5770 }, { "epoch": 2.613677536231884, "grad_norm": 0.09136319482525512, "learning_rate": 4.38010971810144e-05, "loss": 0.9512, "step": 5771 }, { "epoch": 2.6141304347826084, "grad_norm": 0.08642771141920168, "learning_rate": 4.3788506911266667e-05, "loss": 0.9324, "step": 5772 }, { "epoch": 2.6145833333333335, "grad_norm": 0.08481731255902371, "learning_rate": 4.37759162627759e-05, "loss": 0.936, "step": 5773 }, { "epoch": 2.615036231884058, "grad_norm": 0.08744964023476698, "learning_rate": 4.376332523680079e-05, "loss": 0.9251, "step": 5774 }, { "epoch": 2.6154891304347827, "grad_norm": 0.09280858734920826, "learning_rate": 4.3750733834600114e-05, "loss": 0.9291, "step": 5775 }, { "epoch": 2.6159420289855073, "grad_norm": 0.08554653983044934, "learning_rate": 4.373814205743261e-05, "loss": 0.9376, "step": 5776 }, { "epoch": 2.616394927536232, "grad_norm": 0.08597781983536891, "learning_rate": 4.372554990655713e-05, "loss": 0.9443, "step": 5777 }, { "epoch": 2.6168478260869565, "grad_norm": 0.09371756933582367, "learning_rate": 4.3712957383232524e-05, "loss": 0.9409, "step": 5778 }, { "epoch": 2.617300724637681, "grad_norm": 0.1252484146799567, "learning_rate": 4.370036448871769e-05, "loss": 0.9192, "step": 5779 }, { "epoch": 2.6177536231884058, "grad_norm": 0.10999464745164199, "learning_rate": 4.368777122427155e-05, "loss": 0.9339, "step": 5780 }, { "epoch": 2.6182065217391304, "grad_norm": 0.07767772507562831, "learning_rate": 4.367517759115307e-05, "loss": 0.9341, "step": 5781 }, { "epoch": 2.618659420289855, "grad_norm": 0.07702243354257156, "learning_rate": 4.3662583590621266e-05, "loss": 0.9511, "step": 5782 }, { "epoch": 2.6191123188405796, "grad_norm": 0.07237145084742694, "learning_rate": 4.364998922393518e-05, "loss": 0.9424, "step": 5783 }, { "epoch": 2.619565217391304, "grad_norm": 0.0680722178349424, "learning_rate": 4.3637394492353886e-05, "loss": 0.9434, "step": 5784 }, { "epoch": 2.620018115942029, "grad_norm": 0.08561459143615938, "learning_rate": 4.36247993971365e-05, "loss": 0.9402, "step": 5785 }, { "epoch": 2.620471014492754, "grad_norm": 0.08959648685202615, "learning_rate": 4.3612203939542176e-05, "loss": 0.9282, "step": 5786 }, { "epoch": 2.6209239130434785, "grad_norm": 0.08361485821328409, "learning_rate": 4.3599608120830094e-05, "loss": 0.9332, "step": 5787 }, { "epoch": 2.621376811594203, "grad_norm": 0.0824858351131669, "learning_rate": 4.358701194225949e-05, "loss": 0.9019, "step": 5788 }, { "epoch": 2.6218297101449277, "grad_norm": 0.08850353512081242, "learning_rate": 4.357441540508962e-05, "loss": 0.9422, "step": 5789 }, { "epoch": 2.6222826086956523, "grad_norm": 0.07737290960463111, "learning_rate": 4.3561818510579774e-05, "loss": 0.9065, "step": 5790 }, { "epoch": 2.622735507246377, "grad_norm": 0.07865545462254017, "learning_rate": 4.3549221259989286e-05, "loss": 0.9502, "step": 5791 }, { "epoch": 2.6231884057971016, "grad_norm": 0.0830415759507816, "learning_rate": 4.353662365457753e-05, "loss": 0.9597, "step": 5792 }, { "epoch": 2.623641304347826, "grad_norm": 0.07180361616050089, "learning_rate": 4.35240256956039e-05, "loss": 0.9319, "step": 5793 }, { "epoch": 2.6240942028985508, "grad_norm": 0.09224766429199185, "learning_rate": 4.351142738432783e-05, "loss": 0.9462, "step": 5794 }, { "epoch": 2.6245471014492754, "grad_norm": 0.08598760382726536, "learning_rate": 4.3498828722008805e-05, "loss": 0.9266, "step": 5795 }, { "epoch": 2.625, "grad_norm": 0.07363755042418983, "learning_rate": 4.348622970990634e-05, "loss": 0.9299, "step": 5796 }, { "epoch": 2.6254528985507246, "grad_norm": 0.09159950625643994, "learning_rate": 4.347363034927995e-05, "loss": 0.9438, "step": 5797 }, { "epoch": 2.6259057971014492, "grad_norm": 0.10299115319452876, "learning_rate": 4.346103064138923e-05, "loss": 0.9078, "step": 5798 }, { "epoch": 2.626358695652174, "grad_norm": 0.07890551750676439, "learning_rate": 4.3448430587493796e-05, "loss": 0.9296, "step": 5799 }, { "epoch": 2.6268115942028984, "grad_norm": 0.0883320300748188, "learning_rate": 4.3435830188853304e-05, "loss": 0.93, "step": 5800 }, { "epoch": 2.627264492753623, "grad_norm": 0.10695841337008893, "learning_rate": 4.342322944672741e-05, "loss": 0.9328, "step": 5801 }, { "epoch": 2.6277173913043477, "grad_norm": 0.0801330631723727, "learning_rate": 4.3410628362375844e-05, "loss": 0.9406, "step": 5802 }, { "epoch": 2.6281702898550723, "grad_norm": 0.11461416372745889, "learning_rate": 4.339802693705836e-05, "loss": 0.9344, "step": 5803 }, { "epoch": 2.628623188405797, "grad_norm": 0.1043181464503204, "learning_rate": 4.3385425172034744e-05, "loss": 0.9372, "step": 5804 }, { "epoch": 2.6290760869565215, "grad_norm": 0.08413506679308404, "learning_rate": 4.3372823068564805e-05, "loss": 0.93, "step": 5805 }, { "epoch": 2.629528985507246, "grad_norm": 0.11192708626400578, "learning_rate": 4.3360220627908383e-05, "loss": 0.9275, "step": 5806 }, { "epoch": 2.629981884057971, "grad_norm": 0.09213508730800976, "learning_rate": 4.334761785132539e-05, "loss": 0.9597, "step": 5807 }, { "epoch": 2.630434782608696, "grad_norm": 0.11085299068486099, "learning_rate": 4.333501474007574e-05, "loss": 0.9331, "step": 5808 }, { "epoch": 2.6308876811594204, "grad_norm": 0.08875355518295151, "learning_rate": 4.332241129541939e-05, "loss": 0.9445, "step": 5809 }, { "epoch": 2.631340579710145, "grad_norm": 0.09723928524111398, "learning_rate": 4.330980751861631e-05, "loss": 0.9404, "step": 5810 }, { "epoch": 2.6317934782608696, "grad_norm": 0.08696908423474649, "learning_rate": 4.3297203410926526e-05, "loss": 0.9376, "step": 5811 }, { "epoch": 2.6322463768115942, "grad_norm": 0.10915002714344393, "learning_rate": 4.3284598973610085e-05, "loss": 0.9472, "step": 5812 }, { "epoch": 2.632699275362319, "grad_norm": 0.08541811567835554, "learning_rate": 4.32719942079271e-05, "loss": 0.923, "step": 5813 }, { "epoch": 2.6331521739130435, "grad_norm": 0.10858123296758854, "learning_rate": 4.3259389115137654e-05, "loss": 0.9389, "step": 5814 }, { "epoch": 2.633605072463768, "grad_norm": 0.10866490522294715, "learning_rate": 4.3246783696501905e-05, "loss": 0.9336, "step": 5815 }, { "epoch": 2.6340579710144927, "grad_norm": 0.09488172437033698, "learning_rate": 4.3234177953280055e-05, "loss": 0.9132, "step": 5816 }, { "epoch": 2.6345108695652173, "grad_norm": 0.08878691404593492, "learning_rate": 4.322157188673232e-05, "loss": 0.9556, "step": 5817 }, { "epoch": 2.634963768115942, "grad_norm": 0.0935467405738749, "learning_rate": 4.3208965498118924e-05, "loss": 0.9337, "step": 5818 }, { "epoch": 2.6354166666666665, "grad_norm": 0.12176829975390809, "learning_rate": 4.3196358788700164e-05, "loss": 0.9471, "step": 5819 }, { "epoch": 2.6358695652173916, "grad_norm": 0.10924168289758536, "learning_rate": 4.3183751759736354e-05, "loss": 0.9304, "step": 5820 }, { "epoch": 2.636322463768116, "grad_norm": 0.09499194490990691, "learning_rate": 4.317114441248784e-05, "loss": 0.9299, "step": 5821 }, { "epoch": 2.636775362318841, "grad_norm": 0.10244312914835231, "learning_rate": 4.315853674821498e-05, "loss": 0.9202, "step": 5822 }, { "epoch": 2.6372282608695654, "grad_norm": 0.14647718673276766, "learning_rate": 4.3145928768178205e-05, "loss": 0.9338, "step": 5823 }, { "epoch": 2.63768115942029, "grad_norm": 0.1241310196168657, "learning_rate": 4.313332047363795e-05, "loss": 0.9344, "step": 5824 }, { "epoch": 2.6381340579710146, "grad_norm": 0.07252374477730572, "learning_rate": 4.312071186585469e-05, "loss": 0.9476, "step": 5825 }, { "epoch": 2.6385869565217392, "grad_norm": 0.102460129285139, "learning_rate": 4.31081029460889e-05, "loss": 0.9241, "step": 5826 }, { "epoch": 2.639039855072464, "grad_norm": 0.11967088177236679, "learning_rate": 4.3095493715601144e-05, "loss": 0.9193, "step": 5827 }, { "epoch": 2.6394927536231885, "grad_norm": 0.10192349455236116, "learning_rate": 4.308288417565196e-05, "loss": 0.9435, "step": 5828 }, { "epoch": 2.639945652173913, "grad_norm": 0.08058853388828313, "learning_rate": 4.307027432750198e-05, "loss": 0.9539, "step": 5829 }, { "epoch": 2.6403985507246377, "grad_norm": 0.0892260345504647, "learning_rate": 4.305766417241181e-05, "loss": 0.9424, "step": 5830 }, { "epoch": 2.6408514492753623, "grad_norm": 0.07542724149386255, "learning_rate": 4.30450537116421e-05, "loss": 0.9347, "step": 5831 }, { "epoch": 2.641304347826087, "grad_norm": 0.10555002076580806, "learning_rate": 4.303244294645354e-05, "loss": 0.946, "step": 5832 }, { "epoch": 2.6417572463768115, "grad_norm": 0.10680421194500674, "learning_rate": 4.3019831878106855e-05, "loss": 0.9302, "step": 5833 }, { "epoch": 2.642210144927536, "grad_norm": 0.09086537532364662, "learning_rate": 4.30072205078628e-05, "loss": 0.9581, "step": 5834 }, { "epoch": 2.6426630434782608, "grad_norm": 0.08740364842350638, "learning_rate": 4.299460883698214e-05, "loss": 0.9491, "step": 5835 }, { "epoch": 2.6431159420289854, "grad_norm": 0.0802276596926069, "learning_rate": 4.298199686672569e-05, "loss": 0.9419, "step": 5836 }, { "epoch": 2.64356884057971, "grad_norm": 0.0924081672600781, "learning_rate": 4.2969384598354284e-05, "loss": 0.9187, "step": 5837 }, { "epoch": 2.6440217391304346, "grad_norm": 0.07793038307750245, "learning_rate": 4.295677203312881e-05, "loss": 0.9233, "step": 5838 }, { "epoch": 2.644474637681159, "grad_norm": 0.09219603245905221, "learning_rate": 4.294415917231013e-05, "loss": 0.9313, "step": 5839 }, { "epoch": 2.644927536231884, "grad_norm": 0.0883019129149995, "learning_rate": 4.293154601715921e-05, "loss": 0.9334, "step": 5840 }, { "epoch": 2.6453804347826084, "grad_norm": 0.09196566913804749, "learning_rate": 4.291893256893697e-05, "loss": 0.9277, "step": 5841 }, { "epoch": 2.6458333333333335, "grad_norm": 0.07674453883787553, "learning_rate": 4.290631882890443e-05, "loss": 0.9271, "step": 5842 }, { "epoch": 2.646286231884058, "grad_norm": 0.08620374116644473, "learning_rate": 4.2893704798322586e-05, "loss": 0.9257, "step": 5843 }, { "epoch": 2.6467391304347827, "grad_norm": 0.10432707392234729, "learning_rate": 4.288109047845249e-05, "loss": 0.9334, "step": 5844 }, { "epoch": 2.6471920289855073, "grad_norm": 0.09679284217338437, "learning_rate": 4.286847587055521e-05, "loss": 0.946, "step": 5845 }, { "epoch": 2.647644927536232, "grad_norm": 0.0895981889355159, "learning_rate": 4.285586097589186e-05, "loss": 0.9359, "step": 5846 }, { "epoch": 2.6480978260869565, "grad_norm": 0.07624835131071364, "learning_rate": 4.2843245795723554e-05, "loss": 0.9423, "step": 5847 }, { "epoch": 2.648550724637681, "grad_norm": 0.09779717490974277, "learning_rate": 4.283063033131147e-05, "loss": 0.9348, "step": 5848 }, { "epoch": 2.6490036231884058, "grad_norm": 0.09907472179238437, "learning_rate": 4.2818014583916766e-05, "loss": 0.9158, "step": 5849 }, { "epoch": 2.6494565217391304, "grad_norm": 0.07793590553038843, "learning_rate": 4.2805398554800695e-05, "loss": 0.9387, "step": 5850 }, { "epoch": 2.649909420289855, "grad_norm": 0.07935427069147251, "learning_rate": 4.279278224522449e-05, "loss": 0.9435, "step": 5851 }, { "epoch": 2.6503623188405796, "grad_norm": 0.07994066896522573, "learning_rate": 4.27801656564494e-05, "loss": 0.9228, "step": 5852 }, { "epoch": 2.650815217391304, "grad_norm": 0.07456940689619765, "learning_rate": 4.276754878973676e-05, "loss": 0.93, "step": 5853 }, { "epoch": 2.651268115942029, "grad_norm": 0.08163137016534297, "learning_rate": 4.275493164634787e-05, "loss": 0.9471, "step": 5854 }, { "epoch": 2.651721014492754, "grad_norm": 0.0837185276246777, "learning_rate": 4.274231422754411e-05, "loss": 0.9301, "step": 5855 }, { "epoch": 2.6521739130434785, "grad_norm": 0.08691285557755934, "learning_rate": 4.272969653458685e-05, "loss": 0.9284, "step": 5856 }, { "epoch": 2.652626811594203, "grad_norm": 0.07053586996956508, "learning_rate": 4.2717078568737485e-05, "loss": 0.931, "step": 5857 }, { "epoch": 2.6530797101449277, "grad_norm": 0.08170644309303178, "learning_rate": 4.2704460331257484e-05, "loss": 0.9164, "step": 5858 }, { "epoch": 2.6535326086956523, "grad_norm": 0.06837602188845336, "learning_rate": 4.26918418234083e-05, "loss": 0.9694, "step": 5859 }, { "epoch": 2.653985507246377, "grad_norm": 0.08272976478424571, "learning_rate": 4.267922304645142e-05, "loss": 0.9286, "step": 5860 }, { "epoch": 2.6544384057971016, "grad_norm": 0.0777091794918059, "learning_rate": 4.266660400164837e-05, "loss": 0.9461, "step": 5861 }, { "epoch": 2.654891304347826, "grad_norm": 0.08167022871290774, "learning_rate": 4.265398469026067e-05, "loss": 0.9453, "step": 5862 }, { "epoch": 2.6553442028985508, "grad_norm": 0.07067605657040846, "learning_rate": 4.264136511354995e-05, "loss": 0.9182, "step": 5863 }, { "epoch": 2.6557971014492754, "grad_norm": 0.0754782642798268, "learning_rate": 4.262874527277775e-05, "loss": 0.9365, "step": 5864 }, { "epoch": 2.65625, "grad_norm": 0.08608382707721911, "learning_rate": 4.261612516920573e-05, "loss": 0.9334, "step": 5865 }, { "epoch": 2.6567028985507246, "grad_norm": 0.09251421876466681, "learning_rate": 4.260350480409552e-05, "loss": 0.9147, "step": 5866 }, { "epoch": 2.6571557971014492, "grad_norm": 0.09604957836295921, "learning_rate": 4.259088417870882e-05, "loss": 0.9205, "step": 5867 }, { "epoch": 2.657608695652174, "grad_norm": 0.09305725474463639, "learning_rate": 4.257826329430733e-05, "loss": 0.9334, "step": 5868 }, { "epoch": 2.6580615942028984, "grad_norm": 0.07763955113142205, "learning_rate": 4.256564215215275e-05, "loss": 0.9252, "step": 5869 }, { "epoch": 2.658514492753623, "grad_norm": 0.09311954827405687, "learning_rate": 4.2553020753506875e-05, "loss": 0.9526, "step": 5870 }, { "epoch": 2.6589673913043477, "grad_norm": 0.09968899365374595, "learning_rate": 4.2540399099631465e-05, "loss": 0.9331, "step": 5871 }, { "epoch": 2.6594202898550723, "grad_norm": 0.10986974300958177, "learning_rate": 4.252777719178833e-05, "loss": 0.93, "step": 5872 }, { "epoch": 2.659873188405797, "grad_norm": 0.0913633235948142, "learning_rate": 4.2515155031239306e-05, "loss": 0.9347, "step": 5873 }, { "epoch": 2.6603260869565215, "grad_norm": 0.07885753214791758, "learning_rate": 4.250253261924624e-05, "loss": 0.9315, "step": 5874 }, { "epoch": 2.660778985507246, "grad_norm": 0.103267874584999, "learning_rate": 4.248990995707104e-05, "loss": 0.9422, "step": 5875 }, { "epoch": 2.661231884057971, "grad_norm": 0.09450821919749361, "learning_rate": 4.2477287045975585e-05, "loss": 0.929, "step": 5876 }, { "epoch": 2.661684782608696, "grad_norm": 0.07735154456057525, "learning_rate": 4.246466388722182e-05, "loss": 0.9282, "step": 5877 }, { "epoch": 2.6621376811594204, "grad_norm": 0.1097261121289293, "learning_rate": 4.2452040482071694e-05, "loss": 0.9372, "step": 5878 }, { "epoch": 2.662590579710145, "grad_norm": 0.13613230522795405, "learning_rate": 4.2439416831787204e-05, "loss": 0.9177, "step": 5879 }, { "epoch": 2.6630434782608696, "grad_norm": 0.10462635667142779, "learning_rate": 4.242679293763035e-05, "loss": 0.9242, "step": 5880 }, { "epoch": 2.6634963768115942, "grad_norm": 0.09023318404542972, "learning_rate": 4.241416880086316e-05, "loss": 0.9261, "step": 5881 }, { "epoch": 2.663949275362319, "grad_norm": 0.07313251439258844, "learning_rate": 4.240154442274768e-05, "loss": 0.9356, "step": 5882 }, { "epoch": 2.6644021739130435, "grad_norm": 0.09395186490925463, "learning_rate": 4.2388919804545993e-05, "loss": 0.9379, "step": 5883 }, { "epoch": 2.664855072463768, "grad_norm": 0.07937655550091592, "learning_rate": 4.2376294947520226e-05, "loss": 0.9383, "step": 5884 }, { "epoch": 2.6653079710144927, "grad_norm": 0.08347630379836178, "learning_rate": 4.2363669852932475e-05, "loss": 0.949, "step": 5885 }, { "epoch": 2.6657608695652173, "grad_norm": 0.08000824792896279, "learning_rate": 4.235104452204492e-05, "loss": 0.9416, "step": 5886 }, { "epoch": 2.666213768115942, "grad_norm": 0.07724686282793197, "learning_rate": 4.233841895611971e-05, "loss": 0.9439, "step": 5887 }, { "epoch": 2.6666666666666665, "grad_norm": 0.08182265793303635, "learning_rate": 4.2325793156419035e-05, "loss": 0.9428, "step": 5888 }, { "epoch": 2.6671195652173916, "grad_norm": 0.09318203081295101, "learning_rate": 4.2313167124205154e-05, "loss": 0.9399, "step": 5889 }, { "epoch": 2.667572463768116, "grad_norm": 0.0769281518693663, "learning_rate": 4.230054086074028e-05, "loss": 0.9174, "step": 5890 }, { "epoch": 2.668025362318841, "grad_norm": 0.07568774563638551, "learning_rate": 4.228791436728669e-05, "loss": 0.9339, "step": 5891 }, { "epoch": 2.6684782608695654, "grad_norm": 0.10013853655627926, "learning_rate": 4.227528764510668e-05, "loss": 0.9234, "step": 5892 }, { "epoch": 2.66893115942029, "grad_norm": 0.08893386452568579, "learning_rate": 4.226266069546256e-05, "loss": 0.9448, "step": 5893 }, { "epoch": 2.6693840579710146, "grad_norm": 0.07421501932738714, "learning_rate": 4.225003351961666e-05, "loss": 0.9309, "step": 5894 }, { "epoch": 2.6698369565217392, "grad_norm": 0.09697655113767678, "learning_rate": 4.223740611883134e-05, "loss": 0.9413, "step": 5895 }, { "epoch": 2.670289855072464, "grad_norm": 0.11684313638279964, "learning_rate": 4.2224778494368986e-05, "loss": 0.9459, "step": 5896 }, { "epoch": 2.6707427536231885, "grad_norm": 0.1090335457393001, "learning_rate": 4.2212150647492e-05, "loss": 0.9395, "step": 5897 }, { "epoch": 2.671195652173913, "grad_norm": 0.07042340391960274, "learning_rate": 4.219952257946282e-05, "loss": 0.9249, "step": 5898 }, { "epoch": 2.6716485507246377, "grad_norm": 0.07205963696995969, "learning_rate": 4.218689429154387e-05, "loss": 0.9197, "step": 5899 }, { "epoch": 2.6721014492753623, "grad_norm": 0.09928788021375953, "learning_rate": 4.2174265784997635e-05, "loss": 0.9331, "step": 5900 }, { "epoch": 2.672554347826087, "grad_norm": 0.104645268579504, "learning_rate": 4.216163706108661e-05, "loss": 0.9384, "step": 5901 }, { "epoch": 2.6730072463768115, "grad_norm": 0.07577585088774537, "learning_rate": 4.2149008121073305e-05, "loss": 0.9388, "step": 5902 }, { "epoch": 2.673460144927536, "grad_norm": 0.11222064659003696, "learning_rate": 4.213637896622023e-05, "loss": 0.9484, "step": 5903 }, { "epoch": 2.6739130434782608, "grad_norm": 0.1104649044205937, "learning_rate": 4.2123749597789984e-05, "loss": 0.9163, "step": 5904 }, { "epoch": 2.6743659420289854, "grad_norm": 0.1014255955056946, "learning_rate": 4.211112001704513e-05, "loss": 0.9264, "step": 5905 }, { "epoch": 2.67481884057971, "grad_norm": 0.08553056196103852, "learning_rate": 4.209849022524825e-05, "loss": 0.931, "step": 5906 }, { "epoch": 2.6752717391304346, "grad_norm": 0.09730427994014658, "learning_rate": 4.2085860223661984e-05, "loss": 0.928, "step": 5907 }, { "epoch": 2.675724637681159, "grad_norm": 0.12917161636168994, "learning_rate": 4.207323001354896e-05, "loss": 0.9345, "step": 5908 }, { "epoch": 2.676177536231884, "grad_norm": 0.13313548273451578, "learning_rate": 4.206059959617185e-05, "loss": 0.9245, "step": 5909 }, { "epoch": 2.6766304347826084, "grad_norm": 0.06783327656998366, "learning_rate": 4.204796897279334e-05, "loss": 0.923, "step": 5910 }, { "epoch": 2.6770833333333335, "grad_norm": 0.09033090725121672, "learning_rate": 4.203533814467611e-05, "loss": 0.9536, "step": 5911 }, { "epoch": 2.677536231884058, "grad_norm": 0.08880231444057604, "learning_rate": 4.202270711308291e-05, "loss": 0.927, "step": 5912 }, { "epoch": 2.6779891304347827, "grad_norm": 0.09863774658813933, "learning_rate": 4.2010075879276474e-05, "loss": 0.9299, "step": 5913 }, { "epoch": 2.6784420289855073, "grad_norm": 0.07995876655315362, "learning_rate": 4.199744444451957e-05, "loss": 0.9459, "step": 5914 }, { "epoch": 2.678894927536232, "grad_norm": 0.09400828626136823, "learning_rate": 4.198481281007498e-05, "loss": 0.945, "step": 5915 }, { "epoch": 2.6793478260869565, "grad_norm": 0.08909932585702789, "learning_rate": 4.1972180977205495e-05, "loss": 0.9369, "step": 5916 }, { "epoch": 2.679800724637681, "grad_norm": 0.09964914536048572, "learning_rate": 4.195954894717396e-05, "loss": 0.9144, "step": 5917 }, { "epoch": 2.6802536231884058, "grad_norm": 0.07629241256590955, "learning_rate": 4.1946916721243215e-05, "loss": 0.9337, "step": 5918 }, { "epoch": 2.6807065217391304, "grad_norm": 0.11250749090155676, "learning_rate": 4.193428430067612e-05, "loss": 0.9459, "step": 5919 }, { "epoch": 2.681159420289855, "grad_norm": 0.11322595357958282, "learning_rate": 4.1921651686735555e-05, "loss": 0.9747, "step": 5920 }, { "epoch": 2.6816123188405796, "grad_norm": 0.10634705206406073, "learning_rate": 4.190901888068443e-05, "loss": 0.9318, "step": 5921 }, { "epoch": 2.682065217391304, "grad_norm": 0.10090898217455757, "learning_rate": 4.189638588378566e-05, "loss": 0.928, "step": 5922 }, { "epoch": 2.682518115942029, "grad_norm": 0.0850045008800531, "learning_rate": 4.18837526973022e-05, "loss": 0.9076, "step": 5923 }, { "epoch": 2.682971014492754, "grad_norm": 0.09200588140383686, "learning_rate": 4.187111932249698e-05, "loss": 0.9349, "step": 5924 }, { "epoch": 2.6834239130434785, "grad_norm": 0.11242613638134871, "learning_rate": 4.1858485760632996e-05, "loss": 0.9355, "step": 5925 }, { "epoch": 2.683876811594203, "grad_norm": 0.11581901229336714, "learning_rate": 4.184585201297325e-05, "loss": 0.9151, "step": 5926 }, { "epoch": 2.6843297101449277, "grad_norm": 0.10964135305877788, "learning_rate": 4.183321808078076e-05, "loss": 0.9291, "step": 5927 }, { "epoch": 2.6847826086956523, "grad_norm": 0.08359883784568199, "learning_rate": 4.1820583965318555e-05, "loss": 0.9293, "step": 5928 }, { "epoch": 2.685235507246377, "grad_norm": 0.1128919662564389, "learning_rate": 4.180794966784968e-05, "loss": 0.9371, "step": 5929 }, { "epoch": 2.6856884057971016, "grad_norm": 0.13125575655833477, "learning_rate": 4.17953151896372e-05, "loss": 0.9268, "step": 5930 }, { "epoch": 2.686141304347826, "grad_norm": 0.13488732469297068, "learning_rate": 4.1782680531944234e-05, "loss": 0.925, "step": 5931 }, { "epoch": 2.6865942028985508, "grad_norm": 0.09394791739961289, "learning_rate": 4.1770045696033866e-05, "loss": 0.93, "step": 5932 }, { "epoch": 2.6870471014492754, "grad_norm": 0.11530707435648112, "learning_rate": 4.1757410683169214e-05, "loss": 0.9214, "step": 5933 }, { "epoch": 2.6875, "grad_norm": 0.12416078638315514, "learning_rate": 4.174477549461345e-05, "loss": 0.9361, "step": 5934 }, { "epoch": 2.6879528985507246, "grad_norm": 0.10632885652004806, "learning_rate": 4.17321401316297e-05, "loss": 0.9194, "step": 5935 }, { "epoch": 2.6884057971014492, "grad_norm": 0.07513746114042368, "learning_rate": 4.171950459548116e-05, "loss": 0.8987, "step": 5936 }, { "epoch": 2.688858695652174, "grad_norm": 0.1094711447660381, "learning_rate": 4.170686888743102e-05, "loss": 0.9058, "step": 5937 }, { "epoch": 2.6893115942028984, "grad_norm": 0.08209138730301466, "learning_rate": 4.169423300874249e-05, "loss": 0.9145, "step": 5938 }, { "epoch": 2.689764492753623, "grad_norm": 0.09289506104162759, "learning_rate": 4.168159696067882e-05, "loss": 0.9139, "step": 5939 }, { "epoch": 2.6902173913043477, "grad_norm": 0.08005438484453362, "learning_rate": 4.166896074450322e-05, "loss": 0.9057, "step": 5940 }, { "epoch": 2.6906702898550723, "grad_norm": 0.0837474063588873, "learning_rate": 4.165632436147897e-05, "loss": 0.936, "step": 5941 }, { "epoch": 2.691123188405797, "grad_norm": 0.09717682709352485, "learning_rate": 4.1643687812869344e-05, "loss": 0.9374, "step": 5942 }, { "epoch": 2.6915760869565215, "grad_norm": 0.08140672057383135, "learning_rate": 4.163105109993765e-05, "loss": 0.9202, "step": 5943 }, { "epoch": 2.692028985507246, "grad_norm": 0.09348428250809512, "learning_rate": 4.16184142239472e-05, "loss": 0.9486, "step": 5944 }, { "epoch": 2.692481884057971, "grad_norm": 0.07004708628172547, "learning_rate": 4.1605777186161305e-05, "loss": 0.9315, "step": 5945 }, { "epoch": 2.692934782608696, "grad_norm": 0.09536339319724774, "learning_rate": 4.159313998784332e-05, "loss": 0.9345, "step": 5946 }, { "epoch": 2.6933876811594204, "grad_norm": 0.07922784960035487, "learning_rate": 4.1580502630256614e-05, "loss": 0.92, "step": 5947 }, { "epoch": 2.693840579710145, "grad_norm": 0.07530114049357202, "learning_rate": 4.1567865114664554e-05, "loss": 0.9364, "step": 5948 }, { "epoch": 2.6942934782608696, "grad_norm": 0.0860620145370956, "learning_rate": 4.155522744233053e-05, "loss": 0.9363, "step": 5949 }, { "epoch": 2.6947463768115942, "grad_norm": 0.09068455756624252, "learning_rate": 4.154258961451795e-05, "loss": 0.9224, "step": 5950 }, { "epoch": 2.695199275362319, "grad_norm": 0.08264227814416682, "learning_rate": 4.152995163249025e-05, "loss": 0.9342, "step": 5951 }, { "epoch": 2.6956521739130435, "grad_norm": 0.13294207399682303, "learning_rate": 4.151731349751086e-05, "loss": 0.9035, "step": 5952 }, { "epoch": 2.696105072463768, "grad_norm": 0.09455473951506507, "learning_rate": 4.150467521084323e-05, "loss": 0.9339, "step": 5953 }, { "epoch": 2.6965579710144927, "grad_norm": 0.09561189207784182, "learning_rate": 4.1492036773750835e-05, "loss": 0.9533, "step": 5954 }, { "epoch": 2.6970108695652173, "grad_norm": 0.10721985224498554, "learning_rate": 4.147939818749716e-05, "loss": 0.9421, "step": 5955 }, { "epoch": 2.697463768115942, "grad_norm": 0.08112391076624006, "learning_rate": 4.146675945334572e-05, "loss": 0.9525, "step": 5956 }, { "epoch": 2.6979166666666665, "grad_norm": 0.09616237639089274, "learning_rate": 4.145412057256e-05, "loss": 0.9273, "step": 5957 }, { "epoch": 2.6983695652173916, "grad_norm": 0.09912876736692072, "learning_rate": 4.144148154640354e-05, "loss": 0.9196, "step": 5958 }, { "epoch": 2.698822463768116, "grad_norm": 0.09999867390079274, "learning_rate": 4.142884237613988e-05, "loss": 0.937, "step": 5959 }, { "epoch": 2.699275362318841, "grad_norm": 0.09732240641637807, "learning_rate": 4.1416203063032594e-05, "loss": 0.9308, "step": 5960 }, { "epoch": 2.6997282608695654, "grad_norm": 0.08412036319149142, "learning_rate": 4.140356360834524e-05, "loss": 0.934, "step": 5961 }, { "epoch": 2.70018115942029, "grad_norm": 0.08053915531252216, "learning_rate": 4.13909240133414e-05, "loss": 0.9407, "step": 5962 }, { "epoch": 2.7006340579710146, "grad_norm": 0.07946600194163929, "learning_rate": 4.1378284279284685e-05, "loss": 0.9168, "step": 5963 }, { "epoch": 2.7010869565217392, "grad_norm": 0.09045421372858141, "learning_rate": 4.136564440743872e-05, "loss": 0.9413, "step": 5964 }, { "epoch": 2.701539855072464, "grad_norm": 0.13369601772972936, "learning_rate": 4.135300439906711e-05, "loss": 0.9251, "step": 5965 }, { "epoch": 2.7019927536231885, "grad_norm": 0.11366257206685017, "learning_rate": 4.134036425543351e-05, "loss": 0.9155, "step": 5966 }, { "epoch": 2.702445652173913, "grad_norm": 0.08698743553113018, "learning_rate": 4.132772397780157e-05, "loss": 0.9257, "step": 5967 }, { "epoch": 2.7028985507246377, "grad_norm": 0.10018917019860277, "learning_rate": 4.1315083567434954e-05, "loss": 0.9255, "step": 5968 }, { "epoch": 2.7033514492753623, "grad_norm": 0.10815832222327844, "learning_rate": 4.130244302559736e-05, "loss": 0.9125, "step": 5969 }, { "epoch": 2.703804347826087, "grad_norm": 0.09990534698718925, "learning_rate": 4.128980235355247e-05, "loss": 0.9483, "step": 5970 }, { "epoch": 2.7042572463768115, "grad_norm": 0.12065999496246288, "learning_rate": 4.127716155256399e-05, "loss": 0.9434, "step": 5971 }, { "epoch": 2.704710144927536, "grad_norm": 0.07922666745934823, "learning_rate": 4.126452062389565e-05, "loss": 0.9201, "step": 5972 }, { "epoch": 2.7051630434782608, "grad_norm": 0.09247942446755153, "learning_rate": 4.125187956881117e-05, "loss": 0.9471, "step": 5973 }, { "epoch": 2.7056159420289854, "grad_norm": 0.0994155522453105, "learning_rate": 4.123923838857431e-05, "loss": 0.93, "step": 5974 }, { "epoch": 2.70606884057971, "grad_norm": 0.10243120304421115, "learning_rate": 4.1226597084448834e-05, "loss": 0.9396, "step": 5975 }, { "epoch": 2.7065217391304346, "grad_norm": 0.08020131882007285, "learning_rate": 4.12139556576985e-05, "loss": 0.9318, "step": 5976 }, { "epoch": 2.706974637681159, "grad_norm": 0.08548483357828918, "learning_rate": 4.1201314109587096e-05, "loss": 0.9062, "step": 5977 }, { "epoch": 2.707427536231884, "grad_norm": 0.10711207609424638, "learning_rate": 4.118867244137841e-05, "loss": 0.9339, "step": 5978 }, { "epoch": 2.7078804347826084, "grad_norm": 0.10854206563831106, "learning_rate": 4.117603065433626e-05, "loss": 0.939, "step": 5979 }, { "epoch": 2.7083333333333335, "grad_norm": 0.08859850172199542, "learning_rate": 4.116338874972446e-05, "loss": 0.9234, "step": 5980 }, { "epoch": 2.708786231884058, "grad_norm": 0.07986324586765467, "learning_rate": 4.1150746728806844e-05, "loss": 0.9481, "step": 5981 }, { "epoch": 2.7092391304347827, "grad_norm": 0.09764470329116373, "learning_rate": 4.113810459284726e-05, "loss": 0.9187, "step": 5982 }, { "epoch": 2.7096920289855073, "grad_norm": 0.11330542969467357, "learning_rate": 4.112546234310955e-05, "loss": 0.9386, "step": 5983 }, { "epoch": 2.710144927536232, "grad_norm": 0.0954929272570336, "learning_rate": 4.111281998085758e-05, "loss": 0.9268, "step": 5984 }, { "epoch": 2.7105978260869565, "grad_norm": 0.09323670691477003, "learning_rate": 4.110017750735524e-05, "loss": 0.9559, "step": 5985 }, { "epoch": 2.711050724637681, "grad_norm": 0.10471909823017932, "learning_rate": 4.108753492386642e-05, "loss": 0.9262, "step": 5986 }, { "epoch": 2.7115036231884058, "grad_norm": 0.12858746516725103, "learning_rate": 4.1074892231655e-05, "loss": 0.9422, "step": 5987 }, { "epoch": 2.7119565217391304, "grad_norm": 0.09130010221604286, "learning_rate": 4.10622494319849e-05, "loss": 0.9505, "step": 5988 }, { "epoch": 2.712409420289855, "grad_norm": 0.08138872530454105, "learning_rate": 4.104960652612004e-05, "loss": 0.9234, "step": 5989 }, { "epoch": 2.7128623188405796, "grad_norm": 0.08664786269164033, "learning_rate": 4.103696351532437e-05, "loss": 0.922, "step": 5990 }, { "epoch": 2.713315217391304, "grad_norm": 0.10364086204004401, "learning_rate": 4.102432040086179e-05, "loss": 0.9196, "step": 5991 }, { "epoch": 2.713768115942029, "grad_norm": 0.09203650194546875, "learning_rate": 4.101167718399628e-05, "loss": 0.9215, "step": 5992 }, { "epoch": 2.714221014492754, "grad_norm": 0.07573904841271709, "learning_rate": 4.099903386599181e-05, "loss": 0.9364, "step": 5993 }, { "epoch": 2.7146739130434785, "grad_norm": 0.10566580442210692, "learning_rate": 4.098639044811234e-05, "loss": 0.9326, "step": 5994 }, { "epoch": 2.715126811594203, "grad_norm": 0.13542493443798626, "learning_rate": 4.0973746931621855e-05, "loss": 0.9556, "step": 5995 }, { "epoch": 2.7155797101449277, "grad_norm": 0.11110422790012747, "learning_rate": 4.096110331778434e-05, "loss": 0.9609, "step": 5996 }, { "epoch": 2.7160326086956523, "grad_norm": 0.08653733727123153, "learning_rate": 4.09484596078638e-05, "loss": 0.9238, "step": 5997 }, { "epoch": 2.716485507246377, "grad_norm": 0.09187231047605796, "learning_rate": 4.093581580312427e-05, "loss": 0.9253, "step": 5998 }, { "epoch": 2.7169384057971016, "grad_norm": 0.10009743221158815, "learning_rate": 4.092317190482972e-05, "loss": 0.9414, "step": 5999 }, { "epoch": 2.717391304347826, "grad_norm": 0.11497195665069522, "learning_rate": 4.091052791424423e-05, "loss": 0.9421, "step": 6000 }, { "epoch": 2.7178442028985508, "grad_norm": 0.10070798607666145, "learning_rate": 4.089788383263182e-05, "loss": 0.9376, "step": 6001 }, { "epoch": 2.7182971014492754, "grad_norm": 0.07266069947651076, "learning_rate": 4.088523966125654e-05, "loss": 0.9478, "step": 6002 }, { "epoch": 2.71875, "grad_norm": 0.09293515253516915, "learning_rate": 4.087259540138245e-05, "loss": 0.9235, "step": 6003 }, { "epoch": 2.7192028985507246, "grad_norm": 0.10417967953021993, "learning_rate": 4.085995105427361e-05, "loss": 0.9235, "step": 6004 }, { "epoch": 2.7196557971014492, "grad_norm": 0.10843207697937303, "learning_rate": 4.0847306621194096e-05, "loss": 0.9333, "step": 6005 }, { "epoch": 2.720108695652174, "grad_norm": 0.08455203996077891, "learning_rate": 4.0834662103408005e-05, "loss": 0.9586, "step": 6006 }, { "epoch": 2.7205615942028984, "grad_norm": 0.07451262726302185, "learning_rate": 4.082201750217943e-05, "loss": 0.9413, "step": 6007 }, { "epoch": 2.721014492753623, "grad_norm": 0.10691418537619916, "learning_rate": 4.080937281877245e-05, "loss": 0.9525, "step": 6008 }, { "epoch": 2.7214673913043477, "grad_norm": 0.10108904115418937, "learning_rate": 4.0796728054451186e-05, "loss": 0.9351, "step": 6009 }, { "epoch": 2.7219202898550723, "grad_norm": 0.08161817520901885, "learning_rate": 4.078408321047977e-05, "loss": 0.9384, "step": 6010 }, { "epoch": 2.722373188405797, "grad_norm": 0.09911433609089873, "learning_rate": 4.0771438288122294e-05, "loss": 0.9317, "step": 6011 }, { "epoch": 2.7228260869565215, "grad_norm": 0.11784119279369744, "learning_rate": 4.075879328864293e-05, "loss": 0.9424, "step": 6012 }, { "epoch": 2.723278985507246, "grad_norm": 0.11075177092943661, "learning_rate": 4.0746148213305784e-05, "loss": 0.9484, "step": 6013 }, { "epoch": 2.723731884057971, "grad_norm": 0.10098481547587998, "learning_rate": 4.073350306337502e-05, "loss": 0.9304, "step": 6014 }, { "epoch": 2.724184782608696, "grad_norm": 0.11120220986440403, "learning_rate": 4.072085784011479e-05, "loss": 0.9396, "step": 6015 }, { "epoch": 2.7246376811594204, "grad_norm": 0.1028467573422869, "learning_rate": 4.070821254478927e-05, "loss": 0.9344, "step": 6016 }, { "epoch": 2.725090579710145, "grad_norm": 0.09548380998494298, "learning_rate": 4.0695567178662607e-05, "loss": 0.9302, "step": 6017 }, { "epoch": 2.7255434782608696, "grad_norm": 0.08248443796952527, "learning_rate": 4.0682921742999e-05, "loss": 0.9567, "step": 6018 }, { "epoch": 2.7259963768115942, "grad_norm": 0.07858960807474029, "learning_rate": 4.067027623906262e-05, "loss": 0.9137, "step": 6019 }, { "epoch": 2.726449275362319, "grad_norm": 0.08242599235435166, "learning_rate": 4.0657630668117665e-05, "loss": 0.9214, "step": 6020 }, { "epoch": 2.7269021739130435, "grad_norm": 0.08318795365351976, "learning_rate": 4.064498503142833e-05, "loss": 0.9178, "step": 6021 }, { "epoch": 2.727355072463768, "grad_norm": 0.08180923142043844, "learning_rate": 4.063233933025881e-05, "loss": 0.9201, "step": 6022 }, { "epoch": 2.7278079710144927, "grad_norm": 0.08595731333895826, "learning_rate": 4.061969356587333e-05, "loss": 0.9356, "step": 6023 }, { "epoch": 2.7282608695652173, "grad_norm": 0.08374715117777187, "learning_rate": 4.060704773953611e-05, "loss": 0.9254, "step": 6024 }, { "epoch": 2.728713768115942, "grad_norm": 0.08490329917941507, "learning_rate": 4.059440185251136e-05, "loss": 0.9211, "step": 6025 }, { "epoch": 2.7291666666666665, "grad_norm": 0.08432714014014724, "learning_rate": 4.058175590606332e-05, "loss": 0.934, "step": 6026 }, { "epoch": 2.7296195652173916, "grad_norm": 0.08778581577667698, "learning_rate": 4.056910990145621e-05, "loss": 0.9376, "step": 6027 }, { "epoch": 2.730072463768116, "grad_norm": 0.09186732842033092, "learning_rate": 4.055646383995429e-05, "loss": 0.9614, "step": 6028 }, { "epoch": 2.730525362318841, "grad_norm": 0.0710372383040639, "learning_rate": 4.054381772282179e-05, "loss": 0.9359, "step": 6029 }, { "epoch": 2.7309782608695654, "grad_norm": 0.1001881704706878, "learning_rate": 4.053117155132298e-05, "loss": 0.9318, "step": 6030 }, { "epoch": 2.73143115942029, "grad_norm": 0.10217867493593134, "learning_rate": 4.05185253267221e-05, "loss": 0.9344, "step": 6031 }, { "epoch": 2.7318840579710146, "grad_norm": 0.08692302748532886, "learning_rate": 4.0505879050283424e-05, "loss": 0.9327, "step": 6032 }, { "epoch": 2.7323369565217392, "grad_norm": 0.08497503006416897, "learning_rate": 4.049323272327122e-05, "loss": 0.9429, "step": 6033 }, { "epoch": 2.732789855072464, "grad_norm": 0.10359534343200817, "learning_rate": 4.048058634694976e-05, "loss": 0.9444, "step": 6034 }, { "epoch": 2.7332427536231885, "grad_norm": 0.08001193435072754, "learning_rate": 4.046793992258332e-05, "loss": 0.9305, "step": 6035 }, { "epoch": 2.733695652173913, "grad_norm": 0.08269695461349383, "learning_rate": 4.045529345143617e-05, "loss": 0.9406, "step": 6036 }, { "epoch": 2.7341485507246377, "grad_norm": 0.11168096642138413, "learning_rate": 4.0442646934772635e-05, "loss": 0.9378, "step": 6037 }, { "epoch": 2.7346014492753623, "grad_norm": 0.10086763917681395, "learning_rate": 4.0430000373856976e-05, "loss": 0.93, "step": 6038 }, { "epoch": 2.735054347826087, "grad_norm": 0.08906447294205282, "learning_rate": 4.04173537699535e-05, "loss": 0.9114, "step": 6039 }, { "epoch": 2.7355072463768115, "grad_norm": 0.08794287203680382, "learning_rate": 4.04047071243265e-05, "loss": 0.9334, "step": 6040 }, { "epoch": 2.735960144927536, "grad_norm": 0.10095433463900409, "learning_rate": 4.039206043824029e-05, "loss": 0.9301, "step": 6041 }, { "epoch": 2.7364130434782608, "grad_norm": 0.08974706575572212, "learning_rate": 4.0379413712959176e-05, "loss": 0.9413, "step": 6042 }, { "epoch": 2.7368659420289854, "grad_norm": 0.07863371193702673, "learning_rate": 4.0366766949747484e-05, "loss": 0.9302, "step": 6043 }, { "epoch": 2.73731884057971, "grad_norm": 0.10508040038820576, "learning_rate": 4.03541201498695e-05, "loss": 0.9439, "step": 6044 }, { "epoch": 2.7377717391304346, "grad_norm": 0.13532419670703813, "learning_rate": 4.034147331458958e-05, "loss": 0.927, "step": 6045 }, { "epoch": 2.738224637681159, "grad_norm": 0.09018918383480645, "learning_rate": 4.0328826445172027e-05, "loss": 0.9268, "step": 6046 }, { "epoch": 2.738677536231884, "grad_norm": 0.08952836667880894, "learning_rate": 4.0316179542881164e-05, "loss": 0.9435, "step": 6047 }, { "epoch": 2.7391304347826084, "grad_norm": 0.12210826329150436, "learning_rate": 4.0303532608981336e-05, "loss": 0.936, "step": 6048 }, { "epoch": 2.7395833333333335, "grad_norm": 0.09584442322809775, "learning_rate": 4.029088564473688e-05, "loss": 0.9373, "step": 6049 }, { "epoch": 2.740036231884058, "grad_norm": 0.0917073299200752, "learning_rate": 4.027823865141211e-05, "loss": 0.9491, "step": 6050 }, { "epoch": 2.7404891304347827, "grad_norm": 0.11497565909046703, "learning_rate": 4.0265591630271393e-05, "loss": 0.9013, "step": 6051 }, { "epoch": 2.7409420289855073, "grad_norm": 0.12124827417067073, "learning_rate": 4.025294458257905e-05, "loss": 0.9266, "step": 6052 }, { "epoch": 2.741394927536232, "grad_norm": 0.08324611644468471, "learning_rate": 4.024029750959944e-05, "loss": 0.9265, "step": 6053 }, { "epoch": 2.7418478260869565, "grad_norm": 0.09255000866067223, "learning_rate": 4.0227650412596916e-05, "loss": 0.9268, "step": 6054 }, { "epoch": 2.742300724637681, "grad_norm": 0.10328806037076514, "learning_rate": 4.021500329283582e-05, "loss": 0.9348, "step": 6055 }, { "epoch": 2.7427536231884058, "grad_norm": 0.11483173083983318, "learning_rate": 4.020235615158049e-05, "loss": 0.945, "step": 6056 }, { "epoch": 2.7432065217391304, "grad_norm": 0.1045361502713755, "learning_rate": 4.0189708990095295e-05, "loss": 0.9488, "step": 6057 }, { "epoch": 2.743659420289855, "grad_norm": 0.07435545167912912, "learning_rate": 4.017706180964461e-05, "loss": 0.938, "step": 6058 }, { "epoch": 2.7441123188405796, "grad_norm": 0.11004632754983766, "learning_rate": 4.016441461149276e-05, "loss": 0.9357, "step": 6059 }, { "epoch": 2.744565217391304, "grad_norm": 0.11743176194107645, "learning_rate": 4.015176739690413e-05, "loss": 0.9193, "step": 6060 }, { "epoch": 2.745018115942029, "grad_norm": 0.10542202413200188, "learning_rate": 4.0139120167143065e-05, "loss": 0.9385, "step": 6061 }, { "epoch": 2.745471014492754, "grad_norm": 0.09859671635552096, "learning_rate": 4.012647292347394e-05, "loss": 0.9389, "step": 6062 }, { "epoch": 2.7459239130434785, "grad_norm": 0.12744611691894966, "learning_rate": 4.0113825667161114e-05, "loss": 0.9299, "step": 6063 }, { "epoch": 2.746376811594203, "grad_norm": 0.10149464894511358, "learning_rate": 4.010117839946896e-05, "loss": 0.9294, "step": 6064 }, { "epoch": 2.7468297101449277, "grad_norm": 0.0859820171780499, "learning_rate": 4.008853112166183e-05, "loss": 0.9557, "step": 6065 }, { "epoch": 2.7472826086956523, "grad_norm": 0.09845277842938446, "learning_rate": 4.0075883835004124e-05, "loss": 0.9331, "step": 6066 }, { "epoch": 2.747735507246377, "grad_norm": 0.08570784731182404, "learning_rate": 4.006323654076017e-05, "loss": 0.9276, "step": 6067 }, { "epoch": 2.7481884057971016, "grad_norm": 0.10047943714266566, "learning_rate": 4.005058924019436e-05, "loss": 0.9342, "step": 6068 }, { "epoch": 2.748641304347826, "grad_norm": 0.08990948187341292, "learning_rate": 4.0037941934571075e-05, "loss": 0.9171, "step": 6069 }, { "epoch": 2.7490942028985508, "grad_norm": 0.08769884674627526, "learning_rate": 4.002529462515468e-05, "loss": 0.9482, "step": 6070 }, { "epoch": 2.7495471014492754, "grad_norm": 0.09149431254615162, "learning_rate": 4.001264731320953e-05, "loss": 0.9046, "step": 6071 }, { "epoch": 2.75, "grad_norm": 0.07847475594458256, "learning_rate": 4e-05, "loss": 0.9454, "step": 6072 }, { "epoch": 2.7504528985507246, "grad_norm": 0.08080060769683617, "learning_rate": 3.998735268679049e-05, "loss": 0.9355, "step": 6073 }, { "epoch": 2.7509057971014492, "grad_norm": 0.0766127599795585, "learning_rate": 3.997470537484533e-05, "loss": 0.9278, "step": 6074 }, { "epoch": 2.751358695652174, "grad_norm": 0.07247796334482502, "learning_rate": 3.9962058065428945e-05, "loss": 0.9222, "step": 6075 }, { "epoch": 2.7518115942028984, "grad_norm": 0.07342220226122095, "learning_rate": 3.9949410759805645e-05, "loss": 0.9373, "step": 6076 }, { "epoch": 2.752264492753623, "grad_norm": 0.08444045827095799, "learning_rate": 3.993676345923985e-05, "loss": 0.9462, "step": 6077 }, { "epoch": 2.7527173913043477, "grad_norm": 0.08545795491502362, "learning_rate": 3.992411616499589e-05, "loss": 0.9452, "step": 6078 }, { "epoch": 2.7531702898550723, "grad_norm": 0.08099904398926713, "learning_rate": 3.991146887833818e-05, "loss": 0.8991, "step": 6079 }, { "epoch": 2.753623188405797, "grad_norm": 0.06287716196950353, "learning_rate": 3.9898821600531054e-05, "loss": 0.9546, "step": 6080 }, { "epoch": 2.7540760869565215, "grad_norm": 0.07703433549837037, "learning_rate": 3.9886174332838906e-05, "loss": 0.9271, "step": 6081 }, { "epoch": 2.754528985507246, "grad_norm": 0.09376529989253492, "learning_rate": 3.987352707652607e-05, "loss": 0.9331, "step": 6082 }, { "epoch": 2.754981884057971, "grad_norm": 0.07101730642523313, "learning_rate": 3.9860879832856955e-05, "loss": 0.9395, "step": 6083 }, { "epoch": 2.755434782608696, "grad_norm": 0.09248440191750493, "learning_rate": 3.984823260309589e-05, "loss": 0.9443, "step": 6084 }, { "epoch": 2.7558876811594204, "grad_norm": 0.09898415019126695, "learning_rate": 3.983558538850726e-05, "loss": 0.9492, "step": 6085 }, { "epoch": 2.756340579710145, "grad_norm": 0.07337887748732015, "learning_rate": 3.98229381903554e-05, "loss": 0.9237, "step": 6086 }, { "epoch": 2.7567934782608696, "grad_norm": 0.10174419729307993, "learning_rate": 3.981029100990471e-05, "loss": 0.9483, "step": 6087 }, { "epoch": 2.7572463768115942, "grad_norm": 0.10368488417152566, "learning_rate": 3.9797643848419524e-05, "loss": 0.9452, "step": 6088 }, { "epoch": 2.757699275362319, "grad_norm": 0.07527426716631488, "learning_rate": 3.978499670716421e-05, "loss": 0.9511, "step": 6089 }, { "epoch": 2.7581521739130435, "grad_norm": 0.11111227747250094, "learning_rate": 3.977234958740309e-05, "loss": 0.9318, "step": 6090 }, { "epoch": 2.758605072463768, "grad_norm": 0.11991883414634734, "learning_rate": 3.975970249040057e-05, "loss": 0.9473, "step": 6091 }, { "epoch": 2.7590579710144927, "grad_norm": 0.11035222116864046, "learning_rate": 3.974705541742096e-05, "loss": 0.9337, "step": 6092 }, { "epoch": 2.7595108695652173, "grad_norm": 0.07664524750888214, "learning_rate": 3.973440836972863e-05, "loss": 0.9338, "step": 6093 }, { "epoch": 2.759963768115942, "grad_norm": 0.09203312583289672, "learning_rate": 3.97217613485879e-05, "loss": 0.9171, "step": 6094 }, { "epoch": 2.7604166666666665, "grad_norm": 0.10124740408551021, "learning_rate": 3.970911435526314e-05, "loss": 0.9472, "step": 6095 }, { "epoch": 2.7608695652173916, "grad_norm": 0.10445808473436057, "learning_rate": 3.969646739101868e-05, "loss": 0.9393, "step": 6096 }, { "epoch": 2.761322463768116, "grad_norm": 0.09259640808222225, "learning_rate": 3.9683820457118856e-05, "loss": 0.9321, "step": 6097 }, { "epoch": 2.761775362318841, "grad_norm": 0.11154610967327988, "learning_rate": 3.9671173554827994e-05, "loss": 0.9058, "step": 6098 }, { "epoch": 2.7622282608695654, "grad_norm": 0.10383390438279666, "learning_rate": 3.965852668541044e-05, "loss": 0.9298, "step": 6099 }, { "epoch": 2.76268115942029, "grad_norm": 0.09063117919032253, "learning_rate": 3.964587985013051e-05, "loss": 0.9273, "step": 6100 }, { "epoch": 2.7631340579710146, "grad_norm": 0.09795237481682975, "learning_rate": 3.963323305025254e-05, "loss": 0.9227, "step": 6101 }, { "epoch": 2.7635869565217392, "grad_norm": 0.10506386394938393, "learning_rate": 3.962058628704083e-05, "loss": 0.9662, "step": 6102 }, { "epoch": 2.764039855072464, "grad_norm": 0.09009626014930701, "learning_rate": 3.960793956175972e-05, "loss": 0.9429, "step": 6103 }, { "epoch": 2.7644927536231885, "grad_norm": 0.08364628484667025, "learning_rate": 3.959529287567351e-05, "loss": 0.936, "step": 6104 }, { "epoch": 2.764945652173913, "grad_norm": 0.09828258810687099, "learning_rate": 3.958264623004653e-05, "loss": 0.9357, "step": 6105 }, { "epoch": 2.7653985507246377, "grad_norm": 0.09756631715290477, "learning_rate": 3.956999962614304e-05, "loss": 0.9448, "step": 6106 }, { "epoch": 2.7658514492753623, "grad_norm": 0.08822923316347747, "learning_rate": 3.955735306522738e-05, "loss": 0.8973, "step": 6107 }, { "epoch": 2.766304347826087, "grad_norm": 0.08724828038895886, "learning_rate": 3.9544706548563837e-05, "loss": 0.9327, "step": 6108 }, { "epoch": 2.7667572463768115, "grad_norm": 0.08245030598207133, "learning_rate": 3.953206007741671e-05, "loss": 0.9361, "step": 6109 }, { "epoch": 2.767210144927536, "grad_norm": 0.09144839410832543, "learning_rate": 3.9519413653050255e-05, "loss": 0.9245, "step": 6110 }, { "epoch": 2.7676630434782608, "grad_norm": 0.10125282065368557, "learning_rate": 3.95067672767288e-05, "loss": 0.9445, "step": 6111 }, { "epoch": 2.7681159420289854, "grad_norm": 0.06464208362366672, "learning_rate": 3.949412094971659e-05, "loss": 0.9398, "step": 6112 }, { "epoch": 2.76856884057971, "grad_norm": 0.09845205118221105, "learning_rate": 3.948147467327792e-05, "loss": 0.9332, "step": 6113 }, { "epoch": 2.7690217391304346, "grad_norm": 0.08455464359639818, "learning_rate": 3.946882844867703e-05, "loss": 0.9091, "step": 6114 }, { "epoch": 2.769474637681159, "grad_norm": 0.06924619889856504, "learning_rate": 3.945618227717823e-05, "loss": 0.9314, "step": 6115 }, { "epoch": 2.769927536231884, "grad_norm": 0.09048814692721432, "learning_rate": 3.944353616004572e-05, "loss": 0.9253, "step": 6116 }, { "epoch": 2.7703804347826084, "grad_norm": 0.08201734705064753, "learning_rate": 3.9430890098543815e-05, "loss": 0.908, "step": 6117 }, { "epoch": 2.7708333333333335, "grad_norm": 0.0822973918775641, "learning_rate": 3.9418244093936694e-05, "loss": 0.9548, "step": 6118 }, { "epoch": 2.771286231884058, "grad_norm": 0.08333067221820024, "learning_rate": 3.940559814748866e-05, "loss": 0.9033, "step": 6119 }, { "epoch": 2.7717391304347827, "grad_norm": 0.07707394677415101, "learning_rate": 3.93929522604639e-05, "loss": 0.9091, "step": 6120 }, { "epoch": 2.7721920289855073, "grad_norm": 0.09308138363722522, "learning_rate": 3.9380306434126675e-05, "loss": 0.9235, "step": 6121 }, { "epoch": 2.772644927536232, "grad_norm": 0.09849634679644789, "learning_rate": 3.9367660669741196e-05, "loss": 0.9281, "step": 6122 }, { "epoch": 2.7730978260869565, "grad_norm": 0.11025946118682592, "learning_rate": 3.9355014968571696e-05, "loss": 0.9239, "step": 6123 }, { "epoch": 2.773550724637681, "grad_norm": 0.06455278791829626, "learning_rate": 3.934236933188235e-05, "loss": 0.9079, "step": 6124 }, { "epoch": 2.7740036231884058, "grad_norm": 0.09974340178198869, "learning_rate": 3.9329723760937395e-05, "loss": 0.9243, "step": 6125 }, { "epoch": 2.7744565217391304, "grad_norm": 0.08562158322752038, "learning_rate": 3.9317078257001014e-05, "loss": 0.9256, "step": 6126 }, { "epoch": 2.774909420289855, "grad_norm": 0.07888482810573716, "learning_rate": 3.9304432821337414e-05, "loss": 0.9236, "step": 6127 }, { "epoch": 2.7753623188405796, "grad_norm": 0.07939050757300206, "learning_rate": 3.9291787455210746e-05, "loss": 0.9168, "step": 6128 }, { "epoch": 2.775815217391304, "grad_norm": 0.07949960290458129, "learning_rate": 3.9279142159885224e-05, "loss": 0.9264, "step": 6129 }, { "epoch": 2.776268115942029, "grad_norm": 0.07924754633458893, "learning_rate": 3.9266496936624994e-05, "loss": 0.9606, "step": 6130 }, { "epoch": 2.776721014492754, "grad_norm": 0.07419396183988602, "learning_rate": 3.9253851786694236e-05, "loss": 0.9384, "step": 6131 }, { "epoch": 2.7771739130434785, "grad_norm": 0.07281972659668882, "learning_rate": 3.924120671135708e-05, "loss": 0.9379, "step": 6132 }, { "epoch": 2.777626811594203, "grad_norm": 0.07734746761418652, "learning_rate": 3.922856171187771e-05, "loss": 0.9289, "step": 6133 }, { "epoch": 2.7780797101449277, "grad_norm": 0.08954699434224897, "learning_rate": 3.9215916789520246e-05, "loss": 0.9145, "step": 6134 }, { "epoch": 2.7785326086956523, "grad_norm": 0.07685257171923741, "learning_rate": 3.920327194554883e-05, "loss": 0.9394, "step": 6135 }, { "epoch": 2.778985507246377, "grad_norm": 0.0745816922713181, "learning_rate": 3.919062718122755e-05, "loss": 0.9544, "step": 6136 }, { "epoch": 2.7794384057971016, "grad_norm": 0.08478519314183115, "learning_rate": 3.917798249782058e-05, "loss": 0.9081, "step": 6137 }, { "epoch": 2.779891304347826, "grad_norm": 0.11534157414638403, "learning_rate": 3.9165337896591995e-05, "loss": 0.9316, "step": 6138 }, { "epoch": 2.7803442028985508, "grad_norm": 0.10224414674857749, "learning_rate": 3.915269337880591e-05, "loss": 0.9364, "step": 6139 }, { "epoch": 2.7807971014492754, "grad_norm": 0.08066950299187436, "learning_rate": 3.914004894572639e-05, "loss": 0.9264, "step": 6140 }, { "epoch": 2.78125, "grad_norm": 0.07349316582967984, "learning_rate": 3.912740459861756e-05, "loss": 0.9326, "step": 6141 }, { "epoch": 2.7817028985507246, "grad_norm": 0.09692111662814275, "learning_rate": 3.911476033874347e-05, "loss": 0.9532, "step": 6142 }, { "epoch": 2.7821557971014492, "grad_norm": 0.09365902676277953, "learning_rate": 3.91021161673682e-05, "loss": 0.9107, "step": 6143 }, { "epoch": 2.782608695652174, "grad_norm": 0.07627528838250433, "learning_rate": 3.908947208575577e-05, "loss": 0.9301, "step": 6144 }, { "epoch": 2.7830615942028984, "grad_norm": 0.08289578484712143, "learning_rate": 3.9076828095170286e-05, "loss": 0.9232, "step": 6145 }, { "epoch": 2.783514492753623, "grad_norm": 0.11139075027902495, "learning_rate": 3.9064184196875753e-05, "loss": 0.9411, "step": 6146 }, { "epoch": 2.7839673913043477, "grad_norm": 0.11378263946260375, "learning_rate": 3.9051540392136213e-05, "loss": 0.9432, "step": 6147 }, { "epoch": 2.7844202898550723, "grad_norm": 0.0740215188200862, "learning_rate": 3.903889668221567e-05, "loss": 0.9568, "step": 6148 }, { "epoch": 2.784873188405797, "grad_norm": 0.09322938940013517, "learning_rate": 3.902625306837816e-05, "loss": 0.9187, "step": 6149 }, { "epoch": 2.7853260869565215, "grad_norm": 0.08818810677019528, "learning_rate": 3.901360955188767e-05, "loss": 0.9534, "step": 6150 }, { "epoch": 2.785778985507246, "grad_norm": 0.07428258721684239, "learning_rate": 3.9000966134008203e-05, "loss": 0.9505, "step": 6151 }, { "epoch": 2.786231884057971, "grad_norm": 0.08018205085709867, "learning_rate": 3.898832281600371e-05, "loss": 0.9198, "step": 6152 }, { "epoch": 2.786684782608696, "grad_norm": 0.09482915828390696, "learning_rate": 3.897567959913822e-05, "loss": 0.9304, "step": 6153 }, { "epoch": 2.7871376811594204, "grad_norm": 0.09470029986505482, "learning_rate": 3.8963036484675645e-05, "loss": 0.9377, "step": 6154 }, { "epoch": 2.787590579710145, "grad_norm": 0.06409076090255016, "learning_rate": 3.895039347387997e-05, "loss": 0.937, "step": 6155 }, { "epoch": 2.7880434782608696, "grad_norm": 0.08570763970342005, "learning_rate": 3.8937750568015106e-05, "loss": 0.9396, "step": 6156 }, { "epoch": 2.7884963768115942, "grad_norm": 0.08572767777389216, "learning_rate": 3.892510776834501e-05, "loss": 0.9328, "step": 6157 }, { "epoch": 2.788949275362319, "grad_norm": 0.0932627647899418, "learning_rate": 3.891246507613359e-05, "loss": 0.9309, "step": 6158 }, { "epoch": 2.7894021739130435, "grad_norm": 0.07952105998359675, "learning_rate": 3.889982249264477e-05, "loss": 0.9389, "step": 6159 }, { "epoch": 2.789855072463768, "grad_norm": 0.07708196013685491, "learning_rate": 3.888718001914242e-05, "loss": 0.9347, "step": 6160 }, { "epoch": 2.7903079710144927, "grad_norm": 0.072099042565223, "learning_rate": 3.8874537656890466e-05, "loss": 0.9324, "step": 6161 }, { "epoch": 2.7907608695652173, "grad_norm": 0.07893229783622892, "learning_rate": 3.886189540715275e-05, "loss": 0.9223, "step": 6162 }, { "epoch": 2.791213768115942, "grad_norm": 0.09120615001625906, "learning_rate": 3.884925327119316e-05, "loss": 0.9354, "step": 6163 }, { "epoch": 2.7916666666666665, "grad_norm": 0.08428166163034262, "learning_rate": 3.8836611250275546e-05, "loss": 0.9061, "step": 6164 }, { "epoch": 2.7921195652173916, "grad_norm": 0.07694442559767702, "learning_rate": 3.882396934566375e-05, "loss": 0.9383, "step": 6165 }, { "epoch": 2.792572463768116, "grad_norm": 0.0913677954327703, "learning_rate": 3.8811327558621596e-05, "loss": 0.9273, "step": 6166 }, { "epoch": 2.793025362318841, "grad_norm": 0.0931257157997827, "learning_rate": 3.879868589041292e-05, "loss": 0.9496, "step": 6167 }, { "epoch": 2.7934782608695654, "grad_norm": 0.09156384931578687, "learning_rate": 3.878604434230151e-05, "loss": 0.9282, "step": 6168 }, { "epoch": 2.79393115942029, "grad_norm": 0.07396902327879995, "learning_rate": 3.877340291555118e-05, "loss": 0.9148, "step": 6169 }, { "epoch": 2.7943840579710146, "grad_norm": 0.09826714617635293, "learning_rate": 3.876076161142568e-05, "loss": 0.9358, "step": 6170 }, { "epoch": 2.7948369565217392, "grad_norm": 0.0995774257473436, "learning_rate": 3.8748120431188834e-05, "loss": 0.9296, "step": 6171 }, { "epoch": 2.795289855072464, "grad_norm": 0.09978818394624119, "learning_rate": 3.873547937610436e-05, "loss": 0.9417, "step": 6172 }, { "epoch": 2.7957427536231885, "grad_norm": 0.06716110855338588, "learning_rate": 3.8722838447436025e-05, "loss": 0.9067, "step": 6173 }, { "epoch": 2.796195652173913, "grad_norm": 0.09071260492511247, "learning_rate": 3.8710197646447535e-05, "loss": 0.9072, "step": 6174 }, { "epoch": 2.7966485507246377, "grad_norm": 0.07643388897301223, "learning_rate": 3.8697556974402653e-05, "loss": 0.9383, "step": 6175 }, { "epoch": 2.7971014492753623, "grad_norm": 0.08581103672684513, "learning_rate": 3.868491643256505e-05, "loss": 0.9465, "step": 6176 }, { "epoch": 2.797554347826087, "grad_norm": 0.07953657538569903, "learning_rate": 3.867227602219844e-05, "loss": 0.9327, "step": 6177 }, { "epoch": 2.7980072463768115, "grad_norm": 0.0765054868873541, "learning_rate": 3.86596357445665e-05, "loss": 0.929, "step": 6178 }, { "epoch": 2.798460144927536, "grad_norm": 0.10640013967137915, "learning_rate": 3.8646995600932895e-05, "loss": 0.9327, "step": 6179 }, { "epoch": 2.7989130434782608, "grad_norm": 0.08956376615915679, "learning_rate": 3.8634355592561286e-05, "loss": 0.9225, "step": 6180 }, { "epoch": 2.7993659420289854, "grad_norm": 0.08848158020242583, "learning_rate": 3.862171572071532e-05, "loss": 0.9399, "step": 6181 }, { "epoch": 2.79981884057971, "grad_norm": 0.07556482279570105, "learning_rate": 3.86090759866586e-05, "loss": 0.9224, "step": 6182 }, { "epoch": 2.8002717391304346, "grad_norm": 0.10255862439133387, "learning_rate": 3.8596436391654774e-05, "loss": 0.9275, "step": 6183 }, { "epoch": 2.800724637681159, "grad_norm": 0.08655773323091044, "learning_rate": 3.858379693696742e-05, "loss": 0.9525, "step": 6184 }, { "epoch": 2.801177536231884, "grad_norm": 0.08089681031753465, "learning_rate": 3.857115762386014e-05, "loss": 0.9179, "step": 6185 }, { "epoch": 2.8016304347826084, "grad_norm": 0.08404066884868479, "learning_rate": 3.8558518453596466e-05, "loss": 0.9201, "step": 6186 }, { "epoch": 2.8020833333333335, "grad_norm": 0.07749123733518895, "learning_rate": 3.854587942744002e-05, "loss": 0.9132, "step": 6187 }, { "epoch": 2.802536231884058, "grad_norm": 0.09494924557373169, "learning_rate": 3.85332405466543e-05, "loss": 0.9431, "step": 6188 }, { "epoch": 2.8029891304347827, "grad_norm": 0.07443953602998368, "learning_rate": 3.8520601812502844e-05, "loss": 0.9191, "step": 6189 }, { "epoch": 2.8034420289855073, "grad_norm": 0.09698926573510108, "learning_rate": 3.8507963226249165e-05, "loss": 0.9501, "step": 6190 }, { "epoch": 2.803894927536232, "grad_norm": 0.14682784066905444, "learning_rate": 3.849532478915678e-05, "loss": 0.931, "step": 6191 }, { "epoch": 2.8043478260869565, "grad_norm": 0.12982032084714407, "learning_rate": 3.848268650248915e-05, "loss": 0.9378, "step": 6192 }, { "epoch": 2.804800724637681, "grad_norm": 0.08409740305123582, "learning_rate": 3.847004836750976e-05, "loss": 0.9413, "step": 6193 }, { "epoch": 2.8052536231884058, "grad_norm": 0.15461095190551877, "learning_rate": 3.845741038548205e-05, "loss": 0.9225, "step": 6194 }, { "epoch": 2.8057065217391304, "grad_norm": 0.17777627367034574, "learning_rate": 3.844477255766948e-05, "loss": 0.9373, "step": 6195 }, { "epoch": 2.806159420289855, "grad_norm": 0.12216394418245335, "learning_rate": 3.843213488533545e-05, "loss": 0.9429, "step": 6196 }, { "epoch": 2.8066123188405796, "grad_norm": 0.0988897424217214, "learning_rate": 3.84194973697434e-05, "loss": 0.9607, "step": 6197 }, { "epoch": 2.807065217391304, "grad_norm": 0.15730705541723194, "learning_rate": 3.8406860012156695e-05, "loss": 0.919, "step": 6198 }, { "epoch": 2.807518115942029, "grad_norm": 0.12004100251258609, "learning_rate": 3.83942228138387e-05, "loss": 0.9192, "step": 6199 }, { "epoch": 2.807971014492754, "grad_norm": 0.07605371737292364, "learning_rate": 3.8381585776052824e-05, "loss": 0.9228, "step": 6200 }, { "epoch": 2.8084239130434785, "grad_norm": 0.11027044205243672, "learning_rate": 3.836894890006236e-05, "loss": 0.9439, "step": 6201 }, { "epoch": 2.808876811594203, "grad_norm": 0.13560843753110527, "learning_rate": 3.8356312187130676e-05, "loss": 0.9139, "step": 6202 }, { "epoch": 2.8093297101449277, "grad_norm": 0.09323200443190667, "learning_rate": 3.8343675638521044e-05, "loss": 0.9503, "step": 6203 }, { "epoch": 2.8097826086956523, "grad_norm": 0.08004148938393281, "learning_rate": 3.8331039255496805e-05, "loss": 0.9065, "step": 6204 }, { "epoch": 2.810235507246377, "grad_norm": 0.09441558440451364, "learning_rate": 3.83184030393212e-05, "loss": 0.9318, "step": 6205 }, { "epoch": 2.8106884057971016, "grad_norm": 0.10329016870872847, "learning_rate": 3.830576699125753e-05, "loss": 0.9256, "step": 6206 }, { "epoch": 2.811141304347826, "grad_norm": 0.10666365077453117, "learning_rate": 3.8293131112568984e-05, "loss": 0.9394, "step": 6207 }, { "epoch": 2.8115942028985508, "grad_norm": 0.08181267551200583, "learning_rate": 3.828049540451886e-05, "loss": 0.9623, "step": 6208 }, { "epoch": 2.8120471014492754, "grad_norm": 0.07815035014710527, "learning_rate": 3.8267859868370305e-05, "loss": 0.9358, "step": 6209 }, { "epoch": 2.8125, "grad_norm": 0.09646444690424534, "learning_rate": 3.825522450538657e-05, "loss": 0.9325, "step": 6210 }, { "epoch": 2.8129528985507246, "grad_norm": 0.085409778277259, "learning_rate": 3.824258931683079e-05, "loss": 0.9204, "step": 6211 }, { "epoch": 2.8134057971014492, "grad_norm": 0.08964733394011243, "learning_rate": 3.8229954303966154e-05, "loss": 0.938, "step": 6212 }, { "epoch": 2.813858695652174, "grad_norm": 0.07679689148559188, "learning_rate": 3.821731946805577e-05, "loss": 0.9342, "step": 6213 }, { "epoch": 2.8143115942028984, "grad_norm": 0.07850534842762681, "learning_rate": 3.820468481036281e-05, "loss": 0.9154, "step": 6214 }, { "epoch": 2.814764492753623, "grad_norm": 0.07459072498519545, "learning_rate": 3.8192050332150336e-05, "loss": 0.9329, "step": 6215 }, { "epoch": 2.8152173913043477, "grad_norm": 0.08149617241205297, "learning_rate": 3.8179416034681465e-05, "loss": 0.9446, "step": 6216 }, { "epoch": 2.8156702898550723, "grad_norm": 0.09378028120202447, "learning_rate": 3.8166781919219246e-05, "loss": 0.9457, "step": 6217 }, { "epoch": 2.816123188405797, "grad_norm": 0.09125977887059315, "learning_rate": 3.815414798702676e-05, "loss": 0.9433, "step": 6218 }, { "epoch": 2.8165760869565215, "grad_norm": 0.08103154131925029, "learning_rate": 3.814151423936702e-05, "loss": 0.9289, "step": 6219 }, { "epoch": 2.817028985507246, "grad_norm": 0.08890627210428007, "learning_rate": 3.812888067750304e-05, "loss": 0.9367, "step": 6220 }, { "epoch": 2.817481884057971, "grad_norm": 0.09322796554250765, "learning_rate": 3.8116247302697816e-05, "loss": 0.9113, "step": 6221 }, { "epoch": 2.817934782608696, "grad_norm": 0.07553625415179797, "learning_rate": 3.8103614116214354e-05, "loss": 0.9345, "step": 6222 }, { "epoch": 2.8183876811594204, "grad_norm": 0.07252071018794055, "learning_rate": 3.809098111931558e-05, "loss": 0.9152, "step": 6223 }, { "epoch": 2.818840579710145, "grad_norm": 0.09425556393981198, "learning_rate": 3.807834831326446e-05, "loss": 0.9161, "step": 6224 }, { "epoch": 2.8192934782608696, "grad_norm": 0.08884937507141277, "learning_rate": 3.8065715699323884e-05, "loss": 0.9362, "step": 6225 }, { "epoch": 2.8197463768115942, "grad_norm": 0.08994657212101319, "learning_rate": 3.80530832787568e-05, "loss": 0.9394, "step": 6226 }, { "epoch": 2.820199275362319, "grad_norm": 0.09482191779410815, "learning_rate": 3.804045105282605e-05, "loss": 0.9549, "step": 6227 }, { "epoch": 2.8206521739130435, "grad_norm": 0.07349648300100144, "learning_rate": 3.8027819022794525e-05, "loss": 0.9236, "step": 6228 }, { "epoch": 2.821105072463768, "grad_norm": 0.0930818401650112, "learning_rate": 3.801518718992504e-05, "loss": 0.9346, "step": 6229 }, { "epoch": 2.8215579710144927, "grad_norm": 0.07597640430310361, "learning_rate": 3.800255555548045e-05, "loss": 0.9352, "step": 6230 }, { "epoch": 2.8220108695652173, "grad_norm": 0.08976940436498883, "learning_rate": 3.798992412072354e-05, "loss": 0.9456, "step": 6231 }, { "epoch": 2.822463768115942, "grad_norm": 0.07810144768106703, "learning_rate": 3.797729288691711e-05, "loss": 0.9191, "step": 6232 }, { "epoch": 2.8229166666666665, "grad_norm": 0.08142788247999996, "learning_rate": 3.79646618553239e-05, "loss": 0.9434, "step": 6233 }, { "epoch": 2.8233695652173916, "grad_norm": 0.07719762269671515, "learning_rate": 3.795203102720669e-05, "loss": 0.9305, "step": 6234 }, { "epoch": 2.823822463768116, "grad_norm": 0.07695879197307974, "learning_rate": 3.7939400403828167e-05, "loss": 0.9337, "step": 6235 }, { "epoch": 2.824275362318841, "grad_norm": 0.09022881723271119, "learning_rate": 3.792676998645106e-05, "loss": 0.951, "step": 6236 }, { "epoch": 2.8247282608695654, "grad_norm": 0.07558169366403884, "learning_rate": 3.791413977633803e-05, "loss": 0.9123, "step": 6237 }, { "epoch": 2.82518115942029, "grad_norm": 0.08973034767691228, "learning_rate": 3.790150977475177e-05, "loss": 0.9314, "step": 6238 }, { "epoch": 2.8256340579710146, "grad_norm": 0.08050521800228241, "learning_rate": 3.788887998295489e-05, "loss": 0.9106, "step": 6239 }, { "epoch": 2.8260869565217392, "grad_norm": 0.10401272041333966, "learning_rate": 3.7876250402210036e-05, "loss": 0.9199, "step": 6240 }, { "epoch": 2.826539855072464, "grad_norm": 0.12202767429328422, "learning_rate": 3.7863621033779774e-05, "loss": 0.9341, "step": 6241 }, { "epoch": 2.8269927536231885, "grad_norm": 0.10473600507818642, "learning_rate": 3.785099187892672e-05, "loss": 0.9231, "step": 6242 }, { "epoch": 2.827445652173913, "grad_norm": 0.08797178613477263, "learning_rate": 3.7838362938913406e-05, "loss": 0.9228, "step": 6243 }, { "epoch": 2.8278985507246377, "grad_norm": 0.07490659231335955, "learning_rate": 3.7825734215002386e-05, "loss": 0.9342, "step": 6244 }, { "epoch": 2.8283514492753623, "grad_norm": 0.09088335724288701, "learning_rate": 3.7813105708456134e-05, "loss": 0.9305, "step": 6245 }, { "epoch": 2.828804347826087, "grad_norm": 0.09515967258500051, "learning_rate": 3.78004774205372e-05, "loss": 0.9208, "step": 6246 }, { "epoch": 2.8292572463768115, "grad_norm": 0.07751894643707868, "learning_rate": 3.778784935250801e-05, "loss": 0.9315, "step": 6247 }, { "epoch": 2.829710144927536, "grad_norm": 0.08397193965052449, "learning_rate": 3.7775221505631034e-05, "loss": 0.9378, "step": 6248 }, { "epoch": 2.8301630434782608, "grad_norm": 0.09404685995083775, "learning_rate": 3.776259388116867e-05, "loss": 0.9234, "step": 6249 }, { "epoch": 2.8306159420289854, "grad_norm": 0.09597758341567714, "learning_rate": 3.774996648038336e-05, "loss": 0.9169, "step": 6250 }, { "epoch": 2.83106884057971, "grad_norm": 0.07476304956800153, "learning_rate": 3.773733930453745e-05, "loss": 0.9016, "step": 6251 }, { "epoch": 2.8315217391304346, "grad_norm": 0.07433094319827455, "learning_rate": 3.7724712354893335e-05, "loss": 0.934, "step": 6252 }, { "epoch": 2.831974637681159, "grad_norm": 0.09459680822885025, "learning_rate": 3.771208563271332e-05, "loss": 0.957, "step": 6253 }, { "epoch": 2.832427536231884, "grad_norm": 0.10011726252107622, "learning_rate": 3.769945913925974e-05, "loss": 0.9217, "step": 6254 }, { "epoch": 2.8328804347826084, "grad_norm": 0.08086233160882031, "learning_rate": 3.768683287579486e-05, "loss": 0.9045, "step": 6255 }, { "epoch": 2.8333333333333335, "grad_norm": 0.08930842801440418, "learning_rate": 3.767420684358097e-05, "loss": 0.9324, "step": 6256 }, { "epoch": 2.833786231884058, "grad_norm": 0.07490128167846412, "learning_rate": 3.766158104388031e-05, "loss": 0.9255, "step": 6257 }, { "epoch": 2.8342391304347827, "grad_norm": 0.08426587976646667, "learning_rate": 3.76489554779551e-05, "loss": 0.9339, "step": 6258 }, { "epoch": 2.8346920289855073, "grad_norm": 0.083656363891533, "learning_rate": 3.763633014706752e-05, "loss": 0.95, "step": 6259 }, { "epoch": 2.835144927536232, "grad_norm": 0.10410744732106297, "learning_rate": 3.762370505247978e-05, "loss": 0.9366, "step": 6260 }, { "epoch": 2.8355978260869565, "grad_norm": 0.09482694939953586, "learning_rate": 3.7611080195454006e-05, "loss": 0.9433, "step": 6261 }, { "epoch": 2.836050724637681, "grad_norm": 0.11550344690246293, "learning_rate": 3.7598455577252336e-05, "loss": 0.9398, "step": 6262 }, { "epoch": 2.8365036231884058, "grad_norm": 0.0816798351035658, "learning_rate": 3.758583119913685e-05, "loss": 0.9432, "step": 6263 }, { "epoch": 2.8369565217391304, "grad_norm": 0.09893285790323986, "learning_rate": 3.757320706236966e-05, "loss": 0.9482, "step": 6264 }, { "epoch": 2.837409420289855, "grad_norm": 0.08628571007911673, "learning_rate": 3.75605831682128e-05, "loss": 0.913, "step": 6265 }, { "epoch": 2.8378623188405796, "grad_norm": 0.09533868494729582, "learning_rate": 3.754795951792832e-05, "loss": 0.9528, "step": 6266 }, { "epoch": 2.838315217391304, "grad_norm": 0.07847975557484704, "learning_rate": 3.7535336112778186e-05, "loss": 0.9134, "step": 6267 }, { "epoch": 2.838768115942029, "grad_norm": 0.08437639770627858, "learning_rate": 3.752271295402443e-05, "loss": 0.9672, "step": 6268 }, { "epoch": 2.839221014492754, "grad_norm": 0.08765994568998053, "learning_rate": 3.751009004292897e-05, "loss": 0.9482, "step": 6269 }, { "epoch": 2.8396739130434785, "grad_norm": 0.08615030639965496, "learning_rate": 3.7497467380753766e-05, "loss": 0.9347, "step": 6270 }, { "epoch": 2.840126811594203, "grad_norm": 0.09796628408551653, "learning_rate": 3.7484844968760693e-05, "loss": 0.9365, "step": 6271 }, { "epoch": 2.8405797101449277, "grad_norm": 0.09078889732296906, "learning_rate": 3.7472222808211676e-05, "loss": 0.9479, "step": 6272 }, { "epoch": 2.8410326086956523, "grad_norm": 0.08090528549551652, "learning_rate": 3.745960090036854e-05, "loss": 0.9412, "step": 6273 }, { "epoch": 2.841485507246377, "grad_norm": 0.09503988290951158, "learning_rate": 3.7446979246493145e-05, "loss": 0.9679, "step": 6274 }, { "epoch": 2.8419384057971016, "grad_norm": 0.07838446838406435, "learning_rate": 3.743435784784725e-05, "loss": 0.9372, "step": 6275 }, { "epoch": 2.842391304347826, "grad_norm": 0.07849347032090816, "learning_rate": 3.742173670569269e-05, "loss": 0.9435, "step": 6276 }, { "epoch": 2.8428442028985508, "grad_norm": 0.0734805589238776, "learning_rate": 3.7409115821291184e-05, "loss": 0.9285, "step": 6277 }, { "epoch": 2.8432971014492754, "grad_norm": 0.08189976121039592, "learning_rate": 3.739649519590449e-05, "loss": 0.9583, "step": 6278 }, { "epoch": 2.84375, "grad_norm": 0.08968590390945994, "learning_rate": 3.738387483079428e-05, "loss": 0.9422, "step": 6279 }, { "epoch": 2.8442028985507246, "grad_norm": 0.08209116086068383, "learning_rate": 3.737125472722226e-05, "loss": 0.9415, "step": 6280 }, { "epoch": 2.8446557971014492, "grad_norm": 1.0461682942972181, "learning_rate": 3.7358634886450064e-05, "loss": 0.9335, "step": 6281 }, { "epoch": 2.845108695652174, "grad_norm": 0.08502948885969908, "learning_rate": 3.7346015309739334e-05, "loss": 0.9261, "step": 6282 }, { "epoch": 2.8455615942028984, "grad_norm": 0.10266481677797464, "learning_rate": 3.733339599835164e-05, "loss": 0.9491, "step": 6283 }, { "epoch": 2.846014492753623, "grad_norm": 0.10468863060633413, "learning_rate": 3.732077695354859e-05, "loss": 0.9494, "step": 6284 }, { "epoch": 2.8464673913043477, "grad_norm": 0.12196922046895521, "learning_rate": 3.7308158176591714e-05, "loss": 0.9461, "step": 6285 }, { "epoch": 2.8469202898550723, "grad_norm": 0.09837834127542942, "learning_rate": 3.729553966874253e-05, "loss": 0.9409, "step": 6286 }, { "epoch": 2.847373188405797, "grad_norm": 0.09153626542597182, "learning_rate": 3.7282921431262515e-05, "loss": 0.9449, "step": 6287 }, { "epoch": 2.8478260869565215, "grad_norm": 0.09009461982513417, "learning_rate": 3.727030346541317e-05, "loss": 0.9448, "step": 6288 }, { "epoch": 2.848278985507246, "grad_norm": 0.10623417920713431, "learning_rate": 3.72576857724559e-05, "loss": 0.9408, "step": 6289 }, { "epoch": 2.848731884057971, "grad_norm": 0.24283100820889084, "learning_rate": 3.724506835365214e-05, "loss": 0.9683, "step": 6290 }, { "epoch": 2.849184782608696, "grad_norm": 0.07686310675331104, "learning_rate": 3.723245121026325e-05, "loss": 0.9531, "step": 6291 }, { "epoch": 2.8496376811594204, "grad_norm": 0.11228553714955121, "learning_rate": 3.7219834343550606e-05, "loss": 0.9396, "step": 6292 }, { "epoch": 2.850090579710145, "grad_norm": 0.11506689677065841, "learning_rate": 3.720721775477551e-05, "loss": 0.9411, "step": 6293 }, { "epoch": 2.8505434782608696, "grad_norm": 0.07490106248018484, "learning_rate": 3.719460144519932e-05, "loss": 0.931, "step": 6294 }, { "epoch": 2.8509963768115942, "grad_norm": 0.09229658465206679, "learning_rate": 3.7181985416083234e-05, "loss": 0.9452, "step": 6295 }, { "epoch": 2.851449275362319, "grad_norm": 0.10734820441608184, "learning_rate": 3.716936966868855e-05, "loss": 0.9286, "step": 6296 }, { "epoch": 2.8519021739130435, "grad_norm": 0.09711475807898211, "learning_rate": 3.7156754204276446e-05, "loss": 0.9336, "step": 6297 }, { "epoch": 2.852355072463768, "grad_norm": 0.09044125022349075, "learning_rate": 3.714413902410815e-05, "loss": 0.9233, "step": 6298 }, { "epoch": 2.8528079710144927, "grad_norm": 0.08682985260695128, "learning_rate": 3.7131524129444796e-05, "loss": 0.9474, "step": 6299 }, { "epoch": 2.8532608695652173, "grad_norm": 0.09292056516631661, "learning_rate": 3.711890952154752e-05, "loss": 0.9469, "step": 6300 }, { "epoch": 2.853713768115942, "grad_norm": 0.08139780465011166, "learning_rate": 3.7106295201677414e-05, "loss": 0.9298, "step": 6301 }, { "epoch": 2.8541666666666665, "grad_norm": 0.10685332262774268, "learning_rate": 3.709368117109558e-05, "loss": 0.9249, "step": 6302 }, { "epoch": 2.8546195652173916, "grad_norm": 0.1275066589137705, "learning_rate": 3.708106743106303e-05, "loss": 0.953, "step": 6303 }, { "epoch": 2.855072463768116, "grad_norm": 0.08304687190462307, "learning_rate": 3.7068453982840813e-05, "loss": 0.9324, "step": 6304 }, { "epoch": 2.855525362318841, "grad_norm": 0.09117923831910987, "learning_rate": 3.7055840827689874e-05, "loss": 0.9251, "step": 6305 }, { "epoch": 2.8559782608695654, "grad_norm": 0.08638195606738472, "learning_rate": 3.704322796687121e-05, "loss": 0.9509, "step": 6306 }, { "epoch": 2.85643115942029, "grad_norm": 0.08489931144077004, "learning_rate": 3.703061540164572e-05, "loss": 0.9277, "step": 6307 }, { "epoch": 2.8568840579710146, "grad_norm": 0.10368116912726498, "learning_rate": 3.7018003133274324e-05, "loss": 0.9128, "step": 6308 }, { "epoch": 2.8573369565217392, "grad_norm": 0.07125349282163318, "learning_rate": 3.7005391163017866e-05, "loss": 0.9308, "step": 6309 }, { "epoch": 2.857789855072464, "grad_norm": 0.09672556904815725, "learning_rate": 3.6992779492137206e-05, "loss": 0.9264, "step": 6310 }, { "epoch": 2.8582427536231885, "grad_norm": 0.07861113944580876, "learning_rate": 3.6980168121893145e-05, "loss": 0.9392, "step": 6311 }, { "epoch": 2.858695652173913, "grad_norm": 0.07681506347656554, "learning_rate": 3.696755705354647e-05, "loss": 0.9338, "step": 6312 }, { "epoch": 2.8591485507246377, "grad_norm": 0.08289933655657737, "learning_rate": 3.6954946288357906e-05, "loss": 0.954, "step": 6313 }, { "epoch": 2.8596014492753623, "grad_norm": 0.07263175449800127, "learning_rate": 3.6942335827588204e-05, "loss": 0.9309, "step": 6314 }, { "epoch": 2.860054347826087, "grad_norm": 0.08188592939866894, "learning_rate": 3.692972567249803e-05, "loss": 0.9361, "step": 6315 }, { "epoch": 2.8605072463768115, "grad_norm": 0.08783420918025103, "learning_rate": 3.691711582434804e-05, "loss": 0.9472, "step": 6316 }, { "epoch": 2.860960144927536, "grad_norm": 0.07493218551319693, "learning_rate": 3.690450628439886e-05, "loss": 0.933, "step": 6317 }, { "epoch": 2.8614130434782608, "grad_norm": 0.09144075486373185, "learning_rate": 3.689189705391111e-05, "loss": 0.9401, "step": 6318 }, { "epoch": 2.8618659420289854, "grad_norm": 0.08595442125115192, "learning_rate": 3.687928813414533e-05, "loss": 0.9396, "step": 6319 }, { "epoch": 2.86231884057971, "grad_norm": 0.07752878748475688, "learning_rate": 3.686667952636206e-05, "loss": 0.9078, "step": 6320 }, { "epoch": 2.8627717391304346, "grad_norm": 0.07082326319491877, "learning_rate": 3.685407123182181e-05, "loss": 0.9363, "step": 6321 }, { "epoch": 2.863224637681159, "grad_norm": 0.06307735329658885, "learning_rate": 3.6841463251785024e-05, "loss": 0.9531, "step": 6322 }, { "epoch": 2.863677536231884, "grad_norm": 0.07100530198219876, "learning_rate": 3.682885558751218e-05, "loss": 0.9208, "step": 6323 }, { "epoch": 2.8641304347826084, "grad_norm": 0.06379646726480835, "learning_rate": 3.681624824026366e-05, "loss": 0.9183, "step": 6324 }, { "epoch": 2.8645833333333335, "grad_norm": 0.06643951681285756, "learning_rate": 3.6803641211299856e-05, "loss": 0.9156, "step": 6325 }, { "epoch": 2.865036231884058, "grad_norm": 0.0720226144787103, "learning_rate": 3.679103450188109e-05, "loss": 0.9358, "step": 6326 }, { "epoch": 2.8654891304347827, "grad_norm": 0.0947979914713689, "learning_rate": 3.67784281132677e-05, "loss": 0.9356, "step": 6327 }, { "epoch": 2.8659420289855073, "grad_norm": 0.07245115421393863, "learning_rate": 3.676582204671995e-05, "loss": 0.9294, "step": 6328 }, { "epoch": 2.866394927536232, "grad_norm": 0.08680103531498373, "learning_rate": 3.675321630349811e-05, "loss": 0.9233, "step": 6329 }, { "epoch": 2.8668478260869565, "grad_norm": 0.0682628323481197, "learning_rate": 3.6740610884862366e-05, "loss": 0.9489, "step": 6330 }, { "epoch": 2.867300724637681, "grad_norm": 0.08836764494093048, "learning_rate": 3.672800579207293e-05, "loss": 0.9089, "step": 6331 }, { "epoch": 2.8677536231884058, "grad_norm": 0.08056238046037985, "learning_rate": 3.671540102638993e-05, "loss": 0.9274, "step": 6332 }, { "epoch": 2.8682065217391304, "grad_norm": 0.08091717836429572, "learning_rate": 3.6702796589073494e-05, "loss": 0.9317, "step": 6333 }, { "epoch": 2.868659420289855, "grad_norm": 0.09455580646897849, "learning_rate": 3.66901924813837e-05, "loss": 0.932, "step": 6334 }, { "epoch": 2.8691123188405796, "grad_norm": 0.07873570418246524, "learning_rate": 3.667758870458063e-05, "loss": 0.9491, "step": 6335 }, { "epoch": 2.869565217391304, "grad_norm": 0.0754011647012761, "learning_rate": 3.6664985259924265e-05, "loss": 0.972, "step": 6336 }, { "epoch": 2.870018115942029, "grad_norm": 0.08487015031852126, "learning_rate": 3.6652382148674624e-05, "loss": 0.9394, "step": 6337 }, { "epoch": 2.870471014492754, "grad_norm": 0.07434841795082056, "learning_rate": 3.663977937209162e-05, "loss": 0.9415, "step": 6338 }, { "epoch": 2.8709239130434785, "grad_norm": 0.08552914462031717, "learning_rate": 3.662717693143522e-05, "loss": 0.956, "step": 6339 }, { "epoch": 2.871376811594203, "grad_norm": 0.08870736684425329, "learning_rate": 3.661457482796527e-05, "loss": 0.9525, "step": 6340 }, { "epoch": 2.8718297101449277, "grad_norm": 0.07240762912033676, "learning_rate": 3.660197306294165e-05, "loss": 0.9359, "step": 6341 }, { "epoch": 2.8722826086956523, "grad_norm": 0.09287716410108064, "learning_rate": 3.658937163762416e-05, "loss": 0.9134, "step": 6342 }, { "epoch": 2.872735507246377, "grad_norm": 0.07813107167460503, "learning_rate": 3.657677055327261e-05, "loss": 0.9356, "step": 6343 }, { "epoch": 2.8731884057971016, "grad_norm": 0.0762598883653968, "learning_rate": 3.656416981114671e-05, "loss": 0.9506, "step": 6344 }, { "epoch": 2.873641304347826, "grad_norm": 0.08495603871671974, "learning_rate": 3.655156941250621e-05, "loss": 0.9251, "step": 6345 }, { "epoch": 2.8740942028985508, "grad_norm": 0.08881014046065278, "learning_rate": 3.653896935861078e-05, "loss": 0.9352, "step": 6346 }, { "epoch": 2.8745471014492754, "grad_norm": 0.08876558081950309, "learning_rate": 3.652636965072007e-05, "loss": 0.9112, "step": 6347 }, { "epoch": 2.875, "grad_norm": 0.08847009759195422, "learning_rate": 3.6513770290093674e-05, "loss": 0.9262, "step": 6348 }, { "epoch": 2.8754528985507246, "grad_norm": 0.0750343555739269, "learning_rate": 3.65011712779912e-05, "loss": 0.9347, "step": 6349 }, { "epoch": 2.8759057971014492, "grad_norm": 0.10156747928800722, "learning_rate": 3.648857261567218e-05, "loss": 0.9446, "step": 6350 }, { "epoch": 2.876358695652174, "grad_norm": 0.08903891346795438, "learning_rate": 3.6475974304396126e-05, "loss": 0.9547, "step": 6351 }, { "epoch": 2.8768115942028984, "grad_norm": 0.07647749417249813, "learning_rate": 3.646337634542248e-05, "loss": 0.9532, "step": 6352 }, { "epoch": 2.877264492753623, "grad_norm": 0.08138420495784138, "learning_rate": 3.645077874001073e-05, "loss": 0.9417, "step": 6353 }, { "epoch": 2.8777173913043477, "grad_norm": 0.06658327275164368, "learning_rate": 3.643818148942024e-05, "loss": 0.9289, "step": 6354 }, { "epoch": 2.8781702898550723, "grad_norm": 0.09395950804209685, "learning_rate": 3.64255845949104e-05, "loss": 0.9225, "step": 6355 }, { "epoch": 2.878623188405797, "grad_norm": 0.08374220081684047, "learning_rate": 3.6412988057740514e-05, "loss": 0.951, "step": 6356 }, { "epoch": 2.8790760869565215, "grad_norm": 0.06673767796378484, "learning_rate": 3.640039187916992e-05, "loss": 0.941, "step": 6357 }, { "epoch": 2.879528985507246, "grad_norm": 0.08344517441845342, "learning_rate": 3.6387796060457844e-05, "loss": 0.949, "step": 6358 }, { "epoch": 2.879981884057971, "grad_norm": 0.15809431361628062, "learning_rate": 3.637520060286352e-05, "loss": 0.9225, "step": 6359 }, { "epoch": 2.880434782608696, "grad_norm": 0.07813644095289096, "learning_rate": 3.636260550764613e-05, "loss": 0.9229, "step": 6360 }, { "epoch": 2.8808876811594204, "grad_norm": 0.08691858689767823, "learning_rate": 3.635001077606484e-05, "loss": 0.9173, "step": 6361 }, { "epoch": 2.881340579710145, "grad_norm": 0.06922696528085678, "learning_rate": 3.633741640937875e-05, "loss": 0.938, "step": 6362 }, { "epoch": 2.8817934782608696, "grad_norm": 0.08014422916169195, "learning_rate": 3.632482240884695e-05, "loss": 0.9406, "step": 6363 }, { "epoch": 2.8822463768115942, "grad_norm": 0.1024290801908595, "learning_rate": 3.6312228775728465e-05, "loss": 0.928, "step": 6364 }, { "epoch": 2.882699275362319, "grad_norm": 0.08268290588150505, "learning_rate": 3.629963551128233e-05, "loss": 0.918, "step": 6365 }, { "epoch": 2.8831521739130435, "grad_norm": 0.06933923872154246, "learning_rate": 3.628704261676749e-05, "loss": 0.928, "step": 6366 }, { "epoch": 2.883605072463768, "grad_norm": 0.1020411877709442, "learning_rate": 3.627445009344289e-05, "loss": 0.9291, "step": 6367 }, { "epoch": 2.8840579710144927, "grad_norm": 0.11576401988266151, "learning_rate": 3.6261857942567396e-05, "loss": 0.9346, "step": 6368 }, { "epoch": 2.8845108695652173, "grad_norm": 0.08809490691104213, "learning_rate": 3.624926616539991e-05, "loss": 0.9141, "step": 6369 }, { "epoch": 2.884963768115942, "grad_norm": 0.07658613275424153, "learning_rate": 3.623667476319922e-05, "loss": 0.9238, "step": 6370 }, { "epoch": 2.8854166666666665, "grad_norm": 0.07099258079113674, "learning_rate": 3.622408373722412e-05, "loss": 0.9482, "step": 6371 }, { "epoch": 2.8858695652173916, "grad_norm": 0.09334769992582986, "learning_rate": 3.621149308873334e-05, "loss": 0.9464, "step": 6372 }, { "epoch": 2.886322463768116, "grad_norm": 0.07794798897334874, "learning_rate": 3.619890281898562e-05, "loss": 0.9347, "step": 6373 }, { "epoch": 2.886775362318841, "grad_norm": 0.08190584911158211, "learning_rate": 3.61863129292396e-05, "loss": 0.9292, "step": 6374 }, { "epoch": 2.8872282608695654, "grad_norm": 0.09610549242563055, "learning_rate": 3.6173723420753925e-05, "loss": 0.9346, "step": 6375 }, { "epoch": 2.88768115942029, "grad_norm": 0.08923430567228861, "learning_rate": 3.6161134294787166e-05, "loss": 0.9503, "step": 6376 }, { "epoch": 2.8881340579710146, "grad_norm": 0.07444881380718553, "learning_rate": 3.614854555259791e-05, "loss": 0.9338, "step": 6377 }, { "epoch": 2.8885869565217392, "grad_norm": 0.08828801580701505, "learning_rate": 3.613595719544465e-05, "loss": 0.9351, "step": 6378 }, { "epoch": 2.889039855072464, "grad_norm": 0.10106984317171673, "learning_rate": 3.612336922458588e-05, "loss": 0.9342, "step": 6379 }, { "epoch": 2.8894927536231885, "grad_norm": 0.09691753314317375, "learning_rate": 3.611078164128002e-05, "loss": 0.9376, "step": 6380 }, { "epoch": 2.889945652173913, "grad_norm": 0.08747086328565427, "learning_rate": 3.6098194446785495e-05, "loss": 0.9258, "step": 6381 }, { "epoch": 2.8903985507246377, "grad_norm": 0.07457733949684385, "learning_rate": 3.608560764236064e-05, "loss": 0.9451, "step": 6382 }, { "epoch": 2.8908514492753623, "grad_norm": 0.08131509834130518, "learning_rate": 3.607302122926381e-05, "loss": 0.929, "step": 6383 }, { "epoch": 2.891304347826087, "grad_norm": 0.08387613209419281, "learning_rate": 3.6060435208753256e-05, "loss": 0.9386, "step": 6384 }, { "epoch": 2.8917572463768115, "grad_norm": 0.08455573929088758, "learning_rate": 3.6047849582087246e-05, "loss": 0.9274, "step": 6385 }, { "epoch": 2.892210144927536, "grad_norm": 0.07477428489819572, "learning_rate": 3.6035264350523964e-05, "loss": 0.9625, "step": 6386 }, { "epoch": 2.8926630434782608, "grad_norm": 0.0768420983804245, "learning_rate": 3.6022679515321597e-05, "loss": 0.9172, "step": 6387 }, { "epoch": 2.8931159420289854, "grad_norm": 0.08646466033720855, "learning_rate": 3.601009507773826e-05, "loss": 0.9299, "step": 6388 }, { "epoch": 2.89356884057971, "grad_norm": 0.07357345911104217, "learning_rate": 3.599751103903205e-05, "loss": 0.9251, "step": 6389 }, { "epoch": 2.8940217391304346, "grad_norm": 0.0665983046191124, "learning_rate": 3.5984927400460984e-05, "loss": 0.917, "step": 6390 }, { "epoch": 2.894474637681159, "grad_norm": 0.08727117545212233, "learning_rate": 3.59723441632831e-05, "loss": 0.9267, "step": 6391 }, { "epoch": 2.894927536231884, "grad_norm": 0.09380545429743557, "learning_rate": 3.595976132875636e-05, "loss": 0.9663, "step": 6392 }, { "epoch": 2.8953804347826084, "grad_norm": 0.09118693291008985, "learning_rate": 3.5947178898138674e-05, "loss": 0.9239, "step": 6393 }, { "epoch": 2.8958333333333335, "grad_norm": 0.06607475686912317, "learning_rate": 3.5934596872687924e-05, "loss": 0.9272, "step": 6394 }, { "epoch": 2.896286231884058, "grad_norm": 0.07331515984515206, "learning_rate": 3.592201525366199e-05, "loss": 0.9215, "step": 6395 }, { "epoch": 2.8967391304347827, "grad_norm": 0.06655280070545096, "learning_rate": 3.590943404231863e-05, "loss": 0.9439, "step": 6396 }, { "epoch": 2.8971920289855073, "grad_norm": 0.06851628321974516, "learning_rate": 3.5896853239915655e-05, "loss": 0.9284, "step": 6397 }, { "epoch": 2.897644927536232, "grad_norm": 0.07416602711618012, "learning_rate": 3.588427284771074e-05, "loss": 0.9202, "step": 6398 }, { "epoch": 2.8980978260869565, "grad_norm": 0.07201807130960688, "learning_rate": 3.587169286696161e-05, "loss": 0.9242, "step": 6399 }, { "epoch": 2.898550724637681, "grad_norm": 0.08502166224682331, "learning_rate": 3.585911329892587e-05, "loss": 0.9167, "step": 6400 }, { "epoch": 2.8990036231884058, "grad_norm": 0.07974764824341064, "learning_rate": 3.584653414486115e-05, "loss": 0.9153, "step": 6401 }, { "epoch": 2.8994565217391304, "grad_norm": 0.07222881019072817, "learning_rate": 3.583395540602497e-05, "loss": 0.9362, "step": 6402 }, { "epoch": 2.899909420289855, "grad_norm": 0.07354245096841289, "learning_rate": 3.5821377083674894e-05, "loss": 0.9356, "step": 6403 }, { "epoch": 2.9003623188405796, "grad_norm": 0.0781762512842451, "learning_rate": 3.580879917906837e-05, "loss": 0.9352, "step": 6404 }, { "epoch": 2.900815217391304, "grad_norm": 0.07442647636542785, "learning_rate": 3.579622169346283e-05, "loss": 0.9158, "step": 6405 }, { "epoch": 2.901268115942029, "grad_norm": 0.08448940600092836, "learning_rate": 3.578364462811565e-05, "loss": 0.9402, "step": 6406 }, { "epoch": 2.901721014492754, "grad_norm": 0.07579698853814984, "learning_rate": 3.5771067984284225e-05, "loss": 0.9365, "step": 6407 }, { "epoch": 2.9021739130434785, "grad_norm": 0.07217794708058363, "learning_rate": 3.575849176322582e-05, "loss": 0.9337, "step": 6408 }, { "epoch": 2.902626811594203, "grad_norm": 0.07950784041242726, "learning_rate": 3.5745915966197724e-05, "loss": 0.9346, "step": 6409 }, { "epoch": 2.9030797101449277, "grad_norm": 0.07771636691550571, "learning_rate": 3.5733340594457136e-05, "loss": 0.9294, "step": 6410 }, { "epoch": 2.9035326086956523, "grad_norm": 0.06689208970761104, "learning_rate": 3.572076564926127e-05, "loss": 0.9206, "step": 6411 }, { "epoch": 2.903985507246377, "grad_norm": 0.07099889490444347, "learning_rate": 3.570819113186723e-05, "loss": 0.923, "step": 6412 }, { "epoch": 2.9044384057971016, "grad_norm": 0.07130629629589447, "learning_rate": 3.5695617043532124e-05, "loss": 0.9341, "step": 6413 }, { "epoch": 2.904891304347826, "grad_norm": 0.06489300648777917, "learning_rate": 3.568304338551299e-05, "loss": 0.9141, "step": 6414 }, { "epoch": 2.9053442028985508, "grad_norm": 0.07869751598368584, "learning_rate": 3.567047015906687e-05, "loss": 0.9229, "step": 6415 }, { "epoch": 2.9057971014492754, "grad_norm": 0.07163600472666352, "learning_rate": 3.5657897365450706e-05, "loss": 0.939, "step": 6416 }, { "epoch": 2.90625, "grad_norm": 0.08133673919470663, "learning_rate": 3.564532500592143e-05, "loss": 0.9331, "step": 6417 }, { "epoch": 2.9067028985507246, "grad_norm": 0.07455006302147607, "learning_rate": 3.5632753081735895e-05, "loss": 0.9325, "step": 6418 }, { "epoch": 2.9071557971014492, "grad_norm": 0.08087095777918551, "learning_rate": 3.5620181594150966e-05, "loss": 0.9366, "step": 6419 }, { "epoch": 2.907608695652174, "grad_norm": 0.08769298543705165, "learning_rate": 3.560761054442342e-05, "loss": 0.9498, "step": 6420 }, { "epoch": 2.9080615942028984, "grad_norm": 0.07227138567285525, "learning_rate": 3.559503993381001e-05, "loss": 0.9317, "step": 6421 }, { "epoch": 2.908514492753623, "grad_norm": 0.06923342703819024, "learning_rate": 3.558246976356743e-05, "loss": 0.9515, "step": 6422 }, { "epoch": 2.9089673913043477, "grad_norm": 0.07021012438378818, "learning_rate": 3.556990003495236e-05, "loss": 0.934, "step": 6423 }, { "epoch": 2.9094202898550723, "grad_norm": 0.06673786361335166, "learning_rate": 3.55573307492214e-05, "loss": 0.9171, "step": 6424 }, { "epoch": 2.909873188405797, "grad_norm": 0.06834820623617213, "learning_rate": 3.5544761907631127e-05, "loss": 0.9445, "step": 6425 }, { "epoch": 2.9103260869565215, "grad_norm": 0.0732729124579207, "learning_rate": 3.553219351143805e-05, "loss": 0.9201, "step": 6426 }, { "epoch": 2.910778985507246, "grad_norm": 0.06910481078525195, "learning_rate": 3.551962556189869e-05, "loss": 0.9014, "step": 6427 }, { "epoch": 2.911231884057971, "grad_norm": 0.07965214271173793, "learning_rate": 3.5507058060269446e-05, "loss": 0.9152, "step": 6428 }, { "epoch": 2.911684782608696, "grad_norm": 0.07249516843099477, "learning_rate": 3.5494491007806736e-05, "loss": 0.9076, "step": 6429 }, { "epoch": 2.9121376811594204, "grad_norm": 0.07616120020411901, "learning_rate": 3.5481924405766894e-05, "loss": 0.9317, "step": 6430 }, { "epoch": 2.912590579710145, "grad_norm": 0.07687873336187592, "learning_rate": 3.546935825540624e-05, "loss": 0.9424, "step": 6431 }, { "epoch": 2.9130434782608696, "grad_norm": 0.09040567467319761, "learning_rate": 3.5456792557981e-05, "loss": 0.9458, "step": 6432 }, { "epoch": 2.9134963768115942, "grad_norm": 0.09685653232396237, "learning_rate": 3.544422731474743e-05, "loss": 0.9226, "step": 6433 }, { "epoch": 2.913949275362319, "grad_norm": 0.06920209497423949, "learning_rate": 3.5431662526961664e-05, "loss": 0.9285, "step": 6434 }, { "epoch": 2.9144021739130435, "grad_norm": 0.09256039221849502, "learning_rate": 3.541909819587984e-05, "loss": 0.9451, "step": 6435 }, { "epoch": 2.914855072463768, "grad_norm": 0.07451094192964554, "learning_rate": 3.540653432275801e-05, "loss": 0.932, "step": 6436 }, { "epoch": 2.9153079710144927, "grad_norm": 0.0823155502519729, "learning_rate": 3.539397090885225e-05, "loss": 0.9158, "step": 6437 }, { "epoch": 2.9157608695652173, "grad_norm": 0.08410087215999666, "learning_rate": 3.5381407955418496e-05, "loss": 0.9252, "step": 6438 }, { "epoch": 2.916213768115942, "grad_norm": 0.07787066037052724, "learning_rate": 3.536884546371272e-05, "loss": 0.9522, "step": 6439 }, { "epoch": 2.9166666666666665, "grad_norm": 0.08452181807914455, "learning_rate": 3.535628343499079e-05, "loss": 0.9327, "step": 6440 }, { "epoch": 2.9171195652173916, "grad_norm": 0.08181205181096246, "learning_rate": 3.534372187050857e-05, "loss": 0.933, "step": 6441 }, { "epoch": 2.917572463768116, "grad_norm": 0.07097186305563859, "learning_rate": 3.533116077152185e-05, "loss": 0.9504, "step": 6442 }, { "epoch": 2.918025362318841, "grad_norm": 0.0692217258472148, "learning_rate": 3.53186001392864e-05, "loss": 0.9442, "step": 6443 }, { "epoch": 2.9184782608695654, "grad_norm": 0.08654347628742334, "learning_rate": 3.530603997505789e-05, "loss": 0.9341, "step": 6444 }, { "epoch": 2.91893115942029, "grad_norm": 0.06943223708056266, "learning_rate": 3.529348028009202e-05, "loss": 0.9341, "step": 6445 }, { "epoch": 2.9193840579710146, "grad_norm": 0.07956377294369874, "learning_rate": 3.5280921055644387e-05, "loss": 0.9274, "step": 6446 }, { "epoch": 2.9198369565217392, "grad_norm": 0.0835782179820856, "learning_rate": 3.526836230297054e-05, "loss": 0.9441, "step": 6447 }, { "epoch": 2.920289855072464, "grad_norm": 0.06747272698159901, "learning_rate": 3.525580402332603e-05, "loss": 0.9301, "step": 6448 }, { "epoch": 2.9207427536231885, "grad_norm": 0.08232926649609074, "learning_rate": 3.5243246217966284e-05, "loss": 0.9345, "step": 6449 }, { "epoch": 2.921195652173913, "grad_norm": 0.0709001429059477, "learning_rate": 3.523068888814677e-05, "loss": 0.9369, "step": 6450 }, { "epoch": 2.9216485507246377, "grad_norm": 0.08749102788574867, "learning_rate": 3.521813203512284e-05, "loss": 0.9334, "step": 6451 }, { "epoch": 2.9221014492753623, "grad_norm": 0.06884315765060786, "learning_rate": 3.520557566014983e-05, "loss": 0.9325, "step": 6452 }, { "epoch": 2.922554347826087, "grad_norm": 0.07649243117447309, "learning_rate": 3.5193019764483005e-05, "loss": 0.9657, "step": 6453 }, { "epoch": 2.9230072463768115, "grad_norm": 0.0883015728433411, "learning_rate": 3.5180464349377626e-05, "loss": 0.9286, "step": 6454 }, { "epoch": 2.923460144927536, "grad_norm": 0.07132341557765166, "learning_rate": 3.516790941608886e-05, "loss": 0.9422, "step": 6455 }, { "epoch": 2.9239130434782608, "grad_norm": 0.07686206009399582, "learning_rate": 3.5155354965871846e-05, "loss": 0.9394, "step": 6456 }, { "epoch": 2.9243659420289854, "grad_norm": 0.09499493717209237, "learning_rate": 3.5142800999981665e-05, "loss": 0.946, "step": 6457 }, { "epoch": 2.92481884057971, "grad_norm": 0.09454844769475983, "learning_rate": 3.513024751967338e-05, "loss": 0.9511, "step": 6458 }, { "epoch": 2.9252717391304346, "grad_norm": 0.07259823842904285, "learning_rate": 3.511769452620196e-05, "loss": 0.9309, "step": 6459 }, { "epoch": 2.925724637681159, "grad_norm": 0.0917869827099076, "learning_rate": 3.5105142020822366e-05, "loss": 0.9181, "step": 6460 }, { "epoch": 2.926177536231884, "grad_norm": 0.09031927610572466, "learning_rate": 3.509259000478945e-05, "loss": 0.9317, "step": 6461 }, { "epoch": 2.9266304347826084, "grad_norm": 0.08847259664112632, "learning_rate": 3.5080038479358124e-05, "loss": 0.9411, "step": 6462 }, { "epoch": 2.9270833333333335, "grad_norm": 0.08255725249341697, "learning_rate": 3.506748744578312e-05, "loss": 0.9122, "step": 6463 }, { "epoch": 2.927536231884058, "grad_norm": 0.07921844819808901, "learning_rate": 3.5054936905319236e-05, "loss": 0.9467, "step": 6464 }, { "epoch": 2.9279891304347827, "grad_norm": 0.07594832821115828, "learning_rate": 3.5042386859221125e-05, "loss": 0.9042, "step": 6465 }, { "epoch": 2.9284420289855073, "grad_norm": 0.07323539921767115, "learning_rate": 3.5029837308743464e-05, "loss": 0.9214, "step": 6466 }, { "epoch": 2.928894927536232, "grad_norm": 0.07506341573960805, "learning_rate": 3.501728825514085e-05, "loss": 0.9316, "step": 6467 }, { "epoch": 2.9293478260869565, "grad_norm": 0.07578837408848954, "learning_rate": 3.5004739699667816e-05, "loss": 0.9433, "step": 6468 }, { "epoch": 2.929800724637681, "grad_norm": 0.08444141565183551, "learning_rate": 3.499219164357886e-05, "loss": 0.9403, "step": 6469 }, { "epoch": 2.9302536231884058, "grad_norm": 0.07427633174729356, "learning_rate": 3.497964408812846e-05, "loss": 0.9218, "step": 6470 }, { "epoch": 2.9307065217391304, "grad_norm": 0.08277995857737437, "learning_rate": 3.496709703457098e-05, "loss": 0.9083, "step": 6471 }, { "epoch": 2.931159420289855, "grad_norm": 0.0685255542662525, "learning_rate": 3.495455048416081e-05, "loss": 0.9205, "step": 6472 }, { "epoch": 2.9316123188405796, "grad_norm": 0.08645580478849567, "learning_rate": 3.4942004438152196e-05, "loss": 0.9408, "step": 6473 }, { "epoch": 2.932065217391304, "grad_norm": 0.07101236297623505, "learning_rate": 3.492945889779943e-05, "loss": 0.928, "step": 6474 }, { "epoch": 2.932518115942029, "grad_norm": 0.070421498620704, "learning_rate": 3.4916913864356675e-05, "loss": 0.9317, "step": 6475 }, { "epoch": 2.932971014492754, "grad_norm": 0.06779575623913757, "learning_rate": 3.4904369339078117e-05, "loss": 0.9583, "step": 6476 }, { "epoch": 2.9334239130434785, "grad_norm": 0.07140202592344663, "learning_rate": 3.4891825323217824e-05, "loss": 0.9508, "step": 6477 }, { "epoch": 2.933876811594203, "grad_norm": 0.07230513397048437, "learning_rate": 3.4879281818029846e-05, "loss": 0.934, "step": 6478 }, { "epoch": 2.9343297101449277, "grad_norm": 0.06687036355225481, "learning_rate": 3.486673882476817e-05, "loss": 0.9364, "step": 6479 }, { "epoch": 2.9347826086956523, "grad_norm": 0.06555464038205935, "learning_rate": 3.485419634468676e-05, "loss": 0.9373, "step": 6480 }, { "epoch": 2.935235507246377, "grad_norm": 0.06514041441364819, "learning_rate": 3.4841654379039495e-05, "loss": 0.9123, "step": 6481 }, { "epoch": 2.9356884057971016, "grad_norm": 0.08264443375003415, "learning_rate": 3.482911292908022e-05, "loss": 0.9328, "step": 6482 }, { "epoch": 2.936141304347826, "grad_norm": 0.08017275605904282, "learning_rate": 3.48165719960627e-05, "loss": 0.9302, "step": 6483 }, { "epoch": 2.9365942028985508, "grad_norm": 0.07530087891578906, "learning_rate": 3.4804031581240716e-05, "loss": 0.9225, "step": 6484 }, { "epoch": 2.9370471014492754, "grad_norm": 0.08646380178153237, "learning_rate": 3.479149168586791e-05, "loss": 0.9347, "step": 6485 }, { "epoch": 2.9375, "grad_norm": 0.07978115664494718, "learning_rate": 3.477895231119795e-05, "loss": 0.9266, "step": 6486 }, { "epoch": 2.9379528985507246, "grad_norm": 0.07523957521222213, "learning_rate": 3.476641345848438e-05, "loss": 0.9467, "step": 6487 }, { "epoch": 2.9384057971014492, "grad_norm": 0.07984149605402338, "learning_rate": 3.475387512898077e-05, "loss": 0.931, "step": 6488 }, { "epoch": 2.938858695652174, "grad_norm": 0.10076498915338249, "learning_rate": 3.474133732394056e-05, "loss": 0.9292, "step": 6489 }, { "epoch": 2.9393115942028984, "grad_norm": 0.06351165640438387, "learning_rate": 3.472880004461721e-05, "loss": 0.9423, "step": 6490 }, { "epoch": 2.939764492753623, "grad_norm": 0.08919175594159363, "learning_rate": 3.471626329226405e-05, "loss": 0.9391, "step": 6491 }, { "epoch": 2.9402173913043477, "grad_norm": 0.09721067293919629, "learning_rate": 3.470372706813444e-05, "loss": 0.9151, "step": 6492 }, { "epoch": 2.9406702898550723, "grad_norm": 0.07414844091958245, "learning_rate": 3.4691191373481614e-05, "loss": 0.9042, "step": 6493 }, { "epoch": 2.941123188405797, "grad_norm": 0.10036537013369198, "learning_rate": 3.4678656209558814e-05, "loss": 0.9306, "step": 6494 }, { "epoch": 2.9415760869565215, "grad_norm": 0.10569483769382036, "learning_rate": 3.4666121577619164e-05, "loss": 0.9401, "step": 6495 }, { "epoch": 2.942028985507246, "grad_norm": 0.07894622773369869, "learning_rate": 3.465358747891581e-05, "loss": 0.9295, "step": 6496 }, { "epoch": 2.942481884057971, "grad_norm": 0.07411606736814988, "learning_rate": 3.464105391470178e-05, "loss": 0.9348, "step": 6497 }, { "epoch": 2.942934782608696, "grad_norm": 0.08712650936783531, "learning_rate": 3.462852088623009e-05, "loss": 0.9342, "step": 6498 }, { "epoch": 2.9433876811594204, "grad_norm": 0.08584856848977761, "learning_rate": 3.461598839475367e-05, "loss": 0.9268, "step": 6499 }, { "epoch": 2.943840579710145, "grad_norm": 0.07041253813116233, "learning_rate": 3.460345644152543e-05, "loss": 0.9568, "step": 6500 }, { "epoch": 2.9442934782608696, "grad_norm": 0.07056085656056964, "learning_rate": 3.4590925027798194e-05, "loss": 0.9223, "step": 6501 }, { "epoch": 2.9447463768115942, "grad_norm": 0.0816809808120361, "learning_rate": 3.457839415482476e-05, "loss": 0.9213, "step": 6502 }, { "epoch": 2.945199275362319, "grad_norm": 0.08295381483930728, "learning_rate": 3.456586382385784e-05, "loss": 0.9189, "step": 6503 }, { "epoch": 2.9456521739130435, "grad_norm": 0.08629208571689448, "learning_rate": 3.4553334036150146e-05, "loss": 0.9257, "step": 6504 }, { "epoch": 2.946105072463768, "grad_norm": 0.07167973043886122, "learning_rate": 3.454080479295427e-05, "loss": 0.9366, "step": 6505 }, { "epoch": 2.9465579710144927, "grad_norm": 0.07360591746024818, "learning_rate": 3.4528276095522786e-05, "loss": 0.9301, "step": 6506 }, { "epoch": 2.9470108695652173, "grad_norm": 0.07628869933830722, "learning_rate": 3.45157479451082e-05, "loss": 0.9222, "step": 6507 }, { "epoch": 2.947463768115942, "grad_norm": 0.07430928749319235, "learning_rate": 3.450322034296301e-05, "loss": 0.9305, "step": 6508 }, { "epoch": 2.9479166666666665, "grad_norm": 0.07066265203540033, "learning_rate": 3.4490693290339576e-05, "loss": 0.9368, "step": 6509 }, { "epoch": 2.9483695652173916, "grad_norm": 0.07685716253193312, "learning_rate": 3.447816678849028e-05, "loss": 0.9399, "step": 6510 }, { "epoch": 2.948822463768116, "grad_norm": 0.08690124201149356, "learning_rate": 3.446564083866737e-05, "loss": 0.9376, "step": 6511 }, { "epoch": 2.949275362318841, "grad_norm": 0.08441690098834236, "learning_rate": 3.4453115442123136e-05, "loss": 0.9348, "step": 6512 }, { "epoch": 2.9497282608695654, "grad_norm": 0.07717880101132012, "learning_rate": 3.4440590600109734e-05, "loss": 0.9095, "step": 6513 }, { "epoch": 2.95018115942029, "grad_norm": 0.08840148523662615, "learning_rate": 3.442806631387931e-05, "loss": 0.9282, "step": 6514 }, { "epoch": 2.9506340579710146, "grad_norm": 0.06712267495755478, "learning_rate": 3.441554258468391e-05, "loss": 0.9245, "step": 6515 }, { "epoch": 2.9510869565217392, "grad_norm": 0.08073400201754483, "learning_rate": 3.440301941377558e-05, "loss": 0.9244, "step": 6516 }, { "epoch": 2.951539855072464, "grad_norm": 0.0700426683597466, "learning_rate": 3.4390496802406255e-05, "loss": 0.9357, "step": 6517 }, { "epoch": 2.9519927536231885, "grad_norm": 0.092346069781503, "learning_rate": 3.437797475182787e-05, "loss": 0.9281, "step": 6518 }, { "epoch": 2.952445652173913, "grad_norm": 0.09007568054412127, "learning_rate": 3.436545326329225e-05, "loss": 0.9531, "step": 6519 }, { "epoch": 2.9528985507246377, "grad_norm": 0.07688496604202134, "learning_rate": 3.4352932338051204e-05, "loss": 0.9369, "step": 6520 }, { "epoch": 2.9533514492753623, "grad_norm": 0.08787175568956969, "learning_rate": 3.4340411977356445e-05, "loss": 0.9377, "step": 6521 }, { "epoch": 2.953804347826087, "grad_norm": 0.06920711768112428, "learning_rate": 3.432789218245968e-05, "loss": 0.9113, "step": 6522 }, { "epoch": 2.9542572463768115, "grad_norm": 0.07543077631250414, "learning_rate": 3.431537295461252e-05, "loss": 0.9001, "step": 6523 }, { "epoch": 2.954710144927536, "grad_norm": 0.07692244941737489, "learning_rate": 3.430285429506654e-05, "loss": 0.9225, "step": 6524 }, { "epoch": 2.9551630434782608, "grad_norm": 0.07945919285306824, "learning_rate": 3.429033620507323e-05, "loss": 0.9666, "step": 6525 }, { "epoch": 2.9556159420289854, "grad_norm": 0.07682990337859452, "learning_rate": 3.427781868588406e-05, "loss": 0.9245, "step": 6526 }, { "epoch": 2.95606884057971, "grad_norm": 0.0789400730663778, "learning_rate": 3.426530173875043e-05, "loss": 0.9402, "step": 6527 }, { "epoch": 2.9565217391304346, "grad_norm": 0.08514705328899493, "learning_rate": 3.425278536492368e-05, "loss": 0.9365, "step": 6528 }, { "epoch": 2.956974637681159, "grad_norm": 0.07528329629594543, "learning_rate": 3.424026956565506e-05, "loss": 0.9206, "step": 6529 }, { "epoch": 2.957427536231884, "grad_norm": 0.08003003181698204, "learning_rate": 3.422775434219584e-05, "loss": 0.9081, "step": 6530 }, { "epoch": 2.9578804347826084, "grad_norm": 0.08562996939845181, "learning_rate": 3.421523969579716e-05, "loss": 0.9588, "step": 6531 }, { "epoch": 2.9583333333333335, "grad_norm": 0.07983359419007424, "learning_rate": 3.4202725627710136e-05, "loss": 0.9179, "step": 6532 }, { "epoch": 2.958786231884058, "grad_norm": 0.09694559306046625, "learning_rate": 3.4190212139185805e-05, "loss": 0.9193, "step": 6533 }, { "epoch": 2.9592391304347827, "grad_norm": 0.07479615543197327, "learning_rate": 3.417769923147519e-05, "loss": 0.954, "step": 6534 }, { "epoch": 2.9596920289855073, "grad_norm": 0.09529717188837893, "learning_rate": 3.41651869058292e-05, "loss": 0.9208, "step": 6535 }, { "epoch": 2.960144927536232, "grad_norm": 0.0893450948741887, "learning_rate": 3.415267516349873e-05, "loss": 0.9286, "step": 6536 }, { "epoch": 2.9605978260869565, "grad_norm": 0.078689269371243, "learning_rate": 3.414016400573457e-05, "loss": 0.929, "step": 6537 }, { "epoch": 2.961050724637681, "grad_norm": 0.08047731926585953, "learning_rate": 3.412765343378752e-05, "loss": 0.9404, "step": 6538 }, { "epoch": 2.9615036231884058, "grad_norm": 0.11355504877088812, "learning_rate": 3.411514344890826e-05, "loss": 0.9068, "step": 6539 }, { "epoch": 2.9619565217391304, "grad_norm": 0.08416162882704165, "learning_rate": 3.410263405234743e-05, "loss": 0.9326, "step": 6540 }, { "epoch": 2.962409420289855, "grad_norm": 0.09133089558718482, "learning_rate": 3.4090125245355605e-05, "loss": 0.9186, "step": 6541 }, { "epoch": 2.9628623188405796, "grad_norm": 0.09453863571386499, "learning_rate": 3.407761702918334e-05, "loss": 0.9417, "step": 6542 }, { "epoch": 2.963315217391304, "grad_norm": 0.10907255370100688, "learning_rate": 3.4065109405081084e-05, "loss": 0.9181, "step": 6543 }, { "epoch": 2.963768115942029, "grad_norm": 0.0771550584970186, "learning_rate": 3.405260237429925e-05, "loss": 0.9398, "step": 6544 }, { "epoch": 2.964221014492754, "grad_norm": 0.0792087676324483, "learning_rate": 3.404009593808817e-05, "loss": 0.9395, "step": 6545 }, { "epoch": 2.9646739130434785, "grad_norm": 0.08620516961658527, "learning_rate": 3.4027590097698154e-05, "loss": 0.9302, "step": 6546 }, { "epoch": 2.965126811594203, "grad_norm": 0.08144543075582297, "learning_rate": 3.401508485437941e-05, "loss": 0.9216, "step": 6547 }, { "epoch": 2.9655797101449277, "grad_norm": 0.07116798739810655, "learning_rate": 3.400258020938213e-05, "loss": 0.9407, "step": 6548 }, { "epoch": 2.9660326086956523, "grad_norm": 0.08080389993513042, "learning_rate": 3.3990076163956386e-05, "loss": 0.9199, "step": 6549 }, { "epoch": 2.966485507246377, "grad_norm": 0.0772326192131343, "learning_rate": 3.397757271935228e-05, "loss": 0.9309, "step": 6550 }, { "epoch": 2.9669384057971016, "grad_norm": 0.07567755337285878, "learning_rate": 3.3965069876819755e-05, "loss": 0.9482, "step": 6551 }, { "epoch": 2.967391304347826, "grad_norm": 0.10210375226289335, "learning_rate": 3.395256763760877e-05, "loss": 0.9574, "step": 6552 }, { "epoch": 2.9678442028985508, "grad_norm": 0.07811702711473749, "learning_rate": 3.394006600296916e-05, "loss": 0.948, "step": 6553 }, { "epoch": 2.9682971014492754, "grad_norm": 0.07491482542676696, "learning_rate": 3.392756497415078e-05, "loss": 0.9227, "step": 6554 }, { "epoch": 2.96875, "grad_norm": 0.0721366368356893, "learning_rate": 3.3915064552403336e-05, "loss": 0.9305, "step": 6555 }, { "epoch": 2.9692028985507246, "grad_norm": 0.07248485857386928, "learning_rate": 3.3902564738976545e-05, "loss": 0.9239, "step": 6556 }, { "epoch": 2.9696557971014492, "grad_norm": 0.06851507072648144, "learning_rate": 3.389006553512e-05, "loss": 0.9546, "step": 6557 }, { "epoch": 2.970108695652174, "grad_norm": 0.06620334067879868, "learning_rate": 3.3877566942083294e-05, "loss": 0.9427, "step": 6558 }, { "epoch": 2.9705615942028984, "grad_norm": 0.06938603338843848, "learning_rate": 3.386506896111592e-05, "loss": 0.9283, "step": 6559 }, { "epoch": 2.971014492753623, "grad_norm": 0.0902145961263006, "learning_rate": 3.3852571593467335e-05, "loss": 0.9438, "step": 6560 }, { "epoch": 2.9714673913043477, "grad_norm": 0.09405818200704365, "learning_rate": 3.384007484038689e-05, "loss": 0.9254, "step": 6561 }, { "epoch": 2.9719202898550723, "grad_norm": 0.06865332346983481, "learning_rate": 3.382757870312395e-05, "loss": 0.9175, "step": 6562 }, { "epoch": 2.972373188405797, "grad_norm": 0.06377075660622175, "learning_rate": 3.381508318292772e-05, "loss": 0.9255, "step": 6563 }, { "epoch": 2.9728260869565215, "grad_norm": 0.06593315095423345, "learning_rate": 3.3802588281047444e-05, "loss": 0.914, "step": 6564 }, { "epoch": 2.973278985507246, "grad_norm": 0.06955751199571868, "learning_rate": 3.379009399873223e-05, "loss": 0.9218, "step": 6565 }, { "epoch": 2.973731884057971, "grad_norm": 0.07459736179409025, "learning_rate": 3.377760033723117e-05, "loss": 0.9386, "step": 6566 }, { "epoch": 2.974184782608696, "grad_norm": 0.07051163002099314, "learning_rate": 3.376510729779325e-05, "loss": 0.955, "step": 6567 }, { "epoch": 2.9746376811594204, "grad_norm": 0.06489257542228125, "learning_rate": 3.3752614881667444e-05, "loss": 0.9442, "step": 6568 }, { "epoch": 2.975090579710145, "grad_norm": 0.07031431856946328, "learning_rate": 3.374012309010264e-05, "loss": 0.9284, "step": 6569 }, { "epoch": 2.9755434782608696, "grad_norm": 0.07485679067291114, "learning_rate": 3.3727631924347625e-05, "loss": 0.9532, "step": 6570 }, { "epoch": 2.9759963768115942, "grad_norm": 0.0949953479431596, "learning_rate": 3.3715141385651205e-05, "loss": 0.9392, "step": 6571 }, { "epoch": 2.976449275362319, "grad_norm": 0.06550404563196405, "learning_rate": 3.370265147526204e-05, "loss": 0.9274, "step": 6572 }, { "epoch": 2.9769021739130435, "grad_norm": 0.0916394356764105, "learning_rate": 3.3690162194428804e-05, "loss": 0.9456, "step": 6573 }, { "epoch": 2.977355072463768, "grad_norm": 0.07015710553587906, "learning_rate": 3.3677673544400046e-05, "loss": 0.9246, "step": 6574 }, { "epoch": 2.9778079710144927, "grad_norm": 0.092650384197442, "learning_rate": 3.3665185526424285e-05, "loss": 0.951, "step": 6575 }, { "epoch": 2.9782608695652173, "grad_norm": 0.07311701927143663, "learning_rate": 3.3652698141749946e-05, "loss": 0.9386, "step": 6576 }, { "epoch": 2.978713768115942, "grad_norm": 0.07929900283242255, "learning_rate": 3.364021139162546e-05, "loss": 0.9369, "step": 6577 }, { "epoch": 2.9791666666666665, "grad_norm": 0.07412913295315232, "learning_rate": 3.3627725277299103e-05, "loss": 0.9421, "step": 6578 }, { "epoch": 2.9796195652173916, "grad_norm": 0.08169379709919819, "learning_rate": 3.361523980001916e-05, "loss": 0.9152, "step": 6579 }, { "epoch": 2.980072463768116, "grad_norm": 0.07404542148824872, "learning_rate": 3.3602754961033794e-05, "loss": 0.9231, "step": 6580 }, { "epoch": 2.980525362318841, "grad_norm": 0.07575707297512038, "learning_rate": 3.359027076159116e-05, "loss": 0.9368, "step": 6581 }, { "epoch": 2.9809782608695654, "grad_norm": 0.09238938788689191, "learning_rate": 3.357778720293932e-05, "loss": 0.9267, "step": 6582 }, { "epoch": 2.98143115942029, "grad_norm": 0.06630049717109857, "learning_rate": 3.3565304286326274e-05, "loss": 0.9272, "step": 6583 }, { "epoch": 2.9818840579710146, "grad_norm": 0.07596717443579135, "learning_rate": 3.355282201299994e-05, "loss": 0.9123, "step": 6584 }, { "epoch": 2.9823369565217392, "grad_norm": 0.0684885031027947, "learning_rate": 3.354034038420821e-05, "loss": 0.9198, "step": 6585 }, { "epoch": 2.982789855072464, "grad_norm": 0.06551683490575608, "learning_rate": 3.35278594011989e-05, "loss": 0.9271, "step": 6586 }, { "epoch": 2.9832427536231885, "grad_norm": 0.08155205919243004, "learning_rate": 3.351537906521973e-05, "loss": 0.928, "step": 6587 }, { "epoch": 2.983695652173913, "grad_norm": 0.07019670700091217, "learning_rate": 3.350289937751839e-05, "loss": 0.9166, "step": 6588 }, { "epoch": 2.9841485507246377, "grad_norm": 0.07703547584517802, "learning_rate": 3.34904203393425e-05, "loss": 0.9413, "step": 6589 }, { "epoch": 2.9846014492753623, "grad_norm": 0.09455392626792126, "learning_rate": 3.34779419519396e-05, "loss": 0.9187, "step": 6590 }, { "epoch": 2.985054347826087, "grad_norm": 0.0880914832469021, "learning_rate": 3.346546421655719e-05, "loss": 0.9101, "step": 6591 }, { "epoch": 2.9855072463768115, "grad_norm": 0.07781252600328431, "learning_rate": 3.345298713444265e-05, "loss": 0.9184, "step": 6592 }, { "epoch": 2.985960144927536, "grad_norm": 0.105176313498199, "learning_rate": 3.344051070684339e-05, "loss": 0.9284, "step": 6593 }, { "epoch": 2.9864130434782608, "grad_norm": 0.09581927613146973, "learning_rate": 3.3428034935006655e-05, "loss": 0.934, "step": 6594 }, { "epoch": 2.9868659420289854, "grad_norm": 0.07505764152438965, "learning_rate": 3.341555982017969e-05, "loss": 0.9372, "step": 6595 }, { "epoch": 2.98731884057971, "grad_norm": 0.09701252279575669, "learning_rate": 3.3403085363609626e-05, "loss": 0.9511, "step": 6596 }, { "epoch": 2.9877717391304346, "grad_norm": 0.07441213836601515, "learning_rate": 3.339061156654359e-05, "loss": 0.9517, "step": 6597 }, { "epoch": 2.988224637681159, "grad_norm": 0.07500002977833473, "learning_rate": 3.337813843022859e-05, "loss": 0.9241, "step": 6598 }, { "epoch": 2.988677536231884, "grad_norm": 0.07455413115726786, "learning_rate": 3.336566595591158e-05, "loss": 0.9464, "step": 6599 }, { "epoch": 2.9891304347826084, "grad_norm": 0.06778266727741, "learning_rate": 3.3353194144839444e-05, "loss": 0.9239, "step": 6600 }, { "epoch": 2.9895833333333335, "grad_norm": 0.06902992004104301, "learning_rate": 3.3340722998259036e-05, "loss": 0.9222, "step": 6601 }, { "epoch": 2.990036231884058, "grad_norm": 0.06766363910450049, "learning_rate": 3.33282525174171e-05, "loss": 0.9417, "step": 6602 }, { "epoch": 2.9904891304347827, "grad_norm": 0.06782829299148034, "learning_rate": 3.331578270356034e-05, "loss": 0.9147, "step": 6603 }, { "epoch": 2.9909420289855073, "grad_norm": 0.0696497823727761, "learning_rate": 3.3303313557935357e-05, "loss": 0.9377, "step": 6604 }, { "epoch": 2.991394927536232, "grad_norm": 0.07558377363714237, "learning_rate": 3.329084508178874e-05, "loss": 0.9482, "step": 6605 }, { "epoch": 2.9918478260869565, "grad_norm": 0.07894412985237083, "learning_rate": 3.327837727636696e-05, "loss": 0.947, "step": 6606 }, { "epoch": 2.992300724637681, "grad_norm": 0.07162351968553328, "learning_rate": 3.326591014291648e-05, "loss": 0.9353, "step": 6607 }, { "epoch": 2.9927536231884058, "grad_norm": 0.07249442180489118, "learning_rate": 3.325344368268361e-05, "loss": 0.9137, "step": 6608 }, { "epoch": 2.9932065217391304, "grad_norm": 0.08719693572327276, "learning_rate": 3.324097789691468e-05, "loss": 0.9364, "step": 6609 }, { "epoch": 2.993659420289855, "grad_norm": 0.06227423852383193, "learning_rate": 3.322851278685589e-05, "loss": 0.9458, "step": 6610 }, { "epoch": 2.9941123188405796, "grad_norm": 0.0727367905084411, "learning_rate": 3.3216048353753424e-05, "loss": 0.9499, "step": 6611 }, { "epoch": 2.994565217391304, "grad_norm": 0.06788224283597968, "learning_rate": 3.3203584598853335e-05, "loss": 0.9251, "step": 6612 }, { "epoch": 2.995018115942029, "grad_norm": 0.07812106073434393, "learning_rate": 3.319112152340168e-05, "loss": 0.9427, "step": 6613 }, { "epoch": 2.995471014492754, "grad_norm": 0.07525561158769047, "learning_rate": 3.317865912864438e-05, "loss": 0.9297, "step": 6614 }, { "epoch": 2.9959239130434785, "grad_norm": 0.06948176369092471, "learning_rate": 3.3166197415827345e-05, "loss": 0.9477, "step": 6615 }, { "epoch": 2.996376811594203, "grad_norm": 0.07189686062031118, "learning_rate": 3.3153736386196384e-05, "loss": 0.9346, "step": 6616 }, { "epoch": 2.9968297101449277, "grad_norm": 0.07463412322555558, "learning_rate": 3.314127604099725e-05, "loss": 0.957, "step": 6617 }, { "epoch": 2.9972826086956523, "grad_norm": 0.13283483729763723, "learning_rate": 3.31288163814756e-05, "loss": 0.9383, "step": 6618 }, { "epoch": 2.997735507246377, "grad_norm": 0.06445182257774607, "learning_rate": 3.3116357408877084e-05, "loss": 0.945, "step": 6619 }, { "epoch": 2.9981884057971016, "grad_norm": 0.07103629320122516, "learning_rate": 3.3103899124447216e-05, "loss": 0.9138, "step": 6620 }, { "epoch": 2.998641304347826, "grad_norm": 0.07836679707677222, "learning_rate": 3.309144152943149e-05, "loss": 0.9292, "step": 6621 }, { "epoch": 2.9990942028985508, "grad_norm": 0.3844647573530185, "learning_rate": 3.3078984625075284e-05, "loss": 0.9296, "step": 6622 }, { "epoch": 2.9995471014492754, "grad_norm": 0.08171565735025053, "learning_rate": 3.306652841262397e-05, "loss": 0.9258, "step": 6623 }, { "epoch": 3.0, "grad_norm": 0.07647294789596736, "learning_rate": 3.305407289332279e-05, "loss": 0.9464, "step": 6624 }, { "epoch": 3.0004528985507246, "grad_norm": 0.08772537846940139, "learning_rate": 3.304161806841695e-05, "loss": 0.9234, "step": 6625 }, { "epoch": 3.0009057971014492, "grad_norm": 0.10741146957059605, "learning_rate": 3.302916393915157e-05, "loss": 0.9291, "step": 6626 }, { "epoch": 3.001358695652174, "grad_norm": 0.07515367549412115, "learning_rate": 3.3016710506771734e-05, "loss": 0.9211, "step": 6627 }, { "epoch": 3.0018115942028984, "grad_norm": 0.07745724259492648, "learning_rate": 3.300425777252241e-05, "loss": 0.9251, "step": 6628 }, { "epoch": 3.002264492753623, "grad_norm": 0.08581157626348367, "learning_rate": 3.299180573764853e-05, "loss": 0.9577, "step": 6629 }, { "epoch": 3.0027173913043477, "grad_norm": 0.06913120208395734, "learning_rate": 3.297935440339491e-05, "loss": 0.9373, "step": 6630 }, { "epoch": 3.0031702898550723, "grad_norm": 0.07628518674960466, "learning_rate": 3.296690377100637e-05, "loss": 0.9294, "step": 6631 }, { "epoch": 3.003623188405797, "grad_norm": 0.06934449157428434, "learning_rate": 3.295445384172761e-05, "loss": 0.918, "step": 6632 }, { "epoch": 3.004076086956522, "grad_norm": 0.0728925496378932, "learning_rate": 3.294200461680326e-05, "loss": 0.9001, "step": 6633 }, { "epoch": 3.0045289855072466, "grad_norm": 0.06495323876491126, "learning_rate": 3.292955609747788e-05, "loss": 0.937, "step": 6634 }, { "epoch": 3.004981884057971, "grad_norm": 0.08176422536655609, "learning_rate": 3.2917108284996004e-05, "loss": 0.9475, "step": 6635 }, { "epoch": 3.005434782608696, "grad_norm": 0.06366517638050075, "learning_rate": 3.2904661180602024e-05, "loss": 0.9199, "step": 6636 }, { "epoch": 3.0058876811594204, "grad_norm": 0.0807404956101849, "learning_rate": 3.2892214785540315e-05, "loss": 0.9288, "step": 6637 }, { "epoch": 3.006340579710145, "grad_norm": 0.08046495813459309, "learning_rate": 3.287976910105514e-05, "loss": 0.9375, "step": 6638 }, { "epoch": 3.0067934782608696, "grad_norm": 0.06449093695685237, "learning_rate": 3.2867324128390756e-05, "loss": 0.9338, "step": 6639 }, { "epoch": 3.0072463768115942, "grad_norm": 0.08055697625742515, "learning_rate": 3.285487986879127e-05, "loss": 0.9084, "step": 6640 }, { "epoch": 3.007699275362319, "grad_norm": 0.08272191464688897, "learning_rate": 3.284243632350077e-05, "loss": 0.9201, "step": 6641 }, { "epoch": 3.0081521739130435, "grad_norm": 0.07804075430687776, "learning_rate": 3.282999349376324e-05, "loss": 0.9414, "step": 6642 }, { "epoch": 3.008605072463768, "grad_norm": 0.07892290235653407, "learning_rate": 3.2817551380822633e-05, "loss": 0.9228, "step": 6643 }, { "epoch": 3.0090579710144927, "grad_norm": 0.07864085505917123, "learning_rate": 3.2805109985922796e-05, "loss": 0.9265, "step": 6644 }, { "epoch": 3.0095108695652173, "grad_norm": 0.07340565074340381, "learning_rate": 3.2792669310307515e-05, "loss": 0.9284, "step": 6645 }, { "epoch": 3.009963768115942, "grad_norm": 0.07152876936245867, "learning_rate": 3.278022935522049e-05, "loss": 0.9408, "step": 6646 }, { "epoch": 3.0104166666666665, "grad_norm": 0.07243665783785716, "learning_rate": 3.2767790121905396e-05, "loss": 0.946, "step": 6647 }, { "epoch": 3.010869565217391, "grad_norm": 0.0670055599914299, "learning_rate": 3.2755351611605765e-05, "loss": 0.9247, "step": 6648 }, { "epoch": 3.0113224637681157, "grad_norm": 0.07345701517514906, "learning_rate": 3.2742913825565145e-05, "loss": 0.9224, "step": 6649 }, { "epoch": 3.011775362318841, "grad_norm": 0.08985184338270799, "learning_rate": 3.27304767650269e-05, "loss": 0.9306, "step": 6650 }, { "epoch": 3.0122282608695654, "grad_norm": 0.08493967754963279, "learning_rate": 3.2718040431234425e-05, "loss": 0.9144, "step": 6651 }, { "epoch": 3.01268115942029, "grad_norm": 0.06823368052609015, "learning_rate": 3.2705604825430976e-05, "loss": 0.9062, "step": 6652 }, { "epoch": 3.0131340579710146, "grad_norm": 0.0861481643877053, "learning_rate": 3.269316994885978e-05, "loss": 0.9387, "step": 6653 }, { "epoch": 3.0135869565217392, "grad_norm": 0.09799197449688081, "learning_rate": 3.2680735802763966e-05, "loss": 0.9088, "step": 6654 }, { "epoch": 3.014039855072464, "grad_norm": 0.09528905090090252, "learning_rate": 3.2668302388386594e-05, "loss": 0.918, "step": 6655 }, { "epoch": 3.0144927536231885, "grad_norm": 0.08707117109907239, "learning_rate": 3.265586970697064e-05, "loss": 0.9289, "step": 6656 }, { "epoch": 3.014945652173913, "grad_norm": 0.08398842109236171, "learning_rate": 3.264343775975904e-05, "loss": 0.9227, "step": 6657 }, { "epoch": 3.0153985507246377, "grad_norm": 0.09742034292307696, "learning_rate": 3.2631006547994616e-05, "loss": 0.9244, "step": 6658 }, { "epoch": 3.0158514492753623, "grad_norm": 0.07371175874428254, "learning_rate": 3.261857607292015e-05, "loss": 0.9389, "step": 6659 }, { "epoch": 3.016304347826087, "grad_norm": 0.07743584974327548, "learning_rate": 3.260614633577832e-05, "loss": 0.929, "step": 6660 }, { "epoch": 3.0167572463768115, "grad_norm": 0.09834338559167226, "learning_rate": 3.2593717337811775e-05, "loss": 0.9182, "step": 6661 }, { "epoch": 3.017210144927536, "grad_norm": 0.07892078380670342, "learning_rate": 3.258128908026303e-05, "loss": 0.9207, "step": 6662 }, { "epoch": 3.0176630434782608, "grad_norm": 0.07444843484128619, "learning_rate": 3.256886156437458e-05, "loss": 0.9138, "step": 6663 }, { "epoch": 3.0181159420289854, "grad_norm": 0.07457698127119529, "learning_rate": 3.2556434791388804e-05, "loss": 0.9253, "step": 6664 }, { "epoch": 3.01856884057971, "grad_norm": 0.07519056285869322, "learning_rate": 3.254400876254805e-05, "loss": 0.9247, "step": 6665 }, { "epoch": 3.0190217391304346, "grad_norm": 0.06857800087622207, "learning_rate": 3.253158347909455e-05, "loss": 0.9222, "step": 6666 }, { "epoch": 3.019474637681159, "grad_norm": 0.06889505041613422, "learning_rate": 3.2519158942270496e-05, "loss": 0.918, "step": 6667 }, { "epoch": 3.0199275362318843, "grad_norm": 0.07918946294655342, "learning_rate": 3.250673515331795e-05, "loss": 0.9523, "step": 6668 }, { "epoch": 3.020380434782609, "grad_norm": 0.07196491244182869, "learning_rate": 3.249431211347899e-05, "loss": 0.9143, "step": 6669 }, { "epoch": 3.0208333333333335, "grad_norm": 0.07448011007495495, "learning_rate": 3.248188982399553e-05, "loss": 0.9431, "step": 6670 }, { "epoch": 3.021286231884058, "grad_norm": 0.37063995221760715, "learning_rate": 3.2469468286109465e-05, "loss": 0.917, "step": 6671 }, { "epoch": 3.0217391304347827, "grad_norm": 0.07969290055894106, "learning_rate": 3.245704750106257e-05, "loss": 0.9125, "step": 6672 }, { "epoch": 3.0221920289855073, "grad_norm": 0.0725588177792968, "learning_rate": 3.244462747009661e-05, "loss": 0.913, "step": 6673 }, { "epoch": 3.022644927536232, "grad_norm": 0.06236475079104522, "learning_rate": 3.2432208194453204e-05, "loss": 0.9179, "step": 6674 }, { "epoch": 3.0230978260869565, "grad_norm": 0.07710793703548417, "learning_rate": 3.241978967537395e-05, "loss": 0.9033, "step": 6675 }, { "epoch": 3.023550724637681, "grad_norm": 0.07758154905741392, "learning_rate": 3.240737191410031e-05, "loss": 0.9205, "step": 6676 }, { "epoch": 3.0240036231884058, "grad_norm": 0.06260320628603784, "learning_rate": 3.239495491187375e-05, "loss": 0.9279, "step": 6677 }, { "epoch": 3.0244565217391304, "grad_norm": 0.08245227645449667, "learning_rate": 3.238253866993559e-05, "loss": 0.9443, "step": 6678 }, { "epoch": 3.024909420289855, "grad_norm": 0.07497515092259975, "learning_rate": 3.2370123189527125e-05, "loss": 0.9219, "step": 6679 }, { "epoch": 3.0253623188405796, "grad_norm": 0.08007357517966567, "learning_rate": 3.2357708471889514e-05, "loss": 0.9134, "step": 6680 }, { "epoch": 3.025815217391304, "grad_norm": 0.07071878977065366, "learning_rate": 3.234529451826391e-05, "loss": 0.9275, "step": 6681 }, { "epoch": 3.026268115942029, "grad_norm": 0.07951030040200398, "learning_rate": 3.233288132989134e-05, "loss": 0.9218, "step": 6682 }, { "epoch": 3.0267210144927534, "grad_norm": 0.06152284290798771, "learning_rate": 3.2320468908012776e-05, "loss": 0.9155, "step": 6683 }, { "epoch": 3.027173913043478, "grad_norm": 0.0701649562519155, "learning_rate": 3.2308057253869095e-05, "loss": 0.9179, "step": 6684 }, { "epoch": 3.027626811594203, "grad_norm": 0.06772645584794276, "learning_rate": 3.229564636870113e-05, "loss": 0.9255, "step": 6685 }, { "epoch": 3.0280797101449277, "grad_norm": 0.09312484656702821, "learning_rate": 3.2283236253749605e-05, "loss": 0.9269, "step": 6686 }, { "epoch": 3.0285326086956523, "grad_norm": 0.07208609169642327, "learning_rate": 3.227082691025518e-05, "loss": 0.9175, "step": 6687 }, { "epoch": 3.028985507246377, "grad_norm": 0.07617165408817016, "learning_rate": 3.225841833945843e-05, "loss": 0.9291, "step": 6688 }, { "epoch": 3.0294384057971016, "grad_norm": 0.07728819744427859, "learning_rate": 3.224601054259987e-05, "loss": 0.9093, "step": 6689 }, { "epoch": 3.029891304347826, "grad_norm": 0.06384160729139468, "learning_rate": 3.2233603520919913e-05, "loss": 0.9201, "step": 6690 }, { "epoch": 3.0303442028985508, "grad_norm": 0.07001919068576296, "learning_rate": 3.2221197275658945e-05, "loss": 0.9282, "step": 6691 }, { "epoch": 3.0307971014492754, "grad_norm": 0.05802937127256486, "learning_rate": 3.2208791808057185e-05, "loss": 0.9127, "step": 6692 }, { "epoch": 3.03125, "grad_norm": 0.07272352155052682, "learning_rate": 3.219638711935488e-05, "loss": 0.9266, "step": 6693 }, { "epoch": 3.0317028985507246, "grad_norm": 0.06445477069817375, "learning_rate": 3.218398321079211e-05, "loss": 0.9262, "step": 6694 }, { "epoch": 3.0321557971014492, "grad_norm": 0.06643643243434358, "learning_rate": 3.21715800836089e-05, "loss": 0.9177, "step": 6695 }, { "epoch": 3.032608695652174, "grad_norm": 0.05881435720078657, "learning_rate": 3.215917773904528e-05, "loss": 0.9284, "step": 6696 }, { "epoch": 3.0330615942028984, "grad_norm": 0.06598372591970862, "learning_rate": 3.214677617834105e-05, "loss": 0.9197, "step": 6697 }, { "epoch": 3.033514492753623, "grad_norm": 0.0860366608692671, "learning_rate": 3.2134375402736065e-05, "loss": 0.931, "step": 6698 }, { "epoch": 3.0339673913043477, "grad_norm": 0.07293493885418043, "learning_rate": 3.212197541347002e-05, "loss": 0.9197, "step": 6699 }, { "epoch": 3.0344202898550723, "grad_norm": 0.06358802630529929, "learning_rate": 3.210957621178259e-05, "loss": 0.9355, "step": 6700 }, { "epoch": 3.034873188405797, "grad_norm": 0.08454102959586463, "learning_rate": 3.209717779891332e-05, "loss": 0.9374, "step": 6701 }, { "epoch": 3.035326086956522, "grad_norm": 0.07794940884068398, "learning_rate": 3.208478017610173e-05, "loss": 0.927, "step": 6702 }, { "epoch": 3.0357789855072466, "grad_norm": 0.06256953234584406, "learning_rate": 3.207238334458718e-05, "loss": 0.9349, "step": 6703 }, { "epoch": 3.036231884057971, "grad_norm": 0.07526688560933414, "learning_rate": 3.205998730560905e-05, "loss": 0.9401, "step": 6704 }, { "epoch": 3.036684782608696, "grad_norm": 0.06289716937019203, "learning_rate": 3.204759206040657e-05, "loss": 0.9127, "step": 6705 }, { "epoch": 3.0371376811594204, "grad_norm": 0.08275669892018366, "learning_rate": 3.203519761021891e-05, "loss": 0.9167, "step": 6706 }, { "epoch": 3.037590579710145, "grad_norm": 0.0646289193081566, "learning_rate": 3.202280395628515e-05, "loss": 0.9206, "step": 6707 }, { "epoch": 3.0380434782608696, "grad_norm": 0.0708808795023347, "learning_rate": 3.2010411099844345e-05, "loss": 0.9276, "step": 6708 }, { "epoch": 3.0384963768115942, "grad_norm": 0.06716271512198166, "learning_rate": 3.199801904213539e-05, "loss": 0.9454, "step": 6709 }, { "epoch": 3.038949275362319, "grad_norm": 0.08196193771645731, "learning_rate": 3.198562778439715e-05, "loss": 0.9129, "step": 6710 }, { "epoch": 3.0394021739130435, "grad_norm": 0.05847944045933433, "learning_rate": 3.19732373278684e-05, "loss": 0.9156, "step": 6711 }, { "epoch": 3.039855072463768, "grad_norm": 0.08296222623019703, "learning_rate": 3.1960847673787844e-05, "loss": 0.9245, "step": 6712 }, { "epoch": 3.0403079710144927, "grad_norm": 0.08037544765165279, "learning_rate": 3.194845882339407e-05, "loss": 0.9224, "step": 6713 }, { "epoch": 3.0407608695652173, "grad_norm": 0.07973834408338218, "learning_rate": 3.193607077792565e-05, "loss": 0.9114, "step": 6714 }, { "epoch": 3.041213768115942, "grad_norm": 0.07018575783795047, "learning_rate": 3.192368353862098e-05, "loss": 0.9352, "step": 6715 }, { "epoch": 3.0416666666666665, "grad_norm": 0.0769025639364925, "learning_rate": 3.191129710671849e-05, "loss": 0.919, "step": 6716 }, { "epoch": 3.042119565217391, "grad_norm": 0.08516492897989368, "learning_rate": 3.189891148345644e-05, "loss": 0.9088, "step": 6717 }, { "epoch": 3.0425724637681157, "grad_norm": 0.06858720803286258, "learning_rate": 3.188652667007305e-05, "loss": 0.9215, "step": 6718 }, { "epoch": 3.043025362318841, "grad_norm": 0.08561460975087899, "learning_rate": 3.187414266780643e-05, "loss": 0.9267, "step": 6719 }, { "epoch": 3.0434782608695654, "grad_norm": 0.0638092726630195, "learning_rate": 3.1861759477894656e-05, "loss": 0.9516, "step": 6720 }, { "epoch": 3.04393115942029, "grad_norm": 0.07906071271099788, "learning_rate": 3.184937710157568e-05, "loss": 0.9151, "step": 6721 }, { "epoch": 3.0443840579710146, "grad_norm": 0.07351058334664115, "learning_rate": 3.18369955400874e-05, "loss": 0.9057, "step": 6722 }, { "epoch": 3.0448369565217392, "grad_norm": 0.07169501769052596, "learning_rate": 3.182461479466759e-05, "loss": 0.9335, "step": 6723 }, { "epoch": 3.045289855072464, "grad_norm": 0.08429725135256413, "learning_rate": 3.1812234866554014e-05, "loss": 0.9228, "step": 6724 }, { "epoch": 3.0457427536231885, "grad_norm": 0.09504400373906527, "learning_rate": 3.179985575698429e-05, "loss": 0.9346, "step": 6725 }, { "epoch": 3.046195652173913, "grad_norm": 0.08051427983539775, "learning_rate": 3.178747746719598e-05, "loss": 0.9454, "step": 6726 }, { "epoch": 3.0466485507246377, "grad_norm": 0.08454904202085756, "learning_rate": 3.1775099998426545e-05, "loss": 0.903, "step": 6727 }, { "epoch": 3.0471014492753623, "grad_norm": 0.07527514774466934, "learning_rate": 3.1762723351913414e-05, "loss": 0.9478, "step": 6728 }, { "epoch": 3.047554347826087, "grad_norm": 0.07187886470091359, "learning_rate": 3.175034752889388e-05, "loss": 0.9252, "step": 6729 }, { "epoch": 3.0480072463768115, "grad_norm": 0.07291600472095915, "learning_rate": 3.173797253060518e-05, "loss": 0.9263, "step": 6730 }, { "epoch": 3.048460144927536, "grad_norm": 0.09377273366426059, "learning_rate": 3.172559835828444e-05, "loss": 0.9066, "step": 6731 }, { "epoch": 3.0489130434782608, "grad_norm": 0.07369126662094246, "learning_rate": 3.171322501316876e-05, "loss": 0.9211, "step": 6732 }, { "epoch": 3.0493659420289854, "grad_norm": 0.07014416140015936, "learning_rate": 3.1700852496495106e-05, "loss": 0.9263, "step": 6733 }, { "epoch": 3.04981884057971, "grad_norm": 0.0765400478349008, "learning_rate": 3.168848080950038e-05, "loss": 0.9063, "step": 6734 }, { "epoch": 3.0502717391304346, "grad_norm": 0.06768320312366058, "learning_rate": 3.167610995342139e-05, "loss": 0.9348, "step": 6735 }, { "epoch": 3.050724637681159, "grad_norm": 0.09646542091816156, "learning_rate": 3.166373992949489e-05, "loss": 0.914, "step": 6736 }, { "epoch": 3.0511775362318843, "grad_norm": 0.07083840692464771, "learning_rate": 3.16513707389575e-05, "loss": 0.9228, "step": 6737 }, { "epoch": 3.051630434782609, "grad_norm": 0.09724590413645827, "learning_rate": 3.1639002383045834e-05, "loss": 0.9343, "step": 6738 }, { "epoch": 3.0520833333333335, "grad_norm": 0.10827609472372325, "learning_rate": 3.162663486299632e-05, "loss": 0.9251, "step": 6739 }, { "epoch": 3.052536231884058, "grad_norm": 0.062009534235191, "learning_rate": 3.161426818004541e-05, "loss": 0.9203, "step": 6740 }, { "epoch": 3.0529891304347827, "grad_norm": 0.08282676936164315, "learning_rate": 3.160190233542938e-05, "loss": 0.9133, "step": 6741 }, { "epoch": 3.0534420289855073, "grad_norm": 0.09535523745586764, "learning_rate": 3.15895373303845e-05, "loss": 0.9388, "step": 6742 }, { "epoch": 3.053894927536232, "grad_norm": 0.09118897680213031, "learning_rate": 3.157717316614689e-05, "loss": 0.9275, "step": 6743 }, { "epoch": 3.0543478260869565, "grad_norm": 0.07457740982895271, "learning_rate": 3.156480984395263e-05, "loss": 0.9318, "step": 6744 }, { "epoch": 3.054800724637681, "grad_norm": 0.11357848332727287, "learning_rate": 3.1552447365037686e-05, "loss": 0.909, "step": 6745 }, { "epoch": 3.0552536231884058, "grad_norm": 0.08888954471983572, "learning_rate": 3.1540085730637975e-05, "loss": 0.9217, "step": 6746 }, { "epoch": 3.0557065217391304, "grad_norm": 0.07809504546931173, "learning_rate": 3.15277249419893e-05, "loss": 0.9168, "step": 6747 }, { "epoch": 3.056159420289855, "grad_norm": 0.11746346832615895, "learning_rate": 3.151536500032739e-05, "loss": 0.9319, "step": 6748 }, { "epoch": 3.0566123188405796, "grad_norm": 0.09245493190036538, "learning_rate": 3.150300590688787e-05, "loss": 0.9509, "step": 6749 }, { "epoch": 3.057065217391304, "grad_norm": 0.08524551002046474, "learning_rate": 3.1490647662906336e-05, "loss": 0.9395, "step": 6750 }, { "epoch": 3.057518115942029, "grad_norm": 0.10211756434398346, "learning_rate": 3.147829026961822e-05, "loss": 0.9216, "step": 6751 }, { "epoch": 3.0579710144927534, "grad_norm": 0.11003034190532565, "learning_rate": 3.146593372825895e-05, "loss": 0.9303, "step": 6752 }, { "epoch": 3.058423913043478, "grad_norm": 0.07425926028354608, "learning_rate": 3.145357804006379e-05, "loss": 0.9294, "step": 6753 }, { "epoch": 3.058876811594203, "grad_norm": 0.09996795876642028, "learning_rate": 3.144122320626799e-05, "loss": 0.9054, "step": 6754 }, { "epoch": 3.0593297101449277, "grad_norm": 0.10692306873339674, "learning_rate": 3.142886922810666e-05, "loss": 0.912, "step": 6755 }, { "epoch": 3.0597826086956523, "grad_norm": 0.06897472515443086, "learning_rate": 3.141651610681486e-05, "loss": 0.9189, "step": 6756 }, { "epoch": 3.060235507246377, "grad_norm": 0.07873626469515227, "learning_rate": 3.140416384362754e-05, "loss": 0.9336, "step": 6757 }, { "epoch": 3.0606884057971016, "grad_norm": 0.0846868171142201, "learning_rate": 3.1391812439779594e-05, "loss": 0.9417, "step": 6758 }, { "epoch": 3.061141304347826, "grad_norm": 0.06313875194092473, "learning_rate": 3.1379461896505784e-05, "loss": 0.9383, "step": 6759 }, { "epoch": 3.0615942028985508, "grad_norm": 0.07501683063839314, "learning_rate": 3.136711221504085e-05, "loss": 0.9153, "step": 6760 }, { "epoch": 3.0620471014492754, "grad_norm": 0.07058598183701863, "learning_rate": 3.135476339661936e-05, "loss": 0.9227, "step": 6761 }, { "epoch": 3.0625, "grad_norm": 0.0652894185345416, "learning_rate": 3.134241544247589e-05, "loss": 0.9377, "step": 6762 }, { "epoch": 3.0629528985507246, "grad_norm": 0.07404192274377468, "learning_rate": 3.133006835384486e-05, "loss": 0.9427, "step": 6763 }, { "epoch": 3.0634057971014492, "grad_norm": 0.07628919254665979, "learning_rate": 3.1317722131960645e-05, "loss": 0.9361, "step": 6764 }, { "epoch": 3.063858695652174, "grad_norm": 0.06433255123010458, "learning_rate": 3.13053767780575e-05, "loss": 0.9339, "step": 6765 }, { "epoch": 3.0643115942028984, "grad_norm": 0.08963490715687124, "learning_rate": 3.129303229336962e-05, "loss": 0.9508, "step": 6766 }, { "epoch": 3.064764492753623, "grad_norm": 0.07259658713236576, "learning_rate": 3.1280688679131096e-05, "loss": 0.934, "step": 6767 }, { "epoch": 3.0652173913043477, "grad_norm": 0.0807444309543707, "learning_rate": 3.126834593657595e-05, "loss": 0.9176, "step": 6768 }, { "epoch": 3.0656702898550723, "grad_norm": 0.09695066347724748, "learning_rate": 3.125600406693807e-05, "loss": 0.9318, "step": 6769 }, { "epoch": 3.066123188405797, "grad_norm": 0.0707285191995139, "learning_rate": 3.124366307145135e-05, "loss": 0.9102, "step": 6770 }, { "epoch": 3.066576086956522, "grad_norm": 0.07290957094088374, "learning_rate": 3.1231322951349494e-05, "loss": 0.9215, "step": 6771 }, { "epoch": 3.0670289855072466, "grad_norm": 0.0927167067768593, "learning_rate": 3.121898370786619e-05, "loss": 0.9498, "step": 6772 }, { "epoch": 3.067481884057971, "grad_norm": 0.08071581383910813, "learning_rate": 3.120664534223498e-05, "loss": 0.9336, "step": 6773 }, { "epoch": 3.067934782608696, "grad_norm": 0.07365883522938631, "learning_rate": 3.1194307855689386e-05, "loss": 0.9057, "step": 6774 }, { "epoch": 3.0683876811594204, "grad_norm": 0.09121309940277507, "learning_rate": 3.118197124946278e-05, "loss": 0.917, "step": 6775 }, { "epoch": 3.068840579710145, "grad_norm": 0.08332102106382913, "learning_rate": 3.116963552478849e-05, "loss": 0.9554, "step": 6776 }, { "epoch": 3.0692934782608696, "grad_norm": 0.06128732592003105, "learning_rate": 3.1157300682899715e-05, "loss": 0.911, "step": 6777 }, { "epoch": 3.0697463768115942, "grad_norm": 0.07118266179522618, "learning_rate": 3.1144966725029616e-05, "loss": 0.9186, "step": 6778 }, { "epoch": 3.070199275362319, "grad_norm": 0.06845848240982383, "learning_rate": 3.113263365241122e-05, "loss": 0.9243, "step": 6779 }, { "epoch": 3.0706521739130435, "grad_norm": 0.06662911216626524, "learning_rate": 3.11203014662775e-05, "loss": 0.9489, "step": 6780 }, { "epoch": 3.071105072463768, "grad_norm": 0.09173207836615237, "learning_rate": 3.110797016786129e-05, "loss": 0.9196, "step": 6781 }, { "epoch": 3.0715579710144927, "grad_norm": 0.09114964795272618, "learning_rate": 3.1095639758395416e-05, "loss": 0.9263, "step": 6782 }, { "epoch": 3.0720108695652173, "grad_norm": 0.07226054717297256, "learning_rate": 3.1083310239112524e-05, "loss": 0.932, "step": 6783 }, { "epoch": 3.072463768115942, "grad_norm": 0.0935326943881245, "learning_rate": 3.107098161124526e-05, "loss": 0.9207, "step": 6784 }, { "epoch": 3.0729166666666665, "grad_norm": 0.06965926677617841, "learning_rate": 3.1058653876026105e-05, "loss": 0.9266, "step": 6785 }, { "epoch": 3.073369565217391, "grad_norm": 0.09482795731525684, "learning_rate": 3.104632703468749e-05, "loss": 0.9317, "step": 6786 }, { "epoch": 3.0738224637681157, "grad_norm": 0.07772784162973047, "learning_rate": 3.103400108846174e-05, "loss": 0.9054, "step": 6787 }, { "epoch": 3.074275362318841, "grad_norm": 0.07335673228290493, "learning_rate": 3.102167603858113e-05, "loss": 0.9307, "step": 6788 }, { "epoch": 3.0747282608695654, "grad_norm": 0.0757640369263076, "learning_rate": 3.1009351886277774e-05, "loss": 0.9096, "step": 6789 }, { "epoch": 3.07518115942029, "grad_norm": 0.0846045431437604, "learning_rate": 3.0997028632783774e-05, "loss": 0.9337, "step": 6790 }, { "epoch": 3.0756340579710146, "grad_norm": 0.08098994707360896, "learning_rate": 3.098470627933107e-05, "loss": 0.9163, "step": 6791 }, { "epoch": 3.0760869565217392, "grad_norm": 0.07276138912127499, "learning_rate": 3.097238482715158e-05, "loss": 0.9138, "step": 6792 }, { "epoch": 3.076539855072464, "grad_norm": 0.0796161012052949, "learning_rate": 3.0960064277477074e-05, "loss": 0.9223, "step": 6793 }, { "epoch": 3.0769927536231885, "grad_norm": 0.06894962160374991, "learning_rate": 3.094774463153928e-05, "loss": 0.9322, "step": 6794 }, { "epoch": 3.077445652173913, "grad_norm": 0.0864215351626429, "learning_rate": 3.093542589056978e-05, "loss": 0.9198, "step": 6795 }, { "epoch": 3.0778985507246377, "grad_norm": 0.07500790264729441, "learning_rate": 3.0923108055800136e-05, "loss": 0.9383, "step": 6796 }, { "epoch": 3.0783514492753623, "grad_norm": 0.07859728715437105, "learning_rate": 3.091079112846175e-05, "loss": 0.9395, "step": 6797 }, { "epoch": 3.078804347826087, "grad_norm": 0.079339385450752, "learning_rate": 3.0898475109785994e-05, "loss": 0.9283, "step": 6798 }, { "epoch": 3.0792572463768115, "grad_norm": 0.0775168343707445, "learning_rate": 3.0886160001004086e-05, "loss": 0.9154, "step": 6799 }, { "epoch": 3.079710144927536, "grad_norm": 0.08015923677236626, "learning_rate": 3.087384580334722e-05, "loss": 0.9335, "step": 6800 }, { "epoch": 3.0801630434782608, "grad_norm": 0.07976924093477757, "learning_rate": 3.086153251804644e-05, "loss": 0.9206, "step": 6801 }, { "epoch": 3.0806159420289854, "grad_norm": 0.0664554479593683, "learning_rate": 3.0849220146332745e-05, "loss": 0.9306, "step": 6802 }, { "epoch": 3.08106884057971, "grad_norm": 0.0708275458625479, "learning_rate": 3.083690868943699e-05, "loss": 0.9213, "step": 6803 }, { "epoch": 3.0815217391304346, "grad_norm": 0.06961131639386393, "learning_rate": 3.082459814859002e-05, "loss": 0.9053, "step": 6804 }, { "epoch": 3.081974637681159, "grad_norm": 0.07190000523961666, "learning_rate": 3.08122885250225e-05, "loss": 0.9251, "step": 6805 }, { "epoch": 3.0824275362318843, "grad_norm": 0.0626168661856088, "learning_rate": 3.0799979819965067e-05, "loss": 0.9233, "step": 6806 }, { "epoch": 3.082880434782609, "grad_norm": 0.07823254900460966, "learning_rate": 3.0787672034648215e-05, "loss": 0.9264, "step": 6807 }, { "epoch": 3.0833333333333335, "grad_norm": 0.06280048239153338, "learning_rate": 3.07753651703024e-05, "loss": 0.9381, "step": 6808 }, { "epoch": 3.083786231884058, "grad_norm": 0.07549390868303138, "learning_rate": 3.076305922815793e-05, "loss": 0.9156, "step": 6809 }, { "epoch": 3.0842391304347827, "grad_norm": 0.07229499333006927, "learning_rate": 3.075075420944509e-05, "loss": 0.9347, "step": 6810 }, { "epoch": 3.0846920289855073, "grad_norm": 0.07414670655806756, "learning_rate": 3.073845011539398e-05, "loss": 0.9098, "step": 6811 }, { "epoch": 3.085144927536232, "grad_norm": 0.07062954366793166, "learning_rate": 3.072614694723472e-05, "loss": 0.9344, "step": 6812 }, { "epoch": 3.0855978260869565, "grad_norm": 0.07600426768850131, "learning_rate": 3.0713844706197224e-05, "loss": 0.9291, "step": 6813 }, { "epoch": 3.086050724637681, "grad_norm": 0.06668904949734705, "learning_rate": 3.07015433935114e-05, "loss": 0.931, "step": 6814 }, { "epoch": 3.0865036231884058, "grad_norm": 0.07929646042570569, "learning_rate": 3.0689243010406994e-05, "loss": 0.9165, "step": 6815 }, { "epoch": 3.0869565217391304, "grad_norm": 0.06984204644727644, "learning_rate": 3.0676943558113745e-05, "loss": 0.9068, "step": 6816 }, { "epoch": 3.087409420289855, "grad_norm": 0.09229866990149195, "learning_rate": 3.066464503786121e-05, "loss": 0.9229, "step": 6817 }, { "epoch": 3.0878623188405796, "grad_norm": 0.07458001732186108, "learning_rate": 3.065234745087891e-05, "loss": 0.9003, "step": 6818 }, { "epoch": 3.088315217391304, "grad_norm": 0.07519959622034783, "learning_rate": 3.064005079839625e-05, "loss": 0.9143, "step": 6819 }, { "epoch": 3.088768115942029, "grad_norm": 0.07747996905890817, "learning_rate": 3.0627755081642526e-05, "loss": 0.9246, "step": 6820 }, { "epoch": 3.0892210144927534, "grad_norm": 0.08165366754606292, "learning_rate": 3.0615460301846997e-05, "loss": 0.9456, "step": 6821 }, { "epoch": 3.089673913043478, "grad_norm": 0.08143519358203291, "learning_rate": 3.060316646023876e-05, "loss": 0.9325, "step": 6822 }, { "epoch": 3.090126811594203, "grad_norm": 0.0874904439082527, "learning_rate": 3.059087355804688e-05, "loss": 0.93, "step": 6823 }, { "epoch": 3.0905797101449277, "grad_norm": 0.09361287877274328, "learning_rate": 3.057858159650025e-05, "loss": 0.9359, "step": 6824 }, { "epoch": 3.0910326086956523, "grad_norm": 0.07009823023802815, "learning_rate": 3.0566290576827775e-05, "loss": 0.9187, "step": 6825 }, { "epoch": 3.091485507246377, "grad_norm": 0.07566516368430508, "learning_rate": 3.0554000500258166e-05, "loss": 0.9083, "step": 6826 }, { "epoch": 3.0919384057971016, "grad_norm": 0.08470859855313165, "learning_rate": 3.054171136802011e-05, "loss": 0.9299, "step": 6827 }, { "epoch": 3.092391304347826, "grad_norm": 0.07262898977132004, "learning_rate": 3.052942318134213e-05, "loss": 0.9174, "step": 6828 }, { "epoch": 3.0928442028985508, "grad_norm": 0.0733997528863634, "learning_rate": 3.051713594145275e-05, "loss": 0.911, "step": 6829 }, { "epoch": 3.0932971014492754, "grad_norm": 0.0840383221821758, "learning_rate": 3.050484964958029e-05, "loss": 0.9455, "step": 6830 }, { "epoch": 3.09375, "grad_norm": 0.09167357786944794, "learning_rate": 3.0492564306953083e-05, "loss": 0.9075, "step": 6831 }, { "epoch": 3.0942028985507246, "grad_norm": 0.08539502682104917, "learning_rate": 3.0480279914799276e-05, "loss": 0.9212, "step": 6832 }, { "epoch": 3.0946557971014492, "grad_norm": 0.09167675245814248, "learning_rate": 3.0467996474346978e-05, "loss": 0.9266, "step": 6833 }, { "epoch": 3.095108695652174, "grad_norm": 0.0874279903261356, "learning_rate": 3.045571398682416e-05, "loss": 0.9452, "step": 6834 }, { "epoch": 3.0955615942028984, "grad_norm": 0.07169651220834838, "learning_rate": 3.0443432453458758e-05, "loss": 0.9195, "step": 6835 }, { "epoch": 3.096014492753623, "grad_norm": 0.08509713834728837, "learning_rate": 3.043115187547855e-05, "loss": 0.8944, "step": 6836 }, { "epoch": 3.0964673913043477, "grad_norm": 0.0684334963602767, "learning_rate": 3.0418872254111254e-05, "loss": 0.9064, "step": 6837 }, { "epoch": 3.0969202898550723, "grad_norm": 0.07687182607785108, "learning_rate": 3.040659359058447e-05, "loss": 0.9368, "step": 6838 }, { "epoch": 3.097373188405797, "grad_norm": 0.06669538555898162, "learning_rate": 3.0394315886125746e-05, "loss": 0.9282, "step": 6839 }, { "epoch": 3.097826086956522, "grad_norm": 0.06248213832888063, "learning_rate": 3.0382039141962473e-05, "loss": 0.9382, "step": 6840 }, { "epoch": 3.0982789855072466, "grad_norm": 0.06072513242803257, "learning_rate": 3.0369763359321992e-05, "loss": 0.918, "step": 6841 }, { "epoch": 3.098731884057971, "grad_norm": 0.08576780691471889, "learning_rate": 3.035748853943152e-05, "loss": 0.9295, "step": 6842 }, { "epoch": 3.099184782608696, "grad_norm": 0.06101309669480446, "learning_rate": 3.0345214683518204e-05, "loss": 0.9097, "step": 6843 }, { "epoch": 3.0996376811594204, "grad_norm": 0.07653400524803791, "learning_rate": 3.0332941792809077e-05, "loss": 0.9394, "step": 6844 }, { "epoch": 3.100090579710145, "grad_norm": 0.08189084312105306, "learning_rate": 3.032066986853108e-05, "loss": 0.9192, "step": 6845 }, { "epoch": 3.1005434782608696, "grad_norm": 0.06639228131550243, "learning_rate": 3.0308398911911047e-05, "loss": 0.9222, "step": 6846 }, { "epoch": 3.1009963768115942, "grad_norm": 0.07959387707652339, "learning_rate": 3.029612892417574e-05, "loss": 0.9275, "step": 6847 }, { "epoch": 3.101449275362319, "grad_norm": 0.073187956103968, "learning_rate": 3.0283859906551795e-05, "loss": 0.9647, "step": 6848 }, { "epoch": 3.1019021739130435, "grad_norm": 0.06746233451903565, "learning_rate": 3.0271591860265783e-05, "loss": 0.9173, "step": 6849 }, { "epoch": 3.102355072463768, "grad_norm": 0.08382804046799058, "learning_rate": 3.0259324786544128e-05, "loss": 0.9182, "step": 6850 }, { "epoch": 3.1028079710144927, "grad_norm": 0.06999727121983573, "learning_rate": 3.0247058686613227e-05, "loss": 0.9253, "step": 6851 }, { "epoch": 3.1032608695652173, "grad_norm": 0.07118006431374688, "learning_rate": 3.023479356169932e-05, "loss": 0.9054, "step": 6852 }, { "epoch": 3.103713768115942, "grad_norm": 0.08148727776888447, "learning_rate": 3.022252941302858e-05, "loss": 0.9192, "step": 6853 }, { "epoch": 3.1041666666666665, "grad_norm": 0.07618039371378804, "learning_rate": 3.0210266241827047e-05, "loss": 0.9238, "step": 6854 }, { "epoch": 3.104619565217391, "grad_norm": 0.0687410839884647, "learning_rate": 3.0198004049320735e-05, "loss": 0.9011, "step": 6855 }, { "epoch": 3.1050724637681157, "grad_norm": 0.07900813221915029, "learning_rate": 3.0185742836735483e-05, "loss": 0.9279, "step": 6856 }, { "epoch": 3.105525362318841, "grad_norm": 0.08388704129487289, "learning_rate": 3.0173482605297075e-05, "loss": 0.9277, "step": 6857 }, { "epoch": 3.1059782608695654, "grad_norm": 0.07503781773069063, "learning_rate": 3.016122335623117e-05, "loss": 0.8903, "step": 6858 }, { "epoch": 3.10643115942029, "grad_norm": 0.09521351051447109, "learning_rate": 3.014896509076337e-05, "loss": 0.9228, "step": 6859 }, { "epoch": 3.1068840579710146, "grad_norm": 0.1096444552049261, "learning_rate": 3.013670781011913e-05, "loss": 0.9253, "step": 6860 }, { "epoch": 3.1073369565217392, "grad_norm": 0.07427873024358388, "learning_rate": 3.0124451515523857e-05, "loss": 0.9373, "step": 6861 }, { "epoch": 3.107789855072464, "grad_norm": 0.09830261390366908, "learning_rate": 3.0112196208202787e-05, "loss": 0.924, "step": 6862 }, { "epoch": 3.1082427536231885, "grad_norm": 0.11586711587589281, "learning_rate": 3.0099941889381154e-05, "loss": 0.9214, "step": 6863 }, { "epoch": 3.108695652173913, "grad_norm": 0.06837971222169678, "learning_rate": 3.008768856028401e-05, "loss": 0.9115, "step": 6864 }, { "epoch": 3.1091485507246377, "grad_norm": 0.1042535510914859, "learning_rate": 3.0075436222136357e-05, "loss": 0.9242, "step": 6865 }, { "epoch": 3.1096014492753623, "grad_norm": 0.09814537854047822, "learning_rate": 3.006318487616305e-05, "loss": 0.9286, "step": 6866 }, { "epoch": 3.110054347826087, "grad_norm": 0.07215847802402535, "learning_rate": 3.005093452358892e-05, "loss": 0.9244, "step": 6867 }, { "epoch": 3.1105072463768115, "grad_norm": 0.08428866324334952, "learning_rate": 3.0038685165638617e-05, "loss": 0.919, "step": 6868 }, { "epoch": 3.110960144927536, "grad_norm": 0.07711854038009082, "learning_rate": 3.0026436803536758e-05, "loss": 0.924, "step": 6869 }, { "epoch": 3.1114130434782608, "grad_norm": 0.07324308259077117, "learning_rate": 3.0014189438507793e-05, "loss": 0.9129, "step": 6870 }, { "epoch": 3.1118659420289854, "grad_norm": 0.0851536744754265, "learning_rate": 3.0001943071776155e-05, "loss": 0.9298, "step": 6871 }, { "epoch": 3.11231884057971, "grad_norm": 0.0855906077153072, "learning_rate": 2.9989697704566095e-05, "loss": 0.9319, "step": 6872 }, { "epoch": 3.1127717391304346, "grad_norm": 0.06185089838637693, "learning_rate": 2.9977453338101834e-05, "loss": 0.9237, "step": 6873 }, { "epoch": 3.113224637681159, "grad_norm": 0.09217568349357469, "learning_rate": 2.9965209973607436e-05, "loss": 0.921, "step": 6874 }, { "epoch": 3.1136775362318843, "grad_norm": 0.06598727121608612, "learning_rate": 2.9952967612306906e-05, "loss": 0.9339, "step": 6875 }, { "epoch": 3.114130434782609, "grad_norm": 0.07739535857955669, "learning_rate": 2.9940726255424107e-05, "loss": 0.9323, "step": 6876 }, { "epoch": 3.1145833333333335, "grad_norm": 0.08576404949077819, "learning_rate": 2.9928485904182865e-05, "loss": 0.9155, "step": 6877 }, { "epoch": 3.115036231884058, "grad_norm": 0.0694570118278932, "learning_rate": 2.9916246559806832e-05, "loss": 0.9524, "step": 6878 }, { "epoch": 3.1154891304347827, "grad_norm": 0.08543015308895426, "learning_rate": 2.9904008223519624e-05, "loss": 0.933, "step": 6879 }, { "epoch": 3.1159420289855073, "grad_norm": 0.09018420804504787, "learning_rate": 2.9891770896544695e-05, "loss": 0.9205, "step": 6880 }, { "epoch": 3.116394927536232, "grad_norm": 0.08125490032037697, "learning_rate": 2.9879534580105463e-05, "loss": 0.9283, "step": 6881 }, { "epoch": 3.1168478260869565, "grad_norm": 0.06898176822221827, "learning_rate": 2.9867299275425185e-05, "loss": 0.9358, "step": 6882 }, { "epoch": 3.117300724637681, "grad_norm": 0.08869615727327178, "learning_rate": 2.985506498372707e-05, "loss": 0.916, "step": 6883 }, { "epoch": 3.1177536231884058, "grad_norm": 0.06719753213484776, "learning_rate": 2.984283170623416e-05, "loss": 0.9043, "step": 6884 }, { "epoch": 3.1182065217391304, "grad_norm": 0.07567951889195594, "learning_rate": 2.9830599444169484e-05, "loss": 0.9158, "step": 6885 }, { "epoch": 3.118659420289855, "grad_norm": 0.08231687999922754, "learning_rate": 2.9818368198755886e-05, "loss": 0.917, "step": 6886 }, { "epoch": 3.1191123188405796, "grad_norm": 0.07414191887626252, "learning_rate": 2.9806137971216157e-05, "loss": 0.9251, "step": 6887 }, { "epoch": 3.119565217391304, "grad_norm": 0.06483129060440034, "learning_rate": 2.9793908762772954e-05, "loss": 0.9253, "step": 6888 }, { "epoch": 3.120018115942029, "grad_norm": 0.07495502803747475, "learning_rate": 2.9781680574648885e-05, "loss": 0.9206, "step": 6889 }, { "epoch": 3.1204710144927534, "grad_norm": 0.06976956025930364, "learning_rate": 2.9769453408066388e-05, "loss": 0.9256, "step": 6890 }, { "epoch": 3.120923913043478, "grad_norm": 0.07239685400776792, "learning_rate": 2.9757227264247856e-05, "loss": 0.9408, "step": 6891 }, { "epoch": 3.121376811594203, "grad_norm": 0.057964713272657495, "learning_rate": 2.9745002144415526e-05, "loss": 0.9312, "step": 6892 }, { "epoch": 3.1218297101449277, "grad_norm": 0.06816309931490588, "learning_rate": 2.9732778049791603e-05, "loss": 0.9366, "step": 6893 }, { "epoch": 3.1222826086956523, "grad_norm": 0.06534047004369903, "learning_rate": 2.9720554981598123e-05, "loss": 0.9402, "step": 6894 }, { "epoch": 3.122735507246377, "grad_norm": 0.07148896620211495, "learning_rate": 2.9708332941057053e-05, "loss": 0.9373, "step": 6895 }, { "epoch": 3.1231884057971016, "grad_norm": 0.06831936799281862, "learning_rate": 2.9696111929390234e-05, "loss": 0.9237, "step": 6896 }, { "epoch": 3.123641304347826, "grad_norm": 0.07219199581849062, "learning_rate": 2.9683891947819447e-05, "loss": 0.9402, "step": 6897 }, { "epoch": 3.1240942028985508, "grad_norm": 0.0687354441265843, "learning_rate": 2.967167299756632e-05, "loss": 0.9387, "step": 6898 }, { "epoch": 3.1245471014492754, "grad_norm": 0.06697459317276745, "learning_rate": 2.965945507985242e-05, "loss": 0.9473, "step": 6899 }, { "epoch": 3.125, "grad_norm": 0.07363117332291774, "learning_rate": 2.9647238195899168e-05, "loss": 0.9008, "step": 6900 }, { "epoch": 3.1254528985507246, "grad_norm": 0.07051770853186969, "learning_rate": 2.9635022346927934e-05, "loss": 0.9287, "step": 6901 }, { "epoch": 3.1259057971014492, "grad_norm": 0.0792222778407912, "learning_rate": 2.962280753415993e-05, "loss": 0.9113, "step": 6902 }, { "epoch": 3.126358695652174, "grad_norm": 0.07418026103881842, "learning_rate": 2.9610593758816314e-05, "loss": 0.9293, "step": 6903 }, { "epoch": 3.1268115942028984, "grad_norm": 0.07153533520770654, "learning_rate": 2.9598381022118086e-05, "loss": 0.919, "step": 6904 }, { "epoch": 3.127264492753623, "grad_norm": 0.0817673620653662, "learning_rate": 2.95861693252862e-05, "loss": 0.9175, "step": 6905 }, { "epoch": 3.1277173913043477, "grad_norm": 0.07296592063942131, "learning_rate": 2.9573958669541466e-05, "loss": 0.9468, "step": 6906 }, { "epoch": 3.1281702898550723, "grad_norm": 0.07545595403405783, "learning_rate": 2.9561749056104612e-05, "loss": 0.9363, "step": 6907 }, { "epoch": 3.128623188405797, "grad_norm": 0.09484616237646717, "learning_rate": 2.9549540486196232e-05, "loss": 0.9217, "step": 6908 }, { "epoch": 3.1290760869565215, "grad_norm": 0.07566960233609464, "learning_rate": 2.9537332961036863e-05, "loss": 0.9256, "step": 6909 }, { "epoch": 3.1295289855072466, "grad_norm": 0.0677836956912085, "learning_rate": 2.952512648184689e-05, "loss": 0.9363, "step": 6910 }, { "epoch": 3.129981884057971, "grad_norm": 0.07582066359010285, "learning_rate": 2.9512921049846626e-05, "loss": 0.9257, "step": 6911 }, { "epoch": 3.130434782608696, "grad_norm": 0.0741484607274078, "learning_rate": 2.950071666625625e-05, "loss": 0.9264, "step": 6912 }, { "epoch": 3.1308876811594204, "grad_norm": 0.07660827228144786, "learning_rate": 2.9488513332295883e-05, "loss": 0.9148, "step": 6913 }, { "epoch": 3.131340579710145, "grad_norm": 0.06664234435959082, "learning_rate": 2.9476311049185483e-05, "loss": 0.8904, "step": 6914 }, { "epoch": 3.1317934782608696, "grad_norm": 0.08112842674749984, "learning_rate": 2.946410981814496e-05, "loss": 0.9249, "step": 6915 }, { "epoch": 3.1322463768115942, "grad_norm": 0.07179485202676501, "learning_rate": 2.9451909640394052e-05, "loss": 0.946, "step": 6916 }, { "epoch": 3.132699275362319, "grad_norm": 0.06965351389182288, "learning_rate": 2.943971051715247e-05, "loss": 0.9366, "step": 6917 }, { "epoch": 3.1331521739130435, "grad_norm": 0.06870165549576708, "learning_rate": 2.9427512449639744e-05, "loss": 0.9434, "step": 6918 }, { "epoch": 3.133605072463768, "grad_norm": 0.06867511562264833, "learning_rate": 2.9415315439075375e-05, "loss": 0.908, "step": 6919 }, { "epoch": 3.1340579710144927, "grad_norm": 0.07811184682570904, "learning_rate": 2.9403119486678675e-05, "loss": 0.934, "step": 6920 }, { "epoch": 3.1345108695652173, "grad_norm": 0.06438107049875341, "learning_rate": 2.939092459366893e-05, "loss": 0.9112, "step": 6921 }, { "epoch": 3.134963768115942, "grad_norm": 0.087870443878756, "learning_rate": 2.9378730761265243e-05, "loss": 0.9329, "step": 6922 }, { "epoch": 3.1354166666666665, "grad_norm": 0.07340650954860124, "learning_rate": 2.936653799068669e-05, "loss": 0.9456, "step": 6923 }, { "epoch": 3.135869565217391, "grad_norm": 0.07841612769050846, "learning_rate": 2.9354346283152178e-05, "loss": 0.9375, "step": 6924 }, { "epoch": 3.1363224637681157, "grad_norm": 0.07683026467422541, "learning_rate": 2.9342155639880542e-05, "loss": 0.9195, "step": 6925 }, { "epoch": 3.136775362318841, "grad_norm": 0.07521940187811174, "learning_rate": 2.9329966062090482e-05, "loss": 0.9172, "step": 6926 }, { "epoch": 3.1372282608695654, "grad_norm": 0.06536993286237566, "learning_rate": 2.931777755100064e-05, "loss": 0.9233, "step": 6927 }, { "epoch": 3.13768115942029, "grad_norm": 0.0737121495019695, "learning_rate": 2.9305590107829495e-05, "loss": 0.9352, "step": 6928 }, { "epoch": 3.1381340579710146, "grad_norm": 0.07045638416240968, "learning_rate": 2.9293403733795462e-05, "loss": 0.9339, "step": 6929 }, { "epoch": 3.1385869565217392, "grad_norm": 0.07665006099170808, "learning_rate": 2.92812184301168e-05, "loss": 0.8944, "step": 6930 }, { "epoch": 3.139039855072464, "grad_norm": 0.0773280182372667, "learning_rate": 2.9269034198011743e-05, "loss": 0.9046, "step": 6931 }, { "epoch": 3.1394927536231885, "grad_norm": 0.08045145949064988, "learning_rate": 2.925685103869833e-05, "loss": 0.917, "step": 6932 }, { "epoch": 3.139945652173913, "grad_norm": 0.06358957492228923, "learning_rate": 2.924466895339455e-05, "loss": 0.9267, "step": 6933 }, { "epoch": 3.1403985507246377, "grad_norm": 0.06964325223588622, "learning_rate": 2.9232487943318245e-05, "loss": 0.9014, "step": 6934 }, { "epoch": 3.1408514492753623, "grad_norm": 0.06340676075409438, "learning_rate": 2.9220308009687195e-05, "loss": 0.9097, "step": 6935 }, { "epoch": 3.141304347826087, "grad_norm": 0.07604082742340235, "learning_rate": 2.9208129153719026e-05, "loss": 0.918, "step": 6936 }, { "epoch": 3.1417572463768115, "grad_norm": 0.06901332482254527, "learning_rate": 2.91959513766313e-05, "loss": 0.9189, "step": 6937 }, { "epoch": 3.142210144927536, "grad_norm": 0.07256671606803523, "learning_rate": 2.9183774679641418e-05, "loss": 0.9196, "step": 6938 }, { "epoch": 3.1426630434782608, "grad_norm": 0.07531295005034618, "learning_rate": 2.9171599063966737e-05, "loss": 0.9296, "step": 6939 }, { "epoch": 3.1431159420289854, "grad_norm": 0.07897240228741449, "learning_rate": 2.9159424530824448e-05, "loss": 0.9127, "step": 6940 }, { "epoch": 3.14356884057971, "grad_norm": 0.07196214105965858, "learning_rate": 2.9147251081431675e-05, "loss": 0.9391, "step": 6941 }, { "epoch": 3.1440217391304346, "grad_norm": 0.06749346188207654, "learning_rate": 2.913507871700541e-05, "loss": 0.9454, "step": 6942 }, { "epoch": 3.144474637681159, "grad_norm": 0.08009751458752219, "learning_rate": 2.9122907438762526e-05, "loss": 0.911, "step": 6943 }, { "epoch": 3.1449275362318843, "grad_norm": 0.061522398541281886, "learning_rate": 2.9110737247919835e-05, "loss": 0.9387, "step": 6944 }, { "epoch": 3.145380434782609, "grad_norm": 0.06666093562314163, "learning_rate": 2.9098568145693992e-05, "loss": 0.9221, "step": 6945 }, { "epoch": 3.1458333333333335, "grad_norm": 0.06582545991504207, "learning_rate": 2.9086400133301573e-05, "loss": 0.926, "step": 6946 }, { "epoch": 3.146286231884058, "grad_norm": 0.06800163309688424, "learning_rate": 2.9074233211959008e-05, "loss": 0.9256, "step": 6947 }, { "epoch": 3.1467391304347827, "grad_norm": 0.06877119857149888, "learning_rate": 2.906206738288268e-05, "loss": 0.9157, "step": 6948 }, { "epoch": 3.1471920289855073, "grad_norm": 0.0681252118376829, "learning_rate": 2.9049902647288805e-05, "loss": 0.8992, "step": 6949 }, { "epoch": 3.147644927536232, "grad_norm": 0.07316058513928551, "learning_rate": 2.9037739006393518e-05, "loss": 0.9234, "step": 6950 }, { "epoch": 3.1480978260869565, "grad_norm": 0.0799057810411213, "learning_rate": 2.902557646141282e-05, "loss": 0.9231, "step": 6951 }, { "epoch": 3.148550724637681, "grad_norm": 0.073226502561899, "learning_rate": 2.9013415013562648e-05, "loss": 0.9354, "step": 6952 }, { "epoch": 3.1490036231884058, "grad_norm": 0.07665069748546037, "learning_rate": 2.9001254664058778e-05, "loss": 0.92, "step": 6953 }, { "epoch": 3.1494565217391304, "grad_norm": 0.07187798736984949, "learning_rate": 2.8989095414116918e-05, "loss": 0.9558, "step": 6954 }, { "epoch": 3.149909420289855, "grad_norm": 0.07552465848498036, "learning_rate": 2.897693726495262e-05, "loss": 0.9333, "step": 6955 }, { "epoch": 3.1503623188405796, "grad_norm": 0.07638117630012922, "learning_rate": 2.8964780217781386e-05, "loss": 0.9345, "step": 6956 }, { "epoch": 3.150815217391304, "grad_norm": 0.07491132216944586, "learning_rate": 2.8952624273818556e-05, "loss": 0.9086, "step": 6957 }, { "epoch": 3.151268115942029, "grad_norm": 0.07082392003693444, "learning_rate": 2.894046943427939e-05, "loss": 0.9251, "step": 6958 }, { "epoch": 3.1517210144927534, "grad_norm": 0.0669578384134138, "learning_rate": 2.8928315700379008e-05, "loss": 0.9165, "step": 6959 }, { "epoch": 3.1521739130434785, "grad_norm": 0.08231248040900188, "learning_rate": 2.8916163073332462e-05, "loss": 0.9356, "step": 6960 }, { "epoch": 3.152626811594203, "grad_norm": 0.07159482079374767, "learning_rate": 2.8904011554354637e-05, "loss": 0.9201, "step": 6961 }, { "epoch": 3.1530797101449277, "grad_norm": 0.07526451406338495, "learning_rate": 2.8891861144660383e-05, "loss": 0.9253, "step": 6962 }, { "epoch": 3.1535326086956523, "grad_norm": 0.08364345583982587, "learning_rate": 2.8879711845464365e-05, "loss": 0.9531, "step": 6963 }, { "epoch": 3.153985507246377, "grad_norm": 0.09014320682449081, "learning_rate": 2.8867563657981173e-05, "loss": 0.9354, "step": 6964 }, { "epoch": 3.1544384057971016, "grad_norm": 0.08254341557371632, "learning_rate": 2.8855416583425278e-05, "loss": 0.9174, "step": 6965 }, { "epoch": 3.154891304347826, "grad_norm": 0.07300274263125102, "learning_rate": 2.8843270623011056e-05, "loss": 0.9323, "step": 6966 }, { "epoch": 3.1553442028985508, "grad_norm": 0.07023199715758639, "learning_rate": 2.8831125777952746e-05, "loss": 0.927, "step": 6967 }, { "epoch": 3.1557971014492754, "grad_norm": 0.07179951553545758, "learning_rate": 2.88189820494645e-05, "loss": 0.9076, "step": 6968 }, { "epoch": 3.15625, "grad_norm": 0.0644829459908091, "learning_rate": 2.8806839438760322e-05, "loss": 0.9323, "step": 6969 }, { "epoch": 3.1567028985507246, "grad_norm": 0.07626229097786388, "learning_rate": 2.8794697947054147e-05, "loss": 0.9223, "step": 6970 }, { "epoch": 3.1571557971014492, "grad_norm": 0.08477290241234135, "learning_rate": 2.8782557575559776e-05, "loss": 0.9481, "step": 6971 }, { "epoch": 3.157608695652174, "grad_norm": 0.06333711015816826, "learning_rate": 2.8770418325490907e-05, "loss": 0.9136, "step": 6972 }, { "epoch": 3.1580615942028984, "grad_norm": 0.08975048893145121, "learning_rate": 2.875828019806109e-05, "loss": 0.9045, "step": 6973 }, { "epoch": 3.158514492753623, "grad_norm": 0.07275976355344262, "learning_rate": 2.8746143194483844e-05, "loss": 0.9157, "step": 6974 }, { "epoch": 3.1589673913043477, "grad_norm": 0.07320065909181546, "learning_rate": 2.873400731597248e-05, "loss": 0.9246, "step": 6975 }, { "epoch": 3.1594202898550723, "grad_norm": 0.07197386970476152, "learning_rate": 2.8721872563740264e-05, "loss": 0.9175, "step": 6976 }, { "epoch": 3.159873188405797, "grad_norm": 0.09374351748232376, "learning_rate": 2.870973893900031e-05, "loss": 0.9369, "step": 6977 }, { "epoch": 3.1603260869565215, "grad_norm": 0.07530201496315675, "learning_rate": 2.8697606442965655e-05, "loss": 0.933, "step": 6978 }, { "epoch": 3.1607789855072466, "grad_norm": 0.06411451159118234, "learning_rate": 2.868547507684919e-05, "loss": 0.9144, "step": 6979 }, { "epoch": 3.161231884057971, "grad_norm": 0.08794269456207028, "learning_rate": 2.8673344841863718e-05, "loss": 0.9472, "step": 6980 }, { "epoch": 3.161684782608696, "grad_norm": 0.06282909062793159, "learning_rate": 2.866121573922189e-05, "loss": 0.927, "step": 6981 }, { "epoch": 3.1621376811594204, "grad_norm": 0.07701406317773823, "learning_rate": 2.8649087770136312e-05, "loss": 0.8959, "step": 6982 }, { "epoch": 3.162590579710145, "grad_norm": 0.07258942374714392, "learning_rate": 2.8636960935819406e-05, "loss": 0.9414, "step": 6983 }, { "epoch": 3.1630434782608696, "grad_norm": 0.06966378494961313, "learning_rate": 2.8624835237483535e-05, "loss": 0.9042, "step": 6984 }, { "epoch": 3.1634963768115942, "grad_norm": 0.0779509906412343, "learning_rate": 2.8612710676340892e-05, "loss": 0.9258, "step": 6985 }, { "epoch": 3.163949275362319, "grad_norm": 0.06673233427698573, "learning_rate": 2.8600587253603626e-05, "loss": 0.9242, "step": 6986 }, { "epoch": 3.1644021739130435, "grad_norm": 0.0722337499461973, "learning_rate": 2.8588464970483704e-05, "loss": 0.918, "step": 6987 }, { "epoch": 3.164855072463768, "grad_norm": 0.07722025655027052, "learning_rate": 2.857634382819303e-05, "loss": 0.9004, "step": 6988 }, { "epoch": 3.1653079710144927, "grad_norm": 0.07243001575256734, "learning_rate": 2.8564223827943355e-05, "loss": 0.9343, "step": 6989 }, { "epoch": 3.1657608695652173, "grad_norm": 0.07351861220284668, "learning_rate": 2.8552104970946354e-05, "loss": 0.9362, "step": 6990 }, { "epoch": 3.166213768115942, "grad_norm": 0.09044722901757096, "learning_rate": 2.853998725841356e-05, "loss": 0.922, "step": 6991 }, { "epoch": 3.1666666666666665, "grad_norm": 0.07867471239236556, "learning_rate": 2.8527870691556404e-05, "loss": 0.913, "step": 6992 }, { "epoch": 3.167119565217391, "grad_norm": 0.08180969750741433, "learning_rate": 2.8515755271586175e-05, "loss": 0.94, "step": 6993 }, { "epoch": 3.1675724637681157, "grad_norm": 0.10071990966033804, "learning_rate": 2.8503640999714108e-05, "loss": 0.8983, "step": 6994 }, { "epoch": 3.168025362318841, "grad_norm": 0.0777165536993507, "learning_rate": 2.8491527877151264e-05, "loss": 0.9332, "step": 6995 }, { "epoch": 3.1684782608695654, "grad_norm": 0.07897714917428193, "learning_rate": 2.8479415905108614e-05, "loss": 0.9185, "step": 6996 }, { "epoch": 3.16893115942029, "grad_norm": 0.09093864366148605, "learning_rate": 2.8467305084797005e-05, "loss": 0.9342, "step": 6997 }, { "epoch": 3.1693840579710146, "grad_norm": 0.07476988849776955, "learning_rate": 2.8455195417427192e-05, "loss": 0.9313, "step": 6998 }, { "epoch": 3.1698369565217392, "grad_norm": 0.07584579650207707, "learning_rate": 2.8443086904209783e-05, "loss": 0.9231, "step": 6999 }, { "epoch": 3.170289855072464, "grad_norm": 0.0877240452436976, "learning_rate": 2.8430979546355294e-05, "loss": 0.9366, "step": 7000 }, { "epoch": 3.1707427536231885, "grad_norm": 0.08260937603889196, "learning_rate": 2.8418873345074097e-05, "loss": 0.9129, "step": 7001 }, { "epoch": 3.171195652173913, "grad_norm": 0.07830946464029574, "learning_rate": 2.8406768301576506e-05, "loss": 0.9217, "step": 7002 }, { "epoch": 3.1716485507246377, "grad_norm": 0.06666154801969426, "learning_rate": 2.8394664417072636e-05, "loss": 0.9325, "step": 7003 }, { "epoch": 3.1721014492753623, "grad_norm": 0.07709323341052726, "learning_rate": 2.8382561692772576e-05, "loss": 0.9151, "step": 7004 }, { "epoch": 3.172554347826087, "grad_norm": 0.06887984250862363, "learning_rate": 2.837046012988622e-05, "loss": 0.9223, "step": 7005 }, { "epoch": 3.1730072463768115, "grad_norm": 0.07373285047787732, "learning_rate": 2.8358359729623397e-05, "loss": 0.9306, "step": 7006 }, { "epoch": 3.173460144927536, "grad_norm": 0.07196324225430484, "learning_rate": 2.834626049319379e-05, "loss": 0.921, "step": 7007 }, { "epoch": 3.1739130434782608, "grad_norm": 0.07187962062511259, "learning_rate": 2.8334162421807002e-05, "loss": 0.9487, "step": 7008 }, { "epoch": 3.1743659420289854, "grad_norm": 0.0711941260548851, "learning_rate": 2.8322065516672485e-05, "loss": 0.9284, "step": 7009 }, { "epoch": 3.17481884057971, "grad_norm": 0.06942170618163727, "learning_rate": 2.830996977899958e-05, "loss": 0.9305, "step": 7010 }, { "epoch": 3.1752717391304346, "grad_norm": 0.06922938401966365, "learning_rate": 2.8297875209997507e-05, "loss": 0.9202, "step": 7011 }, { "epoch": 3.175724637681159, "grad_norm": 0.06290886904126307, "learning_rate": 2.828578181087541e-05, "loss": 0.9268, "step": 7012 }, { "epoch": 3.1761775362318843, "grad_norm": 0.0645511661797728, "learning_rate": 2.827368958284226e-05, "loss": 0.8991, "step": 7013 }, { "epoch": 3.176630434782609, "grad_norm": 0.06496929819411716, "learning_rate": 2.8261598527106948e-05, "loss": 0.916, "step": 7014 }, { "epoch": 3.1770833333333335, "grad_norm": 0.06257478321621654, "learning_rate": 2.8249508644878224e-05, "loss": 0.9225, "step": 7015 }, { "epoch": 3.177536231884058, "grad_norm": 0.0735275397597544, "learning_rate": 2.8237419937364756e-05, "loss": 0.9184, "step": 7016 }, { "epoch": 3.1779891304347827, "grad_norm": 0.07508422481447356, "learning_rate": 2.8225332405775043e-05, "loss": 0.9173, "step": 7017 }, { "epoch": 3.1784420289855073, "grad_norm": 0.08566081158369476, "learning_rate": 2.8213246051317517e-05, "loss": 0.9094, "step": 7018 }, { "epoch": 3.178894927536232, "grad_norm": 0.068289078302591, "learning_rate": 2.820116087520045e-05, "loss": 0.9314, "step": 7019 }, { "epoch": 3.1793478260869565, "grad_norm": 0.08357714111446675, "learning_rate": 2.818907687863204e-05, "loss": 0.9256, "step": 7020 }, { "epoch": 3.179800724637681, "grad_norm": 0.07201831517350685, "learning_rate": 2.817699406282032e-05, "loss": 0.9563, "step": 7021 }, { "epoch": 3.1802536231884058, "grad_norm": 0.08178182890961652, "learning_rate": 2.8164912428973248e-05, "loss": 0.9149, "step": 7022 }, { "epoch": 3.1807065217391304, "grad_norm": 0.07207495697853153, "learning_rate": 2.815283197829862e-05, "loss": 0.9216, "step": 7023 }, { "epoch": 3.181159420289855, "grad_norm": 0.08258907762859315, "learning_rate": 2.8140752712004168e-05, "loss": 0.9255, "step": 7024 }, { "epoch": 3.1816123188405796, "grad_norm": 0.06809586660683313, "learning_rate": 2.8128674631297453e-05, "loss": 0.927, "step": 7025 }, { "epoch": 3.182065217391304, "grad_norm": 0.079321637076015, "learning_rate": 2.8116597737385958e-05, "loss": 0.9171, "step": 7026 }, { "epoch": 3.182518115942029, "grad_norm": 0.07321806243303941, "learning_rate": 2.8104522031476996e-05, "loss": 0.9381, "step": 7027 }, { "epoch": 3.1829710144927534, "grad_norm": 0.07622377959375141, "learning_rate": 2.8092447514777834e-05, "loss": 0.9479, "step": 7028 }, { "epoch": 3.1834239130434785, "grad_norm": 0.06458358958767385, "learning_rate": 2.808037418849556e-05, "loss": 0.9287, "step": 7029 }, { "epoch": 3.183876811594203, "grad_norm": 0.06735055783875367, "learning_rate": 2.8068302053837166e-05, "loss": 0.9186, "step": 7030 }, { "epoch": 3.1843297101449277, "grad_norm": 0.0606138068424075, "learning_rate": 2.805623111200951e-05, "loss": 0.9398, "step": 7031 }, { "epoch": 3.1847826086956523, "grad_norm": 0.07027574935018518, "learning_rate": 2.8044161364219378e-05, "loss": 0.9127, "step": 7032 }, { "epoch": 3.185235507246377, "grad_norm": 0.06594957804639426, "learning_rate": 2.8032092811673367e-05, "loss": 0.9175, "step": 7033 }, { "epoch": 3.1856884057971016, "grad_norm": 0.06957612674403978, "learning_rate": 2.8020025455578007e-05, "loss": 0.9241, "step": 7034 }, { "epoch": 3.186141304347826, "grad_norm": 0.07289060437752819, "learning_rate": 2.8007959297139678e-05, "loss": 0.9181, "step": 7035 }, { "epoch": 3.1865942028985508, "grad_norm": 0.06348859542648261, "learning_rate": 2.7995894337564674e-05, "loss": 0.9228, "step": 7036 }, { "epoch": 3.1870471014492754, "grad_norm": 0.06934053155113419, "learning_rate": 2.7983830578059126e-05, "loss": 0.9372, "step": 7037 }, { "epoch": 3.1875, "grad_norm": 0.07901943526808278, "learning_rate": 2.7971768019829083e-05, "loss": 0.9367, "step": 7038 }, { "epoch": 3.1879528985507246, "grad_norm": 0.07434258305026782, "learning_rate": 2.795970666408044e-05, "loss": 0.9181, "step": 7039 }, { "epoch": 3.1884057971014492, "grad_norm": 0.08655724762330815, "learning_rate": 2.7947646512019012e-05, "loss": 0.9266, "step": 7040 }, { "epoch": 3.188858695652174, "grad_norm": 0.06784274878849529, "learning_rate": 2.7935587564850455e-05, "loss": 0.9189, "step": 7041 }, { "epoch": 3.1893115942028984, "grad_norm": 0.08553445566637494, "learning_rate": 2.7923529823780335e-05, "loss": 0.9193, "step": 7042 }, { "epoch": 3.189764492753623, "grad_norm": 0.07262977400670984, "learning_rate": 2.7911473290014064e-05, "loss": 0.9105, "step": 7043 }, { "epoch": 3.1902173913043477, "grad_norm": 0.07268119685515281, "learning_rate": 2.7899417964756973e-05, "loss": 0.9421, "step": 7044 }, { "epoch": 3.1906702898550723, "grad_norm": 0.07737548648283485, "learning_rate": 2.788736384921424e-05, "loss": 0.9534, "step": 7045 }, { "epoch": 3.191123188405797, "grad_norm": 0.06416669940921688, "learning_rate": 2.7875310944590944e-05, "loss": 0.9165, "step": 7046 }, { "epoch": 3.1915760869565215, "grad_norm": 0.0773876698781623, "learning_rate": 2.7863259252092007e-05, "loss": 0.9111, "step": 7047 }, { "epoch": 3.1920289855072466, "grad_norm": 0.08687810824038283, "learning_rate": 2.7851208772922295e-05, "loss": 0.9351, "step": 7048 }, { "epoch": 3.192481884057971, "grad_norm": 0.06403717496903223, "learning_rate": 2.7839159508286474e-05, "loss": 0.9155, "step": 7049 }, { "epoch": 3.192934782608696, "grad_norm": 0.08289975744802396, "learning_rate": 2.7827111459389163e-05, "loss": 0.913, "step": 7050 }, { "epoch": 3.1933876811594204, "grad_norm": 0.08737738158378729, "learning_rate": 2.7815064627434806e-05, "loss": 0.9016, "step": 7051 }, { "epoch": 3.193840579710145, "grad_norm": 0.06518597477863586, "learning_rate": 2.780301901362775e-05, "loss": 0.9563, "step": 7052 }, { "epoch": 3.1942934782608696, "grad_norm": 0.09103148447089157, "learning_rate": 2.7790974619172195e-05, "loss": 0.9371, "step": 7053 }, { "epoch": 3.1947463768115942, "grad_norm": 0.06396717814653759, "learning_rate": 2.7778931445272275e-05, "loss": 0.9011, "step": 7054 }, { "epoch": 3.195199275362319, "grad_norm": 0.08305928494506398, "learning_rate": 2.7766889493131933e-05, "loss": 0.9373, "step": 7055 }, { "epoch": 3.1956521739130435, "grad_norm": 0.09199011604956378, "learning_rate": 2.775484876395504e-05, "loss": 0.9097, "step": 7056 }, { "epoch": 3.196105072463768, "grad_norm": 0.07602074794920284, "learning_rate": 2.7742809258945303e-05, "loss": 0.9318, "step": 7057 }, { "epoch": 3.1965579710144927, "grad_norm": 0.08876755359465724, "learning_rate": 2.7730770979306368e-05, "loss": 0.901, "step": 7058 }, { "epoch": 3.1970108695652173, "grad_norm": 0.07684566666560254, "learning_rate": 2.771873392624169e-05, "loss": 0.9426, "step": 7059 }, { "epoch": 3.197463768115942, "grad_norm": 0.07418493840718357, "learning_rate": 2.7706698100954653e-05, "loss": 0.9072, "step": 7060 }, { "epoch": 3.1979166666666665, "grad_norm": 0.07360742641930419, "learning_rate": 2.769466350464847e-05, "loss": 0.9379, "step": 7061 }, { "epoch": 3.198369565217391, "grad_norm": 0.08828719292667869, "learning_rate": 2.7682630138526293e-05, "loss": 0.9215, "step": 7062 }, { "epoch": 3.1988224637681157, "grad_norm": 0.060606068346363104, "learning_rate": 2.7670598003791086e-05, "loss": 0.9221, "step": 7063 }, { "epoch": 3.199275362318841, "grad_norm": 0.07436601997498389, "learning_rate": 2.765856710164575e-05, "loss": 0.9051, "step": 7064 }, { "epoch": 3.1997282608695654, "grad_norm": 0.06260465510918521, "learning_rate": 2.7646537433292997e-05, "loss": 0.9288, "step": 7065 }, { "epoch": 3.20018115942029, "grad_norm": 0.0688619479335149, "learning_rate": 2.763450899993549e-05, "loss": 0.9369, "step": 7066 }, { "epoch": 3.2006340579710146, "grad_norm": 0.05976791477944066, "learning_rate": 2.7622481802775713e-05, "loss": 0.9175, "step": 7067 }, { "epoch": 3.2010869565217392, "grad_norm": 0.07190068320365882, "learning_rate": 2.7610455843016035e-05, "loss": 0.9033, "step": 7068 }, { "epoch": 3.201539855072464, "grad_norm": 0.0634888661884927, "learning_rate": 2.759843112185873e-05, "loss": 0.9301, "step": 7069 }, { "epoch": 3.2019927536231885, "grad_norm": 0.06140794659073449, "learning_rate": 2.7586407640505898e-05, "loss": 0.9381, "step": 7070 }, { "epoch": 3.202445652173913, "grad_norm": 0.06383911945812058, "learning_rate": 2.7574385400159585e-05, "loss": 0.9163, "step": 7071 }, { "epoch": 3.2028985507246377, "grad_norm": 0.06981329181876784, "learning_rate": 2.756236440202164e-05, "loss": 0.9556, "step": 7072 }, { "epoch": 3.2033514492753623, "grad_norm": 0.0633945467686725, "learning_rate": 2.755034464729385e-05, "loss": 0.9396, "step": 7073 }, { "epoch": 3.203804347826087, "grad_norm": 0.07583398762466925, "learning_rate": 2.7538326137177814e-05, "loss": 0.9033, "step": 7074 }, { "epoch": 3.2042572463768115, "grad_norm": 0.07280709623362124, "learning_rate": 2.7526308872875075e-05, "loss": 0.9283, "step": 7075 }, { "epoch": 3.204710144927536, "grad_norm": 0.061233654456767234, "learning_rate": 2.7514292855587e-05, "loss": 0.9307, "step": 7076 }, { "epoch": 3.2051630434782608, "grad_norm": 0.072092418314668, "learning_rate": 2.7502278086514863e-05, "loss": 0.9376, "step": 7077 }, { "epoch": 3.2056159420289854, "grad_norm": 0.0759449803445023, "learning_rate": 2.749026456685977e-05, "loss": 0.9023, "step": 7078 }, { "epoch": 3.20606884057971, "grad_norm": 0.05998606259127784, "learning_rate": 2.7478252297822767e-05, "loss": 0.9539, "step": 7079 }, { "epoch": 3.2065217391304346, "grad_norm": 0.07663203060229737, "learning_rate": 2.7466241280604716e-05, "loss": 0.9348, "step": 7080 }, { "epoch": 3.206974637681159, "grad_norm": 0.07571878423167526, "learning_rate": 2.7454231516406396e-05, "loss": 0.9142, "step": 7081 }, { "epoch": 3.2074275362318843, "grad_norm": 0.05853325567791361, "learning_rate": 2.7442223006428417e-05, "loss": 0.9253, "step": 7082 }, { "epoch": 3.207880434782609, "grad_norm": 0.08140201431609424, "learning_rate": 2.7430215751871318e-05, "loss": 0.9363, "step": 7083 }, { "epoch": 3.2083333333333335, "grad_norm": 0.08244971424183932, "learning_rate": 2.7418209753935464e-05, "loss": 0.9275, "step": 7084 }, { "epoch": 3.208786231884058, "grad_norm": 0.06580467895990286, "learning_rate": 2.740620501382112e-05, "loss": 0.923, "step": 7085 }, { "epoch": 3.2092391304347827, "grad_norm": 0.08184891322106236, "learning_rate": 2.7394201532728405e-05, "loss": 0.9283, "step": 7086 }, { "epoch": 3.2096920289855073, "grad_norm": 0.06793391848661696, "learning_rate": 2.7382199311857352e-05, "loss": 0.9192, "step": 7087 }, { "epoch": 3.210144927536232, "grad_norm": 0.07526474463280407, "learning_rate": 2.7370198352407826e-05, "loss": 0.917, "step": 7088 }, { "epoch": 3.2105978260869565, "grad_norm": 0.0627914498097439, "learning_rate": 2.7358198655579587e-05, "loss": 0.9336, "step": 7089 }, { "epoch": 3.211050724637681, "grad_norm": 0.0772959860247087, "learning_rate": 2.7346200222572245e-05, "loss": 0.9451, "step": 7090 }, { "epoch": 3.2115036231884058, "grad_norm": 0.06828588767743433, "learning_rate": 2.733420305458534e-05, "loss": 0.9375, "step": 7091 }, { "epoch": 3.2119565217391304, "grad_norm": 0.07222527076155313, "learning_rate": 2.7322207152818205e-05, "loss": 0.9269, "step": 7092 }, { "epoch": 3.212409420289855, "grad_norm": 0.07689293532660102, "learning_rate": 2.731021251847014e-05, "loss": 0.9274, "step": 7093 }, { "epoch": 3.2128623188405796, "grad_norm": 0.0853282045666955, "learning_rate": 2.7298219152740208e-05, "loss": 0.9252, "step": 7094 }, { "epoch": 3.213315217391304, "grad_norm": 0.08046014081915098, "learning_rate": 2.728622705682745e-05, "loss": 0.9327, "step": 7095 }, { "epoch": 3.213768115942029, "grad_norm": 0.08524947148515348, "learning_rate": 2.7274236231930705e-05, "loss": 0.9104, "step": 7096 }, { "epoch": 3.2142210144927534, "grad_norm": 0.10138419422246531, "learning_rate": 2.7262246679248748e-05, "loss": 0.9323, "step": 7097 }, { "epoch": 3.2146739130434785, "grad_norm": 0.07304523416272893, "learning_rate": 2.7250258399980166e-05, "loss": 0.9097, "step": 7098 }, { "epoch": 3.215126811594203, "grad_norm": 0.08567214543742564, "learning_rate": 2.7238271395323467e-05, "loss": 0.9365, "step": 7099 }, { "epoch": 3.2155797101449277, "grad_norm": 0.0783996582945663, "learning_rate": 2.722628566647698e-05, "loss": 0.9304, "step": 7100 }, { "epoch": 3.2160326086956523, "grad_norm": 0.06440215476361055, "learning_rate": 2.7214301214638976e-05, "loss": 0.9131, "step": 7101 }, { "epoch": 3.216485507246377, "grad_norm": 0.0684495891762867, "learning_rate": 2.720231804100753e-05, "loss": 0.9369, "step": 7102 }, { "epoch": 3.2169384057971016, "grad_norm": 0.06586424795504711, "learning_rate": 2.7190336146780637e-05, "loss": 0.9436, "step": 7103 }, { "epoch": 3.217391304347826, "grad_norm": 0.08105789655675738, "learning_rate": 2.717835553315612e-05, "loss": 0.9249, "step": 7104 }, { "epoch": 3.2178442028985508, "grad_norm": 0.05680279364624261, "learning_rate": 2.7166376201331738e-05, "loss": 0.9375, "step": 7105 }, { "epoch": 3.2182971014492754, "grad_norm": 0.08255093160642003, "learning_rate": 2.715439815250505e-05, "loss": 0.9205, "step": 7106 }, { "epoch": 3.21875, "grad_norm": 0.08273185950617143, "learning_rate": 2.7142421387873548e-05, "loss": 0.9185, "step": 7107 }, { "epoch": 3.2192028985507246, "grad_norm": 0.07398567317899588, "learning_rate": 2.7130445908634537e-05, "loss": 0.9219, "step": 7108 }, { "epoch": 3.2196557971014492, "grad_norm": 0.10408942114207902, "learning_rate": 2.7118471715985256e-05, "loss": 0.9146, "step": 7109 }, { "epoch": 3.220108695652174, "grad_norm": 0.10627438621256233, "learning_rate": 2.7106498811122762e-05, "loss": 0.9267, "step": 7110 }, { "epoch": 3.2205615942028984, "grad_norm": 0.08050068999490083, "learning_rate": 2.709452719524403e-05, "loss": 0.9211, "step": 7111 }, { "epoch": 3.221014492753623, "grad_norm": 0.08701597910683048, "learning_rate": 2.7082556869545836e-05, "loss": 0.9267, "step": 7112 }, { "epoch": 3.2214673913043477, "grad_norm": 0.0759826106043173, "learning_rate": 2.7070587835224924e-05, "loss": 0.9311, "step": 7113 }, { "epoch": 3.2219202898550723, "grad_norm": 0.06547601508738113, "learning_rate": 2.705862009347782e-05, "loss": 0.9074, "step": 7114 }, { "epoch": 3.222373188405797, "grad_norm": 0.0745019419693348, "learning_rate": 2.704665364550099e-05, "loss": 0.9172, "step": 7115 }, { "epoch": 3.2228260869565215, "grad_norm": 0.08624430452032128, "learning_rate": 2.70346884924907e-05, "loss": 0.9286, "step": 7116 }, { "epoch": 3.2232789855072466, "grad_norm": 0.06656991532955933, "learning_rate": 2.7022724635643167e-05, "loss": 0.9336, "step": 7117 }, { "epoch": 3.223731884057971, "grad_norm": 0.09784297099647106, "learning_rate": 2.7010762076154403e-05, "loss": 0.9378, "step": 7118 }, { "epoch": 3.224184782608696, "grad_norm": 0.06343153684308564, "learning_rate": 2.699880081522035e-05, "loss": 0.9172, "step": 7119 }, { "epoch": 3.2246376811594204, "grad_norm": 0.08996044274834614, "learning_rate": 2.698684085403676e-05, "loss": 0.9134, "step": 7120 }, { "epoch": 3.225090579710145, "grad_norm": 0.07909270828830968, "learning_rate": 2.697488219379933e-05, "loss": 0.913, "step": 7121 }, { "epoch": 3.2255434782608696, "grad_norm": 0.07276473866459539, "learning_rate": 2.696292483570356e-05, "loss": 0.9122, "step": 7122 }, { "epoch": 3.2259963768115942, "grad_norm": 0.06873436227701926, "learning_rate": 2.695096878094486e-05, "loss": 0.9372, "step": 7123 }, { "epoch": 3.226449275362319, "grad_norm": 0.08889264432709651, "learning_rate": 2.6939014030718476e-05, "loss": 0.9143, "step": 7124 }, { "epoch": 3.2269021739130435, "grad_norm": 0.06449837011656977, "learning_rate": 2.6927060586219565e-05, "loss": 0.9144, "step": 7125 }, { "epoch": 3.227355072463768, "grad_norm": 0.06806660992335246, "learning_rate": 2.6915108448643124e-05, "loss": 0.9443, "step": 7126 }, { "epoch": 3.2278079710144927, "grad_norm": 0.09763445331965409, "learning_rate": 2.690315761918403e-05, "loss": 0.9113, "step": 7127 }, { "epoch": 3.2282608695652173, "grad_norm": 0.06620215517082853, "learning_rate": 2.6891208099037006e-05, "loss": 0.9236, "step": 7128 }, { "epoch": 3.228713768115942, "grad_norm": 0.09221601405405129, "learning_rate": 2.6879259889396697e-05, "loss": 0.9194, "step": 7129 }, { "epoch": 3.2291666666666665, "grad_norm": 0.09953112064507992, "learning_rate": 2.6867312991457563e-05, "loss": 0.9643, "step": 7130 }, { "epoch": 3.229619565217391, "grad_norm": 0.07159751717098038, "learning_rate": 2.6855367406413964e-05, "loss": 0.9373, "step": 7131 }, { "epoch": 3.2300724637681157, "grad_norm": 0.07560615150595615, "learning_rate": 2.68434231354601e-05, "loss": 0.9244, "step": 7132 }, { "epoch": 3.230525362318841, "grad_norm": 0.08417091679561745, "learning_rate": 2.6831480179790096e-05, "loss": 0.9171, "step": 7133 }, { "epoch": 3.2309782608695654, "grad_norm": 0.06175895889001859, "learning_rate": 2.6819538540597867e-05, "loss": 0.921, "step": 7134 }, { "epoch": 3.23143115942029, "grad_norm": 0.07964806418582727, "learning_rate": 2.680759821907727e-05, "loss": 0.9463, "step": 7135 }, { "epoch": 3.2318840579710146, "grad_norm": 0.06331922353114634, "learning_rate": 2.679565921642197e-05, "loss": 0.9324, "step": 7136 }, { "epoch": 3.2323369565217392, "grad_norm": 0.07476759546052918, "learning_rate": 2.6783721533825554e-05, "loss": 0.9195, "step": 7137 }, { "epoch": 3.232789855072464, "grad_norm": 0.08857469862237535, "learning_rate": 2.6771785172481426e-05, "loss": 0.9291, "step": 7138 }, { "epoch": 3.2332427536231885, "grad_norm": 0.06927315285671895, "learning_rate": 2.6759850133582914e-05, "loss": 0.9354, "step": 7139 }, { "epoch": 3.233695652173913, "grad_norm": 0.06960012718815037, "learning_rate": 2.6747916418323154e-05, "loss": 0.9208, "step": 7140 }, { "epoch": 3.2341485507246377, "grad_norm": 0.08412174968873214, "learning_rate": 2.6735984027895193e-05, "loss": 0.921, "step": 7141 }, { "epoch": 3.2346014492753623, "grad_norm": 0.06902328739956808, "learning_rate": 2.672405296349192e-05, "loss": 0.9261, "step": 7142 }, { "epoch": 3.235054347826087, "grad_norm": 0.06898921224407513, "learning_rate": 2.671212322630612e-05, "loss": 0.9112, "step": 7143 }, { "epoch": 3.2355072463768115, "grad_norm": 0.08628833327908388, "learning_rate": 2.6700194817530413e-05, "loss": 0.9091, "step": 7144 }, { "epoch": 3.235960144927536, "grad_norm": 0.06883092370297757, "learning_rate": 2.668826773835731e-05, "loss": 0.9269, "step": 7145 }, { "epoch": 3.2364130434782608, "grad_norm": 0.08757110071619356, "learning_rate": 2.6676341989979162e-05, "loss": 0.9454, "step": 7146 }, { "epoch": 3.2368659420289854, "grad_norm": 0.07922707944375747, "learning_rate": 2.6664417573588233e-05, "loss": 0.929, "step": 7147 }, { "epoch": 3.23731884057971, "grad_norm": 0.07460480771314296, "learning_rate": 2.6652494490376602e-05, "loss": 0.9158, "step": 7148 }, { "epoch": 3.2377717391304346, "grad_norm": 0.0671252273784158, "learning_rate": 2.6640572741536258e-05, "loss": 0.9402, "step": 7149 }, { "epoch": 3.238224637681159, "grad_norm": 0.08925922499926885, "learning_rate": 2.662865232825901e-05, "loss": 0.9368, "step": 7150 }, { "epoch": 3.2386775362318843, "grad_norm": 0.06325354219740394, "learning_rate": 2.6616733251736592e-05, "loss": 0.9169, "step": 7151 }, { "epoch": 3.239130434782609, "grad_norm": 0.07275999730646136, "learning_rate": 2.6604815513160556e-05, "loss": 0.9401, "step": 7152 }, { "epoch": 3.2395833333333335, "grad_norm": 0.07174911171150916, "learning_rate": 2.659289911372234e-05, "loss": 0.9137, "step": 7153 }, { "epoch": 3.240036231884058, "grad_norm": 0.08162186435997189, "learning_rate": 2.6580984054613237e-05, "loss": 0.9251, "step": 7154 }, { "epoch": 3.2404891304347827, "grad_norm": 0.08955221127964513, "learning_rate": 2.6569070337024432e-05, "loss": 0.9523, "step": 7155 }, { "epoch": 3.2409420289855073, "grad_norm": 0.07523036775529411, "learning_rate": 2.6557157962146943e-05, "loss": 0.9149, "step": 7156 }, { "epoch": 3.241394927536232, "grad_norm": 0.07885584409442872, "learning_rate": 2.6545246931171688e-05, "loss": 0.9176, "step": 7157 }, { "epoch": 3.2418478260869565, "grad_norm": 0.09184289149286805, "learning_rate": 2.653333724528939e-05, "loss": 0.9077, "step": 7158 }, { "epoch": 3.242300724637681, "grad_norm": 0.06740414990527556, "learning_rate": 2.6521428905690734e-05, "loss": 0.9303, "step": 7159 }, { "epoch": 3.2427536231884058, "grad_norm": 0.08132526333262088, "learning_rate": 2.650952191356617e-05, "loss": 0.9294, "step": 7160 }, { "epoch": 3.2432065217391304, "grad_norm": 0.06760032908946662, "learning_rate": 2.6497616270106097e-05, "loss": 0.9118, "step": 7161 }, { "epoch": 3.243659420289855, "grad_norm": 0.07400183549894049, "learning_rate": 2.6485711976500703e-05, "loss": 0.939, "step": 7162 }, { "epoch": 3.2441123188405796, "grad_norm": 0.07311903072056966, "learning_rate": 2.6473809033940114e-05, "loss": 0.9232, "step": 7163 }, { "epoch": 3.244565217391304, "grad_norm": 0.06253762110618048, "learning_rate": 2.646190744361426e-05, "loss": 0.9416, "step": 7164 }, { "epoch": 3.245018115942029, "grad_norm": 0.07043649046517038, "learning_rate": 2.6450007206712975e-05, "loss": 0.9243, "step": 7165 }, { "epoch": 3.2454710144927534, "grad_norm": 0.06105585048878049, "learning_rate": 2.6438108324425937e-05, "loss": 0.9274, "step": 7166 }, { "epoch": 3.2459239130434785, "grad_norm": 0.06834897367812819, "learning_rate": 2.6426210797942707e-05, "loss": 0.9382, "step": 7167 }, { "epoch": 3.246376811594203, "grad_norm": 0.06240637231592302, "learning_rate": 2.6414314628452687e-05, "loss": 0.9296, "step": 7168 }, { "epoch": 3.2468297101449277, "grad_norm": 0.06879077269880854, "learning_rate": 2.640241981714517e-05, "loss": 0.9201, "step": 7169 }, { "epoch": 3.2472826086956523, "grad_norm": 0.06107931443363554, "learning_rate": 2.6390526365209275e-05, "loss": 0.9496, "step": 7170 }, { "epoch": 3.247735507246377, "grad_norm": 0.05926754693153859, "learning_rate": 2.6378634273834044e-05, "loss": 0.942, "step": 7171 }, { "epoch": 3.2481884057971016, "grad_norm": 0.05928595280078989, "learning_rate": 2.636674354420832e-05, "loss": 0.9263, "step": 7172 }, { "epoch": 3.248641304347826, "grad_norm": 0.062362161012722545, "learning_rate": 2.6354854177520855e-05, "loss": 0.9242, "step": 7173 }, { "epoch": 3.2490942028985508, "grad_norm": 0.07057497976428466, "learning_rate": 2.6342966174960224e-05, "loss": 0.9192, "step": 7174 }, { "epoch": 3.2495471014492754, "grad_norm": 0.06653229248296105, "learning_rate": 2.633107953771492e-05, "loss": 0.9305, "step": 7175 }, { "epoch": 3.25, "grad_norm": 0.07463341026324935, "learning_rate": 2.6319194266973256e-05, "loss": 0.9281, "step": 7176 }, { "epoch": 3.2504528985507246, "grad_norm": 0.08523034868534803, "learning_rate": 2.630731036392342e-05, "loss": 0.9522, "step": 7177 }, { "epoch": 3.2509057971014492, "grad_norm": 0.06770140867863642, "learning_rate": 2.629542782975345e-05, "loss": 0.909, "step": 7178 }, { "epoch": 3.251358695652174, "grad_norm": 0.07011389558641302, "learning_rate": 2.6283546665651293e-05, "loss": 0.9149, "step": 7179 }, { "epoch": 3.2518115942028984, "grad_norm": 0.0734935807308284, "learning_rate": 2.62716668728047e-05, "loss": 0.9291, "step": 7180 }, { "epoch": 3.252264492753623, "grad_norm": 0.0647672957124468, "learning_rate": 2.625978845240134e-05, "loss": 0.9127, "step": 7181 }, { "epoch": 3.2527173913043477, "grad_norm": 0.06711972483935219, "learning_rate": 2.6247911405628683e-05, "loss": 0.9334, "step": 7182 }, { "epoch": 3.2531702898550723, "grad_norm": 0.07292290381272391, "learning_rate": 2.623603573367413e-05, "loss": 0.9248, "step": 7183 }, { "epoch": 3.253623188405797, "grad_norm": 0.06979024838077684, "learning_rate": 2.6224161437724877e-05, "loss": 0.9171, "step": 7184 }, { "epoch": 3.2540760869565215, "grad_norm": 0.07592999156605, "learning_rate": 2.621228851896805e-05, "loss": 0.9519, "step": 7185 }, { "epoch": 3.2545289855072466, "grad_norm": 0.07618048515798759, "learning_rate": 2.6200416978590582e-05, "loss": 0.9223, "step": 7186 }, { "epoch": 3.254981884057971, "grad_norm": 0.07227001062348122, "learning_rate": 2.618854681777931e-05, "loss": 0.9231, "step": 7187 }, { "epoch": 3.255434782608696, "grad_norm": 0.07480732796195844, "learning_rate": 2.6176678037720877e-05, "loss": 0.9177, "step": 7188 }, { "epoch": 3.2558876811594204, "grad_norm": 0.07337231246049833, "learning_rate": 2.616481063960186e-05, "loss": 0.9236, "step": 7189 }, { "epoch": 3.256340579710145, "grad_norm": 0.06789929114599723, "learning_rate": 2.615294462460865e-05, "loss": 0.9174, "step": 7190 }, { "epoch": 3.2567934782608696, "grad_norm": 0.07794876605449191, "learning_rate": 2.6141079993927504e-05, "loss": 0.9142, "step": 7191 }, { "epoch": 3.2572463768115942, "grad_norm": 0.0748507035426283, "learning_rate": 2.6129216748744553e-05, "loss": 0.9177, "step": 7192 }, { "epoch": 3.257699275362319, "grad_norm": 0.062017069320217645, "learning_rate": 2.611735489024577e-05, "loss": 0.9205, "step": 7193 }, { "epoch": 3.2581521739130435, "grad_norm": 0.08375039207620454, "learning_rate": 2.610549441961704e-05, "loss": 0.9242, "step": 7194 }, { "epoch": 3.258605072463768, "grad_norm": 0.06368835956884591, "learning_rate": 2.6093635338044035e-05, "loss": 0.9037, "step": 7195 }, { "epoch": 3.2590579710144927, "grad_norm": 0.06598368088228328, "learning_rate": 2.6081777646712354e-05, "loss": 0.9168, "step": 7196 }, { "epoch": 3.2595108695652173, "grad_norm": 0.07159217723416339, "learning_rate": 2.6069921346807398e-05, "loss": 0.9143, "step": 7197 }, { "epoch": 3.259963768115942, "grad_norm": 0.07126955643931672, "learning_rate": 2.605806643951449e-05, "loss": 0.9354, "step": 7198 }, { "epoch": 3.2604166666666665, "grad_norm": 0.07216337242706904, "learning_rate": 2.6046212926018774e-05, "loss": 0.9204, "step": 7199 }, { "epoch": 3.260869565217391, "grad_norm": 0.07978940915689278, "learning_rate": 2.6034360807505267e-05, "loss": 0.9413, "step": 7200 }, { "epoch": 3.261322463768116, "grad_norm": 0.07736781186190124, "learning_rate": 2.6022510085158827e-05, "loss": 0.9445, "step": 7201 }, { "epoch": 3.261775362318841, "grad_norm": 0.06526703683546865, "learning_rate": 2.6010660760164222e-05, "loss": 0.9252, "step": 7202 }, { "epoch": 3.2622282608695654, "grad_norm": 0.07566076287780411, "learning_rate": 2.5998812833706014e-05, "loss": 0.9121, "step": 7203 }, { "epoch": 3.26268115942029, "grad_norm": 0.07308167668997531, "learning_rate": 2.5986966306968685e-05, "loss": 0.9459, "step": 7204 }, { "epoch": 3.2631340579710146, "grad_norm": 0.0844087864571283, "learning_rate": 2.5975121181136524e-05, "loss": 0.9245, "step": 7205 }, { "epoch": 3.2635869565217392, "grad_norm": 0.06179238654685076, "learning_rate": 2.5963277457393736e-05, "loss": 0.9289, "step": 7206 }, { "epoch": 3.264039855072464, "grad_norm": 0.08737410585997044, "learning_rate": 2.595143513692434e-05, "loss": 0.9392, "step": 7207 }, { "epoch": 3.2644927536231885, "grad_norm": 0.06000202317930089, "learning_rate": 2.593959422091224e-05, "loss": 0.944, "step": 7208 }, { "epoch": 3.264945652173913, "grad_norm": 0.08899567406825691, "learning_rate": 2.5927754710541167e-05, "loss": 0.9239, "step": 7209 }, { "epoch": 3.2653985507246377, "grad_norm": 0.0808502030209781, "learning_rate": 2.5915916606994778e-05, "loss": 0.9398, "step": 7210 }, { "epoch": 3.2658514492753623, "grad_norm": 0.08556786741937696, "learning_rate": 2.5904079911456516e-05, "loss": 0.9217, "step": 7211 }, { "epoch": 3.266304347826087, "grad_norm": 0.07183183916079867, "learning_rate": 2.5892244625109725e-05, "loss": 0.9291, "step": 7212 }, { "epoch": 3.2667572463768115, "grad_norm": 0.08455583225311152, "learning_rate": 2.588041074913758e-05, "loss": 0.9291, "step": 7213 }, { "epoch": 3.267210144927536, "grad_norm": 0.07868309062465377, "learning_rate": 2.5868578284723164e-05, "loss": 0.9226, "step": 7214 }, { "epoch": 3.2676630434782608, "grad_norm": 0.07190446663402991, "learning_rate": 2.5856747233049363e-05, "loss": 0.9378, "step": 7215 }, { "epoch": 3.2681159420289854, "grad_norm": 0.08442441124615088, "learning_rate": 2.584491759529896e-05, "loss": 0.9211, "step": 7216 }, { "epoch": 3.26856884057971, "grad_norm": 0.0636149906910451, "learning_rate": 2.583308937265456e-05, "loss": 0.9229, "step": 7217 }, { "epoch": 3.2690217391304346, "grad_norm": 0.0846202282919292, "learning_rate": 2.582126256629868e-05, "loss": 0.9424, "step": 7218 }, { "epoch": 3.269474637681159, "grad_norm": 0.0649350260380545, "learning_rate": 2.5809437177413638e-05, "loss": 0.9188, "step": 7219 }, { "epoch": 3.269927536231884, "grad_norm": 0.08306843907093664, "learning_rate": 2.5797613207181664e-05, "loss": 0.905, "step": 7220 }, { "epoch": 3.270380434782609, "grad_norm": 0.07917315487591858, "learning_rate": 2.5785790656784782e-05, "loss": 0.9332, "step": 7221 }, { "epoch": 3.2708333333333335, "grad_norm": 0.06096442437834352, "learning_rate": 2.577396952740495e-05, "loss": 0.913, "step": 7222 }, { "epoch": 3.271286231884058, "grad_norm": 0.07262862130688066, "learning_rate": 2.576214982022392e-05, "loss": 0.9377, "step": 7223 }, { "epoch": 3.2717391304347827, "grad_norm": 0.06499534655140177, "learning_rate": 2.575033153642334e-05, "loss": 0.9356, "step": 7224 }, { "epoch": 3.2721920289855073, "grad_norm": 0.06145760559831832, "learning_rate": 2.573851467718468e-05, "loss": 0.9303, "step": 7225 }, { "epoch": 3.272644927536232, "grad_norm": 0.06420948417451922, "learning_rate": 2.5726699243689332e-05, "loss": 0.9313, "step": 7226 }, { "epoch": 3.2730978260869565, "grad_norm": 0.0601203181420064, "learning_rate": 2.5714885237118453e-05, "loss": 0.9113, "step": 7227 }, { "epoch": 3.273550724637681, "grad_norm": 0.06122092736817414, "learning_rate": 2.570307265865315e-05, "loss": 0.9464, "step": 7228 }, { "epoch": 3.2740036231884058, "grad_norm": 0.06300802639783565, "learning_rate": 2.5691261509474325e-05, "loss": 0.947, "step": 7229 }, { "epoch": 3.2744565217391304, "grad_norm": 0.06784974638793977, "learning_rate": 2.5679451790762766e-05, "loss": 0.9305, "step": 7230 }, { "epoch": 3.274909420289855, "grad_norm": 0.07525772010287268, "learning_rate": 2.5667643503699088e-05, "loss": 0.9203, "step": 7231 }, { "epoch": 3.2753623188405796, "grad_norm": 0.06778706862925574, "learning_rate": 2.5655836649463818e-05, "loss": 0.9234, "step": 7232 }, { "epoch": 3.275815217391304, "grad_norm": 0.07347764355533741, "learning_rate": 2.564403122923728e-05, "loss": 0.9254, "step": 7233 }, { "epoch": 3.276268115942029, "grad_norm": 0.06463190058944267, "learning_rate": 2.5632227244199703e-05, "loss": 0.9224, "step": 7234 }, { "epoch": 3.2767210144927534, "grad_norm": 0.07055484695849092, "learning_rate": 2.562042469553111e-05, "loss": 0.9233, "step": 7235 }, { "epoch": 3.2771739130434785, "grad_norm": 0.0598955076034223, "learning_rate": 2.5608623584411466e-05, "loss": 0.9109, "step": 7236 }, { "epoch": 3.277626811594203, "grad_norm": 0.0615238839454474, "learning_rate": 2.559682391202052e-05, "loss": 0.9345, "step": 7237 }, { "epoch": 3.2780797101449277, "grad_norm": 0.06348597400686688, "learning_rate": 2.558502567953792e-05, "loss": 0.9181, "step": 7238 }, { "epoch": 3.2785326086956523, "grad_norm": 0.056895865974857344, "learning_rate": 2.557322888814313e-05, "loss": 0.9378, "step": 7239 }, { "epoch": 3.278985507246377, "grad_norm": 0.06963086333810044, "learning_rate": 2.5561433539015524e-05, "loss": 0.9295, "step": 7240 }, { "epoch": 3.2794384057971016, "grad_norm": 0.06933907719668336, "learning_rate": 2.5549639633334277e-05, "loss": 0.9368, "step": 7241 }, { "epoch": 3.279891304347826, "grad_norm": 0.06476753850483423, "learning_rate": 2.5537847172278465e-05, "loss": 0.9085, "step": 7242 }, { "epoch": 3.2803442028985508, "grad_norm": 0.06697008346057892, "learning_rate": 2.552605615702697e-05, "loss": 0.9152, "step": 7243 }, { "epoch": 3.2807971014492754, "grad_norm": 0.06328770604444751, "learning_rate": 2.55142665887586e-05, "loss": 0.9274, "step": 7244 }, { "epoch": 3.28125, "grad_norm": 0.06208585029453731, "learning_rate": 2.550247846865194e-05, "loss": 0.9345, "step": 7245 }, { "epoch": 3.2817028985507246, "grad_norm": 0.07618399306921175, "learning_rate": 2.549069179788549e-05, "loss": 0.9346, "step": 7246 }, { "epoch": 3.2821557971014492, "grad_norm": 0.07380669085150865, "learning_rate": 2.5478906577637555e-05, "loss": 0.9191, "step": 7247 }, { "epoch": 3.282608695652174, "grad_norm": 0.05823619130181397, "learning_rate": 2.5467122809086355e-05, "loss": 0.9409, "step": 7248 }, { "epoch": 3.2830615942028984, "grad_norm": 0.07398526019739578, "learning_rate": 2.5455340493409912e-05, "loss": 0.9209, "step": 7249 }, { "epoch": 3.283514492753623, "grad_norm": 0.06337712690051421, "learning_rate": 2.544355963178614e-05, "loss": 0.9422, "step": 7250 }, { "epoch": 3.2839673913043477, "grad_norm": 0.07069602643327422, "learning_rate": 2.5431780225392758e-05, "loss": 0.9427, "step": 7251 }, { "epoch": 3.2844202898550723, "grad_norm": 0.06965041547925556, "learning_rate": 2.5420002275407402e-05, "loss": 0.9082, "step": 7252 }, { "epoch": 3.284873188405797, "grad_norm": 0.06343661372355812, "learning_rate": 2.5408225783007523e-05, "loss": 0.9487, "step": 7253 }, { "epoch": 3.2853260869565215, "grad_norm": 0.05870975109419857, "learning_rate": 2.539645074937044e-05, "loss": 0.9515, "step": 7254 }, { "epoch": 3.2857789855072466, "grad_norm": 0.06057735242231109, "learning_rate": 2.53846771756733e-05, "loss": 0.9366, "step": 7255 }, { "epoch": 3.286231884057971, "grad_norm": 0.05662150875037993, "learning_rate": 2.537290506309316e-05, "loss": 0.9268, "step": 7256 }, { "epoch": 3.286684782608696, "grad_norm": 0.0568911253823112, "learning_rate": 2.536113441280687e-05, "loss": 0.917, "step": 7257 }, { "epoch": 3.2871376811594204, "grad_norm": 0.06359490846013957, "learning_rate": 2.534936522599118e-05, "loss": 0.9125, "step": 7258 }, { "epoch": 3.287590579710145, "grad_norm": 0.05900581404530145, "learning_rate": 2.533759750382265e-05, "loss": 0.9363, "step": 7259 }, { "epoch": 3.2880434782608696, "grad_norm": 0.06770884255799611, "learning_rate": 2.5325831247477747e-05, "loss": 0.9195, "step": 7260 }, { "epoch": 3.2884963768115942, "grad_norm": 0.06189024898973494, "learning_rate": 2.5314066458132737e-05, "loss": 0.9218, "step": 7261 }, { "epoch": 3.288949275362319, "grad_norm": 0.0718132902934166, "learning_rate": 2.5302303136963782e-05, "loss": 0.9234, "step": 7262 }, { "epoch": 3.2894021739130435, "grad_norm": 0.06669183970877171, "learning_rate": 2.5290541285146862e-05, "loss": 0.9083, "step": 7263 }, { "epoch": 3.289855072463768, "grad_norm": 0.07569596149075773, "learning_rate": 2.527878090385785e-05, "loss": 0.9242, "step": 7264 }, { "epoch": 3.2903079710144927, "grad_norm": 0.06177416762216088, "learning_rate": 2.5267021994272435e-05, "loss": 0.9329, "step": 7265 }, { "epoch": 3.2907608695652173, "grad_norm": 0.06918951793056624, "learning_rate": 2.5255264557566187e-05, "loss": 0.9189, "step": 7266 }, { "epoch": 3.291213768115942, "grad_norm": 0.0678429178857842, "learning_rate": 2.524350859491448e-05, "loss": 0.9325, "step": 7267 }, { "epoch": 3.2916666666666665, "grad_norm": 0.06219379250275381, "learning_rate": 2.523175410749263e-05, "loss": 0.9177, "step": 7268 }, { "epoch": 3.292119565217391, "grad_norm": 0.05991772338927805, "learning_rate": 2.5220001096475703e-05, "loss": 0.9246, "step": 7269 }, { "epoch": 3.292572463768116, "grad_norm": 0.0635687463643046, "learning_rate": 2.5208249563038712e-05, "loss": 0.957, "step": 7270 }, { "epoch": 3.293025362318841, "grad_norm": 0.06542195655883812, "learning_rate": 2.5196499508356428e-05, "loss": 0.9339, "step": 7271 }, { "epoch": 3.2934782608695654, "grad_norm": 0.06943531476692888, "learning_rate": 2.518475093360356e-05, "loss": 0.9241, "step": 7272 }, { "epoch": 3.29393115942029, "grad_norm": 0.0673005550390779, "learning_rate": 2.51730038399546e-05, "loss": 0.9169, "step": 7273 }, { "epoch": 3.2943840579710146, "grad_norm": 0.07431211548694656, "learning_rate": 2.5161258228583964e-05, "loss": 0.9288, "step": 7274 }, { "epoch": 3.2948369565217392, "grad_norm": 0.07060448389593318, "learning_rate": 2.514951410066585e-05, "loss": 0.936, "step": 7275 }, { "epoch": 3.295289855072464, "grad_norm": 0.06672890222021365, "learning_rate": 2.5137771457374345e-05, "loss": 0.9147, "step": 7276 }, { "epoch": 3.2957427536231885, "grad_norm": 0.07824284896534042, "learning_rate": 2.5126030299883367e-05, "loss": 0.9219, "step": 7277 }, { "epoch": 3.296195652173913, "grad_norm": 0.061974910016659245, "learning_rate": 2.5114290629366727e-05, "loss": 0.9064, "step": 7278 }, { "epoch": 3.2966485507246377, "grad_norm": 0.08006564149801863, "learning_rate": 2.5102552446998027e-05, "loss": 0.919, "step": 7279 }, { "epoch": 3.2971014492753623, "grad_norm": 0.07035790621138752, "learning_rate": 2.5090815753950777e-05, "loss": 0.9533, "step": 7280 }, { "epoch": 3.297554347826087, "grad_norm": 0.09168986867652561, "learning_rate": 2.5079080551398286e-05, "loss": 0.9199, "step": 7281 }, { "epoch": 3.2980072463768115, "grad_norm": 0.07534079681198742, "learning_rate": 2.5067346840513774e-05, "loss": 0.9157, "step": 7282 }, { "epoch": 3.298460144927536, "grad_norm": 0.06910499125388604, "learning_rate": 2.5055614622470252e-05, "loss": 0.927, "step": 7283 }, { "epoch": 3.2989130434782608, "grad_norm": 0.07563468761649274, "learning_rate": 2.5043883898440626e-05, "loss": 0.9478, "step": 7284 }, { "epoch": 3.2993659420289854, "grad_norm": 0.06748737415305708, "learning_rate": 2.503215466959761e-05, "loss": 0.9509, "step": 7285 }, { "epoch": 3.29981884057971, "grad_norm": 0.06725114623206975, "learning_rate": 2.5020426937113827e-05, "loss": 0.9251, "step": 7286 }, { "epoch": 3.3002717391304346, "grad_norm": 0.06919161312002756, "learning_rate": 2.500870070216169e-05, "loss": 0.9205, "step": 7287 }, { "epoch": 3.300724637681159, "grad_norm": 0.06803879162493545, "learning_rate": 2.499697596591351e-05, "loss": 0.9336, "step": 7288 }, { "epoch": 3.301177536231884, "grad_norm": 0.08039567867635145, "learning_rate": 2.49852527295414e-05, "loss": 0.9348, "step": 7289 }, { "epoch": 3.301630434782609, "grad_norm": 0.07620072930465997, "learning_rate": 2.497353099421738e-05, "loss": 0.9268, "step": 7290 }, { "epoch": 3.3020833333333335, "grad_norm": 0.07568513006581347, "learning_rate": 2.4961810761113282e-05, "loss": 0.9156, "step": 7291 }, { "epoch": 3.302536231884058, "grad_norm": 0.08344347078273091, "learning_rate": 2.4950092031400787e-05, "loss": 0.9323, "step": 7292 }, { "epoch": 3.3029891304347827, "grad_norm": 0.06580055393717946, "learning_rate": 2.4938374806251432e-05, "loss": 0.9328, "step": 7293 }, { "epoch": 3.3034420289855073, "grad_norm": 0.0697076793336959, "learning_rate": 2.4926659086836623e-05, "loss": 0.9244, "step": 7294 }, { "epoch": 3.303894927536232, "grad_norm": 0.06710892478958463, "learning_rate": 2.4914944874327593e-05, "loss": 0.9146, "step": 7295 }, { "epoch": 3.3043478260869565, "grad_norm": 0.06276483422670774, "learning_rate": 2.490323216989543e-05, "loss": 0.9414, "step": 7296 }, { "epoch": 3.304800724637681, "grad_norm": 0.06010814091691937, "learning_rate": 2.4891520974711053e-05, "loss": 0.9171, "step": 7297 }, { "epoch": 3.3052536231884058, "grad_norm": 0.06589221386452311, "learning_rate": 2.4879811289945285e-05, "loss": 0.9277, "step": 7298 }, { "epoch": 3.3057065217391304, "grad_norm": 0.0638755438062963, "learning_rate": 2.486810311676873e-05, "loss": 0.9417, "step": 7299 }, { "epoch": 3.306159420289855, "grad_norm": 0.0646052374324446, "learning_rate": 2.48563964563519e-05, "loss": 0.9082, "step": 7300 }, { "epoch": 3.3066123188405796, "grad_norm": 0.056748588955909215, "learning_rate": 2.484469130986509e-05, "loss": 0.9285, "step": 7301 }, { "epoch": 3.307065217391304, "grad_norm": 0.06601868876613481, "learning_rate": 2.4832987678478526e-05, "loss": 0.9228, "step": 7302 }, { "epoch": 3.307518115942029, "grad_norm": 0.061454260662862956, "learning_rate": 2.482128556336221e-05, "loss": 0.9227, "step": 7303 }, { "epoch": 3.3079710144927534, "grad_norm": 0.06262197759833453, "learning_rate": 2.4809584965686034e-05, "loss": 0.9303, "step": 7304 }, { "epoch": 3.3084239130434785, "grad_norm": 0.0613857131219125, "learning_rate": 2.479788588661971e-05, "loss": 0.9338, "step": 7305 }, { "epoch": 3.308876811594203, "grad_norm": 0.06006708677769847, "learning_rate": 2.4786188327332834e-05, "loss": 0.9288, "step": 7306 }, { "epoch": 3.3093297101449277, "grad_norm": 0.057219579657198344, "learning_rate": 2.4774492288994814e-05, "loss": 0.9092, "step": 7307 }, { "epoch": 3.3097826086956523, "grad_norm": 0.06845425534867687, "learning_rate": 2.4762797772774932e-05, "loss": 0.924, "step": 7308 }, { "epoch": 3.310235507246377, "grad_norm": 0.05656094459466774, "learning_rate": 2.4751104779842288e-05, "loss": 0.9165, "step": 7309 }, { "epoch": 3.3106884057971016, "grad_norm": 0.06400685263092368, "learning_rate": 2.4739413311365877e-05, "loss": 0.9329, "step": 7310 }, { "epoch": 3.311141304347826, "grad_norm": 0.05907410197389319, "learning_rate": 2.4727723368514487e-05, "loss": 0.9357, "step": 7311 }, { "epoch": 3.3115942028985508, "grad_norm": 0.07465261944865459, "learning_rate": 2.4716034952456814e-05, "loss": 0.9045, "step": 7312 }, { "epoch": 3.3120471014492754, "grad_norm": 0.06101076565894094, "learning_rate": 2.4704348064361322e-05, "loss": 0.9211, "step": 7313 }, { "epoch": 3.3125, "grad_norm": 0.08849688027582055, "learning_rate": 2.4692662705396412e-05, "loss": 0.9319, "step": 7314 }, { "epoch": 3.3129528985507246, "grad_norm": 0.06016214555957864, "learning_rate": 2.468097887673027e-05, "loss": 0.9249, "step": 7315 }, { "epoch": 3.3134057971014492, "grad_norm": 0.07767907339973135, "learning_rate": 2.466929657953092e-05, "loss": 0.9127, "step": 7316 }, { "epoch": 3.313858695652174, "grad_norm": 0.05940266522079515, "learning_rate": 2.4657615814966316e-05, "loss": 0.9273, "step": 7317 }, { "epoch": 3.3143115942028984, "grad_norm": 0.07690850700705312, "learning_rate": 2.4645936584204142e-05, "loss": 0.9278, "step": 7318 }, { "epoch": 3.314764492753623, "grad_norm": 0.060821604762223154, "learning_rate": 2.4634258888412036e-05, "loss": 0.9261, "step": 7319 }, { "epoch": 3.3152173913043477, "grad_norm": 0.06434208543877735, "learning_rate": 2.46225827287574e-05, "loss": 0.9456, "step": 7320 }, { "epoch": 3.3156702898550723, "grad_norm": 0.05825315906136469, "learning_rate": 2.461090810640756e-05, "loss": 0.9233, "step": 7321 }, { "epoch": 3.316123188405797, "grad_norm": 0.06444708135638479, "learning_rate": 2.459923502252961e-05, "loss": 0.8898, "step": 7322 }, { "epoch": 3.3165760869565215, "grad_norm": 0.0671608336035971, "learning_rate": 2.4587563478290547e-05, "loss": 0.944, "step": 7323 }, { "epoch": 3.3170289855072466, "grad_norm": 0.05949691944810059, "learning_rate": 2.4575893474857167e-05, "loss": 0.9304, "step": 7324 }, { "epoch": 3.317481884057971, "grad_norm": 0.06568490792335, "learning_rate": 2.4564225013396178e-05, "loss": 0.9335, "step": 7325 }, { "epoch": 3.317934782608696, "grad_norm": 0.059925226621096304, "learning_rate": 2.4552558095074062e-05, "loss": 0.9311, "step": 7326 }, { "epoch": 3.3183876811594204, "grad_norm": 0.07507206912000297, "learning_rate": 2.4540892721057205e-05, "loss": 0.9235, "step": 7327 }, { "epoch": 3.318840579710145, "grad_norm": 0.058193224479366444, "learning_rate": 2.452922889251178e-05, "loss": 0.9134, "step": 7328 }, { "epoch": 3.3192934782608696, "grad_norm": 0.06592239402188112, "learning_rate": 2.4517566610603877e-05, "loss": 0.9186, "step": 7329 }, { "epoch": 3.3197463768115942, "grad_norm": 0.06407128002414123, "learning_rate": 2.450590587649937e-05, "loss": 0.9048, "step": 7330 }, { "epoch": 3.320199275362319, "grad_norm": 0.06706653627942855, "learning_rate": 2.4494246691364e-05, "loss": 0.9276, "step": 7331 }, { "epoch": 3.3206521739130435, "grad_norm": 0.06866635828420248, "learning_rate": 2.4482589056363356e-05, "loss": 0.9208, "step": 7332 }, { "epoch": 3.321105072463768, "grad_norm": 0.05973442064411881, "learning_rate": 2.4470932972662886e-05, "loss": 0.9384, "step": 7333 }, { "epoch": 3.3215579710144927, "grad_norm": 0.06575953418583284, "learning_rate": 2.4459278441427844e-05, "loss": 0.9255, "step": 7334 }, { "epoch": 3.3220108695652173, "grad_norm": 0.06435356345425057, "learning_rate": 2.444762546382337e-05, "loss": 0.9266, "step": 7335 }, { "epoch": 3.322463768115942, "grad_norm": 0.06758353945873631, "learning_rate": 2.443597404101441e-05, "loss": 0.9436, "step": 7336 }, { "epoch": 3.3229166666666665, "grad_norm": 0.06677028648911505, "learning_rate": 2.4424324174165808e-05, "loss": 0.9118, "step": 7337 }, { "epoch": 3.323369565217391, "grad_norm": 0.062061141812603, "learning_rate": 2.4412675864442187e-05, "loss": 0.9201, "step": 7338 }, { "epoch": 3.323822463768116, "grad_norm": 0.06935158344580641, "learning_rate": 2.4401029113008064e-05, "loss": 0.949, "step": 7339 }, { "epoch": 3.324275362318841, "grad_norm": 0.06972029234295063, "learning_rate": 2.438938392102777e-05, "loss": 0.9196, "step": 7340 }, { "epoch": 3.3247282608695654, "grad_norm": 0.0679983923608412, "learning_rate": 2.4377740289665517e-05, "loss": 0.9281, "step": 7341 }, { "epoch": 3.32518115942029, "grad_norm": 0.06458631367436794, "learning_rate": 2.4366098220085314e-05, "loss": 0.9183, "step": 7342 }, { "epoch": 3.3256340579710146, "grad_norm": 0.06165774813639128, "learning_rate": 2.4354457713451055e-05, "loss": 0.9316, "step": 7343 }, { "epoch": 3.3260869565217392, "grad_norm": 0.06903696181059714, "learning_rate": 2.4342818770926437e-05, "loss": 0.9356, "step": 7344 }, { "epoch": 3.326539855072464, "grad_norm": 0.0681843797424961, "learning_rate": 2.433118139367505e-05, "loss": 0.9237, "step": 7345 }, { "epoch": 3.3269927536231885, "grad_norm": 0.06337931059575796, "learning_rate": 2.4319545582860274e-05, "loss": 0.9122, "step": 7346 }, { "epoch": 3.327445652173913, "grad_norm": 0.07600167411519018, "learning_rate": 2.4307911339645394e-05, "loss": 0.9455, "step": 7347 }, { "epoch": 3.3278985507246377, "grad_norm": 0.062071508165965135, "learning_rate": 2.4296278665193462e-05, "loss": 0.92, "step": 7348 }, { "epoch": 3.3283514492753623, "grad_norm": 0.07467571752880481, "learning_rate": 2.4284647560667455e-05, "loss": 0.9268, "step": 7349 }, { "epoch": 3.328804347826087, "grad_norm": 0.0745553921664909, "learning_rate": 2.4273018027230126e-05, "loss": 0.9464, "step": 7350 }, { "epoch": 3.3292572463768115, "grad_norm": 0.05878387796943492, "learning_rate": 2.426139006604411e-05, "loss": 0.9356, "step": 7351 }, { "epoch": 3.329710144927536, "grad_norm": 0.06601173360947472, "learning_rate": 2.424976367827186e-05, "loss": 0.9088, "step": 7352 }, { "epoch": 3.3301630434782608, "grad_norm": 0.06850087722559238, "learning_rate": 2.4238138865075703e-05, "loss": 0.9075, "step": 7353 }, { "epoch": 3.3306159420289854, "grad_norm": 0.06196352891777641, "learning_rate": 2.422651562761777e-05, "loss": 0.9174, "step": 7354 }, { "epoch": 3.33106884057971, "grad_norm": 0.06818432126711974, "learning_rate": 2.421489396706008e-05, "loss": 0.9088, "step": 7355 }, { "epoch": 3.3315217391304346, "grad_norm": 0.07580015209869943, "learning_rate": 2.4203273884564437e-05, "loss": 0.927, "step": 7356 }, { "epoch": 3.331974637681159, "grad_norm": 0.05968388093525316, "learning_rate": 2.419165538129255e-05, "loss": 0.9269, "step": 7357 }, { "epoch": 3.332427536231884, "grad_norm": 0.07933182691527525, "learning_rate": 2.4180038458405913e-05, "loss": 0.9403, "step": 7358 }, { "epoch": 3.332880434782609, "grad_norm": 0.06433546753727971, "learning_rate": 2.4168423117065917e-05, "loss": 0.9134, "step": 7359 }, { "epoch": 3.3333333333333335, "grad_norm": 0.0800642391767151, "learning_rate": 2.4156809358433728e-05, "loss": 0.9506, "step": 7360 }, { "epoch": 3.333786231884058, "grad_norm": 0.06777228817329514, "learning_rate": 2.4145197183670433e-05, "loss": 0.927, "step": 7361 }, { "epoch": 3.3342391304347827, "grad_norm": 0.06467829730680193, "learning_rate": 2.413358659393688e-05, "loss": 0.9247, "step": 7362 }, { "epoch": 3.3346920289855073, "grad_norm": 0.06230407182326409, "learning_rate": 2.4121977590393834e-05, "loss": 0.9403, "step": 7363 }, { "epoch": 3.335144927536232, "grad_norm": 0.06770963634674863, "learning_rate": 2.411037017420184e-05, "loss": 0.9032, "step": 7364 }, { "epoch": 3.3355978260869565, "grad_norm": 0.06411430017149367, "learning_rate": 2.409876434652133e-05, "loss": 0.9285, "step": 7365 }, { "epoch": 3.336050724637681, "grad_norm": 0.07006047829667353, "learning_rate": 2.4087160108512526e-05, "loss": 0.9036, "step": 7366 }, { "epoch": 3.3365036231884058, "grad_norm": 0.07382910832802671, "learning_rate": 2.4075557461335562e-05, "loss": 0.9148, "step": 7367 }, { "epoch": 3.3369565217391304, "grad_norm": 0.06959362173673822, "learning_rate": 2.4063956406150345e-05, "loss": 0.9272, "step": 7368 }, { "epoch": 3.337409420289855, "grad_norm": 0.06501055705730724, "learning_rate": 2.4052356944116665e-05, "loss": 0.9523, "step": 7369 }, { "epoch": 3.3378623188405796, "grad_norm": 0.07926201149577194, "learning_rate": 2.4040759076394115e-05, "loss": 0.9368, "step": 7370 }, { "epoch": 3.338315217391304, "grad_norm": 0.06517439859949185, "learning_rate": 2.402916280414219e-05, "loss": 0.9216, "step": 7371 }, { "epoch": 3.338768115942029, "grad_norm": 0.06444547843272011, "learning_rate": 2.4017568128520158e-05, "loss": 0.9006, "step": 7372 }, { "epoch": 3.3392210144927534, "grad_norm": 0.07140530003792261, "learning_rate": 2.4005975050687172e-05, "loss": 0.9371, "step": 7373 }, { "epoch": 3.3396739130434785, "grad_norm": 0.06201125781369861, "learning_rate": 2.399438357180219e-05, "loss": 0.9341, "step": 7374 }, { "epoch": 3.340126811594203, "grad_norm": 0.07115711780044312, "learning_rate": 2.3982793693024063e-05, "loss": 0.9103, "step": 7375 }, { "epoch": 3.3405797101449277, "grad_norm": 0.06529833218912856, "learning_rate": 2.397120541551142e-05, "loss": 0.9294, "step": 7376 }, { "epoch": 3.3410326086956523, "grad_norm": 0.06131081352886268, "learning_rate": 2.3959618740422785e-05, "loss": 0.9278, "step": 7377 }, { "epoch": 3.341485507246377, "grad_norm": 0.06938348973271441, "learning_rate": 2.3948033668916466e-05, "loss": 0.9332, "step": 7378 }, { "epoch": 3.3419384057971016, "grad_norm": 0.05827533238280592, "learning_rate": 2.393645020215067e-05, "loss": 0.912, "step": 7379 }, { "epoch": 3.342391304347826, "grad_norm": 0.07180793116750991, "learning_rate": 2.39248683412834e-05, "loss": 0.927, "step": 7380 }, { "epoch": 3.3428442028985508, "grad_norm": 0.07188149900562103, "learning_rate": 2.3913288087472514e-05, "loss": 0.9444, "step": 7381 }, { "epoch": 3.3432971014492754, "grad_norm": 0.06121655463552024, "learning_rate": 2.3901709441875696e-05, "loss": 0.911, "step": 7382 }, { "epoch": 3.34375, "grad_norm": 0.07230312678275883, "learning_rate": 2.389013240565052e-05, "loss": 0.933, "step": 7383 }, { "epoch": 3.3442028985507246, "grad_norm": 0.0778980749448399, "learning_rate": 2.3878556979954316e-05, "loss": 0.8787, "step": 7384 }, { "epoch": 3.3446557971014492, "grad_norm": 0.07181283169953892, "learning_rate": 2.3866983165944336e-05, "loss": 0.9335, "step": 7385 }, { "epoch": 3.345108695652174, "grad_norm": 0.08876296072144566, "learning_rate": 2.3855410964777587e-05, "loss": 0.9381, "step": 7386 }, { "epoch": 3.3455615942028984, "grad_norm": 0.06608742306810236, "learning_rate": 2.3843840377611008e-05, "loss": 0.9238, "step": 7387 }, { "epoch": 3.346014492753623, "grad_norm": 0.08109555252302353, "learning_rate": 2.3832271405601303e-05, "loss": 0.9335, "step": 7388 }, { "epoch": 3.3464673913043477, "grad_norm": 0.07674834450136843, "learning_rate": 2.382070404990505e-05, "loss": 0.9465, "step": 7389 }, { "epoch": 3.3469202898550723, "grad_norm": 0.07322668961638257, "learning_rate": 2.3809138311678637e-05, "loss": 0.954, "step": 7390 }, { "epoch": 3.347373188405797, "grad_norm": 0.07773998987498126, "learning_rate": 2.3797574192078337e-05, "loss": 0.9281, "step": 7391 }, { "epoch": 3.3478260869565215, "grad_norm": 0.08173464227295979, "learning_rate": 2.3786011692260216e-05, "loss": 0.9247, "step": 7392 }, { "epoch": 3.3482789855072466, "grad_norm": 0.0776986548554712, "learning_rate": 2.3774450813380207e-05, "loss": 0.9243, "step": 7393 }, { "epoch": 3.348731884057971, "grad_norm": 0.08346876700857708, "learning_rate": 2.3762891556594045e-05, "loss": 0.9448, "step": 7394 }, { "epoch": 3.349184782608696, "grad_norm": 0.070362913916102, "learning_rate": 2.3751333923057357e-05, "loss": 0.9252, "step": 7395 }, { "epoch": 3.3496376811594204, "grad_norm": 0.08823195931791501, "learning_rate": 2.373977791392556e-05, "loss": 0.9323, "step": 7396 }, { "epoch": 3.350090579710145, "grad_norm": 0.06905918550594836, "learning_rate": 2.372822353035394e-05, "loss": 0.9297, "step": 7397 }, { "epoch": 3.3505434782608696, "grad_norm": 0.08146556281578984, "learning_rate": 2.371667077349758e-05, "loss": 0.9319, "step": 7398 }, { "epoch": 3.3509963768115942, "grad_norm": 0.07140678843025883, "learning_rate": 2.3705119644511462e-05, "loss": 0.9351, "step": 7399 }, { "epoch": 3.351449275362319, "grad_norm": 0.06911707220688597, "learning_rate": 2.3693570144550348e-05, "loss": 0.9288, "step": 7400 }, { "epoch": 3.3519021739130435, "grad_norm": 0.06199132357128153, "learning_rate": 2.3682022274768874e-05, "loss": 0.9031, "step": 7401 }, { "epoch": 3.352355072463768, "grad_norm": 0.06383317730518259, "learning_rate": 2.3670476036321475e-05, "loss": 0.921, "step": 7402 }, { "epoch": 3.3528079710144927, "grad_norm": 0.07151211440298516, "learning_rate": 2.3658931430362478e-05, "loss": 0.9226, "step": 7403 }, { "epoch": 3.3532608695652173, "grad_norm": 0.0654006043654492, "learning_rate": 2.364738845804599e-05, "loss": 0.9102, "step": 7404 }, { "epoch": 3.353713768115942, "grad_norm": 0.078348873742451, "learning_rate": 2.3635847120525995e-05, "loss": 0.9162, "step": 7405 }, { "epoch": 3.3541666666666665, "grad_norm": 0.07103394418647212, "learning_rate": 2.3624307418956298e-05, "loss": 0.9305, "step": 7406 }, { "epoch": 3.354619565217391, "grad_norm": 0.07110878680814235, "learning_rate": 2.3612769354490546e-05, "loss": 0.9209, "step": 7407 }, { "epoch": 3.355072463768116, "grad_norm": 0.07593485461310721, "learning_rate": 2.3601232928282185e-05, "loss": 0.9328, "step": 7408 }, { "epoch": 3.355525362318841, "grad_norm": 0.06171736805694523, "learning_rate": 2.3589698141484575e-05, "loss": 0.9461, "step": 7409 }, { "epoch": 3.3559782608695654, "grad_norm": 0.06433243799174848, "learning_rate": 2.3578164995250837e-05, "loss": 0.9235, "step": 7410 }, { "epoch": 3.35643115942029, "grad_norm": 0.06501493386325843, "learning_rate": 2.3566633490733982e-05, "loss": 0.9417, "step": 7411 }, { "epoch": 3.3568840579710146, "grad_norm": 0.0603252143786333, "learning_rate": 2.355510362908679e-05, "loss": 0.9193, "step": 7412 }, { "epoch": 3.3573369565217392, "grad_norm": 0.07146282927333616, "learning_rate": 2.3543575411461966e-05, "loss": 0.9288, "step": 7413 }, { "epoch": 3.357789855072464, "grad_norm": 0.06367003079162474, "learning_rate": 2.3532048839011985e-05, "loss": 0.9132, "step": 7414 }, { "epoch": 3.3582427536231885, "grad_norm": 0.06395783916426545, "learning_rate": 2.3520523912889185e-05, "loss": 0.9473, "step": 7415 }, { "epoch": 3.358695652173913, "grad_norm": 0.0648997546754011, "learning_rate": 2.35090006342457e-05, "loss": 0.922, "step": 7416 }, { "epoch": 3.3591485507246377, "grad_norm": 0.06258932986411901, "learning_rate": 2.349747900423357e-05, "loss": 0.8971, "step": 7417 }, { "epoch": 3.3596014492753623, "grad_norm": 0.3407338950044553, "learning_rate": 2.3485959024004612e-05, "loss": 0.9569, "step": 7418 }, { "epoch": 3.360054347826087, "grad_norm": 0.06924319246461391, "learning_rate": 2.3474440694710507e-05, "loss": 0.9015, "step": 7419 }, { "epoch": 3.3605072463768115, "grad_norm": 0.06475451604554787, "learning_rate": 2.3462924017502734e-05, "loss": 0.928, "step": 7420 }, { "epoch": 3.360960144927536, "grad_norm": 0.06905348625410379, "learning_rate": 2.3451408993532665e-05, "loss": 0.9085, "step": 7421 }, { "epoch": 3.3614130434782608, "grad_norm": 0.07000211840472084, "learning_rate": 2.3439895623951464e-05, "loss": 0.9384, "step": 7422 }, { "epoch": 3.3618659420289854, "grad_norm": 0.06113422650674196, "learning_rate": 2.3428383909910135e-05, "loss": 0.9224, "step": 7423 }, { "epoch": 3.36231884057971, "grad_norm": 0.07085798112601602, "learning_rate": 2.341687385255952e-05, "loss": 0.962, "step": 7424 }, { "epoch": 3.3627717391304346, "grad_norm": 0.05913369509751246, "learning_rate": 2.3405365453050316e-05, "loss": 0.9207, "step": 7425 }, { "epoch": 3.363224637681159, "grad_norm": 0.1922987432633098, "learning_rate": 2.3393858712533017e-05, "loss": 0.9579, "step": 7426 }, { "epoch": 3.363677536231884, "grad_norm": 0.06525678071060478, "learning_rate": 2.3382353632157985e-05, "loss": 0.9335, "step": 7427 }, { "epoch": 3.364130434782609, "grad_norm": 0.07497854559562635, "learning_rate": 2.3370850213075376e-05, "loss": 0.9324, "step": 7428 }, { "epoch": 3.3645833333333335, "grad_norm": 0.07515143490408961, "learning_rate": 2.3359348456435243e-05, "loss": 0.9568, "step": 7429 }, { "epoch": 3.365036231884058, "grad_norm": 0.08055512737029139, "learning_rate": 2.3347848363387402e-05, "loss": 0.9289, "step": 7430 }, { "epoch": 3.3654891304347827, "grad_norm": 0.07083023846859213, "learning_rate": 2.333634993508156e-05, "loss": 0.9378, "step": 7431 }, { "epoch": 3.3659420289855073, "grad_norm": 0.06923005183096252, "learning_rate": 2.3324853172667198e-05, "loss": 0.9103, "step": 7432 }, { "epoch": 3.366394927536232, "grad_norm": 0.06675196694826942, "learning_rate": 2.331335807729371e-05, "loss": 0.9361, "step": 7433 }, { "epoch": 3.3668478260869565, "grad_norm": 0.0757500945566563, "learning_rate": 2.330186465011024e-05, "loss": 0.9357, "step": 7434 }, { "epoch": 3.367300724637681, "grad_norm": 0.06239759052062186, "learning_rate": 2.3290372892265838e-05, "loss": 0.9291, "step": 7435 }, { "epoch": 3.3677536231884058, "grad_norm": 0.07723029049974646, "learning_rate": 2.3278882804909334e-05, "loss": 0.926, "step": 7436 }, { "epoch": 3.3682065217391304, "grad_norm": 0.05800601719371266, "learning_rate": 2.3267394389189414e-05, "loss": 0.9176, "step": 7437 }, { "epoch": 3.368659420289855, "grad_norm": 0.07283708689651255, "learning_rate": 2.3255907646254573e-05, "loss": 0.9115, "step": 7438 }, { "epoch": 3.3691123188405796, "grad_norm": 0.05692828290680373, "learning_rate": 2.3244422577253198e-05, "loss": 0.9251, "step": 7439 }, { "epoch": 3.369565217391304, "grad_norm": 0.07614852058463617, "learning_rate": 2.323293918333345e-05, "loss": 0.9115, "step": 7440 }, { "epoch": 3.370018115942029, "grad_norm": 0.05686164837573893, "learning_rate": 2.3221457465643322e-05, "loss": 0.9286, "step": 7441 }, { "epoch": 3.3704710144927534, "grad_norm": 0.0695481577598154, "learning_rate": 2.32099774253307e-05, "loss": 0.9193, "step": 7442 }, { "epoch": 3.3709239130434785, "grad_norm": 0.06024556927999754, "learning_rate": 2.319849906354322e-05, "loss": 0.915, "step": 7443 }, { "epoch": 3.371376811594203, "grad_norm": 0.06502729506855086, "learning_rate": 2.3187022381428433e-05, "loss": 0.9072, "step": 7444 }, { "epoch": 3.3718297101449277, "grad_norm": 0.06317880431443261, "learning_rate": 2.3175547380133656e-05, "loss": 0.9033, "step": 7445 }, { "epoch": 3.3722826086956523, "grad_norm": 0.06948545120833707, "learning_rate": 2.3164074060806076e-05, "loss": 0.934, "step": 7446 }, { "epoch": 3.372735507246377, "grad_norm": 0.07522118973260815, "learning_rate": 2.3152602424592664e-05, "loss": 0.939, "step": 7447 }, { "epoch": 3.3731884057971016, "grad_norm": 0.07171837095124076, "learning_rate": 2.3141132472640307e-05, "loss": 0.9465, "step": 7448 }, { "epoch": 3.373641304347826, "grad_norm": 0.06113108385534525, "learning_rate": 2.3129664206095637e-05, "loss": 0.9301, "step": 7449 }, { "epoch": 3.3740942028985508, "grad_norm": 0.06386758041772814, "learning_rate": 2.3118197626105183e-05, "loss": 0.9401, "step": 7450 }, { "epoch": 3.3745471014492754, "grad_norm": 0.0777512085625092, "learning_rate": 2.310673273381525e-05, "loss": 0.9246, "step": 7451 }, { "epoch": 3.375, "grad_norm": 0.06194987867396691, "learning_rate": 2.3095269530372032e-05, "loss": 0.8969, "step": 7452 }, { "epoch": 3.3754528985507246, "grad_norm": 0.07230414959961132, "learning_rate": 2.3083808016921485e-05, "loss": 0.9429, "step": 7453 }, { "epoch": 3.3759057971014492, "grad_norm": 0.06860271604323757, "learning_rate": 2.3072348194609476e-05, "loss": 0.9296, "step": 7454 }, { "epoch": 3.376358695652174, "grad_norm": 0.0698135375110878, "learning_rate": 2.3060890064581645e-05, "loss": 0.903, "step": 7455 }, { "epoch": 3.3768115942028984, "grad_norm": 0.06878069806593005, "learning_rate": 2.304943362798347e-05, "loss": 0.942, "step": 7456 }, { "epoch": 3.377264492753623, "grad_norm": 0.06622828411669955, "learning_rate": 2.3037978885960266e-05, "loss": 0.9211, "step": 7457 }, { "epoch": 3.3777173913043477, "grad_norm": 0.06733048612698787, "learning_rate": 2.3026525839657207e-05, "loss": 0.9185, "step": 7458 }, { "epoch": 3.3781702898550723, "grad_norm": 0.06327079110314296, "learning_rate": 2.301507449021924e-05, "loss": 0.9127, "step": 7459 }, { "epoch": 3.378623188405797, "grad_norm": 0.0773938363484948, "learning_rate": 2.300362483879121e-05, "loss": 0.9162, "step": 7460 }, { "epoch": 3.3790760869565215, "grad_norm": 0.05945117285987916, "learning_rate": 2.299217688651771e-05, "loss": 0.9159, "step": 7461 }, { "epoch": 3.3795289855072466, "grad_norm": 0.08620195922219676, "learning_rate": 2.2980730634543264e-05, "loss": 0.9352, "step": 7462 }, { "epoch": 3.379981884057971, "grad_norm": 0.07199310510000398, "learning_rate": 2.2969286084012142e-05, "loss": 0.9186, "step": 7463 }, { "epoch": 3.380434782608696, "grad_norm": 0.07241966028176855, "learning_rate": 2.295784323606848e-05, "loss": 0.9158, "step": 7464 }, { "epoch": 3.3808876811594204, "grad_norm": 0.0759726663120944, "learning_rate": 2.294640209185622e-05, "loss": 0.9219, "step": 7465 }, { "epoch": 3.381340579710145, "grad_norm": 0.06614807611212682, "learning_rate": 2.2934962652519182e-05, "loss": 0.9318, "step": 7466 }, { "epoch": 3.3817934782608696, "grad_norm": 0.06802406639491235, "learning_rate": 2.292352491920095e-05, "loss": 0.9448, "step": 7467 }, { "epoch": 3.3822463768115942, "grad_norm": 0.06067463167974688, "learning_rate": 2.2912088893045015e-05, "loss": 0.926, "step": 7468 }, { "epoch": 3.382699275362319, "grad_norm": 0.057491707279992796, "learning_rate": 2.290065457519461e-05, "loss": 0.9265, "step": 7469 }, { "epoch": 3.3831521739130435, "grad_norm": 0.06997626627336861, "learning_rate": 2.2889221966792877e-05, "loss": 0.9364, "step": 7470 }, { "epoch": 3.383605072463768, "grad_norm": 0.061781016428576765, "learning_rate": 2.287779106898274e-05, "loss": 0.9268, "step": 7471 }, { "epoch": 3.3840579710144927, "grad_norm": 0.06818984642997995, "learning_rate": 2.2866361882906966e-05, "loss": 0.9595, "step": 7472 }, { "epoch": 3.3845108695652173, "grad_norm": 0.06363302822869998, "learning_rate": 2.2854934409708126e-05, "loss": 0.9243, "step": 7473 }, { "epoch": 3.384963768115942, "grad_norm": 0.06753910142319743, "learning_rate": 2.284350865052868e-05, "loss": 0.9311, "step": 7474 }, { "epoch": 3.3854166666666665, "grad_norm": 0.060642463644924406, "learning_rate": 2.2832084606510848e-05, "loss": 0.9167, "step": 7475 }, { "epoch": 3.385869565217391, "grad_norm": 0.06257331094493497, "learning_rate": 2.282066227879673e-05, "loss": 0.9361, "step": 7476 }, { "epoch": 3.386322463768116, "grad_norm": 0.06519298464982974, "learning_rate": 2.2809241668528218e-05, "loss": 0.9604, "step": 7477 }, { "epoch": 3.386775362318841, "grad_norm": 0.054415869261035386, "learning_rate": 2.279782277684707e-05, "loss": 0.9379, "step": 7478 }, { "epoch": 3.3872282608695654, "grad_norm": 0.06763212953303382, "learning_rate": 2.2786405604894837e-05, "loss": 0.9141, "step": 7479 }, { "epoch": 3.38768115942029, "grad_norm": 0.05565546089509831, "learning_rate": 2.2774990153812917e-05, "loss": 0.9173, "step": 7480 }, { "epoch": 3.3881340579710146, "grad_norm": 0.06705715755994489, "learning_rate": 2.2763576424742507e-05, "loss": 0.9347, "step": 7481 }, { "epoch": 3.3885869565217392, "grad_norm": 0.05905293730281881, "learning_rate": 2.2752164418824697e-05, "loss": 0.9137, "step": 7482 }, { "epoch": 3.389039855072464, "grad_norm": 0.06031029527967085, "learning_rate": 2.274075413720032e-05, "loss": 0.9002, "step": 7483 }, { "epoch": 3.3894927536231885, "grad_norm": 0.08093107008059619, "learning_rate": 2.272934558101012e-05, "loss": 0.9199, "step": 7484 }, { "epoch": 3.389945652173913, "grad_norm": 0.054474445488031376, "learning_rate": 2.2717938751394595e-05, "loss": 0.9177, "step": 7485 }, { "epoch": 3.3903985507246377, "grad_norm": 0.0677714736927272, "learning_rate": 2.2706533649494134e-05, "loss": 0.9009, "step": 7486 }, { "epoch": 3.3908514492753623, "grad_norm": 0.05479732573260838, "learning_rate": 2.2695130276448905e-05, "loss": 0.9226, "step": 7487 }, { "epoch": 3.391304347826087, "grad_norm": 0.05906452593864538, "learning_rate": 2.2683728633398926e-05, "loss": 0.9221, "step": 7488 }, { "epoch": 3.3917572463768115, "grad_norm": 0.05848087658025871, "learning_rate": 2.267232872148402e-05, "loss": 0.9296, "step": 7489 }, { "epoch": 3.392210144927536, "grad_norm": 0.05935364888269224, "learning_rate": 2.2660930541843884e-05, "loss": 0.9222, "step": 7490 }, { "epoch": 3.3926630434782608, "grad_norm": 0.056012755001513044, "learning_rate": 2.2649534095617982e-05, "loss": 0.945, "step": 7491 }, { "epoch": 3.3931159420289854, "grad_norm": 0.05725576781097115, "learning_rate": 2.263813938394567e-05, "loss": 0.9435, "step": 7492 }, { "epoch": 3.39356884057971, "grad_norm": 0.05529142582850261, "learning_rate": 2.2626746407966057e-05, "loss": 0.8969, "step": 7493 }, { "epoch": 3.3940217391304346, "grad_norm": 0.05557839585511233, "learning_rate": 2.2615355168818153e-05, "loss": 0.9404, "step": 7494 }, { "epoch": 3.394474637681159, "grad_norm": 0.052178583739265774, "learning_rate": 2.2603965667640724e-05, "loss": 0.8997, "step": 7495 }, { "epoch": 3.394927536231884, "grad_norm": 0.0622911972517291, "learning_rate": 2.2592577905572433e-05, "loss": 0.9367, "step": 7496 }, { "epoch": 3.395380434782609, "grad_norm": 0.05837733735330746, "learning_rate": 2.258119188375171e-05, "loss": 0.9233, "step": 7497 }, { "epoch": 3.3958333333333335, "grad_norm": 0.06123131578417694, "learning_rate": 2.2569807603316836e-05, "loss": 0.9218, "step": 7498 }, { "epoch": 3.396286231884058, "grad_norm": 0.061011523412012506, "learning_rate": 2.2558425065405905e-05, "loss": 0.939, "step": 7499 }, { "epoch": 3.3967391304347827, "grad_norm": 0.05771725563898383, "learning_rate": 2.2547044271156873e-05, "loss": 0.9186, "step": 7500 }, { "epoch": 3.3971920289855073, "grad_norm": 0.057194000416111904, "learning_rate": 2.253566522170746e-05, "loss": 0.9147, "step": 7501 }, { "epoch": 3.397644927536232, "grad_norm": 0.07024917101320764, "learning_rate": 2.2524287918195294e-05, "loss": 0.9182, "step": 7502 }, { "epoch": 3.3980978260869565, "grad_norm": 0.0639153848843653, "learning_rate": 2.2512912361757735e-05, "loss": 0.9201, "step": 7503 }, { "epoch": 3.398550724637681, "grad_norm": 0.06248223454259265, "learning_rate": 2.2501538553532055e-05, "loss": 0.9185, "step": 7504 }, { "epoch": 3.3990036231884058, "grad_norm": 0.06411535486808516, "learning_rate": 2.2490166494655295e-05, "loss": 0.9221, "step": 7505 }, { "epoch": 3.3994565217391304, "grad_norm": 0.0611488851601155, "learning_rate": 2.247879618626434e-05, "loss": 0.9072, "step": 7506 }, { "epoch": 3.399909420289855, "grad_norm": 0.06033980627543244, "learning_rate": 2.2467427629495874e-05, "loss": 0.9214, "step": 7507 }, { "epoch": 3.4003623188405796, "grad_norm": 0.06084788988403923, "learning_rate": 2.2456060825486473e-05, "loss": 0.9357, "step": 7508 }, { "epoch": 3.400815217391304, "grad_norm": 0.05485843782497544, "learning_rate": 2.2444695775372454e-05, "loss": 0.9163, "step": 7509 }, { "epoch": 3.401268115942029, "grad_norm": 0.09171970805591947, "learning_rate": 2.243333248029003e-05, "loss": 0.9196, "step": 7510 }, { "epoch": 3.4017210144927534, "grad_norm": 0.0706990280375, "learning_rate": 2.2421970941375183e-05, "loss": 0.9189, "step": 7511 }, { "epoch": 3.4021739130434785, "grad_norm": 0.0758674337653084, "learning_rate": 2.2410611159763765e-05, "loss": 0.9373, "step": 7512 }, { "epoch": 3.402626811594203, "grad_norm": 0.07204862675957727, "learning_rate": 2.2399253136591427e-05, "loss": 0.9302, "step": 7513 }, { "epoch": 3.4030797101449277, "grad_norm": 0.07318741156464635, "learning_rate": 2.2387896872993645e-05, "loss": 0.9111, "step": 7514 }, { "epoch": 3.4035326086956523, "grad_norm": 0.06675134845503937, "learning_rate": 2.2376542370105706e-05, "loss": 0.904, "step": 7515 }, { "epoch": 3.403985507246377, "grad_norm": 0.06485920643099158, "learning_rate": 2.2365189629062764e-05, "loss": 0.9313, "step": 7516 }, { "epoch": 3.4044384057971016, "grad_norm": 0.06830819142195911, "learning_rate": 2.2353838650999745e-05, "loss": 0.9305, "step": 7517 }, { "epoch": 3.404891304347826, "grad_norm": 0.06801529129733597, "learning_rate": 2.2342489437051452e-05, "loss": 0.9139, "step": 7518 }, { "epoch": 3.4053442028985508, "grad_norm": 0.06366004918656742, "learning_rate": 2.2331141988352462e-05, "loss": 0.9213, "step": 7519 }, { "epoch": 3.4057971014492754, "grad_norm": 0.08371810113908425, "learning_rate": 2.2319796306037224e-05, "loss": 0.925, "step": 7520 }, { "epoch": 3.40625, "grad_norm": 0.061433052088693374, "learning_rate": 2.2308452391239958e-05, "loss": 0.8996, "step": 7521 }, { "epoch": 3.4067028985507246, "grad_norm": 0.07545652433602117, "learning_rate": 2.229711024509474e-05, "loss": 0.9137, "step": 7522 }, { "epoch": 3.4071557971014492, "grad_norm": 0.09641690007942433, "learning_rate": 2.228576986873545e-05, "loss": 0.9363, "step": 7523 }, { "epoch": 3.407608695652174, "grad_norm": 0.056017169582259115, "learning_rate": 2.2274431263295834e-05, "loss": 0.9398, "step": 7524 }, { "epoch": 3.4080615942028984, "grad_norm": 0.08244810688376449, "learning_rate": 2.2263094429909398e-05, "loss": 0.9198, "step": 7525 }, { "epoch": 3.408514492753623, "grad_norm": 0.061955127636126296, "learning_rate": 2.2251759369709536e-05, "loss": 0.9076, "step": 7526 }, { "epoch": 3.4089673913043477, "grad_norm": 0.08202996303202226, "learning_rate": 2.2240426083829396e-05, "loss": 0.9323, "step": 7527 }, { "epoch": 3.4094202898550723, "grad_norm": 0.08434811243700305, "learning_rate": 2.222909457340202e-05, "loss": 0.9264, "step": 7528 }, { "epoch": 3.409873188405797, "grad_norm": 0.06480473993568553, "learning_rate": 2.2217764839560213e-05, "loss": 0.9169, "step": 7529 }, { "epoch": 3.4103260869565215, "grad_norm": 0.08212953745145068, "learning_rate": 2.220643688343664e-05, "loss": 0.9077, "step": 7530 }, { "epoch": 3.4107789855072466, "grad_norm": 0.05405110053683581, "learning_rate": 2.2195110706163753e-05, "loss": 0.9104, "step": 7531 }, { "epoch": 3.411231884057971, "grad_norm": 0.08215510186915399, "learning_rate": 2.218378630887388e-05, "loss": 0.904, "step": 7532 }, { "epoch": 3.411684782608696, "grad_norm": 0.07539997150928653, "learning_rate": 2.2172463692699103e-05, "loss": 0.935, "step": 7533 }, { "epoch": 3.4121376811594204, "grad_norm": 0.07036517386715847, "learning_rate": 2.2161142858771393e-05, "loss": 0.8983, "step": 7534 }, { "epoch": 3.412590579710145, "grad_norm": 0.08146630502033049, "learning_rate": 2.2149823808222486e-05, "loss": 0.9098, "step": 7535 }, { "epoch": 3.4130434782608696, "grad_norm": 0.06449021987956462, "learning_rate": 2.2138506542184e-05, "loss": 0.9187, "step": 7536 }, { "epoch": 3.4134963768115942, "grad_norm": 0.07606053725251744, "learning_rate": 2.21271910617873e-05, "loss": 0.9186, "step": 7537 }, { "epoch": 3.413949275362319, "grad_norm": 0.07258609166540951, "learning_rate": 2.2115877368163643e-05, "loss": 0.8997, "step": 7538 }, { "epoch": 3.4144021739130435, "grad_norm": 0.057335971762000795, "learning_rate": 2.210456546244407e-05, "loss": 0.9599, "step": 7539 }, { "epoch": 3.414855072463768, "grad_norm": 0.08046545210671306, "learning_rate": 2.2093255345759443e-05, "loss": 0.903, "step": 7540 }, { "epoch": 3.4153079710144927, "grad_norm": 0.05847558661280076, "learning_rate": 2.2081947019240443e-05, "loss": 0.9407, "step": 7541 }, { "epoch": 3.4157608695652173, "grad_norm": 0.08859505295076912, "learning_rate": 2.2070640484017606e-05, "loss": 0.916, "step": 7542 }, { "epoch": 3.416213768115942, "grad_norm": 0.08568481035423746, "learning_rate": 2.2059335741221232e-05, "loss": 0.916, "step": 7543 }, { "epoch": 3.4166666666666665, "grad_norm": 0.09062630536817863, "learning_rate": 2.2048032791981515e-05, "loss": 0.9347, "step": 7544 }, { "epoch": 3.417119565217391, "grad_norm": 0.10002953642186597, "learning_rate": 2.2036731637428393e-05, "loss": 0.9314, "step": 7545 }, { "epoch": 3.417572463768116, "grad_norm": 0.07308073657131234, "learning_rate": 2.2025432278691687e-05, "loss": 0.9213, "step": 7546 }, { "epoch": 3.418025362318841, "grad_norm": 0.09564514844186026, "learning_rate": 2.201413471690099e-05, "loss": 0.9203, "step": 7547 }, { "epoch": 3.4184782608695654, "grad_norm": 0.07784930629616608, "learning_rate": 2.200283895318576e-05, "loss": 0.9248, "step": 7548 }, { "epoch": 3.41893115942029, "grad_norm": 0.07911482647949929, "learning_rate": 2.1991544988675218e-05, "loss": 0.9375, "step": 7549 }, { "epoch": 3.4193840579710146, "grad_norm": 0.10222835332278163, "learning_rate": 2.1980252824498475e-05, "loss": 0.9356, "step": 7550 }, { "epoch": 3.4198369565217392, "grad_norm": 0.08783252809515768, "learning_rate": 2.1968962461784398e-05, "loss": 0.9197, "step": 7551 }, { "epoch": 3.420289855072464, "grad_norm": 0.0910203492030103, "learning_rate": 2.1957673901661735e-05, "loss": 0.9102, "step": 7552 }, { "epoch": 3.4207427536231885, "grad_norm": 0.09540078587730316, "learning_rate": 2.1946387145258976e-05, "loss": 0.9352, "step": 7553 }, { "epoch": 3.421195652173913, "grad_norm": 0.07609092692620607, "learning_rate": 2.193510219370453e-05, "loss": 0.9378, "step": 7554 }, { "epoch": 3.4216485507246377, "grad_norm": 0.08851089488458724, "learning_rate": 2.1923819048126535e-05, "loss": 0.913, "step": 7555 }, { "epoch": 3.4221014492753623, "grad_norm": 0.061048730136588875, "learning_rate": 2.1912537709653004e-05, "loss": 0.9365, "step": 7556 }, { "epoch": 3.422554347826087, "grad_norm": 0.09628807780240974, "learning_rate": 2.1901258179411714e-05, "loss": 0.9258, "step": 7557 }, { "epoch": 3.4230072463768115, "grad_norm": 0.05828034195197701, "learning_rate": 2.1889980458530346e-05, "loss": 0.9174, "step": 7558 }, { "epoch": 3.423460144927536, "grad_norm": 0.08049982556998504, "learning_rate": 2.187870454813631e-05, "loss": 0.921, "step": 7559 }, { "epoch": 3.4239130434782608, "grad_norm": 0.0704023758531828, "learning_rate": 2.1867430449356917e-05, "loss": 0.901, "step": 7560 }, { "epoch": 3.4243659420289854, "grad_norm": 0.0730954407809109, "learning_rate": 2.1856158163319215e-05, "loss": 0.9173, "step": 7561 }, { "epoch": 3.42481884057971, "grad_norm": 0.0688756825982194, "learning_rate": 2.184488769115015e-05, "loss": 0.9153, "step": 7562 }, { "epoch": 3.4252717391304346, "grad_norm": 0.07026606481051964, "learning_rate": 2.1833619033976432e-05, "loss": 0.9068, "step": 7563 }, { "epoch": 3.425724637681159, "grad_norm": 0.06437792411148072, "learning_rate": 2.1822352192924603e-05, "loss": 0.9229, "step": 7564 }, { "epoch": 3.426177536231884, "grad_norm": 0.08321288297223019, "learning_rate": 2.1811087169121034e-05, "loss": 0.9161, "step": 7565 }, { "epoch": 3.426630434782609, "grad_norm": 0.0631515076831333, "learning_rate": 2.1799823963691892e-05, "loss": 0.9372, "step": 7566 }, { "epoch": 3.4270833333333335, "grad_norm": 0.08887126081524184, "learning_rate": 2.1788562577763192e-05, "loss": 0.93, "step": 7567 }, { "epoch": 3.427536231884058, "grad_norm": 0.06458364756866297, "learning_rate": 2.1777303012460743e-05, "loss": 0.9258, "step": 7568 }, { "epoch": 3.4279891304347827, "grad_norm": 0.0725554112133368, "learning_rate": 2.17660452689102e-05, "loss": 0.9238, "step": 7569 }, { "epoch": 3.4284420289855073, "grad_norm": 0.06425764617184929, "learning_rate": 2.1754789348236997e-05, "loss": 0.9498, "step": 7570 }, { "epoch": 3.428894927536232, "grad_norm": 0.0743982740395563, "learning_rate": 2.174353525156643e-05, "loss": 0.9201, "step": 7571 }, { "epoch": 3.4293478260869565, "grad_norm": 0.06667111806066978, "learning_rate": 2.1732282980023564e-05, "loss": 0.9147, "step": 7572 }, { "epoch": 3.429800724637681, "grad_norm": 0.060220641512127596, "learning_rate": 2.172103253473332e-05, "loss": 0.9486, "step": 7573 }, { "epoch": 3.4302536231884058, "grad_norm": 0.0582980938833892, "learning_rate": 2.1709783916820405e-05, "loss": 0.9365, "step": 7574 }, { "epoch": 3.4307065217391304, "grad_norm": 0.06332021913179943, "learning_rate": 2.1698537127409392e-05, "loss": 0.9134, "step": 7575 }, { "epoch": 3.431159420289855, "grad_norm": 0.0751046546881015, "learning_rate": 2.1687292167624605e-05, "loss": 0.9129, "step": 7576 }, { "epoch": 3.4316123188405796, "grad_norm": 0.06050769584036525, "learning_rate": 2.167604903859026e-05, "loss": 0.9249, "step": 7577 }, { "epoch": 3.432065217391304, "grad_norm": 0.08082267694905046, "learning_rate": 2.1664807741430306e-05, "loss": 0.9308, "step": 7578 }, { "epoch": 3.432518115942029, "grad_norm": 0.0593746618047403, "learning_rate": 2.1653568277268596e-05, "loss": 0.9297, "step": 7579 }, { "epoch": 3.4329710144927534, "grad_norm": 0.08023329215228135, "learning_rate": 2.1642330647228727e-05, "loss": 0.9187, "step": 7580 }, { "epoch": 3.4334239130434785, "grad_norm": 0.059015197009166576, "learning_rate": 2.163109485243418e-05, "loss": 0.9227, "step": 7581 }, { "epoch": 3.433876811594203, "grad_norm": 0.07605469549483437, "learning_rate": 2.1619860894008155e-05, "loss": 0.9162, "step": 7582 }, { "epoch": 3.4343297101449277, "grad_norm": 0.05970328086656068, "learning_rate": 2.1608628773073783e-05, "loss": 0.9214, "step": 7583 }, { "epoch": 3.4347826086956523, "grad_norm": 0.06427691908807835, "learning_rate": 2.1597398490753917e-05, "loss": 0.9387, "step": 7584 }, { "epoch": 3.435235507246377, "grad_norm": 0.05578391574263989, "learning_rate": 2.1586170048171304e-05, "loss": 0.9425, "step": 7585 }, { "epoch": 3.4356884057971016, "grad_norm": 0.0677096670970516, "learning_rate": 2.1574943446448434e-05, "loss": 0.9201, "step": 7586 }, { "epoch": 3.436141304347826, "grad_norm": 0.05462143540552311, "learning_rate": 2.1563718686707677e-05, "loss": 0.9361, "step": 7587 }, { "epoch": 3.4365942028985508, "grad_norm": 0.0815976894533603, "learning_rate": 2.1552495770071165e-05, "loss": 0.946, "step": 7588 }, { "epoch": 3.4370471014492754, "grad_norm": 0.054745816981838225, "learning_rate": 2.15412746976609e-05, "loss": 0.9168, "step": 7589 }, { "epoch": 3.4375, "grad_norm": 0.0646304355325884, "learning_rate": 2.1530055470598654e-05, "loss": 0.9455, "step": 7590 }, { "epoch": 3.4379528985507246, "grad_norm": 0.056456042404204626, "learning_rate": 2.1518838090006025e-05, "loss": 0.9257, "step": 7591 }, { "epoch": 3.4384057971014492, "grad_norm": 0.060939170717574605, "learning_rate": 2.1507622557004425e-05, "loss": 0.9246, "step": 7592 }, { "epoch": 3.438858695652174, "grad_norm": 0.06422958938176475, "learning_rate": 2.1496408872715123e-05, "loss": 0.9415, "step": 7593 }, { "epoch": 3.4393115942028984, "grad_norm": 0.06802634997466905, "learning_rate": 2.1485197038259127e-05, "loss": 0.9269, "step": 7594 }, { "epoch": 3.439764492753623, "grad_norm": 0.0697148785351221, "learning_rate": 2.147398705475734e-05, "loss": 0.9299, "step": 7595 }, { "epoch": 3.4402173913043477, "grad_norm": 0.06086987915539884, "learning_rate": 2.14627789233304e-05, "loss": 0.9238, "step": 7596 }, { "epoch": 3.4406702898550723, "grad_norm": 0.06315373257772353, "learning_rate": 2.1451572645098856e-05, "loss": 0.9298, "step": 7597 }, { "epoch": 3.441123188405797, "grad_norm": 0.06404292183114296, "learning_rate": 2.1440368221182974e-05, "loss": 0.9091, "step": 7598 }, { "epoch": 3.4415760869565215, "grad_norm": 0.05399330975049902, "learning_rate": 2.14291656527029e-05, "loss": 0.9268, "step": 7599 }, { "epoch": 3.4420289855072466, "grad_norm": 0.08964037846440998, "learning_rate": 2.141796494077854e-05, "loss": 0.9351, "step": 7600 }, { "epoch": 3.442481884057971, "grad_norm": 0.05388192157972954, "learning_rate": 2.140676608652969e-05, "loss": 0.9219, "step": 7601 }, { "epoch": 3.442934782608696, "grad_norm": 0.09247353010468029, "learning_rate": 2.1395569091075887e-05, "loss": 0.913, "step": 7602 }, { "epoch": 3.4433876811594204, "grad_norm": 0.08164880828954198, "learning_rate": 2.1384373955536537e-05, "loss": 0.9329, "step": 7603 }, { "epoch": 3.443840579710145, "grad_norm": 0.06108295650085506, "learning_rate": 2.1373180681030803e-05, "loss": 0.9425, "step": 7604 }, { "epoch": 3.4442934782608696, "grad_norm": 0.08535684387690759, "learning_rate": 2.136198926867773e-05, "loss": 0.9301, "step": 7605 }, { "epoch": 3.4447463768115942, "grad_norm": 0.06595010393646833, "learning_rate": 2.135079971959612e-05, "loss": 0.9189, "step": 7606 }, { "epoch": 3.445199275362319, "grad_norm": 0.08910216886377725, "learning_rate": 2.1339612034904623e-05, "loss": 0.9271, "step": 7607 }, { "epoch": 3.4456521739130435, "grad_norm": 0.07101125775480374, "learning_rate": 2.1328426215721655e-05, "loss": 0.9367, "step": 7608 }, { "epoch": 3.446105072463768, "grad_norm": 0.07778446578939759, "learning_rate": 2.131724226316553e-05, "loss": 0.9312, "step": 7609 }, { "epoch": 3.4465579710144927, "grad_norm": 0.07256894264929178, "learning_rate": 2.1306060178354283e-05, "loss": 0.9274, "step": 7610 }, { "epoch": 3.4470108695652173, "grad_norm": 0.07224278548245362, "learning_rate": 2.1294879962405835e-05, "loss": 0.9136, "step": 7611 }, { "epoch": 3.447463768115942, "grad_norm": 0.0832850430940499, "learning_rate": 2.1283701616437867e-05, "loss": 0.9337, "step": 7612 }, { "epoch": 3.4479166666666665, "grad_norm": 0.060154623779379586, "learning_rate": 2.1272525141567925e-05, "loss": 0.9443, "step": 7613 }, { "epoch": 3.448369565217391, "grad_norm": 0.06722755947746692, "learning_rate": 2.126135053891331e-05, "loss": 0.922, "step": 7614 }, { "epoch": 3.448822463768116, "grad_norm": 0.06460006832698008, "learning_rate": 2.1250177809591186e-05, "loss": 0.9146, "step": 7615 }, { "epoch": 3.449275362318841, "grad_norm": 0.06891265518906933, "learning_rate": 2.1239006954718483e-05, "loss": 0.9232, "step": 7616 }, { "epoch": 3.4497282608695654, "grad_norm": 0.07088745026103377, "learning_rate": 2.1227837975411996e-05, "loss": 0.9122, "step": 7617 }, { "epoch": 3.45018115942029, "grad_norm": 0.06317637365855037, "learning_rate": 2.1216670872788276e-05, "loss": 0.9271, "step": 7618 }, { "epoch": 3.4506340579710146, "grad_norm": 0.07068965633776675, "learning_rate": 2.1205505647963748e-05, "loss": 0.9112, "step": 7619 }, { "epoch": 3.4510869565217392, "grad_norm": 0.0595442471543048, "learning_rate": 2.1194342302054595e-05, "loss": 0.9443, "step": 7620 }, { "epoch": 3.451539855072464, "grad_norm": 0.07881209297779414, "learning_rate": 2.1183180836176847e-05, "loss": 0.9401, "step": 7621 }, { "epoch": 3.4519927536231885, "grad_norm": 0.06487304496064937, "learning_rate": 2.117202125144634e-05, "loss": 0.9278, "step": 7622 }, { "epoch": 3.452445652173913, "grad_norm": 0.061194217648054626, "learning_rate": 2.1160863548978697e-05, "loss": 0.9049, "step": 7623 }, { "epoch": 3.4528985507246377, "grad_norm": 0.06421304615572283, "learning_rate": 2.1149707729889362e-05, "loss": 0.9111, "step": 7624 }, { "epoch": 3.4533514492753623, "grad_norm": 0.05762981641627258, "learning_rate": 2.1138553795293636e-05, "loss": 0.9215, "step": 7625 }, { "epoch": 3.453804347826087, "grad_norm": 0.06876330526510423, "learning_rate": 2.1127401746306554e-05, "loss": 0.9193, "step": 7626 }, { "epoch": 3.4542572463768115, "grad_norm": 0.05553016155013394, "learning_rate": 2.1116251584043047e-05, "loss": 0.9463, "step": 7627 }, { "epoch": 3.454710144927536, "grad_norm": 0.07050850820977687, "learning_rate": 2.1105103309617773e-05, "loss": 0.919, "step": 7628 }, { "epoch": 3.4551630434782608, "grad_norm": 0.056772617232239496, "learning_rate": 2.1093956924145283e-05, "loss": 0.915, "step": 7629 }, { "epoch": 3.4556159420289854, "grad_norm": 0.07377037344816012, "learning_rate": 2.108281242873986e-05, "loss": 0.918, "step": 7630 }, { "epoch": 3.45606884057971, "grad_norm": 0.0646222471807376, "learning_rate": 2.107166982451567e-05, "loss": 0.9269, "step": 7631 }, { "epoch": 3.4565217391304346, "grad_norm": 0.07128011677349792, "learning_rate": 2.1060529112586642e-05, "loss": 0.9316, "step": 7632 }, { "epoch": 3.456974637681159, "grad_norm": 0.06711095562586297, "learning_rate": 2.1049390294066537e-05, "loss": 0.9152, "step": 7633 }, { "epoch": 3.457427536231884, "grad_norm": 0.06384941161770347, "learning_rate": 2.1038253370068894e-05, "loss": 0.928, "step": 7634 }, { "epoch": 3.457880434782609, "grad_norm": 0.07276689568184404, "learning_rate": 2.1027118341707124e-05, "loss": 0.9251, "step": 7635 }, { "epoch": 3.4583333333333335, "grad_norm": 0.06178925113292342, "learning_rate": 2.1015985210094385e-05, "loss": 0.9267, "step": 7636 }, { "epoch": 3.458786231884058, "grad_norm": 0.07378350580720686, "learning_rate": 2.1004853976343698e-05, "loss": 0.9533, "step": 7637 }, { "epoch": 3.4592391304347827, "grad_norm": 0.0641915789744012, "learning_rate": 2.099372464156785e-05, "loss": 0.943, "step": 7638 }, { "epoch": 3.4596920289855073, "grad_norm": 0.07046524451873888, "learning_rate": 2.098259720687948e-05, "loss": 0.9443, "step": 7639 }, { "epoch": 3.460144927536232, "grad_norm": 0.07490823559295043, "learning_rate": 2.0971471673391e-05, "loss": 0.9361, "step": 7640 }, { "epoch": 3.4605978260869565, "grad_norm": 0.05679577527552065, "learning_rate": 2.0960348042214656e-05, "loss": 0.9183, "step": 7641 }, { "epoch": 3.461050724637681, "grad_norm": 0.06747935820242444, "learning_rate": 2.094922631446247e-05, "loss": 0.9342, "step": 7642 }, { "epoch": 3.4615036231884058, "grad_norm": 0.06305952301890029, "learning_rate": 2.0938106491246334e-05, "loss": 0.9326, "step": 7643 }, { "epoch": 3.4619565217391304, "grad_norm": 0.06698017749221999, "learning_rate": 2.092698857367788e-05, "loss": 0.9346, "step": 7644 }, { "epoch": 3.462409420289855, "grad_norm": 0.07548871473097278, "learning_rate": 2.0915872562868627e-05, "loss": 0.9486, "step": 7645 }, { "epoch": 3.4628623188405796, "grad_norm": 0.06693404325678053, "learning_rate": 2.0904758459929812e-05, "loss": 0.9279, "step": 7646 }, { "epoch": 3.463315217391304, "grad_norm": 0.07208905804763076, "learning_rate": 2.0893646265972566e-05, "loss": 0.9205, "step": 7647 }, { "epoch": 3.463768115942029, "grad_norm": 0.06447020262614372, "learning_rate": 2.0882535982107785e-05, "loss": 0.9109, "step": 7648 }, { "epoch": 3.4642210144927534, "grad_norm": 0.06431362601622367, "learning_rate": 2.0871427609446174e-05, "loss": 0.9024, "step": 7649 }, { "epoch": 3.4646739130434785, "grad_norm": 0.06807977512873833, "learning_rate": 2.0860321149098244e-05, "loss": 0.9467, "step": 7650 }, { "epoch": 3.465126811594203, "grad_norm": 0.05975531956926235, "learning_rate": 2.0849216602174347e-05, "loss": 0.9382, "step": 7651 }, { "epoch": 3.4655797101449277, "grad_norm": 0.06572944439530719, "learning_rate": 2.0838113969784605e-05, "loss": 0.9131, "step": 7652 }, { "epoch": 3.4660326086956523, "grad_norm": 0.06295636945397196, "learning_rate": 2.0827013253038985e-05, "loss": 0.9312, "step": 7653 }, { "epoch": 3.466485507246377, "grad_norm": 0.06355687484287212, "learning_rate": 2.0815914453047217e-05, "loss": 0.9423, "step": 7654 }, { "epoch": 3.4669384057971016, "grad_norm": 0.062456523215491506, "learning_rate": 2.0804817570918896e-05, "loss": 0.9198, "step": 7655 }, { "epoch": 3.467391304347826, "grad_norm": 0.06218247773224583, "learning_rate": 2.0793722607763383e-05, "loss": 0.9088, "step": 7656 }, { "epoch": 3.4678442028985508, "grad_norm": 0.06901269009081555, "learning_rate": 2.0782629564689855e-05, "loss": 0.9194, "step": 7657 }, { "epoch": 3.4682971014492754, "grad_norm": 0.0593127342608355, "learning_rate": 2.0771538442807283e-05, "loss": 0.9396, "step": 7658 }, { "epoch": 3.46875, "grad_norm": 0.08046687492563231, "learning_rate": 2.0760449243224504e-05, "loss": 0.9376, "step": 7659 }, { "epoch": 3.4692028985507246, "grad_norm": 0.05623064617055445, "learning_rate": 2.074936196705008e-05, "loss": 0.9281, "step": 7660 }, { "epoch": 3.4696557971014492, "grad_norm": 0.06149822730056286, "learning_rate": 2.0738276615392467e-05, "loss": 0.9167, "step": 7661 }, { "epoch": 3.470108695652174, "grad_norm": 0.0733494088201222, "learning_rate": 2.0727193189359848e-05, "loss": 0.9435, "step": 7662 }, { "epoch": 3.4705615942028984, "grad_norm": 0.06974662372302151, "learning_rate": 2.071611169006028e-05, "loss": 0.933, "step": 7663 }, { "epoch": 3.471014492753623, "grad_norm": 0.07462528512366062, "learning_rate": 2.070503211860158e-05, "loss": 0.9174, "step": 7664 }, { "epoch": 3.4714673913043477, "grad_norm": 0.05689880876822683, "learning_rate": 2.0693954476091398e-05, "loss": 0.93, "step": 7665 }, { "epoch": 3.4719202898550723, "grad_norm": 0.07468309977678932, "learning_rate": 2.0682878763637168e-05, "loss": 0.9212, "step": 7666 }, { "epoch": 3.472373188405797, "grad_norm": 0.06315612289480042, "learning_rate": 2.0671804982346168e-05, "loss": 0.9231, "step": 7667 }, { "epoch": 3.4728260869565215, "grad_norm": 0.06651461933942741, "learning_rate": 2.066073313332544e-05, "loss": 0.9281, "step": 7668 }, { "epoch": 3.4732789855072466, "grad_norm": 0.07814084087997566, "learning_rate": 2.0649663217681882e-05, "loss": 0.9369, "step": 7669 }, { "epoch": 3.473731884057971, "grad_norm": 0.07117916114034091, "learning_rate": 2.0638595236522135e-05, "loss": 0.9484, "step": 7670 }, { "epoch": 3.474184782608696, "grad_norm": 0.062433933773568734, "learning_rate": 2.0627529190952727e-05, "loss": 0.914, "step": 7671 }, { "epoch": 3.4746376811594204, "grad_norm": 0.0734991536999557, "learning_rate": 2.06164650820799e-05, "loss": 0.9295, "step": 7672 }, { "epoch": 3.475090579710145, "grad_norm": 0.05704591568757066, "learning_rate": 2.0605402911009796e-05, "loss": 0.9047, "step": 7673 }, { "epoch": 3.4755434782608696, "grad_norm": 0.0628501647913294, "learning_rate": 2.0594342678848296e-05, "loss": 0.9223, "step": 7674 }, { "epoch": 3.4759963768115942, "grad_norm": 0.06155256163878244, "learning_rate": 2.0583284386701104e-05, "loss": 0.9166, "step": 7675 }, { "epoch": 3.476449275362319, "grad_norm": 0.06334244949088676, "learning_rate": 2.057222803567373e-05, "loss": 0.9456, "step": 7676 }, { "epoch": 3.4769021739130435, "grad_norm": 0.06392530105492823, "learning_rate": 2.056117362687152e-05, "loss": 0.9249, "step": 7677 }, { "epoch": 3.477355072463768, "grad_norm": 0.06955299182998843, "learning_rate": 2.055012116139957e-05, "loss": 0.934, "step": 7678 }, { "epoch": 3.4778079710144927, "grad_norm": 0.062287371747675846, "learning_rate": 2.0539070640362847e-05, "loss": 0.9141, "step": 7679 }, { "epoch": 3.4782608695652173, "grad_norm": 0.0760152011428204, "learning_rate": 2.0528022064866048e-05, "loss": 0.9284, "step": 7680 }, { "epoch": 3.478713768115942, "grad_norm": 0.07541808586360851, "learning_rate": 2.051697543601376e-05, "loss": 0.9239, "step": 7681 }, { "epoch": 3.4791666666666665, "grad_norm": 0.07479122688363392, "learning_rate": 2.050593075491031e-05, "loss": 0.9152, "step": 7682 }, { "epoch": 3.479619565217391, "grad_norm": 0.08581484218995822, "learning_rate": 2.0494888022659856e-05, "loss": 0.9303, "step": 7683 }, { "epoch": 3.480072463768116, "grad_norm": 0.06383181533287272, "learning_rate": 2.0483847240366335e-05, "loss": 0.9247, "step": 7684 }, { "epoch": 3.480525362318841, "grad_norm": 0.07157184595039359, "learning_rate": 2.047280840913355e-05, "loss": 0.9462, "step": 7685 }, { "epoch": 3.4809782608695654, "grad_norm": 0.06921233671606554, "learning_rate": 2.0461771530065035e-05, "loss": 0.9256, "step": 7686 }, { "epoch": 3.48143115942029, "grad_norm": 0.06510419542918042, "learning_rate": 2.0450736604264193e-05, "loss": 0.9383, "step": 7687 }, { "epoch": 3.4818840579710146, "grad_norm": 0.06695174027580511, "learning_rate": 2.0439703632834193e-05, "loss": 0.9412, "step": 7688 }, { "epoch": 3.4823369565217392, "grad_norm": 0.06714710729330453, "learning_rate": 2.042867261687801e-05, "loss": 0.9396, "step": 7689 }, { "epoch": 3.482789855072464, "grad_norm": 0.07349036376342795, "learning_rate": 2.0417643557498444e-05, "loss": 0.9168, "step": 7690 }, { "epoch": 3.4832427536231885, "grad_norm": 0.060392612305382405, "learning_rate": 2.0406616455798084e-05, "loss": 0.9128, "step": 7691 }, { "epoch": 3.483695652173913, "grad_norm": 0.06684731277636732, "learning_rate": 2.0395591312879324e-05, "loss": 0.929, "step": 7692 }, { "epoch": 3.4841485507246377, "grad_norm": 0.06141192096613273, "learning_rate": 2.0384568129844354e-05, "loss": 0.9331, "step": 7693 }, { "epoch": 3.4846014492753623, "grad_norm": 0.06359566816903843, "learning_rate": 2.037354690779521e-05, "loss": 0.9289, "step": 7694 }, { "epoch": 3.485054347826087, "grad_norm": 0.06038894505821374, "learning_rate": 2.0362527647833662e-05, "loss": 0.946, "step": 7695 }, { "epoch": 3.4855072463768115, "grad_norm": 0.06171933074289724, "learning_rate": 2.0351510351061357e-05, "loss": 0.925, "step": 7696 }, { "epoch": 3.485960144927536, "grad_norm": 0.056621656076546824, "learning_rate": 2.034049501857969e-05, "loss": 0.9331, "step": 7697 }, { "epoch": 3.4864130434782608, "grad_norm": 0.06252437259528515, "learning_rate": 2.03294816514899e-05, "loss": 0.9429, "step": 7698 }, { "epoch": 3.4868659420289854, "grad_norm": 0.06500877520579744, "learning_rate": 2.0318470250893002e-05, "loss": 0.9481, "step": 7699 }, { "epoch": 3.48731884057971, "grad_norm": 0.06271838273343959, "learning_rate": 2.0307460817889815e-05, "loss": 0.9199, "step": 7700 }, { "epoch": 3.4877717391304346, "grad_norm": 0.07623056826605688, "learning_rate": 2.0296453353580964e-05, "loss": 0.9006, "step": 7701 }, { "epoch": 3.488224637681159, "grad_norm": 0.07419565726885625, "learning_rate": 2.0285447859066913e-05, "loss": 0.9377, "step": 7702 }, { "epoch": 3.488677536231884, "grad_norm": 0.06285363382427302, "learning_rate": 2.027444433544786e-05, "loss": 0.9395, "step": 7703 }, { "epoch": 3.489130434782609, "grad_norm": 0.0688161285354608, "learning_rate": 2.0263442783823884e-05, "loss": 0.9229, "step": 7704 }, { "epoch": 3.4895833333333335, "grad_norm": 0.06494228851211507, "learning_rate": 2.025244320529479e-05, "loss": 0.9223, "step": 7705 }, { "epoch": 3.490036231884058, "grad_norm": 0.07256654795451181, "learning_rate": 2.0241445600960256e-05, "loss": 0.9204, "step": 7706 }, { "epoch": 3.4904891304347827, "grad_norm": 0.06898219518733067, "learning_rate": 2.0230449971919718e-05, "loss": 0.9333, "step": 7707 }, { "epoch": 3.4909420289855073, "grad_norm": 0.06725302353813492, "learning_rate": 2.0219456319272425e-05, "loss": 0.9198, "step": 7708 }, { "epoch": 3.491394927536232, "grad_norm": 0.07818751545088473, "learning_rate": 2.020846464411741e-05, "loss": 0.9097, "step": 7709 }, { "epoch": 3.4918478260869565, "grad_norm": 0.06098208957272281, "learning_rate": 2.0197474947553562e-05, "loss": 0.9348, "step": 7710 }, { "epoch": 3.492300724637681, "grad_norm": 0.07517096625803608, "learning_rate": 2.0186487230679508e-05, "loss": 0.9013, "step": 7711 }, { "epoch": 3.4927536231884058, "grad_norm": 0.06394681169181855, "learning_rate": 2.017550149459374e-05, "loss": 0.9315, "step": 7712 }, { "epoch": 3.4932065217391304, "grad_norm": 0.07139504998663053, "learning_rate": 2.0164517740394487e-05, "loss": 0.9162, "step": 7713 }, { "epoch": 3.493659420289855, "grad_norm": 0.06970582701947552, "learning_rate": 2.0153535969179837e-05, "loss": 0.933, "step": 7714 }, { "epoch": 3.4941123188405796, "grad_norm": 0.07530988426793658, "learning_rate": 2.014255618204764e-05, "loss": 0.9417, "step": 7715 }, { "epoch": 3.494565217391304, "grad_norm": 0.06421690241978678, "learning_rate": 2.0131578380095575e-05, "loss": 0.9302, "step": 7716 }, { "epoch": 3.495018115942029, "grad_norm": 0.07406184987465055, "learning_rate": 2.0120602564421113e-05, "loss": 0.9353, "step": 7717 }, { "epoch": 3.4954710144927534, "grad_norm": 0.06064039905092348, "learning_rate": 2.0109628736121506e-05, "loss": 0.951, "step": 7718 }, { "epoch": 3.4959239130434785, "grad_norm": 0.07654349991634328, "learning_rate": 2.0098656896293816e-05, "loss": 0.9169, "step": 7719 }, { "epoch": 3.496376811594203, "grad_norm": 0.056719757820874524, "learning_rate": 2.0087687046034955e-05, "loss": 0.9112, "step": 7720 }, { "epoch": 3.4968297101449277, "grad_norm": 0.070037911842399, "learning_rate": 2.007671918644156e-05, "loss": 0.9131, "step": 7721 }, { "epoch": 3.4972826086956523, "grad_norm": 0.06031527774246907, "learning_rate": 2.0065753318610123e-05, "loss": 0.9268, "step": 7722 }, { "epoch": 3.497735507246377, "grad_norm": 0.0636374047440713, "learning_rate": 2.005478944363691e-05, "loss": 0.9227, "step": 7723 }, { "epoch": 3.4981884057971016, "grad_norm": 0.06349254910768402, "learning_rate": 2.0043827562618013e-05, "loss": 0.9384, "step": 7724 }, { "epoch": 3.498641304347826, "grad_norm": 0.06521336008069761, "learning_rate": 2.00328676766493e-05, "loss": 0.924, "step": 7725 }, { "epoch": 3.4990942028985508, "grad_norm": 0.060024758433853666, "learning_rate": 2.0021909786826445e-05, "loss": 0.9378, "step": 7726 }, { "epoch": 3.4995471014492754, "grad_norm": 0.07327737138428046, "learning_rate": 2.0010953894244914e-05, "loss": 0.9292, "step": 7727 }, { "epoch": 3.5, "grad_norm": 0.05679070924382384, "learning_rate": 2.0000000000000012e-05, "loss": 0.8959, "step": 7728 }, { "epoch": 3.5004528985507246, "grad_norm": 0.06532744016887414, "learning_rate": 1.998904810518678e-05, "loss": 0.9041, "step": 7729 }, { "epoch": 3.5009057971014492, "grad_norm": 0.052143834567003854, "learning_rate": 1.9978098210900144e-05, "loss": 0.9125, "step": 7730 }, { "epoch": 3.501358695652174, "grad_norm": 0.06592107820346443, "learning_rate": 1.9967150318234738e-05, "loss": 0.9327, "step": 7731 }, { "epoch": 3.5018115942028984, "grad_norm": 0.058144838081618216, "learning_rate": 1.9956204428285074e-05, "loss": 0.918, "step": 7732 }, { "epoch": 3.502264492753623, "grad_norm": 0.0634328362727443, "learning_rate": 1.994526054214542e-05, "loss": 0.9438, "step": 7733 }, { "epoch": 3.5027173913043477, "grad_norm": 0.06316099852296803, "learning_rate": 1.9934318660909843e-05, "loss": 0.9339, "step": 7734 }, { "epoch": 3.5031702898550723, "grad_norm": 0.06215635442822397, "learning_rate": 1.9923378785672213e-05, "loss": 0.9299, "step": 7735 }, { "epoch": 3.503623188405797, "grad_norm": 0.06378377761281581, "learning_rate": 1.9912440917526233e-05, "loss": 0.9188, "step": 7736 }, { "epoch": 3.5040760869565215, "grad_norm": 0.07019604376328882, "learning_rate": 1.9901505057565358e-05, "loss": 0.9147, "step": 7737 }, { "epoch": 3.504528985507246, "grad_norm": 0.06328013434214531, "learning_rate": 1.9890571206882877e-05, "loss": 0.9095, "step": 7738 }, { "epoch": 3.504981884057971, "grad_norm": 0.07406234862047667, "learning_rate": 1.987963936657185e-05, "loss": 0.9343, "step": 7739 }, { "epoch": 3.505434782608696, "grad_norm": 0.05547256072849451, "learning_rate": 1.986870953772517e-05, "loss": 0.9239, "step": 7740 }, { "epoch": 3.5058876811594204, "grad_norm": 0.07397080469706763, "learning_rate": 1.9857781721435504e-05, "loss": 0.9261, "step": 7741 }, { "epoch": 3.506340579710145, "grad_norm": 0.057434129616939035, "learning_rate": 1.9846855918795315e-05, "loss": 0.9325, "step": 7742 }, { "epoch": 3.5067934782608696, "grad_norm": 0.07771836027850168, "learning_rate": 1.9835932130896855e-05, "loss": 0.9356, "step": 7743 }, { "epoch": 3.5072463768115942, "grad_norm": 0.06676699310809574, "learning_rate": 1.982501035883223e-05, "loss": 0.9371, "step": 7744 }, { "epoch": 3.507699275362319, "grad_norm": 0.06183231984785192, "learning_rate": 1.981409060369328e-05, "loss": 0.9176, "step": 7745 }, { "epoch": 3.5081521739130435, "grad_norm": 0.06901946867867477, "learning_rate": 1.980317286657169e-05, "loss": 0.9275, "step": 7746 }, { "epoch": 3.508605072463768, "grad_norm": 0.05886423235661162, "learning_rate": 1.97922571485589e-05, "loss": 0.9276, "step": 7747 }, { "epoch": 3.5090579710144927, "grad_norm": 0.062005866880149876, "learning_rate": 1.97813434507462e-05, "loss": 0.9427, "step": 7748 }, { "epoch": 3.5095108695652173, "grad_norm": 0.06693693400876038, "learning_rate": 1.9770431774224632e-05, "loss": 0.9264, "step": 7749 }, { "epoch": 3.509963768115942, "grad_norm": 0.05399804532142136, "learning_rate": 1.9759522120085057e-05, "loss": 0.9187, "step": 7750 }, { "epoch": 3.5104166666666665, "grad_norm": 0.06459646085750198, "learning_rate": 1.9748614489418118e-05, "loss": 0.8992, "step": 7751 }, { "epoch": 3.5108695652173916, "grad_norm": 0.06736426347756168, "learning_rate": 1.9737708883314287e-05, "loss": 0.9218, "step": 7752 }, { "epoch": 3.511322463768116, "grad_norm": 0.06789988738751744, "learning_rate": 1.97268053028638e-05, "loss": 0.9384, "step": 7753 }, { "epoch": 3.511775362318841, "grad_norm": 0.06375701119255771, "learning_rate": 1.971590374915672e-05, "loss": 0.9194, "step": 7754 }, { "epoch": 3.5122282608695654, "grad_norm": 0.0650489088726389, "learning_rate": 1.9705004223282875e-05, "loss": 0.9289, "step": 7755 }, { "epoch": 3.51268115942029, "grad_norm": 0.06290748183101011, "learning_rate": 1.9694106726331933e-05, "loss": 0.914, "step": 7756 }, { "epoch": 3.5131340579710146, "grad_norm": 0.06193003754906852, "learning_rate": 1.96832112593933e-05, "loss": 0.9296, "step": 7757 }, { "epoch": 3.5135869565217392, "grad_norm": 0.06243154707166319, "learning_rate": 1.9672317823556265e-05, "loss": 0.929, "step": 7758 }, { "epoch": 3.514039855072464, "grad_norm": 0.0678390919824647, "learning_rate": 1.9661426419909803e-05, "loss": 0.9273, "step": 7759 }, { "epoch": 3.5144927536231885, "grad_norm": 0.05694536643441754, "learning_rate": 1.965053704954278e-05, "loss": 0.9191, "step": 7760 }, { "epoch": 3.514945652173913, "grad_norm": 0.06284788613731072, "learning_rate": 1.9639649713543806e-05, "loss": 0.9328, "step": 7761 }, { "epoch": 3.5153985507246377, "grad_norm": 0.06777471927330028, "learning_rate": 1.962876441300132e-05, "loss": 0.9164, "step": 7762 }, { "epoch": 3.5158514492753623, "grad_norm": 0.06278030969517154, "learning_rate": 1.961788114900353e-05, "loss": 0.9324, "step": 7763 }, { "epoch": 3.516304347826087, "grad_norm": 0.07587791886943443, "learning_rate": 1.960699992263847e-05, "loss": 0.9447, "step": 7764 }, { "epoch": 3.5167572463768115, "grad_norm": 0.05700297993420507, "learning_rate": 1.959612073499393e-05, "loss": 0.9294, "step": 7765 }, { "epoch": 3.517210144927536, "grad_norm": 0.0834757857954449, "learning_rate": 1.9585243587157548e-05, "loss": 0.938, "step": 7766 }, { "epoch": 3.5176630434782608, "grad_norm": 0.06147262445079701, "learning_rate": 1.9574368480216713e-05, "loss": 0.9329, "step": 7767 }, { "epoch": 3.5181159420289854, "grad_norm": 0.06384272918758192, "learning_rate": 1.9563495415258627e-05, "loss": 0.9061, "step": 7768 }, { "epoch": 3.51856884057971, "grad_norm": 0.06746257271445087, "learning_rate": 1.9552624393370275e-05, "loss": 0.9382, "step": 7769 }, { "epoch": 3.5190217391304346, "grad_norm": 0.05673376426505589, "learning_rate": 1.9541755415638485e-05, "loss": 0.9512, "step": 7770 }, { "epoch": 3.519474637681159, "grad_norm": 0.06445770677629513, "learning_rate": 1.9530888483149802e-05, "loss": 0.9305, "step": 7771 }, { "epoch": 3.519927536231884, "grad_norm": 0.058178485272666024, "learning_rate": 1.952002359699065e-05, "loss": 0.9102, "step": 7772 }, { "epoch": 3.5203804347826084, "grad_norm": 0.05627762019289382, "learning_rate": 1.950916075824718e-05, "loss": 0.9183, "step": 7773 }, { "epoch": 3.5208333333333335, "grad_norm": 0.054928042187351075, "learning_rate": 1.9498299968005393e-05, "loss": 0.9222, "step": 7774 }, { "epoch": 3.521286231884058, "grad_norm": 0.0600824525520433, "learning_rate": 1.9487441227351044e-05, "loss": 0.9423, "step": 7775 }, { "epoch": 3.5217391304347827, "grad_norm": 0.05761035241963363, "learning_rate": 1.9476584537369705e-05, "loss": 0.9124, "step": 7776 }, { "epoch": 3.5221920289855073, "grad_norm": 0.0618050351372967, "learning_rate": 1.9465729899146715e-05, "loss": 0.9384, "step": 7777 }, { "epoch": 3.522644927536232, "grad_norm": 0.06567431134301319, "learning_rate": 1.945487731376726e-05, "loss": 0.9511, "step": 7778 }, { "epoch": 3.5230978260869565, "grad_norm": 0.06303636237801545, "learning_rate": 1.9444026782316262e-05, "loss": 0.9347, "step": 7779 }, { "epoch": 3.523550724637681, "grad_norm": 0.06323239997565867, "learning_rate": 1.9433178305878497e-05, "loss": 0.9418, "step": 7780 }, { "epoch": 3.5240036231884058, "grad_norm": 0.07170195510001458, "learning_rate": 1.942233188553847e-05, "loss": 0.9345, "step": 7781 }, { "epoch": 3.5244565217391304, "grad_norm": 0.06136313127022192, "learning_rate": 1.9411487522380546e-05, "loss": 0.9359, "step": 7782 }, { "epoch": 3.524909420289855, "grad_norm": 0.059673467135086235, "learning_rate": 1.9400645217488843e-05, "loss": 0.9122, "step": 7783 }, { "epoch": 3.5253623188405796, "grad_norm": 0.07976571604524159, "learning_rate": 1.938980497194728e-05, "loss": 0.9137, "step": 7784 }, { "epoch": 3.525815217391304, "grad_norm": 0.060941117413417456, "learning_rate": 1.9378966786839556e-05, "loss": 0.9278, "step": 7785 }, { "epoch": 3.526268115942029, "grad_norm": 0.06542586944028886, "learning_rate": 1.936813066324922e-05, "loss": 0.9346, "step": 7786 }, { "epoch": 3.526721014492754, "grad_norm": 0.06483203837749692, "learning_rate": 1.9357296602259535e-05, "loss": 0.9206, "step": 7787 }, { "epoch": 3.5271739130434785, "grad_norm": 0.06830726399692026, "learning_rate": 1.9346464604953638e-05, "loss": 0.9187, "step": 7788 }, { "epoch": 3.527626811594203, "grad_norm": 0.058416737566249856, "learning_rate": 1.9335634672414386e-05, "loss": 0.9369, "step": 7789 }, { "epoch": 3.5280797101449277, "grad_norm": 0.06247825121125787, "learning_rate": 1.9324806805724497e-05, "loss": 0.9128, "step": 7790 }, { "epoch": 3.5285326086956523, "grad_norm": 0.05706055511572939, "learning_rate": 1.9313981005966432e-05, "loss": 0.8983, "step": 7791 }, { "epoch": 3.528985507246377, "grad_norm": 0.06220290247559143, "learning_rate": 1.930315727422247e-05, "loss": 0.924, "step": 7792 }, { "epoch": 3.5294384057971016, "grad_norm": 0.05431350213338475, "learning_rate": 1.9292335611574655e-05, "loss": 0.8984, "step": 7793 }, { "epoch": 3.529891304347826, "grad_norm": 0.0619947301534515, "learning_rate": 1.9281516019104874e-05, "loss": 0.9285, "step": 7794 }, { "epoch": 3.5303442028985508, "grad_norm": 0.05686540762964724, "learning_rate": 1.927069849789476e-05, "loss": 0.9142, "step": 7795 }, { "epoch": 3.5307971014492754, "grad_norm": 0.054045207524414655, "learning_rate": 1.9259883049025778e-05, "loss": 0.9587, "step": 7796 }, { "epoch": 3.53125, "grad_norm": 0.056420780603767456, "learning_rate": 1.9249069673579136e-05, "loss": 0.9313, "step": 7797 }, { "epoch": 3.5317028985507246, "grad_norm": 0.051546492832514695, "learning_rate": 1.92382583726359e-05, "loss": 0.9554, "step": 7798 }, { "epoch": 3.5321557971014492, "grad_norm": 0.05983691148781226, "learning_rate": 1.9227449147276873e-05, "loss": 0.9361, "step": 7799 }, { "epoch": 3.532608695652174, "grad_norm": 0.051294699869444155, "learning_rate": 1.9216641998582666e-05, "loss": 0.9308, "step": 7800 }, { "epoch": 3.5330615942028984, "grad_norm": 0.056162680752425316, "learning_rate": 1.920583692763369e-05, "loss": 0.8962, "step": 7801 }, { "epoch": 3.533514492753623, "grad_norm": 0.061212581819464006, "learning_rate": 1.9195033935510152e-05, "loss": 0.9051, "step": 7802 }, { "epoch": 3.5339673913043477, "grad_norm": 0.061666893271500794, "learning_rate": 1.9184233023292026e-05, "loss": 0.9345, "step": 7803 }, { "epoch": 3.5344202898550723, "grad_norm": 0.052843740344892184, "learning_rate": 1.9173434192059125e-05, "loss": 0.9038, "step": 7804 }, { "epoch": 3.534873188405797, "grad_norm": 0.058294608724173184, "learning_rate": 1.9162637442890997e-05, "loss": 0.9113, "step": 7805 }, { "epoch": 3.5353260869565215, "grad_norm": 0.06279488408385894, "learning_rate": 1.915184277686704e-05, "loss": 0.9108, "step": 7806 }, { "epoch": 3.535778985507246, "grad_norm": 0.061203363793068746, "learning_rate": 1.914105019506638e-05, "loss": 0.9094, "step": 7807 }, { "epoch": 3.536231884057971, "grad_norm": 0.06145090211834956, "learning_rate": 1.9130259698568e-05, "loss": 0.9261, "step": 7808 }, { "epoch": 3.536684782608696, "grad_norm": 0.06552880573046978, "learning_rate": 1.9119471288450632e-05, "loss": 0.9318, "step": 7809 }, { "epoch": 3.5371376811594204, "grad_norm": 0.06122168619052876, "learning_rate": 1.9108684965792806e-05, "loss": 0.9308, "step": 7810 }, { "epoch": 3.537590579710145, "grad_norm": 0.056748934854868836, "learning_rate": 1.909790073167285e-05, "loss": 0.9507, "step": 7811 }, { "epoch": 3.5380434782608696, "grad_norm": 0.07092928545498316, "learning_rate": 1.9087118587168865e-05, "loss": 0.921, "step": 7812 }, { "epoch": 3.5384963768115942, "grad_norm": 0.059800068923115075, "learning_rate": 1.9076338533358788e-05, "loss": 0.9245, "step": 7813 }, { "epoch": 3.538949275362319, "grad_norm": 0.05995783682438833, "learning_rate": 1.9065560571320287e-05, "loss": 0.93, "step": 7814 }, { "epoch": 3.5394021739130435, "grad_norm": 0.05837219613706215, "learning_rate": 1.9054784702130888e-05, "loss": 0.9393, "step": 7815 }, { "epoch": 3.539855072463768, "grad_norm": 0.05816609706153589, "learning_rate": 1.904401092686784e-05, "loss": 0.9198, "step": 7816 }, { "epoch": 3.5403079710144927, "grad_norm": 0.05406640948822592, "learning_rate": 1.9033239246608234e-05, "loss": 0.94, "step": 7817 }, { "epoch": 3.5407608695652173, "grad_norm": 0.06782209287904196, "learning_rate": 1.9022469662428928e-05, "loss": 0.9203, "step": 7818 }, { "epoch": 3.541213768115942, "grad_norm": 0.06484108821508025, "learning_rate": 1.9011702175406576e-05, "loss": 0.9258, "step": 7819 }, { "epoch": 3.5416666666666665, "grad_norm": 0.059787412275967226, "learning_rate": 1.90009367866176e-05, "loss": 0.9227, "step": 7820 }, { "epoch": 3.5421195652173916, "grad_norm": 0.06885223739554251, "learning_rate": 1.8990173497138263e-05, "loss": 0.9141, "step": 7821 }, { "epoch": 3.542572463768116, "grad_norm": 0.06405040312220689, "learning_rate": 1.8979412308044557e-05, "loss": 0.9154, "step": 7822 }, { "epoch": 3.543025362318841, "grad_norm": 0.06888396363472307, "learning_rate": 1.896865322041233e-05, "loss": 0.928, "step": 7823 }, { "epoch": 3.5434782608695654, "grad_norm": 0.05446034338507402, "learning_rate": 1.8957896235317153e-05, "loss": 0.9361, "step": 7824 }, { "epoch": 3.54393115942029, "grad_norm": 0.0712692804294362, "learning_rate": 1.894714135383445e-05, "loss": 0.9275, "step": 7825 }, { "epoch": 3.5443840579710146, "grad_norm": 0.20811987416150135, "learning_rate": 1.8936388577039384e-05, "loss": 0.9159, "step": 7826 }, { "epoch": 3.5448369565217392, "grad_norm": 0.07700338003008513, "learning_rate": 1.892563790600693e-05, "loss": 0.9126, "step": 7827 }, { "epoch": 3.545289855072464, "grad_norm": 0.05178756696692678, "learning_rate": 1.8914889341811835e-05, "loss": 0.9244, "step": 7828 }, { "epoch": 3.5457427536231885, "grad_norm": 0.07220976206753635, "learning_rate": 1.8904142885528683e-05, "loss": 0.9335, "step": 7829 }, { "epoch": 3.546195652173913, "grad_norm": 0.058987713020362105, "learning_rate": 1.8893398538231778e-05, "loss": 0.9234, "step": 7830 }, { "epoch": 3.5466485507246377, "grad_norm": 0.06311947976912223, "learning_rate": 1.8882656300995283e-05, "loss": 0.9242, "step": 7831 }, { "epoch": 3.5471014492753623, "grad_norm": 0.06952843036031699, "learning_rate": 1.8871916174893085e-05, "loss": 0.9285, "step": 7832 }, { "epoch": 3.547554347826087, "grad_norm": 0.06333174464485776, "learning_rate": 1.886117816099892e-05, "loss": 0.9259, "step": 7833 }, { "epoch": 3.5480072463768115, "grad_norm": 0.08437674661772403, "learning_rate": 1.8850442260386276e-05, "loss": 0.9168, "step": 7834 }, { "epoch": 3.548460144927536, "grad_norm": 0.06029247494205065, "learning_rate": 1.8839708474128434e-05, "loss": 0.9146, "step": 7835 }, { "epoch": 3.5489130434782608, "grad_norm": 0.06911326271968828, "learning_rate": 1.8828976803298454e-05, "loss": 0.9144, "step": 7836 }, { "epoch": 3.5493659420289854, "grad_norm": 0.0582970318241178, "learning_rate": 1.881824724896922e-05, "loss": 0.9636, "step": 7837 }, { "epoch": 3.54981884057971, "grad_norm": 0.0654753606551664, "learning_rate": 1.8807519812213365e-05, "loss": 0.919, "step": 7838 }, { "epoch": 3.5502717391304346, "grad_norm": 0.06302024873993546, "learning_rate": 1.8796794494103344e-05, "loss": 0.913, "step": 7839 }, { "epoch": 3.550724637681159, "grad_norm": 0.053569733470170175, "learning_rate": 1.878607129571137e-05, "loss": 0.9414, "step": 7840 }, { "epoch": 3.551177536231884, "grad_norm": 0.06152008661538113, "learning_rate": 1.8775350218109473e-05, "loss": 0.9198, "step": 7841 }, { "epoch": 3.5516304347826084, "grad_norm": 0.055770019055153, "learning_rate": 1.8764631262369453e-05, "loss": 0.9332, "step": 7842 }, { "epoch": 3.5520833333333335, "grad_norm": 0.06948185964522902, "learning_rate": 1.875391442956289e-05, "loss": 0.9177, "step": 7843 }, { "epoch": 3.552536231884058, "grad_norm": 0.06130347128994411, "learning_rate": 1.874319972076116e-05, "loss": 0.9293, "step": 7844 }, { "epoch": 3.5529891304347827, "grad_norm": 0.07000324147342898, "learning_rate": 1.8732487137035452e-05, "loss": 0.9283, "step": 7845 }, { "epoch": 3.5534420289855073, "grad_norm": 0.06013387329780595, "learning_rate": 1.872177667945669e-05, "loss": 0.9154, "step": 7846 }, { "epoch": 3.553894927536232, "grad_norm": 0.06698447154535989, "learning_rate": 1.8711068349095648e-05, "loss": 0.9188, "step": 7847 }, { "epoch": 3.5543478260869565, "grad_norm": 0.07349599927036249, "learning_rate": 1.8700362147022823e-05, "loss": 0.9404, "step": 7848 }, { "epoch": 3.554800724637681, "grad_norm": 0.05441372391333821, "learning_rate": 1.8689658074308557e-05, "loss": 0.9281, "step": 7849 }, { "epoch": 3.5552536231884058, "grad_norm": 0.06651619225693935, "learning_rate": 1.8678956132022934e-05, "loss": 0.9181, "step": 7850 }, { "epoch": 3.5557065217391304, "grad_norm": 0.05243730265146696, "learning_rate": 1.8668256321235865e-05, "loss": 0.9252, "step": 7851 }, { "epoch": 3.556159420289855, "grad_norm": 0.07278416783025508, "learning_rate": 1.8657558643017016e-05, "loss": 0.9204, "step": 7852 }, { "epoch": 3.5566123188405796, "grad_norm": 0.05184512505542573, "learning_rate": 1.8646863098435845e-05, "loss": 0.905, "step": 7853 }, { "epoch": 3.557065217391304, "grad_norm": 0.0690588683449133, "learning_rate": 1.8636169688561595e-05, "loss": 0.92, "step": 7854 }, { "epoch": 3.557518115942029, "grad_norm": 0.05819588398821712, "learning_rate": 1.862547841446333e-05, "loss": 0.9044, "step": 7855 }, { "epoch": 3.557971014492754, "grad_norm": 0.06609429421618956, "learning_rate": 1.8614789277209845e-05, "loss": 0.9248, "step": 7856 }, { "epoch": 3.5584239130434785, "grad_norm": 0.05899626810391458, "learning_rate": 1.8604102277869777e-05, "loss": 0.9022, "step": 7857 }, { "epoch": 3.558876811594203, "grad_norm": 0.06976152715671918, "learning_rate": 1.859341741751149e-05, "loss": 0.9158, "step": 7858 }, { "epoch": 3.5593297101449277, "grad_norm": 0.06162506896921567, "learning_rate": 1.858273469720321e-05, "loss": 0.9335, "step": 7859 }, { "epoch": 3.5597826086956523, "grad_norm": 0.05080475198484468, "learning_rate": 1.8572054118012874e-05, "loss": 0.9246, "step": 7860 }, { "epoch": 3.560235507246377, "grad_norm": 0.06670380110202757, "learning_rate": 1.856137568100825e-05, "loss": 0.9106, "step": 7861 }, { "epoch": 3.5606884057971016, "grad_norm": 0.054740770638849294, "learning_rate": 1.855069938725685e-05, "loss": 0.9468, "step": 7862 }, { "epoch": 3.561141304347826, "grad_norm": 0.05753093488047701, "learning_rate": 1.8540025237826035e-05, "loss": 0.9071, "step": 7863 }, { "epoch": 3.5615942028985508, "grad_norm": 0.05718150475512809, "learning_rate": 1.8529353233782893e-05, "loss": 0.9312, "step": 7864 }, { "epoch": 3.5620471014492754, "grad_norm": 0.05766087797987739, "learning_rate": 1.8518683376194345e-05, "loss": 0.9236, "step": 7865 }, { "epoch": 3.5625, "grad_norm": 0.05761074518277988, "learning_rate": 1.8508015666127043e-05, "loss": 0.9188, "step": 7866 }, { "epoch": 3.5629528985507246, "grad_norm": 0.058200394492603, "learning_rate": 1.849735010464749e-05, "loss": 0.9249, "step": 7867 }, { "epoch": 3.5634057971014492, "grad_norm": 0.06131426768153126, "learning_rate": 1.8486686692821914e-05, "loss": 0.9272, "step": 7868 }, { "epoch": 3.563858695652174, "grad_norm": 0.057523617051002986, "learning_rate": 1.8476025431716368e-05, "loss": 0.922, "step": 7869 }, { "epoch": 3.5643115942028984, "grad_norm": 0.05925025091206768, "learning_rate": 1.8465366322396652e-05, "loss": 0.9494, "step": 7870 }, { "epoch": 3.564764492753623, "grad_norm": 0.05759829805680914, "learning_rate": 1.84547093659284e-05, "loss": 0.9366, "step": 7871 }, { "epoch": 3.5652173913043477, "grad_norm": 0.05546375202310927, "learning_rate": 1.844405456337698e-05, "loss": 0.9179, "step": 7872 }, { "epoch": 3.5656702898550723, "grad_norm": 0.05950908006111839, "learning_rate": 1.8433401915807603e-05, "loss": 0.9151, "step": 7873 }, { "epoch": 3.566123188405797, "grad_norm": 0.055478406714515704, "learning_rate": 1.8422751424285192e-05, "loss": 0.9257, "step": 7874 }, { "epoch": 3.5665760869565215, "grad_norm": 0.05306828536578164, "learning_rate": 1.841210308987453e-05, "loss": 0.9241, "step": 7875 }, { "epoch": 3.567028985507246, "grad_norm": 0.05835924701801643, "learning_rate": 1.840145691364013e-05, "loss": 0.923, "step": 7876 }, { "epoch": 3.567481884057971, "grad_norm": 0.05905589018738497, "learning_rate": 1.839081289664631e-05, "loss": 0.9039, "step": 7877 }, { "epoch": 3.567934782608696, "grad_norm": 0.05831032381234657, "learning_rate": 1.838017103995715e-05, "loss": 0.9096, "step": 7878 }, { "epoch": 3.5683876811594204, "grad_norm": 0.06208619158752966, "learning_rate": 1.8369531344636567e-05, "loss": 0.9192, "step": 7879 }, { "epoch": 3.568840579710145, "grad_norm": 0.06269156718606586, "learning_rate": 1.8358893811748196e-05, "loss": 0.9267, "step": 7880 }, { "epoch": 3.5692934782608696, "grad_norm": 0.05594678956863444, "learning_rate": 1.8348258442355523e-05, "loss": 0.935, "step": 7881 }, { "epoch": 3.5697463768115942, "grad_norm": 0.05420114639671923, "learning_rate": 1.8337625237521744e-05, "loss": 0.926, "step": 7882 }, { "epoch": 3.570199275362319, "grad_norm": 0.05724458800664014, "learning_rate": 1.8326994198309914e-05, "loss": 0.8926, "step": 7883 }, { "epoch": 3.5706521739130435, "grad_norm": 0.05268456322348985, "learning_rate": 1.8316365325782818e-05, "loss": 0.9422, "step": 7884 }, { "epoch": 3.571105072463768, "grad_norm": 0.06596349414979574, "learning_rate": 1.830573862100304e-05, "loss": 0.9341, "step": 7885 }, { "epoch": 3.5715579710144927, "grad_norm": 0.05835904646284141, "learning_rate": 1.829511408503293e-05, "loss": 0.9136, "step": 7886 }, { "epoch": 3.5720108695652173, "grad_norm": 0.06237704469526254, "learning_rate": 1.8284491718934684e-05, "loss": 0.9292, "step": 7887 }, { "epoch": 3.572463768115942, "grad_norm": 0.06726552495531289, "learning_rate": 1.827387152377019e-05, "loss": 0.9431, "step": 7888 }, { "epoch": 3.5729166666666665, "grad_norm": 0.06504434902250725, "learning_rate": 1.826325350060121e-05, "loss": 0.8954, "step": 7889 }, { "epoch": 3.5733695652173916, "grad_norm": 0.07749418060681362, "learning_rate": 1.82526376504892e-05, "loss": 0.9274, "step": 7890 }, { "epoch": 3.573822463768116, "grad_norm": 0.07943776023035999, "learning_rate": 1.824202397449548e-05, "loss": 0.9377, "step": 7891 }, { "epoch": 3.574275362318841, "grad_norm": 0.08512775986943237, "learning_rate": 1.8231412473681094e-05, "loss": 0.9424, "step": 7892 }, { "epoch": 3.5747282608695654, "grad_norm": 0.06917101493208266, "learning_rate": 1.8220803149106908e-05, "loss": 0.9201, "step": 7893 }, { "epoch": 3.57518115942029, "grad_norm": 0.0797601929468811, "learning_rate": 1.8210196001833545e-05, "loss": 0.9448, "step": 7894 }, { "epoch": 3.5756340579710146, "grad_norm": 0.06669653965169538, "learning_rate": 1.8199591032921416e-05, "loss": 0.9436, "step": 7895 }, { "epoch": 3.5760869565217392, "grad_norm": 0.07323439406767712, "learning_rate": 1.8188988243430698e-05, "loss": 0.9233, "step": 7896 }, { "epoch": 3.576539855072464, "grad_norm": 0.06358038450907771, "learning_rate": 1.817838763442141e-05, "loss": 0.9283, "step": 7897 }, { "epoch": 3.5769927536231885, "grad_norm": 0.0794894363339244, "learning_rate": 1.816778920695327e-05, "loss": 0.9369, "step": 7898 }, { "epoch": 3.577445652173913, "grad_norm": 0.05813233631290484, "learning_rate": 1.8157192962085854e-05, "loss": 0.9165, "step": 7899 }, { "epoch": 3.5778985507246377, "grad_norm": 0.08576149517301294, "learning_rate": 1.814659890087846e-05, "loss": 0.9114, "step": 7900 }, { "epoch": 3.5783514492753623, "grad_norm": 0.06255924427970903, "learning_rate": 1.8136007024390215e-05, "loss": 0.9321, "step": 7901 }, { "epoch": 3.578804347826087, "grad_norm": 0.08090394900742048, "learning_rate": 1.8125417333679994e-05, "loss": 0.9261, "step": 7902 }, { "epoch": 3.5792572463768115, "grad_norm": 0.06774976065535955, "learning_rate": 1.8114829829806466e-05, "loss": 0.933, "step": 7903 }, { "epoch": 3.579710144927536, "grad_norm": 0.07387135655275678, "learning_rate": 1.8104244513828067e-05, "loss": 0.9363, "step": 7904 }, { "epoch": 3.5801630434782608, "grad_norm": 0.060290870996785054, "learning_rate": 1.8093661386803053e-05, "loss": 0.919, "step": 7905 }, { "epoch": 3.5806159420289854, "grad_norm": 0.07121859815830202, "learning_rate": 1.808308044978941e-05, "loss": 0.9132, "step": 7906 }, { "epoch": 3.58106884057971, "grad_norm": 0.06268965558028214, "learning_rate": 1.8072501703844958e-05, "loss": 0.915, "step": 7907 }, { "epoch": 3.5815217391304346, "grad_norm": 0.06385519481616937, "learning_rate": 1.8061925150027244e-05, "loss": 0.9354, "step": 7908 }, { "epoch": 3.581974637681159, "grad_norm": 0.05984764412259356, "learning_rate": 1.8051350789393654e-05, "loss": 0.9151, "step": 7909 }, { "epoch": 3.582427536231884, "grad_norm": 0.061630392309684284, "learning_rate": 1.8040778623001306e-05, "loss": 0.9392, "step": 7910 }, { "epoch": 3.5828804347826084, "grad_norm": 0.07440393988480934, "learning_rate": 1.803020865190711e-05, "loss": 0.9242, "step": 7911 }, { "epoch": 3.5833333333333335, "grad_norm": 0.05967409914310834, "learning_rate": 1.8019640877167763e-05, "loss": 0.9393, "step": 7912 }, { "epoch": 3.583786231884058, "grad_norm": 0.07694080579494042, "learning_rate": 1.8009075299839755e-05, "loss": 0.9241, "step": 7913 }, { "epoch": 3.5842391304347827, "grad_norm": 0.05292067600075027, "learning_rate": 1.7998511920979327e-05, "loss": 0.9086, "step": 7914 }, { "epoch": 3.5846920289855073, "grad_norm": 0.07274972318919548, "learning_rate": 1.798795074164254e-05, "loss": 0.9391, "step": 7915 }, { "epoch": 3.585144927536232, "grad_norm": 0.05646219295969367, "learning_rate": 1.7977391762885185e-05, "loss": 0.9265, "step": 7916 }, { "epoch": 3.5855978260869565, "grad_norm": 0.07532486403271088, "learning_rate": 1.7966834985762885e-05, "loss": 0.9284, "step": 7917 }, { "epoch": 3.586050724637681, "grad_norm": 0.06193301521801562, "learning_rate": 1.795628041133101e-05, "loss": 0.9065, "step": 7918 }, { "epoch": 3.5865036231884058, "grad_norm": 0.06931291508327751, "learning_rate": 1.7945728040644713e-05, "loss": 0.9157, "step": 7919 }, { "epoch": 3.5869565217391304, "grad_norm": 0.06955873107326814, "learning_rate": 1.793517787475891e-05, "loss": 0.9166, "step": 7920 }, { "epoch": 3.587409420289855, "grad_norm": 0.06703449940703905, "learning_rate": 1.7924629914728363e-05, "loss": 0.9104, "step": 7921 }, { "epoch": 3.5878623188405796, "grad_norm": 0.06963567456475314, "learning_rate": 1.7914084161607528e-05, "loss": 0.9338, "step": 7922 }, { "epoch": 3.588315217391304, "grad_norm": 0.07278168154980877, "learning_rate": 1.7903540616450713e-05, "loss": 0.935, "step": 7923 }, { "epoch": 3.588768115942029, "grad_norm": 0.06762757135900353, "learning_rate": 1.789299928031194e-05, "loss": 0.9048, "step": 7924 }, { "epoch": 3.589221014492754, "grad_norm": 0.06486861751139208, "learning_rate": 1.7882460154245077e-05, "loss": 0.9315, "step": 7925 }, { "epoch": 3.5896739130434785, "grad_norm": 0.07221055808927851, "learning_rate": 1.7871923239303718e-05, "loss": 0.9343, "step": 7926 }, { "epoch": 3.590126811594203, "grad_norm": 0.06066926010727725, "learning_rate": 1.786138853654126e-05, "loss": 0.9088, "step": 7927 }, { "epoch": 3.5905797101449277, "grad_norm": 0.07289128685190746, "learning_rate": 1.785085604701086e-05, "loss": 0.9013, "step": 7928 }, { "epoch": 3.5910326086956523, "grad_norm": 0.05323023551595176, "learning_rate": 1.784032577176549e-05, "loss": 0.9311, "step": 7929 }, { "epoch": 3.591485507246377, "grad_norm": 0.07149779758261487, "learning_rate": 1.782979771185785e-05, "loss": 0.9442, "step": 7930 }, { "epoch": 3.5919384057971016, "grad_norm": 0.05690918403558952, "learning_rate": 1.7819271868340482e-05, "loss": 0.9493, "step": 7931 }, { "epoch": 3.592391304347826, "grad_norm": 0.07045139684513714, "learning_rate": 1.780874824226564e-05, "loss": 0.9246, "step": 7932 }, { "epoch": 3.5928442028985508, "grad_norm": 0.0664762481808858, "learning_rate": 1.7798226834685416e-05, "loss": 0.933, "step": 7933 }, { "epoch": 3.5932971014492754, "grad_norm": 0.0591245347315767, "learning_rate": 1.778770764665162e-05, "loss": 0.9318, "step": 7934 }, { "epoch": 3.59375, "grad_norm": 0.060458900026153846, "learning_rate": 1.7777190679215923e-05, "loss": 0.9119, "step": 7935 }, { "epoch": 3.5942028985507246, "grad_norm": 0.06574159325521028, "learning_rate": 1.776667593342967e-05, "loss": 0.9164, "step": 7936 }, { "epoch": 3.5946557971014492, "grad_norm": 0.058101595291346134, "learning_rate": 1.7756163410344044e-05, "loss": 0.9127, "step": 7937 }, { "epoch": 3.595108695652174, "grad_norm": 0.06118441653673928, "learning_rate": 1.7745653111010024e-05, "loss": 0.9279, "step": 7938 }, { "epoch": 3.5955615942028984, "grad_norm": 0.07022031360492037, "learning_rate": 1.7735145036478313e-05, "loss": 0.934, "step": 7939 }, { "epoch": 3.596014492753623, "grad_norm": 0.054166611417607376, "learning_rate": 1.772463918779945e-05, "loss": 0.9265, "step": 7940 }, { "epoch": 3.5964673913043477, "grad_norm": 0.06912950070284576, "learning_rate": 1.7714135566023688e-05, "loss": 0.9443, "step": 7941 }, { "epoch": 3.5969202898550723, "grad_norm": 0.06036514015017891, "learning_rate": 1.770363417220112e-05, "loss": 0.9411, "step": 7942 }, { "epoch": 3.597373188405797, "grad_norm": 0.058596346350410874, "learning_rate": 1.7693135007381568e-05, "loss": 0.9243, "step": 7943 }, { "epoch": 3.5978260869565215, "grad_norm": 0.06186054626228904, "learning_rate": 1.7682638072614667e-05, "loss": 0.9361, "step": 7944 }, { "epoch": 3.598278985507246, "grad_norm": 0.054317240175001176, "learning_rate": 1.7672143368949806e-05, "loss": 0.9124, "step": 7945 }, { "epoch": 3.598731884057971, "grad_norm": 0.06408873258781729, "learning_rate": 1.7661650897436155e-05, "loss": 0.9284, "step": 7946 }, { "epoch": 3.599184782608696, "grad_norm": 0.05641475687754891, "learning_rate": 1.7651160659122647e-05, "loss": 0.9389, "step": 7947 }, { "epoch": 3.5996376811594204, "grad_norm": 0.05832856763407222, "learning_rate": 1.764067265505803e-05, "loss": 0.9453, "step": 7948 }, { "epoch": 3.600090579710145, "grad_norm": 0.061340837398298736, "learning_rate": 1.7630186886290794e-05, "loss": 0.9201, "step": 7949 }, { "epoch": 3.6005434782608696, "grad_norm": 0.05780472658720887, "learning_rate": 1.7619703353869235e-05, "loss": 0.932, "step": 7950 }, { "epoch": 3.6009963768115942, "grad_norm": 0.06459580444311608, "learning_rate": 1.7609222058841377e-05, "loss": 0.9378, "step": 7951 }, { "epoch": 3.601449275362319, "grad_norm": 0.05525974376406181, "learning_rate": 1.759874300225509e-05, "loss": 0.9386, "step": 7952 }, { "epoch": 3.6019021739130435, "grad_norm": 0.066188956463271, "learning_rate": 1.758826618515796e-05, "loss": 0.9206, "step": 7953 }, { "epoch": 3.602355072463768, "grad_norm": 0.059189789526258814, "learning_rate": 1.757779160859738e-05, "loss": 0.9396, "step": 7954 }, { "epoch": 3.6028079710144927, "grad_norm": 0.05948095801771432, "learning_rate": 1.7567319273620484e-05, "loss": 0.9229, "step": 7955 }, { "epoch": 3.6032608695652173, "grad_norm": 0.060812021408014505, "learning_rate": 1.7556849181274244e-05, "loss": 0.9311, "step": 7956 }, { "epoch": 3.603713768115942, "grad_norm": 0.05587440677559265, "learning_rate": 1.7546381332605342e-05, "loss": 0.9031, "step": 7957 }, { "epoch": 3.6041666666666665, "grad_norm": 0.07515177305755774, "learning_rate": 1.753591572866029e-05, "loss": 0.914, "step": 7958 }, { "epoch": 3.6046195652173916, "grad_norm": 0.0600747182442838, "learning_rate": 1.752545237048533e-05, "loss": 0.9119, "step": 7959 }, { "epoch": 3.605072463768116, "grad_norm": 0.05572171924765303, "learning_rate": 1.7514991259126518e-05, "loss": 0.9271, "step": 7960 }, { "epoch": 3.605525362318841, "grad_norm": 0.07947701832036944, "learning_rate": 1.7504532395629668e-05, "loss": 0.928, "step": 7961 }, { "epoch": 3.6059782608695654, "grad_norm": 0.05562106429528723, "learning_rate": 1.749407578104036e-05, "loss": 0.8961, "step": 7962 }, { "epoch": 3.60643115942029, "grad_norm": 0.08238512655589696, "learning_rate": 1.7483621416403945e-05, "loss": 0.9328, "step": 7963 }, { "epoch": 3.6068840579710146, "grad_norm": 0.06678563207021398, "learning_rate": 1.7473169302765587e-05, "loss": 0.9062, "step": 7964 }, { "epoch": 3.6073369565217392, "grad_norm": 0.06712481235491832, "learning_rate": 1.7462719441170182e-05, "loss": 0.9223, "step": 7965 }, { "epoch": 3.607789855072464, "grad_norm": 0.06877530134428342, "learning_rate": 1.745227183266244e-05, "loss": 0.9299, "step": 7966 }, { "epoch": 3.6082427536231885, "grad_norm": 0.051344305853640006, "learning_rate": 1.7441826478286797e-05, "loss": 0.9088, "step": 7967 }, { "epoch": 3.608695652173913, "grad_norm": 0.05489451747923489, "learning_rate": 1.7431383379087523e-05, "loss": 0.9379, "step": 7968 }, { "epoch": 3.6091485507246377, "grad_norm": 0.06628811112264309, "learning_rate": 1.7420942536108616e-05, "loss": 0.9312, "step": 7969 }, { "epoch": 3.6096014492753623, "grad_norm": 0.04611286019656767, "learning_rate": 1.7410503950393867e-05, "loss": 0.9266, "step": 7970 }, { "epoch": 3.610054347826087, "grad_norm": 0.06429732958515855, "learning_rate": 1.7400067622986813e-05, "loss": 0.8831, "step": 7971 }, { "epoch": 3.6105072463768115, "grad_norm": 0.05825180113459643, "learning_rate": 1.738963355493083e-05, "loss": 0.9395, "step": 7972 }, { "epoch": 3.610960144927536, "grad_norm": 0.05398576933873629, "learning_rate": 1.737920174726899e-05, "loss": 0.9386, "step": 7973 }, { "epoch": 3.6114130434782608, "grad_norm": 0.06617979585409155, "learning_rate": 1.736877220104422e-05, "loss": 0.9097, "step": 7974 }, { "epoch": 3.6118659420289854, "grad_norm": 0.0542562166935106, "learning_rate": 1.7358344917299134e-05, "loss": 0.9102, "step": 7975 }, { "epoch": 3.61231884057971, "grad_norm": 0.06484959078991805, "learning_rate": 1.7347919897076205e-05, "loss": 0.9223, "step": 7976 }, { "epoch": 3.6127717391304346, "grad_norm": 0.06285522870228173, "learning_rate": 1.733749714141762e-05, "loss": 0.9469, "step": 7977 }, { "epoch": 3.613224637681159, "grad_norm": 0.054814172815441584, "learning_rate": 1.7327076651365355e-05, "loss": 0.9188, "step": 7978 }, { "epoch": 3.613677536231884, "grad_norm": 0.06113834040898386, "learning_rate": 1.731665842796115e-05, "loss": 0.9351, "step": 7979 }, { "epoch": 3.6141304347826084, "grad_norm": 0.06476112978939838, "learning_rate": 1.7306242472246556e-05, "loss": 0.9175, "step": 7980 }, { "epoch": 3.6145833333333335, "grad_norm": 0.053863097106490655, "learning_rate": 1.7295828785262857e-05, "loss": 0.9305, "step": 7981 }, { "epoch": 3.615036231884058, "grad_norm": 0.06387612609108478, "learning_rate": 1.7285417368051142e-05, "loss": 0.9298, "step": 7982 }, { "epoch": 3.6154891304347827, "grad_norm": 0.06269035353776652, "learning_rate": 1.7275008221652237e-05, "loss": 0.9325, "step": 7983 }, { "epoch": 3.6159420289855073, "grad_norm": 0.055109543557592254, "learning_rate": 1.7264601347106783e-05, "loss": 0.921, "step": 7984 }, { "epoch": 3.616394927536232, "grad_norm": 0.06379620178607515, "learning_rate": 1.7254196745455144e-05, "loss": 0.9128, "step": 7985 }, { "epoch": 3.6168478260869565, "grad_norm": 0.061122111405601524, "learning_rate": 1.724379441773752e-05, "loss": 0.9396, "step": 7986 }, { "epoch": 3.617300724637681, "grad_norm": 0.05866770964057657, "learning_rate": 1.7233394364993822e-05, "loss": 0.9226, "step": 7987 }, { "epoch": 3.6177536231884058, "grad_norm": 0.06448184183573313, "learning_rate": 1.722299658826377e-05, "loss": 0.9318, "step": 7988 }, { "epoch": 3.6182065217391304, "grad_norm": 0.052861978464522744, "learning_rate": 1.7212601088586823e-05, "loss": 0.9128, "step": 7989 }, { "epoch": 3.618659420289855, "grad_norm": 0.05875637518131255, "learning_rate": 1.7202207867002273e-05, "loss": 0.9142, "step": 7990 }, { "epoch": 3.6191123188405796, "grad_norm": 0.06584760354829798, "learning_rate": 1.719181692454911e-05, "loss": 0.9041, "step": 7991 }, { "epoch": 3.619565217391304, "grad_norm": 0.05335836501306506, "learning_rate": 1.7181428262266172e-05, "loss": 0.9231, "step": 7992 }, { "epoch": 3.620018115942029, "grad_norm": 0.057057615384106405, "learning_rate": 1.7171041881191995e-05, "loss": 0.8972, "step": 7993 }, { "epoch": 3.620471014492754, "grad_norm": 0.05909373210365424, "learning_rate": 1.716065778236495e-05, "loss": 0.9161, "step": 7994 }, { "epoch": 3.6209239130434785, "grad_norm": 0.05835700169407477, "learning_rate": 1.715027596682314e-05, "loss": 0.9217, "step": 7995 }, { "epoch": 3.621376811594203, "grad_norm": 0.06024235402295962, "learning_rate": 1.7139896435604448e-05, "loss": 0.9239, "step": 7996 }, { "epoch": 3.6218297101449277, "grad_norm": 0.05650045759910753, "learning_rate": 1.7129519189746518e-05, "loss": 0.9122, "step": 7997 }, { "epoch": 3.6222826086956523, "grad_norm": 0.06335504280909932, "learning_rate": 1.7119144230286812e-05, "loss": 0.9321, "step": 7998 }, { "epoch": 3.622735507246377, "grad_norm": 0.055636788768709415, "learning_rate": 1.7108771558262504e-05, "loss": 0.9218, "step": 7999 }, { "epoch": 3.6231884057971016, "grad_norm": 0.06329914323709517, "learning_rate": 1.7098401174710586e-05, "loss": 0.9298, "step": 8000 }, { "epoch": 3.623641304347826, "grad_norm": 0.056182069067060504, "learning_rate": 1.7088033080667782e-05, "loss": 0.9301, "step": 8001 }, { "epoch": 3.6240942028985508, "grad_norm": 0.05997587069075778, "learning_rate": 1.7077667277170635e-05, "loss": 0.9085, "step": 8002 }, { "epoch": 3.6245471014492754, "grad_norm": 0.05465091460826914, "learning_rate": 1.706730376525541e-05, "loss": 0.94, "step": 8003 }, { "epoch": 3.625, "grad_norm": 0.0572259649587528, "learning_rate": 1.7056942545958167e-05, "loss": 0.9267, "step": 8004 }, { "epoch": 3.6254528985507246, "grad_norm": 0.06023843174949513, "learning_rate": 1.7046583620314723e-05, "loss": 0.9479, "step": 8005 }, { "epoch": 3.6259057971014492, "grad_norm": 0.06220738757775257, "learning_rate": 1.70362269893607e-05, "loss": 0.94, "step": 8006 }, { "epoch": 3.626358695652174, "grad_norm": 0.058089508661333306, "learning_rate": 1.702587265413144e-05, "loss": 0.9181, "step": 8007 }, { "epoch": 3.6268115942028984, "grad_norm": 0.06055602851796079, "learning_rate": 1.7015520615662108e-05, "loss": 0.9407, "step": 8008 }, { "epoch": 3.627264492753623, "grad_norm": 0.054381752516167045, "learning_rate": 1.700517087498759e-05, "loss": 0.9164, "step": 8009 }, { "epoch": 3.6277173913043477, "grad_norm": 0.06001807603627594, "learning_rate": 1.6994823433142595e-05, "loss": 0.9021, "step": 8010 }, { "epoch": 3.6281702898550723, "grad_norm": 0.06539607902901269, "learning_rate": 1.6984478291161558e-05, "loss": 0.9434, "step": 8011 }, { "epoch": 3.628623188405797, "grad_norm": 0.05348774092695101, "learning_rate": 1.697413545007869e-05, "loss": 0.9412, "step": 8012 }, { "epoch": 3.6290760869565215, "grad_norm": 0.06722182600409585, "learning_rate": 1.696379491092798e-05, "loss": 0.9125, "step": 8013 }, { "epoch": 3.629528985507246, "grad_norm": 0.06367324945050128, "learning_rate": 1.6953456674743206e-05, "loss": 0.9334, "step": 8014 }, { "epoch": 3.629981884057971, "grad_norm": 0.07223866299960231, "learning_rate": 1.694312074255788e-05, "loss": 0.9273, "step": 8015 }, { "epoch": 3.630434782608696, "grad_norm": 0.06563501082777189, "learning_rate": 1.6932787115405318e-05, "loss": 0.916, "step": 8016 }, { "epoch": 3.6308876811594204, "grad_norm": 0.0721901473261015, "learning_rate": 1.692245579431857e-05, "loss": 0.9309, "step": 8017 }, { "epoch": 3.631340579710145, "grad_norm": 0.06734812077549106, "learning_rate": 1.6912126780330503e-05, "loss": 0.9229, "step": 8018 }, { "epoch": 3.6317934782608696, "grad_norm": 0.057060269278125436, "learning_rate": 1.6901800074473706e-05, "loss": 0.9554, "step": 8019 }, { "epoch": 3.6322463768115942, "grad_norm": 0.0705297799002014, "learning_rate": 1.6891475677780556e-05, "loss": 0.9389, "step": 8020 }, { "epoch": 3.632699275362319, "grad_norm": 0.06200358171401451, "learning_rate": 1.6881153591283185e-05, "loss": 0.9308, "step": 8021 }, { "epoch": 3.6331521739130435, "grad_norm": 0.0642423471221402, "learning_rate": 1.6870833816013545e-05, "loss": 0.934, "step": 8022 }, { "epoch": 3.633605072463768, "grad_norm": 0.06578549119692831, "learning_rate": 1.686051635300328e-05, "loss": 0.9219, "step": 8023 }, { "epoch": 3.6340579710144927, "grad_norm": 0.056437561659849776, "learning_rate": 1.685020120328388e-05, "loss": 0.9194, "step": 8024 }, { "epoch": 3.6345108695652173, "grad_norm": 0.06699909203670316, "learning_rate": 1.6839888367886533e-05, "loss": 0.916, "step": 8025 }, { "epoch": 3.634963768115942, "grad_norm": 0.053888182528474465, "learning_rate": 1.6829577847842263e-05, "loss": 0.9233, "step": 8026 }, { "epoch": 3.6354166666666665, "grad_norm": 0.06656441770182439, "learning_rate": 1.68192696441818e-05, "loss": 0.9211, "step": 8027 }, { "epoch": 3.6358695652173916, "grad_norm": 0.055332125665852855, "learning_rate": 1.6808963757935696e-05, "loss": 0.9442, "step": 8028 }, { "epoch": 3.636322463768116, "grad_norm": 0.0705531302539985, "learning_rate": 1.679866019013423e-05, "loss": 0.9447, "step": 8029 }, { "epoch": 3.636775362318841, "grad_norm": 0.05462453120525425, "learning_rate": 1.6788358941807472e-05, "loss": 0.9082, "step": 8030 }, { "epoch": 3.6372282608695654, "grad_norm": 0.05960475681600616, "learning_rate": 1.677806001398524e-05, "loss": 0.9313, "step": 8031 }, { "epoch": 3.63768115942029, "grad_norm": 0.056180501119353896, "learning_rate": 1.6767763407697165e-05, "loss": 0.9348, "step": 8032 }, { "epoch": 3.6381340579710146, "grad_norm": 0.05653833539334374, "learning_rate": 1.6757469123972578e-05, "loss": 0.9259, "step": 8033 }, { "epoch": 3.6385869565217392, "grad_norm": 0.0591345957118159, "learning_rate": 1.674717716384065e-05, "loss": 0.9208, "step": 8034 }, { "epoch": 3.639039855072464, "grad_norm": 0.058864348813313996, "learning_rate": 1.6736887528330248e-05, "loss": 0.9385, "step": 8035 }, { "epoch": 3.6394927536231885, "grad_norm": 0.061595437580559506, "learning_rate": 1.6726600218470085e-05, "loss": 0.9248, "step": 8036 }, { "epoch": 3.639945652173913, "grad_norm": 0.06492094423597146, "learning_rate": 1.6716315235288574e-05, "loss": 0.9536, "step": 8037 }, { "epoch": 3.6403985507246377, "grad_norm": 0.06268803754238145, "learning_rate": 1.6706032579813927e-05, "loss": 0.9314, "step": 8038 }, { "epoch": 3.6408514492753623, "grad_norm": 0.05753213700364053, "learning_rate": 1.6695752253074097e-05, "loss": 0.9169, "step": 8039 }, { "epoch": 3.641304347826087, "grad_norm": 0.06274817172574942, "learning_rate": 1.6685474256096865e-05, "loss": 0.9439, "step": 8040 }, { "epoch": 3.6417572463768115, "grad_norm": 0.06563291010465336, "learning_rate": 1.6675198589909696e-05, "loss": 0.9374, "step": 8041 }, { "epoch": 3.642210144927536, "grad_norm": 0.06021485913427883, "learning_rate": 1.66649252555399e-05, "loss": 0.94, "step": 8042 }, { "epoch": 3.6426630434782608, "grad_norm": 0.06162292130209845, "learning_rate": 1.6654654254014494e-05, "loss": 0.9224, "step": 8043 }, { "epoch": 3.6431159420289854, "grad_norm": 0.06479591678806183, "learning_rate": 1.6644385586360304e-05, "loss": 0.9398, "step": 8044 }, { "epoch": 3.64356884057971, "grad_norm": 0.05307246472622156, "learning_rate": 1.6634119253603903e-05, "loss": 0.9228, "step": 8045 }, { "epoch": 3.6440217391304346, "grad_norm": 0.05886412760993708, "learning_rate": 1.6623855256771623e-05, "loss": 0.939, "step": 8046 }, { "epoch": 3.644474637681159, "grad_norm": 0.06244444059564103, "learning_rate": 1.6613593596889566e-05, "loss": 0.9164, "step": 8047 }, { "epoch": 3.644927536231884, "grad_norm": 0.05747733041499147, "learning_rate": 1.6603334274983626e-05, "loss": 0.9237, "step": 8048 }, { "epoch": 3.6453804347826084, "grad_norm": 0.05487253020846845, "learning_rate": 1.659307729207942e-05, "loss": 0.8905, "step": 8049 }, { "epoch": 3.6458333333333335, "grad_norm": 0.05464950562240707, "learning_rate": 1.6582822649202382e-05, "loss": 0.9307, "step": 8050 }, { "epoch": 3.646286231884058, "grad_norm": 0.06499532141147668, "learning_rate": 1.657257034737766e-05, "loss": 0.931, "step": 8051 }, { "epoch": 3.6467391304347827, "grad_norm": 0.05676046589326725, "learning_rate": 1.6562320387630216e-05, "loss": 0.9352, "step": 8052 }, { "epoch": 3.6471920289855073, "grad_norm": 0.056081713898561765, "learning_rate": 1.6552072770984744e-05, "loss": 0.934, "step": 8053 }, { "epoch": 3.647644927536232, "grad_norm": 0.054324664373427575, "learning_rate": 1.654182749846571e-05, "loss": 0.9326, "step": 8054 }, { "epoch": 3.6480978260869565, "grad_norm": 0.05618625698431923, "learning_rate": 1.6531584571097344e-05, "loss": 0.9269, "step": 8055 }, { "epoch": 3.648550724637681, "grad_norm": 0.05771560255732399, "learning_rate": 1.6521343989903673e-05, "loss": 0.9436, "step": 8056 }, { "epoch": 3.6490036231884058, "grad_norm": 0.05585338902512025, "learning_rate": 1.6511105755908434e-05, "loss": 0.9395, "step": 8057 }, { "epoch": 3.6494565217391304, "grad_norm": 0.05933635559626731, "learning_rate": 1.6500869870135186e-05, "loss": 0.9217, "step": 8058 }, { "epoch": 3.649909420289855, "grad_norm": 0.05638507064448724, "learning_rate": 1.64906363336072e-05, "loss": 0.9276, "step": 8059 }, { "epoch": 3.6503623188405796, "grad_norm": 0.05733680584998444, "learning_rate": 1.6480405147347567e-05, "loss": 0.9135, "step": 8060 }, { "epoch": 3.650815217391304, "grad_norm": 0.05388693789525262, "learning_rate": 1.64701763123791e-05, "loss": 0.9299, "step": 8061 }, { "epoch": 3.651268115942029, "grad_norm": 0.05973526390090892, "learning_rate": 1.6459949829724397e-05, "loss": 0.9313, "step": 8062 }, { "epoch": 3.651721014492754, "grad_norm": 0.060751478655536824, "learning_rate": 1.6449725700405813e-05, "loss": 0.9291, "step": 8063 }, { "epoch": 3.6521739130434785, "grad_norm": 0.05844115457238079, "learning_rate": 1.6439503925445453e-05, "loss": 0.9301, "step": 8064 }, { "epoch": 3.652626811594203, "grad_norm": 0.06660375413906401, "learning_rate": 1.6429284505865232e-05, "loss": 0.9103, "step": 8065 }, { "epoch": 3.6530797101449277, "grad_norm": 0.05674823773200116, "learning_rate": 1.6419067442686777e-05, "loss": 0.9235, "step": 8066 }, { "epoch": 3.6535326086956523, "grad_norm": 0.06563326933720896, "learning_rate": 1.6408852736931532e-05, "loss": 0.9309, "step": 8067 }, { "epoch": 3.653985507246377, "grad_norm": 0.05607918263216044, "learning_rate": 1.639864038962065e-05, "loss": 0.9142, "step": 8068 }, { "epoch": 3.6544384057971016, "grad_norm": 0.06799838510125122, "learning_rate": 1.63884304017751e-05, "loss": 0.9148, "step": 8069 }, { "epoch": 3.654891304347826, "grad_norm": 0.05320641244506582, "learning_rate": 1.6378222774415565e-05, "loss": 0.9532, "step": 8070 }, { "epoch": 3.6553442028985508, "grad_norm": 0.062274912504362664, "learning_rate": 1.6368017508562545e-05, "loss": 0.9377, "step": 8071 }, { "epoch": 3.6557971014492754, "grad_norm": 0.05883504318201446, "learning_rate": 1.6357814605236262e-05, "loss": 0.9081, "step": 8072 }, { "epoch": 3.65625, "grad_norm": 0.057817519054328026, "learning_rate": 1.6347614065456715e-05, "loss": 0.9317, "step": 8073 }, { "epoch": 3.6567028985507246, "grad_norm": 0.05507632432297807, "learning_rate": 1.633741589024366e-05, "loss": 0.9153, "step": 8074 }, { "epoch": 3.6571557971014492, "grad_norm": 0.05772510963686833, "learning_rate": 1.6327220080616652e-05, "loss": 0.9282, "step": 8075 }, { "epoch": 3.657608695652174, "grad_norm": 0.0604521452343116, "learning_rate": 1.6317026637594945e-05, "loss": 0.9064, "step": 8076 }, { "epoch": 3.6580615942028984, "grad_norm": 0.05999452545370729, "learning_rate": 1.630683556219763e-05, "loss": 0.9185, "step": 8077 }, { "epoch": 3.658514492753623, "grad_norm": 0.0597943036636698, "learning_rate": 1.6296646855443495e-05, "loss": 0.9235, "step": 8078 }, { "epoch": 3.6589673913043477, "grad_norm": 0.06090634347670857, "learning_rate": 1.6286460518351148e-05, "loss": 0.9145, "step": 8079 }, { "epoch": 3.6594202898550723, "grad_norm": 0.062224550417680824, "learning_rate": 1.6276276551938926e-05, "loss": 0.9204, "step": 8080 }, { "epoch": 3.659873188405797, "grad_norm": 0.07223424508924674, "learning_rate": 1.6266094957224924e-05, "loss": 0.9394, "step": 8081 }, { "epoch": 3.6603260869565215, "grad_norm": 0.05945355765907468, "learning_rate": 1.6255915735227006e-05, "loss": 0.9132, "step": 8082 }, { "epoch": 3.660778985507246, "grad_norm": 0.06026918146642293, "learning_rate": 1.6245738886962832e-05, "loss": 0.9135, "step": 8083 }, { "epoch": 3.661231884057971, "grad_norm": 0.1301051026109173, "learning_rate": 1.623556441344977e-05, "loss": 0.9231, "step": 8084 }, { "epoch": 3.661684782608696, "grad_norm": 0.06498391870514587, "learning_rate": 1.6225392315705006e-05, "loss": 0.9435, "step": 8085 }, { "epoch": 3.6621376811594204, "grad_norm": 0.05656183332190103, "learning_rate": 1.621522259474543e-05, "loss": 0.9228, "step": 8086 }, { "epoch": 3.662590579710145, "grad_norm": 0.06626404340910093, "learning_rate": 1.6205055251587757e-05, "loss": 0.9315, "step": 8087 }, { "epoch": 3.6630434782608696, "grad_norm": 0.056201587999139956, "learning_rate": 1.619489028724842e-05, "loss": 0.9112, "step": 8088 }, { "epoch": 3.6634963768115942, "grad_norm": 0.06183590889133021, "learning_rate": 1.618472770274362e-05, "loss": 0.9306, "step": 8089 }, { "epoch": 3.663949275362319, "grad_norm": 0.06680155831361478, "learning_rate": 1.6174567499089318e-05, "loss": 0.9357, "step": 8090 }, { "epoch": 3.6644021739130435, "grad_norm": 0.0635311794858998, "learning_rate": 1.616440967730127e-05, "loss": 0.9189, "step": 8091 }, { "epoch": 3.664855072463768, "grad_norm": 0.06615739961966269, "learning_rate": 1.615425423839494e-05, "loss": 0.9273, "step": 8092 }, { "epoch": 3.6653079710144927, "grad_norm": 0.05984261653153082, "learning_rate": 1.614410118338562e-05, "loss": 0.9379, "step": 8093 }, { "epoch": 3.6657608695652173, "grad_norm": 0.06464916455997508, "learning_rate": 1.613395051328829e-05, "loss": 0.9505, "step": 8094 }, { "epoch": 3.666213768115942, "grad_norm": 0.05648343069891812, "learning_rate": 1.6123802229117762e-05, "loss": 0.9306, "step": 8095 }, { "epoch": 3.6666666666666665, "grad_norm": 0.054854837310419076, "learning_rate": 1.6113656331888563e-05, "loss": 0.926, "step": 8096 }, { "epoch": 3.6671195652173916, "grad_norm": 0.06757841633455308, "learning_rate": 1.610351282261499e-05, "loss": 0.9208, "step": 8097 }, { "epoch": 3.667572463768116, "grad_norm": 0.07267436312979127, "learning_rate": 1.6093371702311088e-05, "loss": 0.9444, "step": 8098 }, { "epoch": 3.668025362318841, "grad_norm": 0.062294530443883306, "learning_rate": 1.608323297199072e-05, "loss": 0.921, "step": 8099 }, { "epoch": 3.6684782608695654, "grad_norm": 0.054360300725843975, "learning_rate": 1.6073096632667437e-05, "loss": 0.9078, "step": 8100 }, { "epoch": 3.66893115942029, "grad_norm": 0.0548188812657846, "learning_rate": 1.606296268535461e-05, "loss": 0.9246, "step": 8101 }, { "epoch": 3.6693840579710146, "grad_norm": 0.056013495138166364, "learning_rate": 1.6052831131065323e-05, "loss": 0.9116, "step": 8102 }, { "epoch": 3.6698369565217392, "grad_norm": 0.05419031805034522, "learning_rate": 1.6042701970812468e-05, "loss": 0.9479, "step": 8103 }, { "epoch": 3.670289855072464, "grad_norm": 0.05513207190351107, "learning_rate": 1.6032575205608664e-05, "loss": 0.9112, "step": 8104 }, { "epoch": 3.6707427536231885, "grad_norm": 0.06061311714168587, "learning_rate": 1.6022450836466294e-05, "loss": 0.9264, "step": 8105 }, { "epoch": 3.671195652173913, "grad_norm": 0.05279438727516655, "learning_rate": 1.6012328864397494e-05, "loss": 0.9182, "step": 8106 }, { "epoch": 3.6716485507246377, "grad_norm": 0.05969834453809159, "learning_rate": 1.6002209290414203e-05, "loss": 0.9291, "step": 8107 }, { "epoch": 3.6721014492753623, "grad_norm": 0.05388615326701192, "learning_rate": 1.5992092115528067e-05, "loss": 0.9392, "step": 8108 }, { "epoch": 3.672554347826087, "grad_norm": 0.05878916449971029, "learning_rate": 1.5981977340750537e-05, "loss": 0.9406, "step": 8109 }, { "epoch": 3.6730072463768115, "grad_norm": 0.06038181931356845, "learning_rate": 1.5971864967092776e-05, "loss": 0.9272, "step": 8110 }, { "epoch": 3.673460144927536, "grad_norm": 0.05521244433907904, "learning_rate": 1.596175499556577e-05, "loss": 0.9235, "step": 8111 }, { "epoch": 3.6739130434782608, "grad_norm": 0.05811456694456294, "learning_rate": 1.595164742718019e-05, "loss": 0.9177, "step": 8112 }, { "epoch": 3.6743659420289854, "grad_norm": 0.052857768746867984, "learning_rate": 1.5941542262946553e-05, "loss": 0.914, "step": 8113 }, { "epoch": 3.67481884057971, "grad_norm": 0.060439031486901565, "learning_rate": 1.5931439503875036e-05, "loss": 0.9246, "step": 8114 }, { "epoch": 3.6752717391304346, "grad_norm": 0.05892328697128165, "learning_rate": 1.592133915097566e-05, "loss": 0.9313, "step": 8115 }, { "epoch": 3.675724637681159, "grad_norm": 0.05732161951757676, "learning_rate": 1.5911241205258153e-05, "loss": 0.936, "step": 8116 }, { "epoch": 3.676177536231884, "grad_norm": 0.05547066956793772, "learning_rate": 1.5901145667732046e-05, "loss": 0.9235, "step": 8117 }, { "epoch": 3.6766304347826084, "grad_norm": 0.05414144391852717, "learning_rate": 1.589105253940659e-05, "loss": 0.9151, "step": 8118 }, { "epoch": 3.6770833333333335, "grad_norm": 0.0556707707370092, "learning_rate": 1.588096182129082e-05, "loss": 0.9191, "step": 8119 }, { "epoch": 3.677536231884058, "grad_norm": 0.059683511733017704, "learning_rate": 1.5870873514393507e-05, "loss": 0.9367, "step": 8120 }, { "epoch": 3.6779891304347827, "grad_norm": 0.05360306712418817, "learning_rate": 1.586078761972322e-05, "loss": 0.9208, "step": 8121 }, { "epoch": 3.6784420289855073, "grad_norm": 0.05430048393441726, "learning_rate": 1.5850704138288244e-05, "loss": 0.9303, "step": 8122 }, { "epoch": 3.678894927536232, "grad_norm": 0.054811308509685186, "learning_rate": 1.584062307109664e-05, "loss": 0.907, "step": 8123 }, { "epoch": 3.6793478260869565, "grad_norm": 0.05472574070678434, "learning_rate": 1.5830544419156223e-05, "loss": 0.9097, "step": 8124 }, { "epoch": 3.679800724637681, "grad_norm": 0.061591699100634306, "learning_rate": 1.5820468183474588e-05, "loss": 0.9318, "step": 8125 }, { "epoch": 3.6802536231884058, "grad_norm": 0.054284583709818535, "learning_rate": 1.5810394365059058e-05, "loss": 0.9149, "step": 8126 }, { "epoch": 3.6807065217391304, "grad_norm": 0.06240507579752003, "learning_rate": 1.5800322964916747e-05, "loss": 0.9252, "step": 8127 }, { "epoch": 3.681159420289855, "grad_norm": 0.057727464138216604, "learning_rate": 1.579025398405448e-05, "loss": 0.9248, "step": 8128 }, { "epoch": 3.6816123188405796, "grad_norm": 0.06611085485817908, "learning_rate": 1.5780187423478902e-05, "loss": 0.9433, "step": 8129 }, { "epoch": 3.682065217391304, "grad_norm": 0.0535290448491184, "learning_rate": 1.5770123284196365e-05, "loss": 0.9209, "step": 8130 }, { "epoch": 3.682518115942029, "grad_norm": 0.0648027124523306, "learning_rate": 1.5760061567212996e-05, "loss": 0.9371, "step": 8131 }, { "epoch": 3.682971014492754, "grad_norm": 0.05409016127712561, "learning_rate": 1.5750002273534676e-05, "loss": 0.9282, "step": 8132 }, { "epoch": 3.6834239130434785, "grad_norm": 0.0535548077762624, "learning_rate": 1.5739945404167064e-05, "loss": 0.9083, "step": 8133 }, { "epoch": 3.683876811594203, "grad_norm": 0.06211356619403498, "learning_rate": 1.5729890960115545e-05, "loss": 0.9199, "step": 8134 }, { "epoch": 3.6843297101449277, "grad_norm": 0.34393299866579746, "learning_rate": 1.5719838942385298e-05, "loss": 0.9595, "step": 8135 }, { "epoch": 3.6847826086956523, "grad_norm": 0.060498188531768006, "learning_rate": 1.5709789351981214e-05, "loss": 0.9346, "step": 8136 }, { "epoch": 3.685235507246377, "grad_norm": 0.06627623805988206, "learning_rate": 1.569974218990799e-05, "loss": 0.9253, "step": 8137 }, { "epoch": 3.6856884057971016, "grad_norm": 0.06730814835144185, "learning_rate": 1.5689697457170053e-05, "loss": 0.9331, "step": 8138 }, { "epoch": 3.686141304347826, "grad_norm": 0.0649782263959563, "learning_rate": 1.567965515477158e-05, "loss": 0.9097, "step": 8139 }, { "epoch": 3.6865942028985508, "grad_norm": 0.061106788978363955, "learning_rate": 1.5669615283716506e-05, "loss": 0.9277, "step": 8140 }, { "epoch": 3.6870471014492754, "grad_norm": 0.0669668006592492, "learning_rate": 1.565957784500856e-05, "loss": 0.9319, "step": 8141 }, { "epoch": 3.6875, "grad_norm": 0.055385611361486696, "learning_rate": 1.5649542839651175e-05, "loss": 0.9273, "step": 8142 }, { "epoch": 3.6879528985507246, "grad_norm": 0.06138705765505473, "learning_rate": 1.5639510268647594e-05, "loss": 0.9135, "step": 8143 }, { "epoch": 3.6884057971014492, "grad_norm": 0.06132956746636293, "learning_rate": 1.5629480133000753e-05, "loss": 0.9192, "step": 8144 }, { "epoch": 3.688858695652174, "grad_norm": 0.05909787232518087, "learning_rate": 1.5619452433713426e-05, "loss": 0.9618, "step": 8145 }, { "epoch": 3.6893115942028984, "grad_norm": 0.0591387873268634, "learning_rate": 1.5609427171788064e-05, "loss": 0.9249, "step": 8146 }, { "epoch": 3.689764492753623, "grad_norm": 0.05761133450564159, "learning_rate": 1.5599404348226923e-05, "loss": 0.9262, "step": 8147 }, { "epoch": 3.6902173913043477, "grad_norm": 0.0527517235155471, "learning_rate": 1.5589383964031982e-05, "loss": 0.9197, "step": 8148 }, { "epoch": 3.6906702898550723, "grad_norm": 0.050117991049757095, "learning_rate": 1.557936602020502e-05, "loss": 0.9138, "step": 8149 }, { "epoch": 3.691123188405797, "grad_norm": 0.05366854693622696, "learning_rate": 1.5569350517747524e-05, "loss": 0.9185, "step": 8150 }, { "epoch": 3.6915760869565215, "grad_norm": 0.05078208824861365, "learning_rate": 1.5559337457660778e-05, "loss": 0.9264, "step": 8151 }, { "epoch": 3.692028985507246, "grad_norm": 0.05509537232918637, "learning_rate": 1.5549326840945787e-05, "loss": 0.9146, "step": 8152 }, { "epoch": 3.692481884057971, "grad_norm": 0.05165203730984248, "learning_rate": 1.553931866860335e-05, "loss": 0.9231, "step": 8153 }, { "epoch": 3.692934782608696, "grad_norm": 0.05742143754998911, "learning_rate": 1.5529312941633988e-05, "loss": 0.9295, "step": 8154 }, { "epoch": 3.6933876811594204, "grad_norm": 0.05153544659249191, "learning_rate": 1.551930966103799e-05, "loss": 0.9358, "step": 8155 }, { "epoch": 3.693840579710145, "grad_norm": 0.05802581169509, "learning_rate": 1.550930882781538e-05, "loss": 0.9405, "step": 8156 }, { "epoch": 3.6942934782608696, "grad_norm": 0.05265612009774735, "learning_rate": 1.549931044296599e-05, "loss": 0.9196, "step": 8157 }, { "epoch": 3.6947463768115942, "grad_norm": 0.05377980728824436, "learning_rate": 1.5489314507489344e-05, "loss": 0.9396, "step": 8158 }, { "epoch": 3.695199275362319, "grad_norm": 0.06032258647082928, "learning_rate": 1.5479321022384787e-05, "loss": 0.901, "step": 8159 }, { "epoch": 3.6956521739130435, "grad_norm": 0.05541169799936949, "learning_rate": 1.5469329988651344e-05, "loss": 0.9217, "step": 8160 }, { "epoch": 3.696105072463768, "grad_norm": 0.05561674212950056, "learning_rate": 1.5459341407287873e-05, "loss": 0.9455, "step": 8161 }, { "epoch": 3.6965579710144927, "grad_norm": 0.05174112477668385, "learning_rate": 1.5449355279292914e-05, "loss": 0.9101, "step": 8162 }, { "epoch": 3.6970108695652173, "grad_norm": 0.06198705859476563, "learning_rate": 1.543937160566482e-05, "loss": 0.9279, "step": 8163 }, { "epoch": 3.697463768115942, "grad_norm": 0.05316109915069456, "learning_rate": 1.5429390387401673e-05, "loss": 0.9308, "step": 8164 }, { "epoch": 3.6979166666666665, "grad_norm": 0.0645711645922038, "learning_rate": 1.5419411625501302e-05, "loss": 0.939, "step": 8165 }, { "epoch": 3.6983695652173916, "grad_norm": 0.061355734542273714, "learning_rate": 1.540943532096129e-05, "loss": 0.9213, "step": 8166 }, { "epoch": 3.698822463768116, "grad_norm": 0.05209994853742443, "learning_rate": 1.5399461474779006e-05, "loss": 0.946, "step": 8167 }, { "epoch": 3.699275362318841, "grad_norm": 0.0616123997220556, "learning_rate": 1.5389490087951532e-05, "loss": 0.9121, "step": 8168 }, { "epoch": 3.6997282608695654, "grad_norm": 0.05762427700782207, "learning_rate": 1.537952116147574e-05, "loss": 0.9214, "step": 8169 }, { "epoch": 3.70018115942029, "grad_norm": 0.05665637905576976, "learning_rate": 1.5369554696348226e-05, "loss": 0.9209, "step": 8170 }, { "epoch": 3.7006340579710146, "grad_norm": 0.055480441853227605, "learning_rate": 1.5359590693565367e-05, "loss": 0.934, "step": 8171 }, { "epoch": 3.7010869565217392, "grad_norm": 0.05291605236609675, "learning_rate": 1.5349629154123275e-05, "loss": 0.9161, "step": 8172 }, { "epoch": 3.701539855072464, "grad_norm": 0.05076314377937719, "learning_rate": 1.5339670079017823e-05, "loss": 0.9136, "step": 8173 }, { "epoch": 3.7019927536231885, "grad_norm": 0.04999783220075941, "learning_rate": 1.532971346924461e-05, "loss": 0.8984, "step": 8174 }, { "epoch": 3.702445652173913, "grad_norm": 0.051989540510414364, "learning_rate": 1.531975932579905e-05, "loss": 0.9276, "step": 8175 }, { "epoch": 3.7028985507246377, "grad_norm": 0.0520606350847536, "learning_rate": 1.530980764967625e-05, "loss": 0.9264, "step": 8176 }, { "epoch": 3.7033514492753623, "grad_norm": 0.054717989227181366, "learning_rate": 1.5299858441871113e-05, "loss": 0.9414, "step": 8177 }, { "epoch": 3.703804347826087, "grad_norm": 0.05859868712532511, "learning_rate": 1.5289911703378258e-05, "loss": 0.9205, "step": 8178 }, { "epoch": 3.7042572463768115, "grad_norm": 0.053820020511481284, "learning_rate": 1.5279967435192097e-05, "loss": 0.9376, "step": 8179 }, { "epoch": 3.704710144927536, "grad_norm": 0.05565773594347285, "learning_rate": 1.5270025638306766e-05, "loss": 0.9375, "step": 8180 }, { "epoch": 3.7051630434782608, "grad_norm": 0.0541805836732464, "learning_rate": 1.5260086313716157e-05, "loss": 0.9239, "step": 8181 }, { "epoch": 3.7056159420289854, "grad_norm": 0.05782377278994152, "learning_rate": 1.5250149462413913e-05, "loss": 0.9169, "step": 8182 }, { "epoch": 3.70606884057971, "grad_norm": 0.05117683476266655, "learning_rate": 1.5240215085393457e-05, "loss": 0.9108, "step": 8183 }, { "epoch": 3.7065217391304346, "grad_norm": 0.06031146892609711, "learning_rate": 1.5230283183647938e-05, "loss": 0.9197, "step": 8184 }, { "epoch": 3.706974637681159, "grad_norm": 0.055462900572628426, "learning_rate": 1.522035375817024e-05, "loss": 0.9251, "step": 8185 }, { "epoch": 3.707427536231884, "grad_norm": 0.05825261576170498, "learning_rate": 1.5210426809953066e-05, "loss": 0.9348, "step": 8186 }, { "epoch": 3.7078804347826084, "grad_norm": 0.058117483717615034, "learning_rate": 1.5200502339988784e-05, "loss": 0.9387, "step": 8187 }, { "epoch": 3.7083333333333335, "grad_norm": 0.054494741695727524, "learning_rate": 1.5190580349269604e-05, "loss": 0.935, "step": 8188 }, { "epoch": 3.708786231884058, "grad_norm": 0.053432046689595544, "learning_rate": 1.5180660838787416e-05, "loss": 0.9172, "step": 8189 }, { "epoch": 3.7092391304347827, "grad_norm": 0.05527664545767234, "learning_rate": 1.517074380953389e-05, "loss": 0.9415, "step": 8190 }, { "epoch": 3.7096920289855073, "grad_norm": 0.05704948397392755, "learning_rate": 1.5160829262500438e-05, "loss": 0.9086, "step": 8191 }, { "epoch": 3.710144927536232, "grad_norm": 0.061158831836473374, "learning_rate": 1.5150917198678263e-05, "loss": 0.9229, "step": 8192 }, { "epoch": 3.7105978260869565, "grad_norm": 0.05071783968602903, "learning_rate": 1.5141007619058257e-05, "loss": 0.9243, "step": 8193 }, { "epoch": 3.711050724637681, "grad_norm": 0.05969003727300566, "learning_rate": 1.5131100524631128e-05, "loss": 0.9425, "step": 8194 }, { "epoch": 3.7115036231884058, "grad_norm": 0.05472545878311389, "learning_rate": 1.5121195916387273e-05, "loss": 0.9124, "step": 8195 }, { "epoch": 3.7119565217391304, "grad_norm": 0.06122203000994831, "learning_rate": 1.5111293795316896e-05, "loss": 0.9341, "step": 8196 }, { "epoch": 3.712409420289855, "grad_norm": 0.0558676596812452, "learning_rate": 1.5101394162409922e-05, "loss": 0.9163, "step": 8197 }, { "epoch": 3.7128623188405796, "grad_norm": 0.05354616501782377, "learning_rate": 1.5091497018656026e-05, "loss": 0.9464, "step": 8198 }, { "epoch": 3.713315217391304, "grad_norm": 0.06132241354827973, "learning_rate": 1.5081602365044629e-05, "loss": 0.9197, "step": 8199 }, { "epoch": 3.713768115942029, "grad_norm": 0.05685406475388608, "learning_rate": 1.5071710202564949e-05, "loss": 0.9248, "step": 8200 }, { "epoch": 3.714221014492754, "grad_norm": 0.06170088154487195, "learning_rate": 1.5061820532205885e-05, "loss": 0.9217, "step": 8201 }, { "epoch": 3.7146739130434785, "grad_norm": 0.060983981038595224, "learning_rate": 1.5051933354956152e-05, "loss": 0.919, "step": 8202 }, { "epoch": 3.715126811594203, "grad_norm": 0.04892677445886916, "learning_rate": 1.5042048671804166e-05, "loss": 0.9459, "step": 8203 }, { "epoch": 3.7155797101449277, "grad_norm": 0.057848749167320516, "learning_rate": 1.5032166483738136e-05, "loss": 0.9307, "step": 8204 }, { "epoch": 3.7160326086956523, "grad_norm": 0.050712603177871116, "learning_rate": 1.5022286791745972e-05, "loss": 0.9096, "step": 8205 }, { "epoch": 3.716485507246377, "grad_norm": 0.05283847721723633, "learning_rate": 1.5012409596815393e-05, "loss": 0.9253, "step": 8206 }, { "epoch": 3.7169384057971016, "grad_norm": 0.05450205499838461, "learning_rate": 1.5002534899933823e-05, "loss": 0.9436, "step": 8207 }, { "epoch": 3.717391304347826, "grad_norm": 0.05320183443172097, "learning_rate": 1.4992662702088451e-05, "loss": 0.9345, "step": 8208 }, { "epoch": 3.7178442028985508, "grad_norm": 0.05618957878246501, "learning_rate": 1.4982793004266203e-05, "loss": 0.9448, "step": 8209 }, { "epoch": 3.7182971014492754, "grad_norm": 0.05810397837163294, "learning_rate": 1.4972925807453793e-05, "loss": 0.9345, "step": 8210 }, { "epoch": 3.71875, "grad_norm": 0.05185085660941824, "learning_rate": 1.4963061112637637e-05, "loss": 0.9311, "step": 8211 }, { "epoch": 3.7192028985507246, "grad_norm": 0.05517700963961898, "learning_rate": 1.495319892080395e-05, "loss": 0.9193, "step": 8212 }, { "epoch": 3.7196557971014492, "grad_norm": 0.05654999785810364, "learning_rate": 1.4943339232938643e-05, "loss": 0.9484, "step": 8213 }, { "epoch": 3.720108695652174, "grad_norm": 0.057053529902174764, "learning_rate": 1.493348205002743e-05, "loss": 0.9292, "step": 8214 }, { "epoch": 3.7205615942028984, "grad_norm": 0.05352120716040301, "learning_rate": 1.492362737305574e-05, "loss": 0.9184, "step": 8215 }, { "epoch": 3.721014492753623, "grad_norm": 0.05826154902415867, "learning_rate": 1.4913775203008759e-05, "loss": 0.9365, "step": 8216 }, { "epoch": 3.7214673913043477, "grad_norm": 0.060813888849877734, "learning_rate": 1.4903925540871406e-05, "loss": 0.9257, "step": 8217 }, { "epoch": 3.7219202898550723, "grad_norm": 0.057139784827633634, "learning_rate": 1.4894078387628406e-05, "loss": 0.9282, "step": 8218 }, { "epoch": 3.722373188405797, "grad_norm": 0.06722115362591122, "learning_rate": 1.4884233744264153e-05, "loss": 0.9284, "step": 8219 }, { "epoch": 3.7228260869565215, "grad_norm": 0.05982008212360623, "learning_rate": 1.4874391611762868e-05, "loss": 0.9128, "step": 8220 }, { "epoch": 3.723278985507246, "grad_norm": 0.060456558664548354, "learning_rate": 1.4864551991108455e-05, "loss": 0.9302, "step": 8221 }, { "epoch": 3.723731884057971, "grad_norm": 0.05442545875840399, "learning_rate": 1.4854714883284623e-05, "loss": 0.9362, "step": 8222 }, { "epoch": 3.724184782608696, "grad_norm": 0.06019570895543269, "learning_rate": 1.484488028927479e-05, "loss": 0.9397, "step": 8223 }, { "epoch": 3.7246376811594204, "grad_norm": 0.05318091891392222, "learning_rate": 1.4835048210062137e-05, "loss": 0.9329, "step": 8224 }, { "epoch": 3.725090579710145, "grad_norm": 0.06052991470956869, "learning_rate": 1.4825218646629576e-05, "loss": 0.9219, "step": 8225 }, { "epoch": 3.7255434782608696, "grad_norm": 0.05978323047864957, "learning_rate": 1.4815391599959816e-05, "loss": 0.9671, "step": 8226 }, { "epoch": 3.7259963768115942, "grad_norm": 0.061419109729248425, "learning_rate": 1.4805567071035252e-05, "loss": 0.9332, "step": 8227 }, { "epoch": 3.726449275362319, "grad_norm": 0.05621547838959602, "learning_rate": 1.4795745060838083e-05, "loss": 0.9233, "step": 8228 }, { "epoch": 3.7269021739130435, "grad_norm": 0.05649267875087561, "learning_rate": 1.4785925570350208e-05, "loss": 0.9113, "step": 8229 }, { "epoch": 3.727355072463768, "grad_norm": 0.05690684891376725, "learning_rate": 1.4776108600553323e-05, "loss": 0.924, "step": 8230 }, { "epoch": 3.7278079710144927, "grad_norm": 0.052829911938357794, "learning_rate": 1.4766294152428832e-05, "loss": 0.9102, "step": 8231 }, { "epoch": 3.7282608695652173, "grad_norm": 0.058763512574070295, "learning_rate": 1.47564822269579e-05, "loss": 0.9296, "step": 8232 }, { "epoch": 3.728713768115942, "grad_norm": 0.05140125410057616, "learning_rate": 1.4746672825121429e-05, "loss": 0.9248, "step": 8233 }, { "epoch": 3.7291666666666665, "grad_norm": 0.05552520001571998, "learning_rate": 1.4736865947900106e-05, "loss": 0.92, "step": 8234 }, { "epoch": 3.7296195652173916, "grad_norm": 0.057783965470356476, "learning_rate": 1.4727061596274315e-05, "loss": 0.9384, "step": 8235 }, { "epoch": 3.730072463768116, "grad_norm": 0.05126350954103511, "learning_rate": 1.4717259771224237e-05, "loss": 0.9283, "step": 8236 }, { "epoch": 3.730525362318841, "grad_norm": 0.060822796487437575, "learning_rate": 1.4707460473729747e-05, "loss": 0.9097, "step": 8237 }, { "epoch": 3.7309782608695654, "grad_norm": 0.050643416190227615, "learning_rate": 1.469766370477053e-05, "loss": 0.8942, "step": 8238 }, { "epoch": 3.73143115942029, "grad_norm": 0.054329023043391496, "learning_rate": 1.4687869465325966e-05, "loss": 0.9203, "step": 8239 }, { "epoch": 3.7318840579710146, "grad_norm": 0.052423927818360715, "learning_rate": 1.4678077756375207e-05, "loss": 0.9279, "step": 8240 }, { "epoch": 3.7323369565217392, "grad_norm": 0.05378298987442879, "learning_rate": 1.466828857889712e-05, "loss": 0.9237, "step": 8241 }, { "epoch": 3.732789855072464, "grad_norm": 0.05600377539399648, "learning_rate": 1.4658501933870382e-05, "loss": 0.9038, "step": 8242 }, { "epoch": 3.7332427536231885, "grad_norm": 0.04880056365787984, "learning_rate": 1.464871782227335e-05, "loss": 0.9554, "step": 8243 }, { "epoch": 3.733695652173913, "grad_norm": 0.059611898714957284, "learning_rate": 1.463893624508418e-05, "loss": 0.9016, "step": 8244 }, { "epoch": 3.7341485507246377, "grad_norm": 0.05291787419205672, "learning_rate": 1.462915720328073e-05, "loss": 0.9381, "step": 8245 }, { "epoch": 3.7346014492753623, "grad_norm": 0.057114997441348035, "learning_rate": 1.4619380697840653e-05, "loss": 0.9369, "step": 8246 }, { "epoch": 3.735054347826087, "grad_norm": 0.05919811665837547, "learning_rate": 1.4609606729741291e-05, "loss": 0.9142, "step": 8247 }, { "epoch": 3.7355072463768115, "grad_norm": 0.05397006940102369, "learning_rate": 1.4599835299959794e-05, "loss": 0.9043, "step": 8248 }, { "epoch": 3.735960144927536, "grad_norm": 0.05473325962223254, "learning_rate": 1.4590066409473011e-05, "loss": 0.9257, "step": 8249 }, { "epoch": 3.7364130434782608, "grad_norm": 0.055805920285213265, "learning_rate": 1.4580300059257554e-05, "loss": 0.929, "step": 8250 }, { "epoch": 3.7368659420289854, "grad_norm": 0.058908065100575134, "learning_rate": 1.4570536250289768e-05, "loss": 0.9216, "step": 8251 }, { "epoch": 3.73731884057971, "grad_norm": 0.055010597233361866, "learning_rate": 1.4560774983545787e-05, "loss": 0.9314, "step": 8252 }, { "epoch": 3.7377717391304346, "grad_norm": 0.05232345237075347, "learning_rate": 1.4551016260001424e-05, "loss": 0.9317, "step": 8253 }, { "epoch": 3.738224637681159, "grad_norm": 0.05676430394368231, "learning_rate": 1.4541260080632307e-05, "loss": 0.9215, "step": 8254 }, { "epoch": 3.738677536231884, "grad_norm": 0.05311722356039652, "learning_rate": 1.453150644641375e-05, "loss": 0.9224, "step": 8255 }, { "epoch": 3.7391304347826084, "grad_norm": 0.05805659537508504, "learning_rate": 1.4521755358320868e-05, "loss": 0.902, "step": 8256 }, { "epoch": 3.7395833333333335, "grad_norm": 0.059079042066569276, "learning_rate": 1.4512006817328472e-05, "loss": 0.9236, "step": 8257 }, { "epoch": 3.740036231884058, "grad_norm": 0.05653252745978652, "learning_rate": 1.4502260824411147e-05, "loss": 0.9435, "step": 8258 }, { "epoch": 3.7404891304347827, "grad_norm": 0.052284599799371306, "learning_rate": 1.4492517380543198e-05, "loss": 0.9202, "step": 8259 }, { "epoch": 3.7409420289855073, "grad_norm": 0.060883756971140554, "learning_rate": 1.4482776486698717e-05, "loss": 0.9126, "step": 8260 }, { "epoch": 3.741394927536232, "grad_norm": 0.05642885421596597, "learning_rate": 1.4473038143851494e-05, "loss": 0.9174, "step": 8261 }, { "epoch": 3.7418478260869565, "grad_norm": 0.06160647830579079, "learning_rate": 1.4463302352975115e-05, "loss": 0.9255, "step": 8262 }, { "epoch": 3.742300724637681, "grad_norm": 0.06310710047308501, "learning_rate": 1.445356911504285e-05, "loss": 0.9649, "step": 8263 }, { "epoch": 3.7427536231884058, "grad_norm": 0.06768576065638522, "learning_rate": 1.4443838431027777e-05, "loss": 0.9258, "step": 8264 }, { "epoch": 3.7432065217391304, "grad_norm": 0.0636540676309701, "learning_rate": 1.4434110301902675e-05, "loss": 0.9338, "step": 8265 }, { "epoch": 3.743659420289855, "grad_norm": 0.056258357851384326, "learning_rate": 1.4424384728640077e-05, "loss": 0.9354, "step": 8266 }, { "epoch": 3.7441123188405796, "grad_norm": 0.0640362128818919, "learning_rate": 1.441466171221225e-05, "loss": 0.9125, "step": 8267 }, { "epoch": 3.744565217391304, "grad_norm": 0.062182894682638865, "learning_rate": 1.4404941253591251e-05, "loss": 0.9225, "step": 8268 }, { "epoch": 3.745018115942029, "grad_norm": 0.059009247972775826, "learning_rate": 1.439522335374882e-05, "loss": 0.9262, "step": 8269 }, { "epoch": 3.745471014492754, "grad_norm": 0.054033144595288594, "learning_rate": 1.4385508013656497e-05, "loss": 0.9527, "step": 8270 }, { "epoch": 3.7459239130434785, "grad_norm": 0.06213434546531259, "learning_rate": 1.4375795234285512e-05, "loss": 0.9406, "step": 8271 }, { "epoch": 3.746376811594203, "grad_norm": 0.06357812898241157, "learning_rate": 1.4366085016606901e-05, "loss": 0.9447, "step": 8272 }, { "epoch": 3.7468297101449277, "grad_norm": 0.062091508311729146, "learning_rate": 1.435637736159139e-05, "loss": 0.9189, "step": 8273 }, { "epoch": 3.7472826086956523, "grad_norm": 0.06448543760347229, "learning_rate": 1.4346672270209468e-05, "loss": 0.9575, "step": 8274 }, { "epoch": 3.747735507246377, "grad_norm": 0.06291071582939738, "learning_rate": 1.4336969743431358e-05, "loss": 0.9333, "step": 8275 }, { "epoch": 3.7481884057971016, "grad_norm": 0.0618881268068554, "learning_rate": 1.4327269782227062e-05, "loss": 0.9465, "step": 8276 }, { "epoch": 3.748641304347826, "grad_norm": 0.06584883712557824, "learning_rate": 1.4317572387566276e-05, "loss": 0.9185, "step": 8277 }, { "epoch": 3.7490942028985508, "grad_norm": 0.06868067372762861, "learning_rate": 1.4307877560418488e-05, "loss": 0.9178, "step": 8278 }, { "epoch": 3.7495471014492754, "grad_norm": 0.05499664473947327, "learning_rate": 1.4298185301752879e-05, "loss": 0.8994, "step": 8279 }, { "epoch": 3.75, "grad_norm": 0.06507242836833872, "learning_rate": 1.4288495612538427e-05, "loss": 0.9262, "step": 8280 }, { "epoch": 3.7504528985507246, "grad_norm": 0.053779681757840224, "learning_rate": 1.4278808493743812e-05, "loss": 0.927, "step": 8281 }, { "epoch": 3.7509057971014492, "grad_norm": 0.061725141239941446, "learning_rate": 1.4269123946337474e-05, "loss": 0.9198, "step": 8282 }, { "epoch": 3.751358695652174, "grad_norm": 0.05531873993944544, "learning_rate": 1.4259441971287573e-05, "loss": 0.9119, "step": 8283 }, { "epoch": 3.7518115942028984, "grad_norm": 0.06036564704153776, "learning_rate": 1.4249762569562058e-05, "loss": 0.9316, "step": 8284 }, { "epoch": 3.752264492753623, "grad_norm": 0.06243320768638082, "learning_rate": 1.4240085742128576e-05, "loss": 0.9501, "step": 8285 }, { "epoch": 3.7527173913043477, "grad_norm": 0.05467502758448178, "learning_rate": 1.4230411489954552e-05, "loss": 0.9148, "step": 8286 }, { "epoch": 3.7531702898550723, "grad_norm": 0.05526529501672485, "learning_rate": 1.4220739814007112e-05, "loss": 0.9095, "step": 8287 }, { "epoch": 3.753623188405797, "grad_norm": 0.06579454894789614, "learning_rate": 1.4211070715253179e-05, "loss": 0.9091, "step": 8288 }, { "epoch": 3.7540760869565215, "grad_norm": 0.055720343542210934, "learning_rate": 1.4201404194659359e-05, "loss": 0.926, "step": 8289 }, { "epoch": 3.754528985507246, "grad_norm": 0.05681737801127166, "learning_rate": 1.4191740253192073e-05, "loss": 0.9273, "step": 8290 }, { "epoch": 3.754981884057971, "grad_norm": 0.06441242393207255, "learning_rate": 1.4182078891817379e-05, "loss": 0.9222, "step": 8291 }, { "epoch": 3.755434782608696, "grad_norm": 0.062348140613134276, "learning_rate": 1.4172420111501186e-05, "loss": 0.9238, "step": 8292 }, { "epoch": 3.7558876811594204, "grad_norm": 0.0609110194585766, "learning_rate": 1.4162763913209068e-05, "loss": 0.9098, "step": 8293 }, { "epoch": 3.756340579710145, "grad_norm": 0.06480447051140983, "learning_rate": 1.4153110297906398e-05, "loss": 0.9347, "step": 8294 }, { "epoch": 3.7567934782608696, "grad_norm": 0.057254592041795775, "learning_rate": 1.4143459266558237e-05, "loss": 0.9286, "step": 8295 }, { "epoch": 3.7572463768115942, "grad_norm": 0.06504197210436602, "learning_rate": 1.4133810820129438e-05, "loss": 0.9176, "step": 8296 }, { "epoch": 3.757699275362319, "grad_norm": 0.05610976041052015, "learning_rate": 1.4124164959584548e-05, "loss": 0.9277, "step": 8297 }, { "epoch": 3.7581521739130435, "grad_norm": 0.057216933656140535, "learning_rate": 1.4114521685887898e-05, "loss": 0.9148, "step": 8298 }, { "epoch": 3.758605072463768, "grad_norm": 0.05706873988936007, "learning_rate": 1.4104881000003543e-05, "loss": 0.9405, "step": 8299 }, { "epoch": 3.7590579710144927, "grad_norm": 0.05194452324273873, "learning_rate": 1.4095242902895265e-05, "loss": 0.9327, "step": 8300 }, { "epoch": 3.7595108695652173, "grad_norm": 0.055351037405826795, "learning_rate": 1.4085607395526589e-05, "loss": 0.9394, "step": 8301 }, { "epoch": 3.759963768115942, "grad_norm": 0.049904599737834635, "learning_rate": 1.4075974478860821e-05, "loss": 0.9286, "step": 8302 }, { "epoch": 3.7604166666666665, "grad_norm": 0.052732220305336745, "learning_rate": 1.406634415386095e-05, "loss": 0.9213, "step": 8303 }, { "epoch": 3.7608695652173916, "grad_norm": 0.051993111695067616, "learning_rate": 1.4056716421489762e-05, "loss": 0.8977, "step": 8304 }, { "epoch": 3.761322463768116, "grad_norm": 0.05751427307238269, "learning_rate": 1.4047091282709735e-05, "loss": 0.9458, "step": 8305 }, { "epoch": 3.761775362318841, "grad_norm": 0.05192026373480652, "learning_rate": 1.4037468738483129e-05, "loss": 0.9291, "step": 8306 }, { "epoch": 3.7622282608695654, "grad_norm": 0.053666588197545606, "learning_rate": 1.4027848789771917e-05, "loss": 0.9102, "step": 8307 }, { "epoch": 3.76268115942029, "grad_norm": 0.05368591802093802, "learning_rate": 1.4018231437537812e-05, "loss": 0.9334, "step": 8308 }, { "epoch": 3.7631340579710146, "grad_norm": 0.049195952762765574, "learning_rate": 1.4008616682742285e-05, "loss": 0.9451, "step": 8309 }, { "epoch": 3.7635869565217392, "grad_norm": 0.04982830748969065, "learning_rate": 1.3999004526346523e-05, "loss": 0.9346, "step": 8310 }, { "epoch": 3.764039855072464, "grad_norm": 0.047489902652020466, "learning_rate": 1.3989394969311491e-05, "loss": 0.9222, "step": 8311 }, { "epoch": 3.7644927536231885, "grad_norm": 0.05660713721189825, "learning_rate": 1.397978801259785e-05, "loss": 0.9262, "step": 8312 }, { "epoch": 3.764945652173913, "grad_norm": 0.04917185848351984, "learning_rate": 1.3970183657166047e-05, "loss": 0.9231, "step": 8313 }, { "epoch": 3.7653985507246377, "grad_norm": 0.05940625074445311, "learning_rate": 1.3960581903976217e-05, "loss": 0.9117, "step": 8314 }, { "epoch": 3.7658514492753623, "grad_norm": 0.053005151607699315, "learning_rate": 1.3950982753988296e-05, "loss": 0.9263, "step": 8315 }, { "epoch": 3.766304347826087, "grad_norm": 0.04836883315133145, "learning_rate": 1.39413862081619e-05, "loss": 0.9078, "step": 8316 }, { "epoch": 3.7667572463768115, "grad_norm": 0.07415643795003223, "learning_rate": 1.3931792267456424e-05, "loss": 0.9233, "step": 8317 }, { "epoch": 3.767210144927536, "grad_norm": 0.05102812816343002, "learning_rate": 1.3922200932830965e-05, "loss": 0.9157, "step": 8318 }, { "epoch": 3.7676630434782608, "grad_norm": 0.05744806093846893, "learning_rate": 1.391261220524442e-05, "loss": 0.9414, "step": 8319 }, { "epoch": 3.7681159420289854, "grad_norm": 0.05203448526357209, "learning_rate": 1.3903026085655356e-05, "loss": 0.9361, "step": 8320 }, { "epoch": 3.76856884057971, "grad_norm": 0.05305121173202063, "learning_rate": 1.3893442575022143e-05, "loss": 0.9241, "step": 8321 }, { "epoch": 3.7690217391304346, "grad_norm": 0.05449891768452933, "learning_rate": 1.3883861674302832e-05, "loss": 0.934, "step": 8322 }, { "epoch": 3.769474637681159, "grad_norm": 0.058003687851705614, "learning_rate": 1.3874283384455267e-05, "loss": 0.9273, "step": 8323 }, { "epoch": 3.769927536231884, "grad_norm": 0.051171166670560125, "learning_rate": 1.3864707706436993e-05, "loss": 0.9022, "step": 8324 }, { "epoch": 3.7703804347826084, "grad_norm": 0.058250591380061945, "learning_rate": 1.3855134641205302e-05, "loss": 0.9467, "step": 8325 }, { "epoch": 3.7708333333333335, "grad_norm": 0.05801181571549195, "learning_rate": 1.3845564189717218e-05, "loss": 0.9227, "step": 8326 }, { "epoch": 3.771286231884058, "grad_norm": 0.04935289622823816, "learning_rate": 1.3835996352929542e-05, "loss": 0.9371, "step": 8327 }, { "epoch": 3.7717391304347827, "grad_norm": 0.053055998035160076, "learning_rate": 1.3826431131798757e-05, "loss": 0.9248, "step": 8328 }, { "epoch": 3.7721920289855073, "grad_norm": 0.05879702696743085, "learning_rate": 1.3816868527281143e-05, "loss": 0.9138, "step": 8329 }, { "epoch": 3.772644927536232, "grad_norm": 0.05147171320074463, "learning_rate": 1.3807308540332659e-05, "loss": 0.9154, "step": 8330 }, { "epoch": 3.7730978260869565, "grad_norm": 0.05442454934042806, "learning_rate": 1.3797751171909055e-05, "loss": 0.9167, "step": 8331 }, { "epoch": 3.773550724637681, "grad_norm": 0.053562592887246976, "learning_rate": 1.3788196422965792e-05, "loss": 0.9439, "step": 8332 }, { "epoch": 3.7740036231884058, "grad_norm": 0.05064687123968551, "learning_rate": 1.3778644294458064e-05, "loss": 0.9389, "step": 8333 }, { "epoch": 3.7744565217391304, "grad_norm": 0.050506052160131674, "learning_rate": 1.376909478734081e-05, "loss": 0.9257, "step": 8334 }, { "epoch": 3.774909420289855, "grad_norm": 0.05661329032258418, "learning_rate": 1.3759547902568726e-05, "loss": 0.9099, "step": 8335 }, { "epoch": 3.7753623188405796, "grad_norm": 0.051236505203293155, "learning_rate": 1.3750003641096207e-05, "loss": 0.9191, "step": 8336 }, { "epoch": 3.775815217391304, "grad_norm": 0.060851243268840595, "learning_rate": 1.3740462003877433e-05, "loss": 0.9135, "step": 8337 }, { "epoch": 3.776268115942029, "grad_norm": 0.04915538742760045, "learning_rate": 1.3730922991866273e-05, "loss": 0.911, "step": 8338 }, { "epoch": 3.776721014492754, "grad_norm": 0.06660066876877777, "learning_rate": 1.372138660601638e-05, "loss": 0.9269, "step": 8339 }, { "epoch": 3.7771739130434785, "grad_norm": 0.05034583383566836, "learning_rate": 1.3711852847281098e-05, "loss": 0.9127, "step": 8340 }, { "epoch": 3.777626811594203, "grad_norm": 0.05721320629144208, "learning_rate": 1.3702321716613556e-05, "loss": 0.9158, "step": 8341 }, { "epoch": 3.7780797101449277, "grad_norm": 0.05591512128265637, "learning_rate": 1.3692793214966584e-05, "loss": 0.9248, "step": 8342 }, { "epoch": 3.7785326086956523, "grad_norm": 0.05411531465999687, "learning_rate": 1.3683267343292763e-05, "loss": 0.9327, "step": 8343 }, { "epoch": 3.778985507246377, "grad_norm": 0.050839478626623566, "learning_rate": 1.3673744102544397e-05, "loss": 0.9005, "step": 8344 }, { "epoch": 3.7794384057971016, "grad_norm": 0.05602377928330534, "learning_rate": 1.3664223493673561e-05, "loss": 0.9285, "step": 8345 }, { "epoch": 3.779891304347826, "grad_norm": 0.05424807835050049, "learning_rate": 1.365470551763202e-05, "loss": 0.9223, "step": 8346 }, { "epoch": 3.7803442028985508, "grad_norm": 0.05137366167074343, "learning_rate": 1.3645190175371331e-05, "loss": 0.9104, "step": 8347 }, { "epoch": 3.7807971014492754, "grad_norm": 0.08176508264482928, "learning_rate": 1.3635677467842734e-05, "loss": 0.9181, "step": 8348 }, { "epoch": 3.78125, "grad_norm": 0.051571518104306, "learning_rate": 1.3626167395997247e-05, "loss": 0.9258, "step": 8349 }, { "epoch": 3.7817028985507246, "grad_norm": 0.059597643584845386, "learning_rate": 1.36166599607856e-05, "loss": 0.9229, "step": 8350 }, { "epoch": 3.7821557971014492, "grad_norm": 0.0494322469996742, "learning_rate": 1.360715516315826e-05, "loss": 0.9331, "step": 8351 }, { "epoch": 3.782608695652174, "grad_norm": 0.05790461790224351, "learning_rate": 1.3597653004065436e-05, "loss": 0.9086, "step": 8352 }, { "epoch": 3.7830615942028984, "grad_norm": 0.05230194434211665, "learning_rate": 1.3588153484457087e-05, "loss": 0.8978, "step": 8353 }, { "epoch": 3.783514492753623, "grad_norm": 0.056071755198834214, "learning_rate": 1.3578656605282875e-05, "loss": 0.9208, "step": 8354 }, { "epoch": 3.7839673913043477, "grad_norm": 0.14605258486761416, "learning_rate": 1.3569162367492247e-05, "loss": 0.9259, "step": 8355 }, { "epoch": 3.7844202898550723, "grad_norm": 0.06155354553514895, "learning_rate": 1.3559670772034324e-05, "loss": 0.9074, "step": 8356 }, { "epoch": 3.784873188405797, "grad_norm": 0.06518939693196307, "learning_rate": 1.3550181819858024e-05, "loss": 0.9359, "step": 8357 }, { "epoch": 3.7853260869565215, "grad_norm": 0.05986509378793195, "learning_rate": 1.3540695511911963e-05, "loss": 0.9273, "step": 8358 }, { "epoch": 3.785778985507246, "grad_norm": 0.07551009535105453, "learning_rate": 1.3531211849144499e-05, "loss": 0.9555, "step": 8359 }, { "epoch": 3.786231884057971, "grad_norm": 0.05130732951065814, "learning_rate": 1.3521730832503722e-05, "loss": 0.927, "step": 8360 }, { "epoch": 3.786684782608696, "grad_norm": 0.06991973633136077, "learning_rate": 1.3512252462937477e-05, "loss": 0.9394, "step": 8361 }, { "epoch": 3.7871376811594204, "grad_norm": 0.05461969290088301, "learning_rate": 1.3502776741393317e-05, "loss": 0.9491, "step": 8362 }, { "epoch": 3.787590579710145, "grad_norm": 0.06180372822512767, "learning_rate": 1.3493303668818567e-05, "loss": 0.9265, "step": 8363 }, { "epoch": 3.7880434782608696, "grad_norm": 0.061930301094045855, "learning_rate": 1.3483833246160241e-05, "loss": 0.9216, "step": 8364 }, { "epoch": 3.7884963768115942, "grad_norm": 0.06049867735054141, "learning_rate": 1.3474365474365136e-05, "loss": 0.9051, "step": 8365 }, { "epoch": 3.788949275362319, "grad_norm": 0.05277887904039159, "learning_rate": 1.3464900354379751e-05, "loss": 0.9271, "step": 8366 }, { "epoch": 3.7894021739130435, "grad_norm": 0.06021416019789049, "learning_rate": 1.3455437887150323e-05, "loss": 0.9302, "step": 8367 }, { "epoch": 3.789855072463768, "grad_norm": 0.050277298015051604, "learning_rate": 1.344597807362282e-05, "loss": 0.9303, "step": 8368 }, { "epoch": 3.7903079710144927, "grad_norm": 0.059995049954024195, "learning_rate": 1.3436520914742981e-05, "loss": 0.9179, "step": 8369 }, { "epoch": 3.7907608695652173, "grad_norm": 0.057723744664403595, "learning_rate": 1.3427066411456226e-05, "loss": 0.942, "step": 8370 }, { "epoch": 3.791213768115942, "grad_norm": 0.06494111449993949, "learning_rate": 1.3417614564707764e-05, "loss": 0.9301, "step": 8371 }, { "epoch": 3.7916666666666665, "grad_norm": 0.06123747590071867, "learning_rate": 1.3408165375442486e-05, "loss": 0.9458, "step": 8372 }, { "epoch": 3.7921195652173916, "grad_norm": 0.05649890529177583, "learning_rate": 1.3398718844605063e-05, "loss": 0.9252, "step": 8373 }, { "epoch": 3.792572463768116, "grad_norm": 0.0611784496428329, "learning_rate": 1.3389274973139874e-05, "loss": 0.9237, "step": 8374 }, { "epoch": 3.793025362318841, "grad_norm": 0.04948673742180207, "learning_rate": 1.337983376199103e-05, "loss": 0.9227, "step": 8375 }, { "epoch": 3.7934782608695654, "grad_norm": 0.06150022186964523, "learning_rate": 1.337039521210238e-05, "loss": 0.9489, "step": 8376 }, { "epoch": 3.79393115942029, "grad_norm": 0.04524386367853949, "learning_rate": 1.3360959324417526e-05, "loss": 0.9064, "step": 8377 }, { "epoch": 3.7943840579710146, "grad_norm": 0.05909705671707934, "learning_rate": 1.3351526099879774e-05, "loss": 0.9347, "step": 8378 }, { "epoch": 3.7948369565217392, "grad_norm": 0.05094898330660005, "learning_rate": 1.3342095539432197e-05, "loss": 0.9217, "step": 8379 }, { "epoch": 3.795289855072464, "grad_norm": 0.053086667438170816, "learning_rate": 1.333266764401756e-05, "loss": 0.9136, "step": 8380 }, { "epoch": 3.7957427536231885, "grad_norm": 0.06046067932950313, "learning_rate": 1.3323242414578407e-05, "loss": 0.9193, "step": 8381 }, { "epoch": 3.796195652173913, "grad_norm": 0.05236541418712256, "learning_rate": 1.331381985205697e-05, "loss": 0.9118, "step": 8382 }, { "epoch": 3.7966485507246377, "grad_norm": 0.05817077305999598, "learning_rate": 1.3304399957395267e-05, "loss": 0.9365, "step": 8383 }, { "epoch": 3.7971014492753623, "grad_norm": 0.05402492347722647, "learning_rate": 1.3294982731535e-05, "loss": 0.9295, "step": 8384 }, { "epoch": 3.797554347826087, "grad_norm": 0.05062348310462133, "learning_rate": 1.3285568175417628e-05, "loss": 0.9085, "step": 8385 }, { "epoch": 3.7980072463768115, "grad_norm": 0.05317345074793525, "learning_rate": 1.3276156289984323e-05, "loss": 0.932, "step": 8386 }, { "epoch": 3.798460144927536, "grad_norm": 0.05037710508778429, "learning_rate": 1.326674707617604e-05, "loss": 0.9315, "step": 8387 }, { "epoch": 3.7989130434782608, "grad_norm": 0.05446783186658322, "learning_rate": 1.3257340534933394e-05, "loss": 0.9049, "step": 8388 }, { "epoch": 3.7993659420289854, "grad_norm": 0.05468958910183411, "learning_rate": 1.324793666719681e-05, "loss": 0.9297, "step": 8389 }, { "epoch": 3.79981884057971, "grad_norm": 0.054320121829030905, "learning_rate": 1.3238535473906375e-05, "loss": 0.9212, "step": 8390 }, { "epoch": 3.8002717391304346, "grad_norm": 0.053926787856949196, "learning_rate": 1.322913695600197e-05, "loss": 0.9235, "step": 8391 }, { "epoch": 3.800724637681159, "grad_norm": 0.05258401467233938, "learning_rate": 1.3219741114423164e-05, "loss": 0.9152, "step": 8392 }, { "epoch": 3.801177536231884, "grad_norm": 0.05613467963685389, "learning_rate": 1.3210347950109274e-05, "loss": 0.9378, "step": 8393 }, { "epoch": 3.8016304347826084, "grad_norm": 0.051504165942147774, "learning_rate": 1.3200957463999343e-05, "loss": 0.9237, "step": 8394 }, { "epoch": 3.8020833333333335, "grad_norm": 0.05437530069183981, "learning_rate": 1.319156965703217e-05, "loss": 0.942, "step": 8395 }, { "epoch": 3.802536231884058, "grad_norm": 0.057876118463507716, "learning_rate": 1.318218453014625e-05, "loss": 0.9185, "step": 8396 }, { "epoch": 3.8029891304347827, "grad_norm": 0.06025661468586792, "learning_rate": 1.317280208427985e-05, "loss": 0.8937, "step": 8397 }, { "epoch": 3.8034420289855073, "grad_norm": 0.06061772665435259, "learning_rate": 1.3163422320370924e-05, "loss": 0.9424, "step": 8398 }, { "epoch": 3.803894927536232, "grad_norm": 0.05194947029145026, "learning_rate": 1.315404523935721e-05, "loss": 0.9329, "step": 8399 }, { "epoch": 3.8043478260869565, "grad_norm": 0.053427552812953014, "learning_rate": 1.3144670842176134e-05, "loss": 0.8983, "step": 8400 }, { "epoch": 3.804800724637681, "grad_norm": 0.058686516295977986, "learning_rate": 1.313529912976487e-05, "loss": 0.9237, "step": 8401 }, { "epoch": 3.8052536231884058, "grad_norm": 0.05315054142322652, "learning_rate": 1.3125930103060304e-05, "loss": 0.919, "step": 8402 }, { "epoch": 3.8057065217391304, "grad_norm": 0.055375570077183155, "learning_rate": 1.3116563762999111e-05, "loss": 0.9074, "step": 8403 }, { "epoch": 3.806159420289855, "grad_norm": 0.0586396904950092, "learning_rate": 1.3107200110517622e-05, "loss": 0.9291, "step": 8404 }, { "epoch": 3.8066123188405796, "grad_norm": 0.049014278892050914, "learning_rate": 1.3097839146551969e-05, "loss": 0.9273, "step": 8405 }, { "epoch": 3.807065217391304, "grad_norm": 0.05425736152407822, "learning_rate": 1.3088480872037952e-05, "loss": 0.9363, "step": 8406 }, { "epoch": 3.807518115942029, "grad_norm": 0.05164313147784249, "learning_rate": 1.3079125287911162e-05, "loss": 0.9273, "step": 8407 }, { "epoch": 3.807971014492754, "grad_norm": 0.05422128012047149, "learning_rate": 1.3069772395106877e-05, "loss": 0.9399, "step": 8408 }, { "epoch": 3.8084239130434785, "grad_norm": 0.05305045540953018, "learning_rate": 1.3060422194560119e-05, "loss": 0.9418, "step": 8409 }, { "epoch": 3.808876811594203, "grad_norm": 0.056631713445176896, "learning_rate": 1.305107468720563e-05, "loss": 0.9378, "step": 8410 }, { "epoch": 3.8093297101449277, "grad_norm": 0.059696684233523124, "learning_rate": 1.3041729873977924e-05, "loss": 0.9376, "step": 8411 }, { "epoch": 3.8097826086956523, "grad_norm": 0.05155126033891387, "learning_rate": 1.3032387755811184e-05, "loss": 0.9323, "step": 8412 }, { "epoch": 3.810235507246377, "grad_norm": 0.06323405266426747, "learning_rate": 1.3023048333639392e-05, "loss": 0.9324, "step": 8413 }, { "epoch": 3.8106884057971016, "grad_norm": 0.05347867723567134, "learning_rate": 1.301371160839619e-05, "loss": 0.9179, "step": 8414 }, { "epoch": 3.811141304347826, "grad_norm": 0.05090390403374754, "learning_rate": 1.3004377581015017e-05, "loss": 0.9316, "step": 8415 }, { "epoch": 3.8115942028985508, "grad_norm": 0.06723442197564677, "learning_rate": 1.2995046252428996e-05, "loss": 0.9318, "step": 8416 }, { "epoch": 3.8120471014492754, "grad_norm": 0.05247307985475618, "learning_rate": 1.2985717623570992e-05, "loss": 0.9186, "step": 8417 }, { "epoch": 3.8125, "grad_norm": 0.054221221432240965, "learning_rate": 1.297639169537359e-05, "loss": 0.9375, "step": 8418 }, { "epoch": 3.8129528985507246, "grad_norm": 0.05905785578214892, "learning_rate": 1.2967068468769154e-05, "loss": 0.932, "step": 8419 }, { "epoch": 3.8134057971014492, "grad_norm": 0.049733817628479386, "learning_rate": 1.29577479446897e-05, "loss": 0.9183, "step": 8420 }, { "epoch": 3.813858695652174, "grad_norm": 0.056481704782920895, "learning_rate": 1.2948430124067054e-05, "loss": 0.9191, "step": 8421 }, { "epoch": 3.8143115942028984, "grad_norm": 0.05043216095543979, "learning_rate": 1.2939115007832697e-05, "loss": 0.933, "step": 8422 }, { "epoch": 3.814764492753623, "grad_norm": 0.054194824045996556, "learning_rate": 1.292980259691791e-05, "loss": 0.9274, "step": 8423 }, { "epoch": 3.8152173913043477, "grad_norm": 0.05738985610805368, "learning_rate": 1.2920492892253642e-05, "loss": 0.896, "step": 8424 }, { "epoch": 3.8156702898550723, "grad_norm": 0.05740824946190387, "learning_rate": 1.2911185894770623e-05, "loss": 0.9441, "step": 8425 }, { "epoch": 3.816123188405797, "grad_norm": 0.05314192522690501, "learning_rate": 1.2901881605399273e-05, "loss": 0.9023, "step": 8426 }, { "epoch": 3.8165760869565215, "grad_norm": 0.052042204235858286, "learning_rate": 1.2892580025069762e-05, "loss": 0.9268, "step": 8427 }, { "epoch": 3.817028985507246, "grad_norm": 0.05297043940219459, "learning_rate": 1.2883281154711971e-05, "loss": 0.9243, "step": 8428 }, { "epoch": 3.817481884057971, "grad_norm": 0.05946943002934491, "learning_rate": 1.2873984995255545e-05, "loss": 0.922, "step": 8429 }, { "epoch": 3.817934782608696, "grad_norm": 0.047983135961735104, "learning_rate": 1.2864691547629807e-05, "loss": 0.9365, "step": 8430 }, { "epoch": 3.8183876811594204, "grad_norm": 0.05611281193708216, "learning_rate": 1.2855400812763872e-05, "loss": 0.9309, "step": 8431 }, { "epoch": 3.818840579710145, "grad_norm": 0.06052002706924069, "learning_rate": 1.2846112791586526e-05, "loss": 0.9405, "step": 8432 }, { "epoch": 3.8192934782608696, "grad_norm": 0.05105600738130084, "learning_rate": 1.2836827485026308e-05, "loss": 0.9297, "step": 8433 }, { "epoch": 3.8197463768115942, "grad_norm": 0.058774924715238985, "learning_rate": 1.2827544894011496e-05, "loss": 0.9245, "step": 8434 }, { "epoch": 3.820199275362319, "grad_norm": 0.056420459477855345, "learning_rate": 1.2818265019470087e-05, "loss": 0.919, "step": 8435 }, { "epoch": 3.8206521739130435, "grad_norm": 0.052470846046348246, "learning_rate": 1.280898786232979e-05, "loss": 0.9291, "step": 8436 }, { "epoch": 3.821105072463768, "grad_norm": 0.05486058620281071, "learning_rate": 1.2799713423518054e-05, "loss": 0.9345, "step": 8437 }, { "epoch": 3.8215579710144927, "grad_norm": 0.05547259547531177, "learning_rate": 1.2790441703962082e-05, "loss": 0.911, "step": 8438 }, { "epoch": 3.8220108695652173, "grad_norm": 0.050444242391544575, "learning_rate": 1.2781172704588757e-05, "loss": 0.9464, "step": 8439 }, { "epoch": 3.822463768115942, "grad_norm": 0.05503814185050626, "learning_rate": 1.2771906426324746e-05, "loss": 0.9359, "step": 8440 }, { "epoch": 3.8229166666666665, "grad_norm": 0.05419521181023756, "learning_rate": 1.2762642870096377e-05, "loss": 0.9076, "step": 8441 }, { "epoch": 3.8233695652173916, "grad_norm": 0.05695151236192348, "learning_rate": 1.2753382036829778e-05, "loss": 0.9287, "step": 8442 }, { "epoch": 3.823822463768116, "grad_norm": 0.053747695165177765, "learning_rate": 1.2744123927450755e-05, "loss": 0.948, "step": 8443 }, { "epoch": 3.824275362318841, "grad_norm": 0.056139800597588226, "learning_rate": 1.273486854288485e-05, "loss": 0.9263, "step": 8444 }, { "epoch": 3.8247282608695654, "grad_norm": 0.05612936367741799, "learning_rate": 1.2725615884057328e-05, "loss": 0.9294, "step": 8445 }, { "epoch": 3.82518115942029, "grad_norm": 0.0552748940198207, "learning_rate": 1.2716365951893224e-05, "loss": 0.9263, "step": 8446 }, { "epoch": 3.8256340579710146, "grad_norm": 0.05787187991761834, "learning_rate": 1.270711874731723e-05, "loss": 0.9358, "step": 8447 }, { "epoch": 3.8260869565217392, "grad_norm": 0.058897227353162274, "learning_rate": 1.2697874271253844e-05, "loss": 0.932, "step": 8448 }, { "epoch": 3.826539855072464, "grad_norm": 0.05048002768207195, "learning_rate": 1.2688632524627215e-05, "loss": 0.9403, "step": 8449 }, { "epoch": 3.8269927536231885, "grad_norm": 0.0585863485944782, "learning_rate": 1.2679393508361284e-05, "loss": 0.9382, "step": 8450 }, { "epoch": 3.827445652173913, "grad_norm": 0.05347602993676602, "learning_rate": 1.2670157223379675e-05, "loss": 0.9138, "step": 8451 }, { "epoch": 3.8278985507246377, "grad_norm": 0.054408155347850644, "learning_rate": 1.2660923670605763e-05, "loss": 0.9413, "step": 8452 }, { "epoch": 3.8283514492753623, "grad_norm": 0.05871624607778704, "learning_rate": 1.2651692850962615e-05, "loss": 0.9196, "step": 8453 }, { "epoch": 3.828804347826087, "grad_norm": 0.05421136773644535, "learning_rate": 1.2642464765373084e-05, "loss": 0.9115, "step": 8454 }, { "epoch": 3.8292572463768115, "grad_norm": 0.060145369918461, "learning_rate": 1.263323941475969e-05, "loss": 0.9362, "step": 8455 }, { "epoch": 3.829710144927536, "grad_norm": 0.057783368609003696, "learning_rate": 1.2624016800044733e-05, "loss": 0.9288, "step": 8456 }, { "epoch": 3.8301630434782608, "grad_norm": 0.05672067881855979, "learning_rate": 1.2614796922150183e-05, "loss": 0.9056, "step": 8457 }, { "epoch": 3.8306159420289854, "grad_norm": 0.05171014276574784, "learning_rate": 1.2605579781997789e-05, "loss": 0.934, "step": 8458 }, { "epoch": 3.83106884057971, "grad_norm": 0.0545765731239741, "learning_rate": 1.2596365380508999e-05, "loss": 0.9215, "step": 8459 }, { "epoch": 3.8315217391304346, "grad_norm": 0.06318031735606038, "learning_rate": 1.2587153718604985e-05, "loss": 0.9314, "step": 8460 }, { "epoch": 3.831974637681159, "grad_norm": 0.05278455894064268, "learning_rate": 1.2577944797206643e-05, "loss": 0.9124, "step": 8461 }, { "epoch": 3.832427536231884, "grad_norm": 0.053499650791867005, "learning_rate": 1.2568738617234626e-05, "loss": 0.9133, "step": 8462 }, { "epoch": 3.8328804347826084, "grad_norm": 0.07437048989471173, "learning_rate": 1.2559535179609262e-05, "loss": 0.9503, "step": 8463 }, { "epoch": 3.8333333333333335, "grad_norm": 0.048712028676895794, "learning_rate": 1.2550334485250661e-05, "loss": 0.9135, "step": 8464 }, { "epoch": 3.833786231884058, "grad_norm": 0.06511271880687083, "learning_rate": 1.254113653507861e-05, "loss": 0.9181, "step": 8465 }, { "epoch": 3.8342391304347827, "grad_norm": 0.05277290787577492, "learning_rate": 1.2531941330012667e-05, "loss": 0.9224, "step": 8466 }, { "epoch": 3.8346920289855073, "grad_norm": 0.05113559833081846, "learning_rate": 1.252274887097206e-05, "loss": 0.9197, "step": 8467 }, { "epoch": 3.835144927536232, "grad_norm": 0.061530214726367743, "learning_rate": 1.2513559158875813e-05, "loss": 0.9119, "step": 8468 }, { "epoch": 3.8355978260869565, "grad_norm": 0.053237215476291004, "learning_rate": 1.2504372194642591e-05, "loss": 0.9545, "step": 8469 }, { "epoch": 3.836050724637681, "grad_norm": 0.05551276089906128, "learning_rate": 1.2495187979190861e-05, "loss": 0.9291, "step": 8470 }, { "epoch": 3.8365036231884058, "grad_norm": 0.04843558312390614, "learning_rate": 1.2486006513438764e-05, "loss": 0.9393, "step": 8471 }, { "epoch": 3.8369565217391304, "grad_norm": 0.05188494736424006, "learning_rate": 1.247682779830421e-05, "loss": 0.9209, "step": 8472 }, { "epoch": 3.837409420289855, "grad_norm": 0.049129568672645306, "learning_rate": 1.246765183470478e-05, "loss": 0.9144, "step": 8473 }, { "epoch": 3.8378623188405796, "grad_norm": 0.05008643203680874, "learning_rate": 1.2458478623557833e-05, "loss": 0.9262, "step": 8474 }, { "epoch": 3.838315217391304, "grad_norm": 0.054206479549939846, "learning_rate": 1.2449308165780414e-05, "loss": 0.9381, "step": 8475 }, { "epoch": 3.838768115942029, "grad_norm": 0.05246589355769453, "learning_rate": 1.2440140462289328e-05, "loss": 0.9245, "step": 8476 }, { "epoch": 3.839221014492754, "grad_norm": 0.06313195304231067, "learning_rate": 1.2430975514001071e-05, "loss": 0.9289, "step": 8477 }, { "epoch": 3.8396739130434785, "grad_norm": 0.0562940958856164, "learning_rate": 1.2421813321831877e-05, "loss": 0.9082, "step": 8478 }, { "epoch": 3.840126811594203, "grad_norm": 0.055339844073939326, "learning_rate": 1.2412653886697696e-05, "loss": 0.9232, "step": 8479 }, { "epoch": 3.8405797101449277, "grad_norm": 0.0637647847625564, "learning_rate": 1.2403497209514237e-05, "loss": 0.9502, "step": 8480 }, { "epoch": 3.8410326086956523, "grad_norm": 0.053750963144513074, "learning_rate": 1.2394343291196878e-05, "loss": 0.9307, "step": 8481 }, { "epoch": 3.841485507246377, "grad_norm": 0.06536381141556004, "learning_rate": 1.2385192132660775e-05, "loss": 0.9156, "step": 8482 }, { "epoch": 3.8419384057971016, "grad_norm": 0.056710400302797524, "learning_rate": 1.2376043734820762e-05, "loss": 0.9426, "step": 8483 }, { "epoch": 3.842391304347826, "grad_norm": 0.05558794742260681, "learning_rate": 1.2366898098591445e-05, "loss": 0.9173, "step": 8484 }, { "epoch": 3.8428442028985508, "grad_norm": 0.05721702354355171, "learning_rate": 1.2357755224887113e-05, "loss": 0.9086, "step": 8485 }, { "epoch": 3.8432971014492754, "grad_norm": 0.05892312687286903, "learning_rate": 1.2348615114621794e-05, "loss": 0.9042, "step": 8486 }, { "epoch": 3.84375, "grad_norm": 0.06279867228420029, "learning_rate": 1.233947776870923e-05, "loss": 0.9267, "step": 8487 }, { "epoch": 3.8442028985507246, "grad_norm": 0.05325273111760386, "learning_rate": 1.2330343188062912e-05, "loss": 0.8992, "step": 8488 }, { "epoch": 3.8446557971014492, "grad_norm": 0.05091712295844323, "learning_rate": 1.2321211373596023e-05, "loss": 0.9153, "step": 8489 }, { "epoch": 3.845108695652174, "grad_norm": 0.052894491634414405, "learning_rate": 1.2312082326221507e-05, "loss": 0.9207, "step": 8490 }, { "epoch": 3.8455615942028984, "grad_norm": 0.04884180263038032, "learning_rate": 1.230295604685198e-05, "loss": 0.946, "step": 8491 }, { "epoch": 3.846014492753623, "grad_norm": 0.05295698606498656, "learning_rate": 1.2293832536399841e-05, "loss": 0.9106, "step": 8492 }, { "epoch": 3.8464673913043477, "grad_norm": 0.04904682010013459, "learning_rate": 1.2284711795777167e-05, "loss": 0.9355, "step": 8493 }, { "epoch": 3.8469202898550723, "grad_norm": 0.05437979530008157, "learning_rate": 1.2275593825895778e-05, "loss": 0.9254, "step": 8494 }, { "epoch": 3.847373188405797, "grad_norm": 0.05726874304797571, "learning_rate": 1.2266478627667189e-05, "loss": 0.9112, "step": 8495 }, { "epoch": 3.8478260869565215, "grad_norm": 0.05349785734078354, "learning_rate": 1.2257366202002693e-05, "loss": 0.9297, "step": 8496 }, { "epoch": 3.848278985507246, "grad_norm": 0.04858192959734888, "learning_rate": 1.2248256549813249e-05, "loss": 0.9263, "step": 8497 }, { "epoch": 3.848731884057971, "grad_norm": 0.045107542091518744, "learning_rate": 1.2239149672009583e-05, "loss": 0.9172, "step": 8498 }, { "epoch": 3.849184782608696, "grad_norm": 0.04895209871549163, "learning_rate": 1.2230045569502105e-05, "loss": 0.9371, "step": 8499 }, { "epoch": 3.8496376811594204, "grad_norm": 0.04924209135948672, "learning_rate": 1.222094424320099e-05, "loss": 0.9227, "step": 8500 }, { "epoch": 3.850090579710145, "grad_norm": 0.04775069030798312, "learning_rate": 1.22118456940161e-05, "loss": 0.9386, "step": 8501 }, { "epoch": 3.8505434782608696, "grad_norm": 0.04946326745334516, "learning_rate": 1.2202749922857024e-05, "loss": 0.9417, "step": 8502 }, { "epoch": 3.8509963768115942, "grad_norm": 0.0511075605437066, "learning_rate": 1.219365693063308e-05, "loss": 0.9218, "step": 8503 }, { "epoch": 3.851449275362319, "grad_norm": 0.051109471352464186, "learning_rate": 1.2184566718253326e-05, "loss": 0.9306, "step": 8504 }, { "epoch": 3.8519021739130435, "grad_norm": 0.05593384895076793, "learning_rate": 1.2175479286626501e-05, "loss": 0.9163, "step": 8505 }, { "epoch": 3.852355072463768, "grad_norm": 0.05040586061579703, "learning_rate": 1.216639463666112e-05, "loss": 0.9283, "step": 8506 }, { "epoch": 3.8528079710144927, "grad_norm": 0.05623315605284067, "learning_rate": 1.2157312769265359e-05, "loss": 0.9073, "step": 8507 }, { "epoch": 3.8532608695652173, "grad_norm": 0.05333595612534265, "learning_rate": 1.214823368534717e-05, "loss": 0.9264, "step": 8508 }, { "epoch": 3.853713768115942, "grad_norm": 0.055384038892333974, "learning_rate": 1.2139157385814198e-05, "loss": 0.9402, "step": 8509 }, { "epoch": 3.8541666666666665, "grad_norm": 0.05618341626650603, "learning_rate": 1.2130083871573812e-05, "loss": 0.9047, "step": 8510 }, { "epoch": 3.8546195652173916, "grad_norm": 0.05470176036989307, "learning_rate": 1.2121013143533089e-05, "loss": 0.9299, "step": 8511 }, { "epoch": 3.855072463768116, "grad_norm": 0.05877680155115007, "learning_rate": 1.2111945202598876e-05, "loss": 0.9237, "step": 8512 }, { "epoch": 3.855525362318841, "grad_norm": 0.050316821004070596, "learning_rate": 1.2102880049677679e-05, "loss": 0.9367, "step": 8513 }, { "epoch": 3.8559782608695654, "grad_norm": 0.05314690659782095, "learning_rate": 1.2093817685675782e-05, "loss": 0.9403, "step": 8514 }, { "epoch": 3.85643115942029, "grad_norm": 0.050855986129492915, "learning_rate": 1.208475811149914e-05, "loss": 0.9074, "step": 8515 }, { "epoch": 3.8568840579710146, "grad_norm": 0.05080883210365074, "learning_rate": 1.2075701328053482e-05, "loss": 0.9242, "step": 8516 }, { "epoch": 3.8573369565217392, "grad_norm": 0.054639742020083315, "learning_rate": 1.2066647336244195e-05, "loss": 0.9274, "step": 8517 }, { "epoch": 3.857789855072464, "grad_norm": 0.04907354484971356, "learning_rate": 1.2057596136976453e-05, "loss": 0.9171, "step": 8518 }, { "epoch": 3.8582427536231885, "grad_norm": 0.05637899217498986, "learning_rate": 1.20485477311551e-05, "loss": 0.9535, "step": 8519 }, { "epoch": 3.858695652173913, "grad_norm": 0.04936217522350031, "learning_rate": 1.2039502119684725e-05, "loss": 0.8978, "step": 8520 }, { "epoch": 3.8591485507246377, "grad_norm": 0.06246865168180816, "learning_rate": 1.2030459303469617e-05, "loss": 0.9132, "step": 8521 }, { "epoch": 3.8596014492753623, "grad_norm": 0.05024000973980356, "learning_rate": 1.2021419283413831e-05, "loss": 0.9158, "step": 8522 }, { "epoch": 3.860054347826087, "grad_norm": 0.0651455116204076, "learning_rate": 1.2012382060421079e-05, "loss": 0.8952, "step": 8523 }, { "epoch": 3.8605072463768115, "grad_norm": 0.05190947381902504, "learning_rate": 1.2003347635394853e-05, "loss": 0.9374, "step": 8524 }, { "epoch": 3.860960144927536, "grad_norm": 0.053463445749449195, "learning_rate": 1.1994316009238317e-05, "loss": 0.904, "step": 8525 }, { "epoch": 3.8614130434782608, "grad_norm": 0.05513984080437051, "learning_rate": 1.1985287182854406e-05, "loss": 0.9452, "step": 8526 }, { "epoch": 3.8618659420289854, "grad_norm": 0.0474062012812674, "learning_rate": 1.1976261157145724e-05, "loss": 0.9183, "step": 8527 }, { "epoch": 3.86231884057971, "grad_norm": 0.05096473117613417, "learning_rate": 1.1967237933014619e-05, "loss": 0.9202, "step": 8528 }, { "epoch": 3.8627717391304346, "grad_norm": 0.05251255569236638, "learning_rate": 1.195821751136315e-05, "loss": 0.9124, "step": 8529 }, { "epoch": 3.863224637681159, "grad_norm": 0.04958644860077454, "learning_rate": 1.1949199893093128e-05, "loss": 0.9296, "step": 8530 }, { "epoch": 3.863677536231884, "grad_norm": 0.053093794228534275, "learning_rate": 1.1940185079106028e-05, "loss": 0.9087, "step": 8531 }, { "epoch": 3.8641304347826084, "grad_norm": 0.051445692753111304, "learning_rate": 1.19311730703031e-05, "loss": 0.9245, "step": 8532 }, { "epoch": 3.8645833333333335, "grad_norm": 0.057670181960108986, "learning_rate": 1.1922163867585268e-05, "loss": 0.9302, "step": 8533 }, { "epoch": 3.865036231884058, "grad_norm": 0.05327550171260109, "learning_rate": 1.191315747185322e-05, "loss": 0.9178, "step": 8534 }, { "epoch": 3.8654891304347827, "grad_norm": 0.04971912542832644, "learning_rate": 1.1904153884007333e-05, "loss": 0.9246, "step": 8535 }, { "epoch": 3.8659420289855073, "grad_norm": 0.0526941653968449, "learning_rate": 1.1895153104947696e-05, "loss": 0.9154, "step": 8536 }, { "epoch": 3.866394927536232, "grad_norm": 0.048845130680297076, "learning_rate": 1.1886155135574132e-05, "loss": 0.9248, "step": 8537 }, { "epoch": 3.8668478260869565, "grad_norm": 0.05319575379095647, "learning_rate": 1.1877159976786197e-05, "loss": 0.9308, "step": 8538 }, { "epoch": 3.867300724637681, "grad_norm": 0.050963268308161246, "learning_rate": 1.1868167629483134e-05, "loss": 0.9311, "step": 8539 }, { "epoch": 3.8677536231884058, "grad_norm": 0.04830203473629808, "learning_rate": 1.1859178094563939e-05, "loss": 0.9302, "step": 8540 }, { "epoch": 3.8682065217391304, "grad_norm": 0.05247646840126707, "learning_rate": 1.1850191372927298e-05, "loss": 0.9374, "step": 8541 }, { "epoch": 3.868659420289855, "grad_norm": 0.05188390868560488, "learning_rate": 1.1841207465471642e-05, "loss": 0.9102, "step": 8542 }, { "epoch": 3.8691123188405796, "grad_norm": 0.05447430474129002, "learning_rate": 1.1832226373095095e-05, "loss": 0.9175, "step": 8543 }, { "epoch": 3.869565217391304, "grad_norm": 0.04755642116786022, "learning_rate": 1.1823248096695515e-05, "loss": 0.9176, "step": 8544 }, { "epoch": 3.870018115942029, "grad_norm": 0.05958057774240255, "learning_rate": 1.1814272637170458e-05, "loss": 0.9404, "step": 8545 }, { "epoch": 3.870471014492754, "grad_norm": 0.04799546259350199, "learning_rate": 1.1805299995417245e-05, "loss": 0.8951, "step": 8546 }, { "epoch": 3.8709239130434785, "grad_norm": 0.053271049044634027, "learning_rate": 1.179633017233286e-05, "loss": 0.928, "step": 8547 }, { "epoch": 3.871376811594203, "grad_norm": 0.0552050245553821, "learning_rate": 1.1787363168814046e-05, "loss": 0.9191, "step": 8548 }, { "epoch": 3.8718297101449277, "grad_norm": 0.047982787176699834, "learning_rate": 1.1778398985757234e-05, "loss": 0.9163, "step": 8549 }, { "epoch": 3.8722826086956523, "grad_norm": 0.06215920659717898, "learning_rate": 1.1769437624058608e-05, "loss": 0.9137, "step": 8550 }, { "epoch": 3.872735507246377, "grad_norm": 0.04966230625313317, "learning_rate": 1.176047908461404e-05, "loss": 0.8983, "step": 8551 }, { "epoch": 3.8731884057971016, "grad_norm": 0.05895432696672567, "learning_rate": 1.1751523368319125e-05, "loss": 0.9115, "step": 8552 }, { "epoch": 3.873641304347826, "grad_norm": 0.05144186903509929, "learning_rate": 1.174257047606917e-05, "loss": 0.9124, "step": 8553 }, { "epoch": 3.8740942028985508, "grad_norm": 0.05419886782179085, "learning_rate": 1.1733620408759237e-05, "loss": 0.9217, "step": 8554 }, { "epoch": 3.8745471014492754, "grad_norm": 0.044312573046210516, "learning_rate": 1.1724673167284046e-05, "loss": 0.9192, "step": 8555 }, { "epoch": 3.875, "grad_norm": 0.057210792932717386, "learning_rate": 1.1715728752538103e-05, "loss": 0.9402, "step": 8556 }, { "epoch": 3.8754528985507246, "grad_norm": 0.056547352350384814, "learning_rate": 1.1706787165415574e-05, "loss": 0.947, "step": 8557 }, { "epoch": 3.8759057971014492, "grad_norm": 0.06031026363586516, "learning_rate": 1.1697848406810354e-05, "loss": 0.9137, "step": 8558 }, { "epoch": 3.876358695652174, "grad_norm": 0.054088722768173905, "learning_rate": 1.168891247761609e-05, "loss": 0.9306, "step": 8559 }, { "epoch": 3.8768115942028984, "grad_norm": 0.05831498070552094, "learning_rate": 1.1679979378726096e-05, "loss": 0.9346, "step": 8560 }, { "epoch": 3.877264492753623, "grad_norm": 0.053237984792168694, "learning_rate": 1.1671049111033454e-05, "loss": 0.9312, "step": 8561 }, { "epoch": 3.8777173913043477, "grad_norm": 0.05575765668433309, "learning_rate": 1.1662121675430922e-05, "loss": 0.9066, "step": 8562 }, { "epoch": 3.8781702898550723, "grad_norm": 0.05596343056004997, "learning_rate": 1.1653197072810993e-05, "loss": 0.9105, "step": 8563 }, { "epoch": 3.878623188405797, "grad_norm": 0.054623707652408385, "learning_rate": 1.1644275304065857e-05, "loss": 0.9072, "step": 8564 }, { "epoch": 3.8790760869565215, "grad_norm": 0.05914511038959764, "learning_rate": 1.1635356370087467e-05, "loss": 0.9239, "step": 8565 }, { "epoch": 3.879528985507246, "grad_norm": 0.05054962400568257, "learning_rate": 1.1626440271767433e-05, "loss": 0.9531, "step": 8566 }, { "epoch": 3.879981884057971, "grad_norm": 0.06337353490365068, "learning_rate": 1.1617527009997143e-05, "loss": 0.9223, "step": 8567 }, { "epoch": 3.880434782608696, "grad_norm": 0.04964713246254109, "learning_rate": 1.160861658566764e-05, "loss": 0.9392, "step": 8568 }, { "epoch": 3.8808876811594204, "grad_norm": 0.06542389375781942, "learning_rate": 1.1599708999669738e-05, "loss": 0.9481, "step": 8569 }, { "epoch": 3.881340579710145, "grad_norm": 0.055657295454770626, "learning_rate": 1.159080425289393e-05, "loss": 0.9303, "step": 8570 }, { "epoch": 3.8817934782608696, "grad_norm": 0.04802648173345213, "learning_rate": 1.1581902346230445e-05, "loss": 0.9191, "step": 8571 }, { "epoch": 3.8822463768115942, "grad_norm": 0.056653730847718624, "learning_rate": 1.1573003280569197e-05, "loss": 0.9152, "step": 8572 }, { "epoch": 3.882699275362319, "grad_norm": 0.051451552938103925, "learning_rate": 1.1564107056799872e-05, "loss": 0.9237, "step": 8573 }, { "epoch": 3.8831521739130435, "grad_norm": 0.06365549493395294, "learning_rate": 1.155521367581181e-05, "loss": 0.9308, "step": 8574 }, { "epoch": 3.883605072463768, "grad_norm": 0.047698978533419964, "learning_rate": 1.1546323138494127e-05, "loss": 0.9118, "step": 8575 }, { "epoch": 3.8840579710144927, "grad_norm": 0.06802122710921696, "learning_rate": 1.1537435445735596e-05, "loss": 0.9158, "step": 8576 }, { "epoch": 3.8845108695652173, "grad_norm": 0.050957945215677744, "learning_rate": 1.1528550598424757e-05, "loss": 0.939, "step": 8577 }, { "epoch": 3.884963768115942, "grad_norm": 0.06260535718103981, "learning_rate": 1.151966859744983e-05, "loss": 0.9406, "step": 8578 }, { "epoch": 3.8854166666666665, "grad_norm": 0.05017599986917725, "learning_rate": 1.1510789443698772e-05, "loss": 0.92, "step": 8579 }, { "epoch": 3.8858695652173916, "grad_norm": 0.048677120520243575, "learning_rate": 1.1501913138059218e-05, "loss": 0.9399, "step": 8580 }, { "epoch": 3.886322463768116, "grad_norm": 0.059902714778705186, "learning_rate": 1.149303968141858e-05, "loss": 0.9364, "step": 8581 }, { "epoch": 3.886775362318841, "grad_norm": 0.047543175641899145, "learning_rate": 1.148416907466393e-05, "loss": 0.9283, "step": 8582 }, { "epoch": 3.8872282608695654, "grad_norm": 0.0594182552095456, "learning_rate": 1.1475301318682095e-05, "loss": 0.9277, "step": 8583 }, { "epoch": 3.88768115942029, "grad_norm": 0.04644011506416385, "learning_rate": 1.1466436414359574e-05, "loss": 0.9334, "step": 8584 }, { "epoch": 3.8881340579710146, "grad_norm": 0.05782289716897645, "learning_rate": 1.1457574362582635e-05, "loss": 0.9178, "step": 8585 }, { "epoch": 3.8885869565217392, "grad_norm": 0.05208081615252967, "learning_rate": 1.1448715164237219e-05, "loss": 0.9243, "step": 8586 }, { "epoch": 3.889039855072464, "grad_norm": 0.06273654412962347, "learning_rate": 1.1439858820208985e-05, "loss": 0.9187, "step": 8587 }, { "epoch": 3.8894927536231885, "grad_norm": 0.04832419665658532, "learning_rate": 1.1431005331383314e-05, "loss": 0.9326, "step": 8588 }, { "epoch": 3.889945652173913, "grad_norm": 0.054543791720577155, "learning_rate": 1.142215469864532e-05, "loss": 0.9049, "step": 8589 }, { "epoch": 3.8903985507246377, "grad_norm": 0.052811341681638264, "learning_rate": 1.1413306922879799e-05, "loss": 0.9052, "step": 8590 }, { "epoch": 3.8908514492753623, "grad_norm": 0.05699840292352247, "learning_rate": 1.1404462004971296e-05, "loss": 0.9493, "step": 8591 }, { "epoch": 3.891304347826087, "grad_norm": 0.05673969405211487, "learning_rate": 1.1395619945804021e-05, "loss": 0.9039, "step": 8592 }, { "epoch": 3.8917572463768115, "grad_norm": 0.04963450346849169, "learning_rate": 1.1386780746261965e-05, "loss": 0.9016, "step": 8593 }, { "epoch": 3.892210144927536, "grad_norm": 0.06210958242640526, "learning_rate": 1.1377944407228783e-05, "loss": 0.9113, "step": 8594 }, { "epoch": 3.8926630434782608, "grad_norm": 0.050017357887010554, "learning_rate": 1.1369110929587848e-05, "loss": 0.9332, "step": 8595 }, { "epoch": 3.8931159420289854, "grad_norm": 0.0543356430709729, "learning_rate": 1.1360280314222255e-05, "loss": 0.9093, "step": 8596 }, { "epoch": 3.89356884057971, "grad_norm": 0.05740712265676919, "learning_rate": 1.135145256201483e-05, "loss": 0.9042, "step": 8597 }, { "epoch": 3.8940217391304346, "grad_norm": 0.04743551905243772, "learning_rate": 1.1342627673848079e-05, "loss": 0.9238, "step": 8598 }, { "epoch": 3.894474637681159, "grad_norm": 0.06666257275789216, "learning_rate": 1.1333805650604264e-05, "loss": 0.9175, "step": 8599 }, { "epoch": 3.894927536231884, "grad_norm": 0.051699766133586456, "learning_rate": 1.1324986493165314e-05, "loss": 0.9237, "step": 8600 }, { "epoch": 3.8953804347826084, "grad_norm": 0.05923815450730659, "learning_rate": 1.1316170202412912e-05, "loss": 0.9277, "step": 8601 }, { "epoch": 3.8958333333333335, "grad_norm": 0.051602377995298654, "learning_rate": 1.130735677922842e-05, "loss": 0.9239, "step": 8602 }, { "epoch": 3.896286231884058, "grad_norm": 0.05365024386573971, "learning_rate": 1.1298546224492948e-05, "loss": 0.9379, "step": 8603 }, { "epoch": 3.8967391304347827, "grad_norm": 0.04913107925103459, "learning_rate": 1.1289738539087289e-05, "loss": 0.9338, "step": 8604 }, { "epoch": 3.8971920289855073, "grad_norm": 0.04872583709287335, "learning_rate": 1.1280933723891968e-05, "loss": 0.9063, "step": 8605 }, { "epoch": 3.897644927536232, "grad_norm": 0.05720398502228319, "learning_rate": 1.1272131779787196e-05, "loss": 0.9207, "step": 8606 }, { "epoch": 3.8980978260869565, "grad_norm": 0.05067631936026844, "learning_rate": 1.126333270765295e-05, "loss": 0.9227, "step": 8607 }, { "epoch": 3.898550724637681, "grad_norm": 0.04977058013228348, "learning_rate": 1.1254536508368852e-05, "loss": 0.9407, "step": 8608 }, { "epoch": 3.8990036231884058, "grad_norm": 0.04682202904269372, "learning_rate": 1.1245743182814305e-05, "loss": 0.9113, "step": 8609 }, { "epoch": 3.8994565217391304, "grad_norm": 0.04975163109739702, "learning_rate": 1.1236952731868369e-05, "loss": 0.9182, "step": 8610 }, { "epoch": 3.899909420289855, "grad_norm": 0.05317816530471291, "learning_rate": 1.1228165156409854e-05, "loss": 0.9425, "step": 8611 }, { "epoch": 3.9003623188405796, "grad_norm": 0.05026570935157706, "learning_rate": 1.1219380457317266e-05, "loss": 0.9149, "step": 8612 }, { "epoch": 3.900815217391304, "grad_norm": 0.05060444767379311, "learning_rate": 1.1210598635468815e-05, "loss": 0.9054, "step": 8613 }, { "epoch": 3.901268115942029, "grad_norm": 0.05315816157723548, "learning_rate": 1.1201819691742433e-05, "loss": 0.9275, "step": 8614 }, { "epoch": 3.901721014492754, "grad_norm": 0.05080873026392057, "learning_rate": 1.1193043627015779e-05, "loss": 0.9126, "step": 8615 }, { "epoch": 3.9021739130434785, "grad_norm": 0.051021666658507714, "learning_rate": 1.118427044216619e-05, "loss": 0.9174, "step": 8616 }, { "epoch": 3.902626811594203, "grad_norm": 0.051122748140907545, "learning_rate": 1.1175500138070761e-05, "loss": 0.9113, "step": 8617 }, { "epoch": 3.9030797101449277, "grad_norm": 0.04905041594141445, "learning_rate": 1.1166732715606248e-05, "loss": 0.9353, "step": 8618 }, { "epoch": 3.9035326086956523, "grad_norm": 0.04952130175793198, "learning_rate": 1.1157968175649163e-05, "loss": 0.9225, "step": 8619 }, { "epoch": 3.903985507246377, "grad_norm": 0.04816613475950183, "learning_rate": 1.1149206519075708e-05, "loss": 0.9368, "step": 8620 }, { "epoch": 3.9044384057971016, "grad_norm": 0.049938810491502526, "learning_rate": 1.1140447746761796e-05, "loss": 0.908, "step": 8621 }, { "epoch": 3.904891304347826, "grad_norm": 0.04952283597760251, "learning_rate": 1.113169185958304e-05, "loss": 0.9383, "step": 8622 }, { "epoch": 3.9053442028985508, "grad_norm": 0.048676319550998715, "learning_rate": 1.1122938858414809e-05, "loss": 0.9236, "step": 8623 }, { "epoch": 3.9057971014492754, "grad_norm": 0.05176490391008929, "learning_rate": 1.1114188744132125e-05, "loss": 0.9357, "step": 8624 }, { "epoch": 3.90625, "grad_norm": 0.04612700392392841, "learning_rate": 1.110544151760978e-05, "loss": 0.9285, "step": 8625 }, { "epoch": 3.9067028985507246, "grad_norm": 0.047573539775783776, "learning_rate": 1.109669717972222e-05, "loss": 0.9332, "step": 8626 }, { "epoch": 3.9071557971014492, "grad_norm": 0.05316716210246292, "learning_rate": 1.108795573134366e-05, "loss": 0.9214, "step": 8627 }, { "epoch": 3.907608695652174, "grad_norm": 0.052074982524985935, "learning_rate": 1.1079217173347976e-05, "loss": 0.9479, "step": 8628 }, { "epoch": 3.9080615942028984, "grad_norm": 0.055628902090114035, "learning_rate": 1.1070481506608783e-05, "loss": 0.9252, "step": 8629 }, { "epoch": 3.908514492753623, "grad_norm": 0.05235985727509426, "learning_rate": 1.1061748731999384e-05, "loss": 0.9058, "step": 8630 }, { "epoch": 3.9089673913043477, "grad_norm": 0.05359651732400794, "learning_rate": 1.1053018850392832e-05, "loss": 0.9024, "step": 8631 }, { "epoch": 3.9094202898550723, "grad_norm": 0.04952923714318974, "learning_rate": 1.1044291862661844e-05, "loss": 0.9133, "step": 8632 }, { "epoch": 3.909873188405797, "grad_norm": 0.04528808066348747, "learning_rate": 1.1035567769678895e-05, "loss": 0.9222, "step": 8633 }, { "epoch": 3.9103260869565215, "grad_norm": 0.047734521780480696, "learning_rate": 1.1026846572316124e-05, "loss": 0.9355, "step": 8634 }, { "epoch": 3.910778985507246, "grad_norm": 0.04750053358293556, "learning_rate": 1.1018128271445425e-05, "loss": 0.9301, "step": 8635 }, { "epoch": 3.911231884057971, "grad_norm": 0.04657151229551419, "learning_rate": 1.1009412867938365e-05, "loss": 0.9248, "step": 8636 }, { "epoch": 3.911684782608696, "grad_norm": 0.04906994029438087, "learning_rate": 1.1000700362666241e-05, "loss": 0.9398, "step": 8637 }, { "epoch": 3.9121376811594204, "grad_norm": 0.053128403177762094, "learning_rate": 1.0991990756500047e-05, "loss": 0.9357, "step": 8638 }, { "epoch": 3.912590579710145, "grad_norm": 0.05403322174979153, "learning_rate": 1.098328405031051e-05, "loss": 0.9506, "step": 8639 }, { "epoch": 3.9130434782608696, "grad_norm": 0.05556690461837141, "learning_rate": 1.0974580244968039e-05, "loss": 0.9209, "step": 8640 }, { "epoch": 3.9134963768115942, "grad_norm": 0.04909607371123229, "learning_rate": 1.0965879341342789e-05, "loss": 0.9357, "step": 8641 }, { "epoch": 3.913949275362319, "grad_norm": 0.05291960609462137, "learning_rate": 1.0957181340304577e-05, "loss": 0.9156, "step": 8642 }, { "epoch": 3.9144021739130435, "grad_norm": 0.047780234081546594, "learning_rate": 1.0948486242722978e-05, "loss": 0.9238, "step": 8643 }, { "epoch": 3.914855072463768, "grad_norm": 0.049286835369022806, "learning_rate": 1.0939794049467234e-05, "loss": 0.9168, "step": 8644 }, { "epoch": 3.9153079710144927, "grad_norm": 0.046916917252608124, "learning_rate": 1.0931104761406353e-05, "loss": 0.9567, "step": 8645 }, { "epoch": 3.9157608695652173, "grad_norm": 0.0500370948774326, "learning_rate": 1.0922418379408973e-05, "loss": 0.9472, "step": 8646 }, { "epoch": 3.916213768115942, "grad_norm": 0.048028396660184054, "learning_rate": 1.091373490434351e-05, "loss": 0.9151, "step": 8647 }, { "epoch": 3.9166666666666665, "grad_norm": 0.05040777611464356, "learning_rate": 1.0905054337078051e-05, "loss": 0.9199, "step": 8648 }, { "epoch": 3.9171195652173916, "grad_norm": 0.050648629239978746, "learning_rate": 1.0896376678480428e-05, "loss": 0.922, "step": 8649 }, { "epoch": 3.917572463768116, "grad_norm": 0.047833628133917686, "learning_rate": 1.0887701929418135e-05, "loss": 0.9193, "step": 8650 }, { "epoch": 3.918025362318841, "grad_norm": 0.053808643208518574, "learning_rate": 1.0879030090758422e-05, "loss": 0.9264, "step": 8651 }, { "epoch": 3.9184782608695654, "grad_norm": 0.04820698342871823, "learning_rate": 1.0870361163368205e-05, "loss": 0.9178, "step": 8652 }, { "epoch": 3.91893115942029, "grad_norm": 0.04990531718709012, "learning_rate": 1.0861695148114157e-05, "loss": 0.9352, "step": 8653 }, { "epoch": 3.9193840579710146, "grad_norm": 0.04886403652341054, "learning_rate": 1.085303204586262e-05, "loss": 0.9286, "step": 8654 }, { "epoch": 3.9198369565217392, "grad_norm": 0.050050077218805736, "learning_rate": 1.0844371857479653e-05, "loss": 0.9119, "step": 8655 }, { "epoch": 3.920289855072464, "grad_norm": 0.04870484280328968, "learning_rate": 1.0835714583831024e-05, "loss": 0.9241, "step": 8656 }, { "epoch": 3.9207427536231885, "grad_norm": 0.050504508698669855, "learning_rate": 1.0827060225782228e-05, "loss": 0.9175, "step": 8657 }, { "epoch": 3.921195652173913, "grad_norm": 0.053850687622541124, "learning_rate": 1.0818408784198447e-05, "loss": 0.9218, "step": 8658 }, { "epoch": 3.9216485507246377, "grad_norm": 0.05300375809018255, "learning_rate": 1.0809760259944588e-05, "loss": 0.9063, "step": 8659 }, { "epoch": 3.9221014492753623, "grad_norm": 0.058120506937742046, "learning_rate": 1.0801114653885238e-05, "loss": 0.9342, "step": 8660 }, { "epoch": 3.922554347826087, "grad_norm": 0.04999379982508067, "learning_rate": 1.0792471966884741e-05, "loss": 0.911, "step": 8661 }, { "epoch": 3.9230072463768115, "grad_norm": 0.05636443647522285, "learning_rate": 1.0783832199807108e-05, "loss": 0.9196, "step": 8662 }, { "epoch": 3.923460144927536, "grad_norm": 0.04879132070753507, "learning_rate": 1.0775195353516064e-05, "loss": 0.8966, "step": 8663 }, { "epoch": 3.9239130434782608, "grad_norm": 0.051650503198863676, "learning_rate": 1.0766561428875036e-05, "loss": 0.9239, "step": 8664 }, { "epoch": 3.9243659420289854, "grad_norm": 0.05295698254090839, "learning_rate": 1.0757930426747202e-05, "loss": 0.9277, "step": 8665 }, { "epoch": 3.92481884057971, "grad_norm": 0.046917503355327246, "learning_rate": 1.0749302347995388e-05, "loss": 0.9214, "step": 8666 }, { "epoch": 3.9252717391304346, "grad_norm": 0.04823000351231603, "learning_rate": 1.0740677193482184e-05, "loss": 0.9237, "step": 8667 }, { "epoch": 3.925724637681159, "grad_norm": 0.04769370889471607, "learning_rate": 1.0732054964069828e-05, "loss": 0.923, "step": 8668 }, { "epoch": 3.926177536231884, "grad_norm": 0.05121833490278573, "learning_rate": 1.0723435660620334e-05, "loss": 0.9229, "step": 8669 }, { "epoch": 3.9266304347826084, "grad_norm": 0.0495282829720821, "learning_rate": 1.0714819283995368e-05, "loss": 0.8996, "step": 8670 }, { "epoch": 3.9270833333333335, "grad_norm": 0.05012066565928563, "learning_rate": 1.0706205835056326e-05, "loss": 0.9083, "step": 8671 }, { "epoch": 3.927536231884058, "grad_norm": 0.04915570801752309, "learning_rate": 1.0697595314664295e-05, "loss": 0.8968, "step": 8672 }, { "epoch": 3.9279891304347827, "grad_norm": 0.0490834963866088, "learning_rate": 1.0688987723680105e-05, "loss": 0.9285, "step": 8673 }, { "epoch": 3.9284420289855073, "grad_norm": 0.05253070523759322, "learning_rate": 1.0680383062964249e-05, "loss": 0.915, "step": 8674 }, { "epoch": 3.928894927536232, "grad_norm": 0.04672005902701664, "learning_rate": 1.0671781333376967e-05, "loss": 0.9344, "step": 8675 }, { "epoch": 3.9293478260869565, "grad_norm": 0.052124212465055046, "learning_rate": 1.0663182535778169e-05, "loss": 0.9085, "step": 8676 }, { "epoch": 3.929800724637681, "grad_norm": 0.04681697187787189, "learning_rate": 1.0654586671027513e-05, "loss": 0.9147, "step": 8677 }, { "epoch": 3.9302536231884058, "grad_norm": 0.04545397419851247, "learning_rate": 1.064599373998433e-05, "loss": 0.9173, "step": 8678 }, { "epoch": 3.9307065217391304, "grad_norm": 0.05437214267341093, "learning_rate": 1.063740374350767e-05, "loss": 0.9143, "step": 8679 }, { "epoch": 3.931159420289855, "grad_norm": 0.05441563217273086, "learning_rate": 1.0628816682456273e-05, "loss": 0.9137, "step": 8680 }, { "epoch": 3.9316123188405796, "grad_norm": 0.046950474575606756, "learning_rate": 1.062023255768863e-05, "loss": 0.9299, "step": 8681 }, { "epoch": 3.932065217391304, "grad_norm": 0.05441276874423417, "learning_rate": 1.061165137006289e-05, "loss": 0.9127, "step": 8682 }, { "epoch": 3.932518115942029, "grad_norm": 0.0487830852141515, "learning_rate": 1.0603073120436923e-05, "loss": 0.9223, "step": 8683 }, { "epoch": 3.932971014492754, "grad_norm": 0.05864067949616435, "learning_rate": 1.059449780966833e-05, "loss": 0.9266, "step": 8684 }, { "epoch": 3.9334239130434785, "grad_norm": 0.049917570080172016, "learning_rate": 1.0585925438614377e-05, "loss": 0.9286, "step": 8685 }, { "epoch": 3.933876811594203, "grad_norm": 0.06293822553497835, "learning_rate": 1.0577356008132083e-05, "loss": 0.9371, "step": 8686 }, { "epoch": 3.9343297101449277, "grad_norm": 0.05304659303020183, "learning_rate": 1.0568789519078129e-05, "loss": 0.9456, "step": 8687 }, { "epoch": 3.9347826086956523, "grad_norm": 0.05615795672516028, "learning_rate": 1.0560225972308925e-05, "loss": 0.9197, "step": 8688 }, { "epoch": 3.935235507246377, "grad_norm": 0.05061821505418653, "learning_rate": 1.055166536868057e-05, "loss": 0.9289, "step": 8689 }, { "epoch": 3.9356884057971016, "grad_norm": 0.056770878956051936, "learning_rate": 1.0543107709048908e-05, "loss": 0.9295, "step": 8690 }, { "epoch": 3.936141304347826, "grad_norm": 0.05054505003027779, "learning_rate": 1.053455299426943e-05, "loss": 0.9221, "step": 8691 }, { "epoch": 3.9365942028985508, "grad_norm": 0.049227343401538344, "learning_rate": 1.0526001225197397e-05, "loss": 0.9318, "step": 8692 }, { "epoch": 3.9370471014492754, "grad_norm": 0.05348499660525952, "learning_rate": 1.051745240268771e-05, "loss": 0.9286, "step": 8693 }, { "epoch": 3.9375, "grad_norm": 0.049721768591416554, "learning_rate": 1.0508906527595042e-05, "loss": 0.9195, "step": 8694 }, { "epoch": 3.9379528985507246, "grad_norm": 0.05544597088546663, "learning_rate": 1.050036360077371e-05, "loss": 0.9114, "step": 8695 }, { "epoch": 3.9384057971014492, "grad_norm": 0.049180479740078255, "learning_rate": 1.0491823623077786e-05, "loss": 0.9434, "step": 8696 }, { "epoch": 3.938858695652174, "grad_norm": 0.049274615927853364, "learning_rate": 1.048328659536101e-05, "loss": 0.9286, "step": 8697 }, { "epoch": 3.9393115942028984, "grad_norm": 0.05071955792320231, "learning_rate": 1.047475251847685e-05, "loss": 0.8933, "step": 8698 }, { "epoch": 3.939764492753623, "grad_norm": 0.05284232045977983, "learning_rate": 1.0466221393278455e-05, "loss": 0.9243, "step": 8699 }, { "epoch": 3.9402173913043477, "grad_norm": 0.049431873829607315, "learning_rate": 1.0457693220618719e-05, "loss": 0.9545, "step": 8700 }, { "epoch": 3.9406702898550723, "grad_norm": 0.054560856445041724, "learning_rate": 1.0449168001350193e-05, "loss": 0.9014, "step": 8701 }, { "epoch": 3.941123188405797, "grad_norm": 0.055671753299437465, "learning_rate": 1.0440645736325186e-05, "loss": 0.9123, "step": 8702 }, { "epoch": 3.9415760869565215, "grad_norm": 0.0469984452553422, "learning_rate": 1.0432126426395648e-05, "loss": 0.9233, "step": 8703 }, { "epoch": 3.942028985507246, "grad_norm": 0.05578632013079203, "learning_rate": 1.04236100724133e-05, "loss": 0.9187, "step": 8704 }, { "epoch": 3.942481884057971, "grad_norm": 0.047367916693442924, "learning_rate": 1.0415096675229521e-05, "loss": 0.9394, "step": 8705 }, { "epoch": 3.942934782608696, "grad_norm": 0.05674965130578537, "learning_rate": 1.0406586235695406e-05, "loss": 0.9046, "step": 8706 }, { "epoch": 3.9433876811594204, "grad_norm": 0.051864945783264484, "learning_rate": 1.0398078754661754e-05, "loss": 0.9064, "step": 8707 }, { "epoch": 3.943840579710145, "grad_norm": 0.05436631893248813, "learning_rate": 1.0389574232979082e-05, "loss": 0.9186, "step": 8708 }, { "epoch": 3.9442934782608696, "grad_norm": 0.05537746718900641, "learning_rate": 1.0381072671497585e-05, "loss": 0.9219, "step": 8709 }, { "epoch": 3.9447463768115942, "grad_norm": 0.05068605102287369, "learning_rate": 1.0372574071067203e-05, "loss": 0.925, "step": 8710 }, { "epoch": 3.945199275362319, "grad_norm": 0.05687709142760279, "learning_rate": 1.0364078432537528e-05, "loss": 0.9211, "step": 8711 }, { "epoch": 3.9456521739130435, "grad_norm": 0.05115182751999106, "learning_rate": 1.0355585756757907e-05, "loss": 0.9288, "step": 8712 }, { "epoch": 3.946105072463768, "grad_norm": 0.05260353917267765, "learning_rate": 1.0347096044577349e-05, "loss": 0.9062, "step": 8713 }, { "epoch": 3.9465579710144927, "grad_norm": 0.06398352005264296, "learning_rate": 1.0338609296844591e-05, "loss": 0.9115, "step": 8714 }, { "epoch": 3.9470108695652173, "grad_norm": 0.045406124103092964, "learning_rate": 1.0330125514408053e-05, "loss": 0.9224, "step": 8715 }, { "epoch": 3.947463768115942, "grad_norm": 0.055814967496775475, "learning_rate": 1.0321644698115892e-05, "loss": 0.9273, "step": 8716 }, { "epoch": 3.9479166666666665, "grad_norm": 0.0521402164015719, "learning_rate": 1.0313166848815931e-05, "loss": 0.9332, "step": 8717 }, { "epoch": 3.9483695652173916, "grad_norm": 0.0517609003311151, "learning_rate": 1.0304691967355733e-05, "loss": 0.912, "step": 8718 }, { "epoch": 3.948822463768116, "grad_norm": 0.05439901438706387, "learning_rate": 1.0296220054582525e-05, "loss": 0.9081, "step": 8719 }, { "epoch": 3.949275362318841, "grad_norm": 0.04678071290627989, "learning_rate": 1.028775111134328e-05, "loss": 0.8975, "step": 8720 }, { "epoch": 3.9497282608695654, "grad_norm": 0.050321308620588305, "learning_rate": 1.0279285138484641e-05, "loss": 0.923, "step": 8721 }, { "epoch": 3.95018115942029, "grad_norm": 0.0478576371181892, "learning_rate": 1.0270822136852959e-05, "loss": 0.9271, "step": 8722 }, { "epoch": 3.9506340579710146, "grad_norm": 0.04882731022990722, "learning_rate": 1.0262362107294286e-05, "loss": 0.9073, "step": 8723 }, { "epoch": 3.9510869565217392, "grad_norm": 0.047875730226374945, "learning_rate": 1.0253905050654409e-05, "loss": 0.9329, "step": 8724 }, { "epoch": 3.951539855072464, "grad_norm": 0.04685791798499427, "learning_rate": 1.0245450967778772e-05, "loss": 0.9388, "step": 8725 }, { "epoch": 3.9519927536231885, "grad_norm": 0.04681643348371889, "learning_rate": 1.0236999859512564e-05, "loss": 0.9101, "step": 8726 }, { "epoch": 3.952445652173913, "grad_norm": 0.050936733465110406, "learning_rate": 1.0228551726700626e-05, "loss": 0.94, "step": 8727 }, { "epoch": 3.9528985507246377, "grad_norm": 0.04840408869118191, "learning_rate": 1.0220106570187568e-05, "loss": 0.9256, "step": 8728 }, { "epoch": 3.9533514492753623, "grad_norm": 0.05257006683819969, "learning_rate": 1.021166439081764e-05, "loss": 0.9366, "step": 8729 }, { "epoch": 3.953804347826087, "grad_norm": 0.04870705549313369, "learning_rate": 1.0203225189434832e-05, "loss": 0.929, "step": 8730 }, { "epoch": 3.9542572463768115, "grad_norm": 0.04782614942475988, "learning_rate": 1.01947889668828e-05, "loss": 0.9169, "step": 8731 }, { "epoch": 3.954710144927536, "grad_norm": 0.04727624277686595, "learning_rate": 1.0186355724004962e-05, "loss": 0.9234, "step": 8732 }, { "epoch": 3.9551630434782608, "grad_norm": 0.04971582370978813, "learning_rate": 1.0177925461644374e-05, "loss": 0.9194, "step": 8733 }, { "epoch": 3.9556159420289854, "grad_norm": 0.050091380071796046, "learning_rate": 1.0169498180643846e-05, "loss": 0.941, "step": 8734 }, { "epoch": 3.95606884057971, "grad_norm": 0.05780397135795655, "learning_rate": 1.0161073881845845e-05, "loss": 0.9216, "step": 8735 }, { "epoch": 3.9565217391304346, "grad_norm": 0.05633135798296623, "learning_rate": 1.0152652566092582e-05, "loss": 0.9199, "step": 8736 }, { "epoch": 3.956974637681159, "grad_norm": 0.05003883562645097, "learning_rate": 1.014423423422593e-05, "loss": 0.9132, "step": 8737 }, { "epoch": 3.957427536231884, "grad_norm": 0.05486427673452699, "learning_rate": 1.01358188870875e-05, "loss": 0.9032, "step": 8738 }, { "epoch": 3.9578804347826084, "grad_norm": 0.0548314646650798, "learning_rate": 1.0127406525518584e-05, "loss": 0.945, "step": 8739 }, { "epoch": 3.9583333333333335, "grad_norm": 0.04891469005468157, "learning_rate": 1.0118997150360169e-05, "loss": 0.9412, "step": 8740 }, { "epoch": 3.958786231884058, "grad_norm": 0.052601775217459845, "learning_rate": 1.011059076245295e-05, "loss": 0.9126, "step": 8741 }, { "epoch": 3.9592391304347827, "grad_norm": 0.06275298758460585, "learning_rate": 1.010218736263735e-05, "loss": 0.9442, "step": 8742 }, { "epoch": 3.9596920289855073, "grad_norm": 0.04489233537380919, "learning_rate": 1.0093786951753443e-05, "loss": 0.9469, "step": 8743 }, { "epoch": 3.960144927536232, "grad_norm": 0.05994720688961265, "learning_rate": 1.0085389530641057e-05, "loss": 0.9117, "step": 8744 }, { "epoch": 3.9605978260869565, "grad_norm": 0.05209687939035548, "learning_rate": 1.0076995100139672e-05, "loss": 0.9025, "step": 8745 }, { "epoch": 3.961050724637681, "grad_norm": 0.04957966561595237, "learning_rate": 1.006860366108852e-05, "loss": 0.9334, "step": 8746 }, { "epoch": 3.9615036231884058, "grad_norm": 0.06071242356974532, "learning_rate": 1.0060215214326483e-05, "loss": 0.9385, "step": 8747 }, { "epoch": 3.9619565217391304, "grad_norm": 0.0518346538377532, "learning_rate": 1.005182976069218e-05, "loss": 0.9177, "step": 8748 }, { "epoch": 3.962409420289855, "grad_norm": 0.04962659809149322, "learning_rate": 1.0043447301023898e-05, "loss": 0.9125, "step": 8749 }, { "epoch": 3.9628623188405796, "grad_norm": 0.047158212990579394, "learning_rate": 1.0035067836159671e-05, "loss": 0.9212, "step": 8750 }, { "epoch": 3.963315217391304, "grad_norm": 0.04590112966422476, "learning_rate": 1.0026691366937187e-05, "loss": 0.9189, "step": 8751 }, { "epoch": 3.963768115942029, "grad_norm": 0.05163778160990102, "learning_rate": 1.001831789419387e-05, "loss": 0.9318, "step": 8752 }, { "epoch": 3.964221014492754, "grad_norm": 0.04826524819634031, "learning_rate": 1.0009947418766818e-05, "loss": 0.9198, "step": 8753 }, { "epoch": 3.9646739130434785, "grad_norm": 0.05575863978553619, "learning_rate": 1.0001579941492854e-05, "loss": 0.9201, "step": 8754 }, { "epoch": 3.965126811594203, "grad_norm": 0.048202967253135, "learning_rate": 9.993215463208479e-06, "loss": 0.9293, "step": 8755 }, { "epoch": 3.9655797101449277, "grad_norm": 0.0475941293056345, "learning_rate": 9.984853984749904e-06, "loss": 0.9278, "step": 8756 }, { "epoch": 3.9660326086956523, "grad_norm": 0.04972191222712633, "learning_rate": 9.976495506953032e-06, "loss": 0.9086, "step": 8757 }, { "epoch": 3.966485507246377, "grad_norm": 0.051427580443807415, "learning_rate": 9.96814003065349e-06, "loss": 0.9283, "step": 8758 }, { "epoch": 3.9669384057971016, "grad_norm": 0.047119171151341024, "learning_rate": 9.959787556686567e-06, "loss": 0.9154, "step": 8759 }, { "epoch": 3.967391304347826, "grad_norm": 0.046072962976770535, "learning_rate": 9.951438085887294e-06, "loss": 0.927, "step": 8760 }, { "epoch": 3.9678442028985508, "grad_norm": 0.04748326133962676, "learning_rate": 9.943091619090359e-06, "loss": 0.907, "step": 8761 }, { "epoch": 3.9682971014492754, "grad_norm": 0.04925431347624281, "learning_rate": 9.934748157130203e-06, "loss": 0.9237, "step": 8762 }, { "epoch": 3.96875, "grad_norm": 0.051031430642313, "learning_rate": 9.92640770084091e-06, "loss": 0.9233, "step": 8763 }, { "epoch": 3.9692028985507246, "grad_norm": 0.05132047678279103, "learning_rate": 9.918070251056297e-06, "loss": 0.9178, "step": 8764 }, { "epoch": 3.9696557971014492, "grad_norm": 0.04753502795508379, "learning_rate": 9.909735808609856e-06, "loss": 0.9186, "step": 8765 }, { "epoch": 3.970108695652174, "grad_norm": 0.05001885058268923, "learning_rate": 9.901404374334823e-06, "loss": 0.9335, "step": 8766 }, { "epoch": 3.9705615942028984, "grad_norm": 0.04925403512395199, "learning_rate": 9.893075949064075e-06, "loss": 0.9136, "step": 8767 }, { "epoch": 3.971014492753623, "grad_norm": 0.04778123186043669, "learning_rate": 9.884750533630249e-06, "loss": 0.9337, "step": 8768 }, { "epoch": 3.9714673913043477, "grad_norm": 0.04475296114613658, "learning_rate": 9.876428128865618e-06, "loss": 0.9501, "step": 8769 }, { "epoch": 3.9719202898550723, "grad_norm": 0.04865982111466945, "learning_rate": 9.868108735602217e-06, "loss": 0.9207, "step": 8770 }, { "epoch": 3.972373188405797, "grad_norm": 0.04911969300135895, "learning_rate": 9.85979235467173e-06, "loss": 0.9392, "step": 8771 }, { "epoch": 3.9728260869565215, "grad_norm": 0.04879863792283871, "learning_rate": 9.851478986905567e-06, "loss": 0.9177, "step": 8772 }, { "epoch": 3.973278985507246, "grad_norm": 0.04771451780502813, "learning_rate": 9.84316863313481e-06, "loss": 0.9428, "step": 8773 }, { "epoch": 3.973731884057971, "grad_norm": 0.0470177212759896, "learning_rate": 9.834861294190282e-06, "loss": 0.9214, "step": 8774 }, { "epoch": 3.974184782608696, "grad_norm": 0.048687420146685596, "learning_rate": 9.826556970902464e-06, "loss": 0.9245, "step": 8775 }, { "epoch": 3.9746376811594204, "grad_norm": 0.04723160065410167, "learning_rate": 9.818255664101568e-06, "loss": 0.9214, "step": 8776 }, { "epoch": 3.975090579710145, "grad_norm": 0.04621134508858864, "learning_rate": 9.80995737461747e-06, "loss": 0.9208, "step": 8777 }, { "epoch": 3.9755434782608696, "grad_norm": 0.04784079885992919, "learning_rate": 9.801662103279783e-06, "loss": 0.9451, "step": 8778 }, { "epoch": 3.9759963768115942, "grad_norm": 0.04974862149923932, "learning_rate": 9.79336985091778e-06, "loss": 0.9116, "step": 8779 }, { "epoch": 3.976449275362319, "grad_norm": 0.04755022652437295, "learning_rate": 9.785080618360468e-06, "loss": 0.9091, "step": 8780 }, { "epoch": 3.9769021739130435, "grad_norm": 0.053899133776210605, "learning_rate": 9.77679440643653e-06, "loss": 0.9447, "step": 8781 }, { "epoch": 3.977355072463768, "grad_norm": 0.04983770892365669, "learning_rate": 9.768511215974343e-06, "loss": 0.9272, "step": 8782 }, { "epoch": 3.9778079710144927, "grad_norm": 0.04938562387278173, "learning_rate": 9.76023104780199e-06, "loss": 0.9072, "step": 8783 }, { "epoch": 3.9782608695652173, "grad_norm": 0.05270589289516349, "learning_rate": 9.751953902747267e-06, "loss": 0.9091, "step": 8784 }, { "epoch": 3.978713768115942, "grad_norm": 0.05165987767683678, "learning_rate": 9.743679781637634e-06, "loss": 0.92, "step": 8785 }, { "epoch": 3.9791666666666665, "grad_norm": 0.05271710177366883, "learning_rate": 9.735408685300287e-06, "loss": 0.9303, "step": 8786 }, { "epoch": 3.9796195652173916, "grad_norm": 0.04931438725534516, "learning_rate": 9.72714061456208e-06, "loss": 0.9154, "step": 8787 }, { "epoch": 3.980072463768116, "grad_norm": 0.05072648961757135, "learning_rate": 9.718875570249615e-06, "loss": 0.9343, "step": 8788 }, { "epoch": 3.980525362318841, "grad_norm": 0.05114695139522048, "learning_rate": 9.710613553189141e-06, "loss": 0.9493, "step": 8789 }, { "epoch": 3.9809782608695654, "grad_norm": 0.05126094861027592, "learning_rate": 9.702354564206625e-06, "loss": 0.916, "step": 8790 }, { "epoch": 3.98143115942029, "grad_norm": 0.04791111800792212, "learning_rate": 9.694098604127724e-06, "loss": 0.922, "step": 8791 }, { "epoch": 3.9818840579710146, "grad_norm": 0.05612816358384224, "learning_rate": 9.68584567377782e-06, "loss": 0.9417, "step": 8792 }, { "epoch": 3.9823369565217392, "grad_norm": 0.05431826242055979, "learning_rate": 9.677595773981951e-06, "loss": 0.9235, "step": 8793 }, { "epoch": 3.982789855072464, "grad_norm": 0.04670979341073655, "learning_rate": 9.669348905564893e-06, "loss": 0.928, "step": 8794 }, { "epoch": 3.9832427536231885, "grad_norm": 0.04817452321626167, "learning_rate": 9.661105069351078e-06, "loss": 0.9244, "step": 8795 }, { "epoch": 3.983695652173913, "grad_norm": 0.053442962584211016, "learning_rate": 9.652864266164674e-06, "loss": 0.9178, "step": 8796 }, { "epoch": 3.9841485507246377, "grad_norm": 0.04856818228303785, "learning_rate": 9.644626496829522e-06, "loss": 0.9076, "step": 8797 }, { "epoch": 3.9846014492753623, "grad_norm": 0.04622583469601212, "learning_rate": 9.636391762169159e-06, "loss": 0.9273, "step": 8798 }, { "epoch": 3.985054347826087, "grad_norm": 0.05074526397473732, "learning_rate": 9.628160063006814e-06, "loss": 0.9364, "step": 8799 }, { "epoch": 3.9855072463768115, "grad_norm": 0.04678642491170682, "learning_rate": 9.619931400165451e-06, "loss": 0.9129, "step": 8800 }, { "epoch": 3.985960144927536, "grad_norm": 0.04826013404158219, "learning_rate": 9.611705774467674e-06, "loss": 0.9399, "step": 8801 }, { "epoch": 3.9864130434782608, "grad_norm": 0.048368387473384435, "learning_rate": 9.603483186735837e-06, "loss": 0.9033, "step": 8802 }, { "epoch": 3.9868659420289854, "grad_norm": 0.05197628971845634, "learning_rate": 9.595263637791938e-06, "loss": 0.9055, "step": 8803 }, { "epoch": 3.98731884057971, "grad_norm": 0.049483817874542396, "learning_rate": 9.587047128457732e-06, "loss": 0.9199, "step": 8804 }, { "epoch": 3.9877717391304346, "grad_norm": 0.04666014788002775, "learning_rate": 9.578833659554614e-06, "loss": 0.9259, "step": 8805 }, { "epoch": 3.988224637681159, "grad_norm": 0.0472621690067734, "learning_rate": 9.570623231903702e-06, "loss": 0.9215, "step": 8806 }, { "epoch": 3.988677536231884, "grad_norm": 0.04753345499768872, "learning_rate": 9.562415846325805e-06, "loss": 0.9326, "step": 8807 }, { "epoch": 3.9891304347826084, "grad_norm": 0.04557073661019809, "learning_rate": 9.554211503641416e-06, "loss": 0.9237, "step": 8808 }, { "epoch": 3.9895833333333335, "grad_norm": 0.04994135669574651, "learning_rate": 9.546010204670759e-06, "loss": 0.9282, "step": 8809 }, { "epoch": 3.990036231884058, "grad_norm": 0.048868295994172695, "learning_rate": 9.53781195023371e-06, "loss": 0.9473, "step": 8810 }, { "epoch": 3.9904891304347827, "grad_norm": 0.05009881298802257, "learning_rate": 9.529616741149885e-06, "loss": 0.9105, "step": 8811 }, { "epoch": 3.9909420289855073, "grad_norm": 0.05134191576427652, "learning_rate": 9.521424578238547e-06, "loss": 0.9251, "step": 8812 }, { "epoch": 3.991394927536232, "grad_norm": 0.04898654817398589, "learning_rate": 9.5132354623187e-06, "loss": 0.9175, "step": 8813 }, { "epoch": 3.9918478260869565, "grad_norm": 0.05303661796825569, "learning_rate": 9.505049394209015e-06, "loss": 0.921, "step": 8814 }, { "epoch": 3.992300724637681, "grad_norm": 0.0495049018283433, "learning_rate": 9.49686637472787e-06, "loss": 0.9318, "step": 8815 }, { "epoch": 3.9927536231884058, "grad_norm": 0.04974184724434075, "learning_rate": 9.488686404693314e-06, "loss": 0.9188, "step": 8816 }, { "epoch": 3.9932065217391304, "grad_norm": 0.04915405570931228, "learning_rate": 9.48050948492314e-06, "loss": 0.9226, "step": 8817 }, { "epoch": 3.993659420289855, "grad_norm": 0.04934344658255101, "learning_rate": 9.472335616234782e-06, "loss": 0.9095, "step": 8818 }, { "epoch": 3.9941123188405796, "grad_norm": 0.050907925625245605, "learning_rate": 9.464164799445421e-06, "loss": 0.9207, "step": 8819 }, { "epoch": 3.994565217391304, "grad_norm": 0.052095385532898025, "learning_rate": 9.455997035371881e-06, "loss": 0.9228, "step": 8820 }, { "epoch": 3.995018115942029, "grad_norm": 0.050747430160477225, "learning_rate": 9.44783232483073e-06, "loss": 0.9258, "step": 8821 }, { "epoch": 3.995471014492754, "grad_norm": 0.04940511304719789, "learning_rate": 9.439670668638188e-06, "loss": 0.9201, "step": 8822 }, { "epoch": 3.9959239130434785, "grad_norm": 0.052192435918718846, "learning_rate": 9.431512067610216e-06, "loss": 0.9103, "step": 8823 }, { "epoch": 3.996376811594203, "grad_norm": 0.05025453841768649, "learning_rate": 9.423356522562398e-06, "loss": 0.9236, "step": 8824 }, { "epoch": 3.9968297101449277, "grad_norm": 0.05525182737553626, "learning_rate": 9.415204034310093e-06, "loss": 0.9239, "step": 8825 }, { "epoch": 3.9972826086956523, "grad_norm": 0.05329380554103527, "learning_rate": 9.407054603668295e-06, "loss": 0.9176, "step": 8826 }, { "epoch": 3.997735507246377, "grad_norm": 0.050476732882470836, "learning_rate": 9.39890823145174e-06, "loss": 0.9382, "step": 8827 }, { "epoch": 3.9981884057971016, "grad_norm": 0.05237453088171902, "learning_rate": 9.390764918474811e-06, "loss": 0.9434, "step": 8828 }, { "epoch": 3.998641304347826, "grad_norm": 0.05071094269223023, "learning_rate": 9.382624665551625e-06, "loss": 0.9075, "step": 8829 }, { "epoch": 3.9990942028985508, "grad_norm": 0.048966922207246195, "learning_rate": 9.374487473495959e-06, "loss": 0.9131, "step": 8830 }, { "epoch": 3.9995471014492754, "grad_norm": 0.05895997452929947, "learning_rate": 9.366353343121321e-06, "loss": 0.9253, "step": 8831 }, { "epoch": 4.0, "grad_norm": 0.0516590586564297, "learning_rate": 9.358222275240884e-06, "loss": 0.9184, "step": 8832 }, { "epoch": 4.000452898550725, "grad_norm": 0.0519320438920315, "learning_rate": 9.350094270667522e-06, "loss": 0.9079, "step": 8833 }, { "epoch": 4.000905797101449, "grad_norm": 0.056914182409634195, "learning_rate": 9.341969330213794e-06, "loss": 0.9355, "step": 8834 }, { "epoch": 4.001358695652174, "grad_norm": 0.05630548750043776, "learning_rate": 9.333847454691983e-06, "loss": 0.9276, "step": 8835 }, { "epoch": 4.0018115942028984, "grad_norm": 0.05430688359935294, "learning_rate": 9.325728644914024e-06, "loss": 0.9116, "step": 8836 }, { "epoch": 4.002264492753623, "grad_norm": 0.05659259097689502, "learning_rate": 9.317612901691597e-06, "loss": 0.8997, "step": 8837 }, { "epoch": 4.002717391304348, "grad_norm": 0.05248591592961795, "learning_rate": 9.309500225836014e-06, "loss": 0.9046, "step": 8838 }, { "epoch": 4.003170289855072, "grad_norm": 0.055092231743737854, "learning_rate": 9.301390618158334e-06, "loss": 0.9209, "step": 8839 }, { "epoch": 4.003623188405797, "grad_norm": 0.05403327767036031, "learning_rate": 9.293284079469288e-06, "loss": 0.933, "step": 8840 }, { "epoch": 4.0040760869565215, "grad_norm": 0.04536993683529677, "learning_rate": 9.285180610579284e-06, "loss": 0.8979, "step": 8841 }, { "epoch": 4.004528985507246, "grad_norm": 0.0502533731532687, "learning_rate": 9.277080212298442e-06, "loss": 0.8973, "step": 8842 }, { "epoch": 4.004981884057971, "grad_norm": 0.049867316456612625, "learning_rate": 9.268982885436584e-06, "loss": 0.9084, "step": 8843 }, { "epoch": 4.005434782608695, "grad_norm": 0.053064823766548, "learning_rate": 9.260888630803188e-06, "loss": 0.9146, "step": 8844 }, { "epoch": 4.00588768115942, "grad_norm": 0.04827478205197272, "learning_rate": 9.252797449207485e-06, "loss": 0.8909, "step": 8845 }, { "epoch": 4.006340579710145, "grad_norm": 0.04899167098258615, "learning_rate": 9.244709341458327e-06, "loss": 0.9293, "step": 8846 }, { "epoch": 4.006793478260869, "grad_norm": 0.05394449429614252, "learning_rate": 9.236624308364326e-06, "loss": 0.9447, "step": 8847 }, { "epoch": 4.007246376811594, "grad_norm": 0.046059449685685656, "learning_rate": 9.228542350733743e-06, "loss": 0.9035, "step": 8848 }, { "epoch": 4.007699275362318, "grad_norm": 0.049723000556253824, "learning_rate": 9.220463469374544e-06, "loss": 0.9198, "step": 8849 }, { "epoch": 4.008152173913044, "grad_norm": 0.046746683271731225, "learning_rate": 9.212387665094371e-06, "loss": 0.9219, "step": 8850 }, { "epoch": 4.0086050724637685, "grad_norm": 0.047891591491082006, "learning_rate": 9.204314938700602e-06, "loss": 0.9344, "step": 8851 }, { "epoch": 4.009057971014493, "grad_norm": 0.04814990402009266, "learning_rate": 9.19624529100026e-06, "loss": 0.9007, "step": 8852 }, { "epoch": 4.009510869565218, "grad_norm": 0.04784598543368741, "learning_rate": 9.188178722800095e-06, "loss": 0.9155, "step": 8853 }, { "epoch": 4.009963768115942, "grad_norm": 0.04845968021578461, "learning_rate": 9.180115234906522e-06, "loss": 0.9402, "step": 8854 }, { "epoch": 4.010416666666667, "grad_norm": 0.05502800374195018, "learning_rate": 9.172054828125678e-06, "loss": 0.9315, "step": 8855 }, { "epoch": 4.010869565217392, "grad_norm": 0.05422426583181074, "learning_rate": 9.163997503263364e-06, "loss": 0.9434, "step": 8856 }, { "epoch": 4.011322463768116, "grad_norm": 0.04746831383326309, "learning_rate": 9.155943261125081e-06, "loss": 0.9374, "step": 8857 }, { "epoch": 4.011775362318841, "grad_norm": 0.04637869043224356, "learning_rate": 9.147892102516019e-06, "loss": 0.9199, "step": 8858 }, { "epoch": 4.012228260869565, "grad_norm": 0.054853019506542276, "learning_rate": 9.13984402824108e-06, "loss": 0.9414, "step": 8859 }, { "epoch": 4.01268115942029, "grad_norm": 0.04667955290650221, "learning_rate": 9.131799039104825e-06, "loss": 0.9152, "step": 8860 }, { "epoch": 4.013134057971015, "grad_norm": 0.06564514855039381, "learning_rate": 9.12375713591155e-06, "loss": 0.9165, "step": 8861 }, { "epoch": 4.013586956521739, "grad_norm": 0.04931515393601894, "learning_rate": 9.115718319465187e-06, "loss": 0.9319, "step": 8862 }, { "epoch": 4.014039855072464, "grad_norm": 0.04805405901504255, "learning_rate": 9.107682590569418e-06, "loss": 0.9023, "step": 8863 }, { "epoch": 4.0144927536231885, "grad_norm": 0.057047637728089726, "learning_rate": 9.099649950027572e-06, "loss": 0.9439, "step": 8864 }, { "epoch": 4.014945652173913, "grad_norm": 0.04727304557939986, "learning_rate": 9.091620398642686e-06, "loss": 0.908, "step": 8865 }, { "epoch": 4.015398550724638, "grad_norm": 0.05243546335091959, "learning_rate": 9.083593937217477e-06, "loss": 0.9349, "step": 8866 }, { "epoch": 4.015851449275362, "grad_norm": 0.05178269104665061, "learning_rate": 9.075570566554383e-06, "loss": 0.9124, "step": 8867 }, { "epoch": 4.016304347826087, "grad_norm": 0.04998379376171153, "learning_rate": 9.067550287455495e-06, "loss": 0.9258, "step": 8868 }, { "epoch": 4.0167572463768115, "grad_norm": 0.052396532693535204, "learning_rate": 9.059533100722628e-06, "loss": 0.9003, "step": 8869 }, { "epoch": 4.017210144927536, "grad_norm": 0.04875695849768856, "learning_rate": 9.051519007157257e-06, "loss": 0.9259, "step": 8870 }, { "epoch": 4.017663043478261, "grad_norm": 0.049330164783535885, "learning_rate": 9.043508007560588e-06, "loss": 0.9132, "step": 8871 }, { "epoch": 4.018115942028985, "grad_norm": 0.050096966816988886, "learning_rate": 9.035500102733463e-06, "loss": 0.9139, "step": 8872 }, { "epoch": 4.01856884057971, "grad_norm": 0.0494516697151903, "learning_rate": 9.027495293476471e-06, "loss": 0.9236, "step": 8873 }, { "epoch": 4.019021739130435, "grad_norm": 0.0476009911300554, "learning_rate": 9.019493580589858e-06, "loss": 0.9481, "step": 8874 }, { "epoch": 4.019474637681159, "grad_norm": 0.04653300673434477, "learning_rate": 9.011494964873564e-06, "loss": 0.9219, "step": 8875 }, { "epoch": 4.019927536231884, "grad_norm": 0.05347954886273975, "learning_rate": 9.003499447127209e-06, "loss": 0.9238, "step": 8876 }, { "epoch": 4.020380434782608, "grad_norm": 0.04872521031977196, "learning_rate": 8.995507028150147e-06, "loss": 0.904, "step": 8877 }, { "epoch": 4.020833333333333, "grad_norm": 0.04414029004966063, "learning_rate": 8.987517708741364e-06, "loss": 0.9508, "step": 8878 }, { "epoch": 4.021286231884058, "grad_norm": 0.050738808402587245, "learning_rate": 8.979531489699594e-06, "loss": 0.9325, "step": 8879 }, { "epoch": 4.021739130434782, "grad_norm": 0.0501475313428188, "learning_rate": 8.971548371823205e-06, "loss": 0.9143, "step": 8880 }, { "epoch": 4.022192028985507, "grad_norm": 0.04676939304847934, "learning_rate": 8.963568355910301e-06, "loss": 0.9315, "step": 8881 }, { "epoch": 4.0226449275362315, "grad_norm": 0.04733238959430386, "learning_rate": 8.955591442758655e-06, "loss": 0.8994, "step": 8882 }, { "epoch": 4.023097826086956, "grad_norm": 0.0492703831194603, "learning_rate": 8.947617633165726e-06, "loss": 0.8999, "step": 8883 }, { "epoch": 4.023550724637682, "grad_norm": 0.046163088321362226, "learning_rate": 8.939646927928659e-06, "loss": 0.9015, "step": 8884 }, { "epoch": 4.024003623188406, "grad_norm": 0.04865646187612968, "learning_rate": 8.931679327844316e-06, "loss": 0.9314, "step": 8885 }, { "epoch": 4.024456521739131, "grad_norm": 0.05067765559712045, "learning_rate": 8.92371483370921e-06, "loss": 0.9108, "step": 8886 }, { "epoch": 4.024909420289855, "grad_norm": 0.05063962858285401, "learning_rate": 8.91575344631959e-06, "loss": 0.92, "step": 8887 }, { "epoch": 4.02536231884058, "grad_norm": 0.049252530115069346, "learning_rate": 8.907795166471342e-06, "loss": 0.9349, "step": 8888 }, { "epoch": 4.025815217391305, "grad_norm": 0.04800851935751081, "learning_rate": 8.899839994960095e-06, "loss": 0.9321, "step": 8889 }, { "epoch": 4.026268115942029, "grad_norm": 0.04671600296648848, "learning_rate": 8.891887932581128e-06, "loss": 0.9099, "step": 8890 }, { "epoch": 4.026721014492754, "grad_norm": 0.046395915255543936, "learning_rate": 8.883938980129417e-06, "loss": 0.9179, "step": 8891 }, { "epoch": 4.0271739130434785, "grad_norm": 0.049064125753490416, "learning_rate": 8.875993138399624e-06, "loss": 0.9127, "step": 8892 }, { "epoch": 4.027626811594203, "grad_norm": 0.04797746614582114, "learning_rate": 8.868050408186125e-06, "loss": 0.9003, "step": 8893 }, { "epoch": 4.028079710144928, "grad_norm": 0.04964622991652426, "learning_rate": 8.86011079028295e-06, "loss": 0.9184, "step": 8894 }, { "epoch": 4.028532608695652, "grad_norm": 0.0510075393322778, "learning_rate": 8.852174285483856e-06, "loss": 0.9318, "step": 8895 }, { "epoch": 4.028985507246377, "grad_norm": 0.04931975682614836, "learning_rate": 8.844240894582245e-06, "loss": 0.9374, "step": 8896 }, { "epoch": 4.0294384057971016, "grad_norm": 0.04783590168521714, "learning_rate": 8.836310618371256e-06, "loss": 0.9438, "step": 8897 }, { "epoch": 4.029891304347826, "grad_norm": 0.062344254431978156, "learning_rate": 8.828383457643675e-06, "loss": 0.935, "step": 8898 }, { "epoch": 4.030344202898551, "grad_norm": 0.05114115560189463, "learning_rate": 8.820459413191993e-06, "loss": 0.9238, "step": 8899 }, { "epoch": 4.030797101449275, "grad_norm": 0.048803672751974975, "learning_rate": 8.812538485808386e-06, "loss": 0.9122, "step": 8900 }, { "epoch": 4.03125, "grad_norm": 0.053646915614780036, "learning_rate": 8.804620676284736e-06, "loss": 0.901, "step": 8901 }, { "epoch": 4.031702898550725, "grad_norm": 0.04597513002720196, "learning_rate": 8.79670598541258e-06, "loss": 0.9265, "step": 8902 }, { "epoch": 4.032155797101449, "grad_norm": 0.046990216705506974, "learning_rate": 8.788794413983179e-06, "loss": 0.8962, "step": 8903 }, { "epoch": 4.032608695652174, "grad_norm": 0.055047411422809944, "learning_rate": 8.780885962787452e-06, "loss": 0.9177, "step": 8904 }, { "epoch": 4.0330615942028984, "grad_norm": 0.04887960495754472, "learning_rate": 8.772980632616037e-06, "loss": 0.9255, "step": 8905 }, { "epoch": 4.033514492753623, "grad_norm": 0.0465611302366153, "learning_rate": 8.76507842425923e-06, "loss": 0.9061, "step": 8906 }, { "epoch": 4.033967391304348, "grad_norm": 0.05917982232670507, "learning_rate": 8.757179338507033e-06, "loss": 0.8955, "step": 8907 }, { "epoch": 4.034420289855072, "grad_norm": 0.046241245597046246, "learning_rate": 8.749283376149108e-06, "loss": 0.9432, "step": 8908 }, { "epoch": 4.034873188405797, "grad_norm": 0.04617899494449223, "learning_rate": 8.741390537974865e-06, "loss": 0.9149, "step": 8909 }, { "epoch": 4.0353260869565215, "grad_norm": 0.05397643745847504, "learning_rate": 8.733500824773325e-06, "loss": 0.9158, "step": 8910 }, { "epoch": 4.035778985507246, "grad_norm": 0.05074255049438988, "learning_rate": 8.725614237333268e-06, "loss": 0.9228, "step": 8911 }, { "epoch": 4.036231884057971, "grad_norm": 0.04609025991637418, "learning_rate": 8.717730776443104e-06, "loss": 0.8997, "step": 8912 }, { "epoch": 4.036684782608695, "grad_norm": 0.04916193260844144, "learning_rate": 8.709850442890974e-06, "loss": 0.9132, "step": 8913 }, { "epoch": 4.03713768115942, "grad_norm": 0.047778232940368044, "learning_rate": 8.701973237464667e-06, "loss": 0.9233, "step": 8914 }, { "epoch": 4.037590579710145, "grad_norm": 0.04805334925099305, "learning_rate": 8.694099160951706e-06, "loss": 0.9159, "step": 8915 }, { "epoch": 4.038043478260869, "grad_norm": 0.04939111923134671, "learning_rate": 8.686228214139256e-06, "loss": 0.937, "step": 8916 }, { "epoch": 4.038496376811594, "grad_norm": 0.04863039455288443, "learning_rate": 8.678360397814196e-06, "loss": 0.9298, "step": 8917 }, { "epoch": 4.038949275362318, "grad_norm": 0.047411220651090484, "learning_rate": 8.670495712763066e-06, "loss": 0.9354, "step": 8918 }, { "epoch": 4.039402173913044, "grad_norm": 0.048066412094739684, "learning_rate": 8.66263415977214e-06, "loss": 0.9346, "step": 8919 }, { "epoch": 4.0398550724637685, "grad_norm": 0.04767516299525429, "learning_rate": 8.654775739627324e-06, "loss": 0.9209, "step": 8920 }, { "epoch": 4.040307971014493, "grad_norm": 0.048368080776197565, "learning_rate": 8.646920453114256e-06, "loss": 0.9425, "step": 8921 }, { "epoch": 4.040760869565218, "grad_norm": 0.04969076776669997, "learning_rate": 8.639068301018226e-06, "loss": 0.9461, "step": 8922 }, { "epoch": 4.041213768115942, "grad_norm": 0.04909044812539219, "learning_rate": 8.631219284124239e-06, "loss": 0.9237, "step": 8923 }, { "epoch": 4.041666666666667, "grad_norm": 0.05013873281581939, "learning_rate": 8.623373403216972e-06, "loss": 0.9474, "step": 8924 }, { "epoch": 4.042119565217392, "grad_norm": 0.0510840858929077, "learning_rate": 8.615530659080788e-06, "loss": 0.9426, "step": 8925 }, { "epoch": 4.042572463768116, "grad_norm": 0.05666058705130271, "learning_rate": 8.607691052499726e-06, "loss": 0.9028, "step": 8926 }, { "epoch": 4.043025362318841, "grad_norm": 0.047162479586804834, "learning_rate": 8.599854584257543e-06, "loss": 0.8978, "step": 8927 }, { "epoch": 4.043478260869565, "grad_norm": 0.0498651375597587, "learning_rate": 8.592021255137646e-06, "loss": 0.9188, "step": 8928 }, { "epoch": 4.04393115942029, "grad_norm": 0.047078833897594576, "learning_rate": 8.584191065923163e-06, "loss": 0.933, "step": 8929 }, { "epoch": 4.044384057971015, "grad_norm": 0.050653744986592586, "learning_rate": 8.576364017396886e-06, "loss": 0.9122, "step": 8930 }, { "epoch": 4.044836956521739, "grad_norm": 0.045992034626243905, "learning_rate": 8.568540110341282e-06, "loss": 0.909, "step": 8931 }, { "epoch": 4.045289855072464, "grad_norm": 0.04495722608656552, "learning_rate": 8.560719345538535e-06, "loss": 0.9284, "step": 8932 }, { "epoch": 4.0457427536231885, "grad_norm": 0.04973075882685732, "learning_rate": 8.552901723770502e-06, "loss": 0.9259, "step": 8933 }, { "epoch": 4.046195652173913, "grad_norm": 0.050894677388741996, "learning_rate": 8.54508724581871e-06, "loss": 0.909, "step": 8934 }, { "epoch": 4.046648550724638, "grad_norm": 0.054282145451227594, "learning_rate": 8.537275912464383e-06, "loss": 0.9216, "step": 8935 }, { "epoch": 4.047101449275362, "grad_norm": 0.05465548841375788, "learning_rate": 8.529467724488452e-06, "loss": 0.9203, "step": 8936 }, { "epoch": 4.047554347826087, "grad_norm": 0.04776766579538778, "learning_rate": 8.52166268267149e-06, "loss": 0.9147, "step": 8937 }, { "epoch": 4.0480072463768115, "grad_norm": 0.056812937495058405, "learning_rate": 8.513860787793801e-06, "loss": 0.9168, "step": 8938 }, { "epoch": 4.048460144927536, "grad_norm": 0.04898418771518502, "learning_rate": 8.506062040635332e-06, "loss": 0.9057, "step": 8939 }, { "epoch": 4.048913043478261, "grad_norm": 0.06179794005320412, "learning_rate": 8.49826644197576e-06, "loss": 0.9322, "step": 8940 }, { "epoch": 4.049365942028985, "grad_norm": 0.05177235408600491, "learning_rate": 8.490473992594413e-06, "loss": 0.9383, "step": 8941 }, { "epoch": 4.04981884057971, "grad_norm": 0.05584639409659749, "learning_rate": 8.482684693270306e-06, "loss": 0.9283, "step": 8942 }, { "epoch": 4.050271739130435, "grad_norm": 0.054122232470730613, "learning_rate": 8.474898544782144e-06, "loss": 0.8975, "step": 8943 }, { "epoch": 4.050724637681159, "grad_norm": 0.05119958281567796, "learning_rate": 8.467115547908348e-06, "loss": 0.9019, "step": 8944 }, { "epoch": 4.051177536231884, "grad_norm": 0.0559190151606559, "learning_rate": 8.459335703426959e-06, "loss": 0.9283, "step": 8945 }, { "epoch": 4.051630434782608, "grad_norm": 0.05051935250481435, "learning_rate": 8.451559012115775e-06, "loss": 0.9327, "step": 8946 }, { "epoch": 4.052083333333333, "grad_norm": 0.05252008326087875, "learning_rate": 8.44378547475222e-06, "loss": 0.9341, "step": 8947 }, { "epoch": 4.052536231884058, "grad_norm": 0.059599561453158374, "learning_rate": 8.436015092113443e-06, "loss": 0.9237, "step": 8948 }, { "epoch": 4.052989130434782, "grad_norm": 0.058648277549994125, "learning_rate": 8.42824786497626e-06, "loss": 0.9275, "step": 8949 }, { "epoch": 4.053442028985507, "grad_norm": 0.07058699257974467, "learning_rate": 8.420483794117165e-06, "loss": 0.9376, "step": 8950 }, { "epoch": 4.0538949275362315, "grad_norm": 0.05271053104127945, "learning_rate": 8.412722880312331e-06, "loss": 0.9447, "step": 8951 }, { "epoch": 4.054347826086956, "grad_norm": 0.057911954853165144, "learning_rate": 8.404965124337665e-06, "loss": 0.9108, "step": 8952 }, { "epoch": 4.054800724637682, "grad_norm": 0.0702642294471535, "learning_rate": 8.397210526968687e-06, "loss": 0.9312, "step": 8953 }, { "epoch": 4.055253623188406, "grad_norm": 0.05247091409889462, "learning_rate": 8.38945908898066e-06, "loss": 0.9168, "step": 8954 }, { "epoch": 4.055706521739131, "grad_norm": 0.061975625256685, "learning_rate": 8.381710811148492e-06, "loss": 0.9356, "step": 8955 }, { "epoch": 4.056159420289855, "grad_norm": 0.07137121732341009, "learning_rate": 8.373965694246809e-06, "loss": 0.9262, "step": 8956 }, { "epoch": 4.05661231884058, "grad_norm": 0.057853243256714246, "learning_rate": 8.366223739049876e-06, "loss": 0.9082, "step": 8957 }, { "epoch": 4.057065217391305, "grad_norm": 0.06221149842523141, "learning_rate": 8.3584849463317e-06, "loss": 0.9194, "step": 8958 }, { "epoch": 4.057518115942029, "grad_norm": 0.05149232756091462, "learning_rate": 8.35074931686592e-06, "loss": 0.9187, "step": 8959 }, { "epoch": 4.057971014492754, "grad_norm": 0.06683241500735466, "learning_rate": 8.343016851425885e-06, "loss": 0.9294, "step": 8960 }, { "epoch": 4.0584239130434785, "grad_norm": 0.05157904012425983, "learning_rate": 8.33528755078461e-06, "loss": 0.9199, "step": 8961 }, { "epoch": 4.058876811594203, "grad_norm": 0.046939857355920156, "learning_rate": 8.327561415714824e-06, "loss": 0.9258, "step": 8962 }, { "epoch": 4.059329710144928, "grad_norm": 0.0661595655303797, "learning_rate": 8.319838446988902e-06, "loss": 0.9268, "step": 8963 }, { "epoch": 4.059782608695652, "grad_norm": 0.05083026986294648, "learning_rate": 8.312118645378943e-06, "loss": 0.9412, "step": 8964 }, { "epoch": 4.060235507246377, "grad_norm": 0.057648055384417696, "learning_rate": 8.30440201165669e-06, "loss": 0.9228, "step": 8965 }, { "epoch": 4.0606884057971016, "grad_norm": 0.06432984282249389, "learning_rate": 8.2966885465936e-06, "loss": 0.9327, "step": 8966 }, { "epoch": 4.061141304347826, "grad_norm": 0.04814764454763028, "learning_rate": 8.2889782509608e-06, "loss": 0.8945, "step": 8967 }, { "epoch": 4.061594202898551, "grad_norm": 0.05352761648668978, "learning_rate": 8.281271125529087e-06, "loss": 0.9165, "step": 8968 }, { "epoch": 4.062047101449275, "grad_norm": 0.04521643594983191, "learning_rate": 8.273567171068957e-06, "loss": 0.9313, "step": 8969 }, { "epoch": 4.0625, "grad_norm": 0.04680005146605212, "learning_rate": 8.265866388350598e-06, "loss": 0.9073, "step": 8970 }, { "epoch": 4.062952898550725, "grad_norm": 0.0485707345226338, "learning_rate": 8.258168778143858e-06, "loss": 0.913, "step": 8971 }, { "epoch": 4.063405797101449, "grad_norm": 0.04549099197948973, "learning_rate": 8.250474341218292e-06, "loss": 0.9188, "step": 8972 }, { "epoch": 4.063858695652174, "grad_norm": 0.048383898898778324, "learning_rate": 8.242783078343106e-06, "loss": 0.9259, "step": 8973 }, { "epoch": 4.0643115942028984, "grad_norm": 0.04481523012719575, "learning_rate": 8.235094990287234e-06, "loss": 0.9139, "step": 8974 }, { "epoch": 4.064764492753623, "grad_norm": 0.04399706971154914, "learning_rate": 8.227410077819247e-06, "loss": 0.9131, "step": 8975 }, { "epoch": 4.065217391304348, "grad_norm": 0.05406052000416312, "learning_rate": 8.219728341707425e-06, "loss": 0.9257, "step": 8976 }, { "epoch": 4.065670289855072, "grad_norm": 0.04622412166486723, "learning_rate": 8.212049782719708e-06, "loss": 0.9069, "step": 8977 }, { "epoch": 4.066123188405797, "grad_norm": 0.04596032232950234, "learning_rate": 8.20437440162376e-06, "loss": 0.9333, "step": 8978 }, { "epoch": 4.0665760869565215, "grad_norm": 0.05360697780450437, "learning_rate": 8.196702199186873e-06, "loss": 0.9328, "step": 8979 }, { "epoch": 4.067028985507246, "grad_norm": 0.04998556946325565, "learning_rate": 8.18903317617608e-06, "loss": 0.9341, "step": 8980 }, { "epoch": 4.067481884057971, "grad_norm": 0.04932942087917784, "learning_rate": 8.181367333358032e-06, "loss": 0.9059, "step": 8981 }, { "epoch": 4.067934782608695, "grad_norm": 0.049936186255880234, "learning_rate": 8.17370467149913e-06, "loss": 0.9262, "step": 8982 }, { "epoch": 4.06838768115942, "grad_norm": 0.05118256977370535, "learning_rate": 8.1660451913654e-06, "loss": 0.9057, "step": 8983 }, { "epoch": 4.068840579710145, "grad_norm": 0.046359896111423236, "learning_rate": 8.15838889372258e-06, "loss": 0.9232, "step": 8984 }, { "epoch": 4.069293478260869, "grad_norm": 0.05740907109390107, "learning_rate": 8.150735779336067e-06, "loss": 0.9114, "step": 8985 }, { "epoch": 4.069746376811594, "grad_norm": 0.04904055235284306, "learning_rate": 8.14308584897098e-06, "loss": 0.912, "step": 8986 }, { "epoch": 4.070199275362318, "grad_norm": 0.04805953219806928, "learning_rate": 8.135439103392074e-06, "loss": 0.9087, "step": 8987 }, { "epoch": 4.070652173913044, "grad_norm": 0.04903175577716336, "learning_rate": 8.127795543363821e-06, "loss": 0.9412, "step": 8988 }, { "epoch": 4.0711050724637685, "grad_norm": 0.04415034133233376, "learning_rate": 8.120155169650346e-06, "loss": 0.9102, "step": 8989 }, { "epoch": 4.071557971014493, "grad_norm": 0.05436185053351668, "learning_rate": 8.112517983015485e-06, "loss": 0.9231, "step": 8990 }, { "epoch": 4.072010869565218, "grad_norm": 0.05140705440116463, "learning_rate": 8.10488398422273e-06, "loss": 0.9082, "step": 8991 }, { "epoch": 4.072463768115942, "grad_norm": 0.04941789235872437, "learning_rate": 8.097253174035268e-06, "loss": 0.9451, "step": 8992 }, { "epoch": 4.072916666666667, "grad_norm": 0.05319809152805272, "learning_rate": 8.089625553215947e-06, "loss": 0.9101, "step": 8993 }, { "epoch": 4.073369565217392, "grad_norm": 0.046175797165821646, "learning_rate": 8.08200112252734e-06, "loss": 0.9153, "step": 8994 }, { "epoch": 4.073822463768116, "grad_norm": 0.05842021636740336, "learning_rate": 8.074379882731645e-06, "loss": 0.9062, "step": 8995 }, { "epoch": 4.074275362318841, "grad_norm": 0.04542615214826501, "learning_rate": 8.066761834590794e-06, "loss": 0.926, "step": 8996 }, { "epoch": 4.074728260869565, "grad_norm": 0.04803631081458058, "learning_rate": 8.059146978866358e-06, "loss": 0.9264, "step": 8997 }, { "epoch": 4.07518115942029, "grad_norm": 0.049282965536774394, "learning_rate": 8.051535316319622e-06, "loss": 0.931, "step": 8998 }, { "epoch": 4.075634057971015, "grad_norm": 0.047439711709869554, "learning_rate": 8.043926847711518e-06, "loss": 0.925, "step": 8999 }, { "epoch": 4.076086956521739, "grad_norm": 0.04786615828080371, "learning_rate": 8.036321573802706e-06, "loss": 0.935, "step": 9000 }, { "epoch": 4.076539855072464, "grad_norm": 0.051088533278667025, "learning_rate": 8.028719495353457e-06, "loss": 0.9279, "step": 9001 }, { "epoch": 4.0769927536231885, "grad_norm": 0.04913606798445405, "learning_rate": 8.021120613123789e-06, "loss": 0.9155, "step": 9002 }, { "epoch": 4.077445652173913, "grad_norm": 0.04404998931275893, "learning_rate": 8.01352492787336e-06, "loss": 0.9292, "step": 9003 }, { "epoch": 4.077898550724638, "grad_norm": 0.0471755439926039, "learning_rate": 8.005932440361541e-06, "loss": 0.9315, "step": 9004 }, { "epoch": 4.078351449275362, "grad_norm": 0.04793075973075758, "learning_rate": 7.998343151347346e-06, "loss": 0.9212, "step": 9005 }, { "epoch": 4.078804347826087, "grad_norm": 0.044151415928908866, "learning_rate": 7.990757061589508e-06, "loss": 0.9262, "step": 9006 }, { "epoch": 4.0792572463768115, "grad_norm": 0.053026688324388685, "learning_rate": 7.983174171846397e-06, "loss": 0.9, "step": 9007 }, { "epoch": 4.079710144927536, "grad_norm": 0.044711421184707244, "learning_rate": 7.975594482876113e-06, "loss": 0.9172, "step": 9008 }, { "epoch": 4.080163043478261, "grad_norm": 0.045272419711416934, "learning_rate": 7.968017995436393e-06, "loss": 0.908, "step": 9009 }, { "epoch": 4.080615942028985, "grad_norm": 0.04634154653113212, "learning_rate": 7.960444710284671e-06, "loss": 0.9317, "step": 9010 }, { "epoch": 4.08106884057971, "grad_norm": 0.05176489531876029, "learning_rate": 7.952874628178055e-06, "loss": 0.9352, "step": 9011 }, { "epoch": 4.081521739130435, "grad_norm": 0.043220653593806994, "learning_rate": 7.945307749873356e-06, "loss": 0.9078, "step": 9012 }, { "epoch": 4.081974637681159, "grad_norm": 0.04713599537915565, "learning_rate": 7.937744076127022e-06, "loss": 0.9306, "step": 9013 }, { "epoch": 4.082427536231884, "grad_norm": 0.047086802418765894, "learning_rate": 7.930183607695237e-06, "loss": 0.9269, "step": 9014 }, { "epoch": 4.082880434782608, "grad_norm": 0.05222484532621395, "learning_rate": 7.922626345333797e-06, "loss": 0.9346, "step": 9015 }, { "epoch": 4.083333333333333, "grad_norm": 0.04397878900850762, "learning_rate": 7.915072289798247e-06, "loss": 0.9211, "step": 9016 }, { "epoch": 4.083786231884058, "grad_norm": 0.0525553226774777, "learning_rate": 7.907521441843764e-06, "loss": 0.9299, "step": 9017 }, { "epoch": 4.084239130434782, "grad_norm": 0.04861709325508121, "learning_rate": 7.899973802225216e-06, "loss": 0.9079, "step": 9018 }, { "epoch": 4.084692028985507, "grad_norm": 0.04340901403361405, "learning_rate": 7.892429371697146e-06, "loss": 0.9364, "step": 9019 }, { "epoch": 4.0851449275362315, "grad_norm": 0.05068995341729933, "learning_rate": 7.8848881510138e-06, "loss": 0.9212, "step": 9020 }, { "epoch": 4.085597826086956, "grad_norm": 0.04618644164868628, "learning_rate": 7.87735014092906e-06, "loss": 0.9312, "step": 9021 }, { "epoch": 4.086050724637682, "grad_norm": 0.047216811279862776, "learning_rate": 7.869815342196547e-06, "loss": 0.9085, "step": 9022 }, { "epoch": 4.086503623188406, "grad_norm": 0.04714095210601361, "learning_rate": 7.862283755569496e-06, "loss": 0.9179, "step": 9023 }, { "epoch": 4.086956521739131, "grad_norm": 0.046953131218174245, "learning_rate": 7.854755381800872e-06, "loss": 0.9112, "step": 9024 }, { "epoch": 4.087409420289855, "grad_norm": 0.04822247126358976, "learning_rate": 7.847230221643296e-06, "loss": 0.9083, "step": 9025 }, { "epoch": 4.08786231884058, "grad_norm": 0.04759610640687854, "learning_rate": 7.83970827584906e-06, "loss": 0.9337, "step": 9026 }, { "epoch": 4.088315217391305, "grad_norm": 0.047609765982826395, "learning_rate": 7.832189545170142e-06, "loss": 0.9059, "step": 9027 }, { "epoch": 4.088768115942029, "grad_norm": 0.04948623744124742, "learning_rate": 7.824674030358222e-06, "loss": 0.9159, "step": 9028 }, { "epoch": 4.089221014492754, "grad_norm": 0.05103222304282763, "learning_rate": 7.81716173216461e-06, "loss": 0.9094, "step": 9029 }, { "epoch": 4.0896739130434785, "grad_norm": 0.04828823638137761, "learning_rate": 7.809652651340353e-06, "loss": 0.9167, "step": 9030 }, { "epoch": 4.090126811594203, "grad_norm": 0.04492799642844877, "learning_rate": 7.802146788636121e-06, "loss": 0.908, "step": 9031 }, { "epoch": 4.090579710144928, "grad_norm": 0.04952315032398976, "learning_rate": 7.794644144802302e-06, "loss": 0.9285, "step": 9032 }, { "epoch": 4.091032608695652, "grad_norm": 0.052681281430338, "learning_rate": 7.787144720588942e-06, "loss": 0.9204, "step": 9033 }, { "epoch": 4.091485507246377, "grad_norm": 0.04775546533275965, "learning_rate": 7.779648516745775e-06, "loss": 0.9361, "step": 9034 }, { "epoch": 4.0919384057971016, "grad_norm": 0.05599027481525119, "learning_rate": 7.772155534022188e-06, "loss": 0.9264, "step": 9035 }, { "epoch": 4.092391304347826, "grad_norm": 0.04944613390134746, "learning_rate": 7.764665773167293e-06, "loss": 0.9242, "step": 9036 }, { "epoch": 4.092844202898551, "grad_norm": 0.05386346879902592, "learning_rate": 7.757179234929833e-06, "loss": 0.9179, "step": 9037 }, { "epoch": 4.093297101449275, "grad_norm": 0.05027363483373286, "learning_rate": 7.749695920058266e-06, "loss": 0.9238, "step": 9038 }, { "epoch": 4.09375, "grad_norm": 0.06997794283480957, "learning_rate": 7.742215829300695e-06, "loss": 0.918, "step": 9039 }, { "epoch": 4.094202898550725, "grad_norm": 0.061557939577371656, "learning_rate": 7.734738963404932e-06, "loss": 0.9241, "step": 9040 }, { "epoch": 4.094655797101449, "grad_norm": 0.04693409038233397, "learning_rate": 7.727265323118445e-06, "loss": 0.9185, "step": 9041 }, { "epoch": 4.095108695652174, "grad_norm": 0.048246657807522485, "learning_rate": 7.719794909188385e-06, "loss": 0.9224, "step": 9042 }, { "epoch": 4.0955615942028984, "grad_norm": 0.04912298544231887, "learning_rate": 7.712327722361567e-06, "loss": 0.9001, "step": 9043 }, { "epoch": 4.096014492753623, "grad_norm": 0.04435475565576238, "learning_rate": 7.704863763384526e-06, "loss": 0.9346, "step": 9044 }, { "epoch": 4.096467391304348, "grad_norm": 0.05051851842607345, "learning_rate": 7.697403033003419e-06, "loss": 0.9395, "step": 9045 }, { "epoch": 4.096920289855072, "grad_norm": 0.04860404791943612, "learning_rate": 7.689945531964125e-06, "loss": 0.9053, "step": 9046 }, { "epoch": 4.097373188405797, "grad_norm": 0.04619491353715604, "learning_rate": 7.682491261012171e-06, "loss": 0.9133, "step": 9047 }, { "epoch": 4.0978260869565215, "grad_norm": 0.043514375522444915, "learning_rate": 7.675040220892782e-06, "loss": 0.9115, "step": 9048 }, { "epoch": 4.098278985507246, "grad_norm": 0.046885563507732965, "learning_rate": 7.667592412350839e-06, "loss": 0.9168, "step": 9049 }, { "epoch": 4.098731884057971, "grad_norm": 0.04902684585897958, "learning_rate": 7.660147836130928e-06, "loss": 0.9292, "step": 9050 }, { "epoch": 4.099184782608695, "grad_norm": 0.04729543481296953, "learning_rate": 7.652706492977282e-06, "loss": 0.9122, "step": 9051 }, { "epoch": 4.09963768115942, "grad_norm": 0.05039656486337042, "learning_rate": 7.645268383633828e-06, "loss": 0.9289, "step": 9052 }, { "epoch": 4.100090579710145, "grad_norm": 0.04842753158708562, "learning_rate": 7.637833508844164e-06, "loss": 0.9279, "step": 9053 }, { "epoch": 4.100543478260869, "grad_norm": 0.04923506084855102, "learning_rate": 7.630401869351556e-06, "loss": 0.9366, "step": 9054 }, { "epoch": 4.100996376811594, "grad_norm": 0.048628553792265224, "learning_rate": 7.622973465898979e-06, "loss": 0.9312, "step": 9055 }, { "epoch": 4.101449275362318, "grad_norm": 0.04405718933628444, "learning_rate": 7.615548299229036e-06, "loss": 0.9106, "step": 9056 }, { "epoch": 4.101902173913044, "grad_norm": 0.04393962074769586, "learning_rate": 7.608126370084061e-06, "loss": 0.916, "step": 9057 }, { "epoch": 4.1023550724637685, "grad_norm": 0.050402508264996156, "learning_rate": 7.600707679206012e-06, "loss": 0.9289, "step": 9058 }, { "epoch": 4.102807971014493, "grad_norm": 0.0464562312702112, "learning_rate": 7.5932922273365665e-06, "loss": 0.9125, "step": 9059 }, { "epoch": 4.103260869565218, "grad_norm": 0.048278710408772366, "learning_rate": 7.585880015217051e-06, "loss": 0.9159, "step": 9060 }, { "epoch": 4.103713768115942, "grad_norm": 0.04875354745578297, "learning_rate": 7.578471043588473e-06, "loss": 0.9056, "step": 9061 }, { "epoch": 4.104166666666667, "grad_norm": 0.046221195049965816, "learning_rate": 7.5710653131915125e-06, "loss": 0.9303, "step": 9062 }, { "epoch": 4.104619565217392, "grad_norm": 0.04673999953346028, "learning_rate": 7.563662824766549e-06, "loss": 0.9149, "step": 9063 }, { "epoch": 4.105072463768116, "grad_norm": 0.04195589460489588, "learning_rate": 7.556263579053605e-06, "loss": 0.905, "step": 9064 }, { "epoch": 4.105525362318841, "grad_norm": 0.04445680950877715, "learning_rate": 7.548867576792411e-06, "loss": 0.9197, "step": 9065 }, { "epoch": 4.105978260869565, "grad_norm": 0.05519037805777705, "learning_rate": 7.541474818722338e-06, "loss": 0.9289, "step": 9066 }, { "epoch": 4.10643115942029, "grad_norm": 0.04749892063085932, "learning_rate": 7.534085305582475e-06, "loss": 0.9198, "step": 9067 }, { "epoch": 4.106884057971015, "grad_norm": 0.045739921363694744, "learning_rate": 7.526699038111549e-06, "loss": 0.9368, "step": 9068 }, { "epoch": 4.107336956521739, "grad_norm": 0.04943889329395335, "learning_rate": 7.5193160170479795e-06, "loss": 0.9109, "step": 9069 }, { "epoch": 4.107789855072464, "grad_norm": 0.044800785437136025, "learning_rate": 7.511936243129847e-06, "loss": 0.9286, "step": 9070 }, { "epoch": 4.1082427536231885, "grad_norm": 0.04853350172842268, "learning_rate": 7.504559717094939e-06, "loss": 0.9328, "step": 9071 }, { "epoch": 4.108695652173913, "grad_norm": 0.05190165574948598, "learning_rate": 7.4971864396806795e-06, "loss": 0.9229, "step": 9072 }, { "epoch": 4.109148550724638, "grad_norm": 0.04767814806190346, "learning_rate": 7.489816411624207e-06, "loss": 0.9057, "step": 9073 }, { "epoch": 4.109601449275362, "grad_norm": 0.04836730478434141, "learning_rate": 7.482449633662296e-06, "loss": 0.9061, "step": 9074 }, { "epoch": 4.110054347826087, "grad_norm": 0.05301850908717192, "learning_rate": 7.475086106531431e-06, "loss": 0.9145, "step": 9075 }, { "epoch": 4.1105072463768115, "grad_norm": 0.04168788666936198, "learning_rate": 7.467725830967753e-06, "loss": 0.9046, "step": 9076 }, { "epoch": 4.110960144927536, "grad_norm": 0.04647000161968196, "learning_rate": 7.460368807707072e-06, "loss": 0.9181, "step": 9077 }, { "epoch": 4.111413043478261, "grad_norm": 0.0471514635572182, "learning_rate": 7.4530150374848745e-06, "loss": 0.9122, "step": 9078 }, { "epoch": 4.111865942028985, "grad_norm": 0.04281431665033956, "learning_rate": 7.445664521036349e-06, "loss": 0.9159, "step": 9079 }, { "epoch": 4.11231884057971, "grad_norm": 0.043414101863579774, "learning_rate": 7.438317259096317e-06, "loss": 0.9103, "step": 9080 }, { "epoch": 4.112771739130435, "grad_norm": 0.04753790580652879, "learning_rate": 7.430973252399321e-06, "loss": 0.9334, "step": 9081 }, { "epoch": 4.113224637681159, "grad_norm": 0.045011573433674075, "learning_rate": 7.4236325016795275e-06, "loss": 0.9211, "step": 9082 }, { "epoch": 4.113677536231884, "grad_norm": 0.04670101413846941, "learning_rate": 7.416295007670822e-06, "loss": 0.9132, "step": 9083 }, { "epoch": 4.114130434782608, "grad_norm": 0.047896276615777285, "learning_rate": 7.408960771106741e-06, "loss": 0.917, "step": 9084 }, { "epoch": 4.114583333333333, "grad_norm": 0.0455918058211403, "learning_rate": 7.401629792720495e-06, "loss": 0.9295, "step": 9085 }, { "epoch": 4.115036231884058, "grad_norm": 0.04855699988944298, "learning_rate": 7.3943020732449635e-06, "loss": 0.9116, "step": 9086 }, { "epoch": 4.115489130434782, "grad_norm": 0.04648834310688952, "learning_rate": 7.3869776134127335e-06, "loss": 0.9274, "step": 9087 }, { "epoch": 4.115942028985507, "grad_norm": 0.04486751054199097, "learning_rate": 7.37965641395602e-06, "loss": 0.9197, "step": 9088 }, { "epoch": 4.1163949275362315, "grad_norm": 0.056058100710910296, "learning_rate": 7.372338475606757e-06, "loss": 0.9398, "step": 9089 }, { "epoch": 4.116847826086956, "grad_norm": 0.04923428402249835, "learning_rate": 7.365023799096511e-06, "loss": 0.9279, "step": 9090 }, { "epoch": 4.117300724637682, "grad_norm": 0.04827973987088484, "learning_rate": 7.357712385156559e-06, "loss": 0.917, "step": 9091 }, { "epoch": 4.117753623188406, "grad_norm": 0.04841665998259195, "learning_rate": 7.350404234517818e-06, "loss": 0.9284, "step": 9092 }, { "epoch": 4.118206521739131, "grad_norm": 0.05156391336650868, "learning_rate": 7.343099347910914e-06, "loss": 0.9032, "step": 9093 }, { "epoch": 4.118659420289855, "grad_norm": 0.05002769470749493, "learning_rate": 7.335797726066114e-06, "loss": 0.9329, "step": 9094 }, { "epoch": 4.11911231884058, "grad_norm": 0.04824115758490432, "learning_rate": 7.328499369713378e-06, "loss": 0.9311, "step": 9095 }, { "epoch": 4.119565217391305, "grad_norm": 0.049138149821439235, "learning_rate": 7.321204279582321e-06, "loss": 0.9011, "step": 9096 }, { "epoch": 4.120018115942029, "grad_norm": 0.05021596508513306, "learning_rate": 7.313912456402268e-06, "loss": 0.9368, "step": 9097 }, { "epoch": 4.120471014492754, "grad_norm": 0.04461362133477805, "learning_rate": 7.306623900902173e-06, "loss": 0.9305, "step": 9098 }, { "epoch": 4.1209239130434785, "grad_norm": 0.04934428960236105, "learning_rate": 7.2993386138107e-06, "loss": 0.9444, "step": 9099 }, { "epoch": 4.121376811594203, "grad_norm": 0.047703237008791644, "learning_rate": 7.292056595856158e-06, "loss": 0.9309, "step": 9100 }, { "epoch": 4.121829710144928, "grad_norm": 0.049459761138776176, "learning_rate": 7.284777847766555e-06, "loss": 0.9145, "step": 9101 }, { "epoch": 4.122282608695652, "grad_norm": 0.046892789208286934, "learning_rate": 7.277502370269553e-06, "loss": 0.9117, "step": 9102 }, { "epoch": 4.122735507246377, "grad_norm": 0.042849349152804965, "learning_rate": 7.270230164092495e-06, "loss": 0.9131, "step": 9103 }, { "epoch": 4.1231884057971016, "grad_norm": 0.045765311709925455, "learning_rate": 7.262961229962382e-06, "loss": 0.909, "step": 9104 }, { "epoch": 4.123641304347826, "grad_norm": 0.04784835777324678, "learning_rate": 7.255695568605921e-06, "loss": 0.9069, "step": 9105 }, { "epoch": 4.124094202898551, "grad_norm": 0.04881180760327453, "learning_rate": 7.24843318074945e-06, "loss": 0.9334, "step": 9106 }, { "epoch": 4.124547101449275, "grad_norm": 0.04705868035481858, "learning_rate": 7.241174067119021e-06, "loss": 0.9281, "step": 9107 }, { "epoch": 4.125, "grad_norm": 0.04183022519477614, "learning_rate": 7.233918228440324e-06, "loss": 0.9199, "step": 9108 }, { "epoch": 4.125452898550725, "grad_norm": 0.045423092150428446, "learning_rate": 7.226665665438757e-06, "loss": 0.9298, "step": 9109 }, { "epoch": 4.125905797101449, "grad_norm": 0.048421937609726254, "learning_rate": 7.219416378839353e-06, "loss": 0.9119, "step": 9110 }, { "epoch": 4.126358695652174, "grad_norm": 0.045996543335509066, "learning_rate": 7.212170369366842e-06, "loss": 0.9149, "step": 9111 }, { "epoch": 4.1268115942028984, "grad_norm": 0.04347351013178872, "learning_rate": 7.204927637745602e-06, "loss": 0.9184, "step": 9112 }, { "epoch": 4.127264492753623, "grad_norm": 0.04570539190000612, "learning_rate": 7.197688184699729e-06, "loss": 0.9065, "step": 9113 }, { "epoch": 4.127717391304348, "grad_norm": 0.05088311386543294, "learning_rate": 7.190452010952933e-06, "loss": 0.9075, "step": 9114 }, { "epoch": 4.128170289855072, "grad_norm": 0.04936628780648191, "learning_rate": 7.183219117228652e-06, "loss": 0.9485, "step": 9115 }, { "epoch": 4.128623188405797, "grad_norm": 0.043824541172179245, "learning_rate": 7.175989504249954e-06, "loss": 0.9234, "step": 9116 }, { "epoch": 4.1290760869565215, "grad_norm": 0.0467405467119152, "learning_rate": 7.168763172739601e-06, "loss": 0.9074, "step": 9117 }, { "epoch": 4.129528985507246, "grad_norm": 0.045326405721668385, "learning_rate": 7.161540123420026e-06, "loss": 0.8983, "step": 9118 }, { "epoch": 4.129981884057971, "grad_norm": 0.04729808875630408, "learning_rate": 7.15432035701332e-06, "loss": 0.9299, "step": 9119 }, { "epoch": 4.130434782608695, "grad_norm": 0.04391791444523515, "learning_rate": 7.147103874241246e-06, "loss": 0.9129, "step": 9120 }, { "epoch": 4.13088768115942, "grad_norm": 0.04591547933585074, "learning_rate": 7.139890675825265e-06, "loss": 0.9195, "step": 9121 }, { "epoch": 4.131340579710145, "grad_norm": 0.04743776503320371, "learning_rate": 7.132680762486477e-06, "loss": 0.9225, "step": 9122 }, { "epoch": 4.131793478260869, "grad_norm": 0.04355278971365922, "learning_rate": 7.125474134945682e-06, "loss": 0.9108, "step": 9123 }, { "epoch": 4.132246376811594, "grad_norm": 0.04649309999008047, "learning_rate": 7.118270793923327e-06, "loss": 0.9241, "step": 9124 }, { "epoch": 4.132699275362318, "grad_norm": 0.05299593253968106, "learning_rate": 7.111070740139556e-06, "loss": 0.9297, "step": 9125 }, { "epoch": 4.133152173913044, "grad_norm": 0.046548088010444695, "learning_rate": 7.1038739743141574e-06, "loss": 0.9413, "step": 9126 }, { "epoch": 4.1336050724637685, "grad_norm": 0.044922948584336465, "learning_rate": 7.096680497166608e-06, "loss": 0.9337, "step": 9127 }, { "epoch": 4.134057971014493, "grad_norm": 0.04560505141349129, "learning_rate": 7.089490309416036e-06, "loss": 0.9244, "step": 9128 }, { "epoch": 4.134510869565218, "grad_norm": 0.04705722142538242, "learning_rate": 7.082303411781284e-06, "loss": 0.9235, "step": 9129 }, { "epoch": 4.134963768115942, "grad_norm": 0.04966207987852858, "learning_rate": 7.075119804980808e-06, "loss": 0.9153, "step": 9130 }, { "epoch": 4.135416666666667, "grad_norm": 0.050120323152460206, "learning_rate": 7.067939489732794e-06, "loss": 0.9477, "step": 9131 }, { "epoch": 4.135869565217392, "grad_norm": 0.05112461249057509, "learning_rate": 7.06076246675504e-06, "loss": 0.9343, "step": 9132 }, { "epoch": 4.136322463768116, "grad_norm": 0.04839906301985513, "learning_rate": 7.053588736765067e-06, "loss": 0.8986, "step": 9133 }, { "epoch": 4.136775362318841, "grad_norm": 0.05360673908591168, "learning_rate": 7.046418300480029e-06, "loss": 0.9338, "step": 9134 }, { "epoch": 4.137228260869565, "grad_norm": 0.046075152223846956, "learning_rate": 7.0392511586167846e-06, "loss": 0.9252, "step": 9135 }, { "epoch": 4.13768115942029, "grad_norm": 0.046100372380921587, "learning_rate": 7.032087311891831e-06, "loss": 0.9222, "step": 9136 }, { "epoch": 4.138134057971015, "grad_norm": 0.05114043588679653, "learning_rate": 7.0249267610213514e-06, "loss": 0.9285, "step": 9137 }, { "epoch": 4.138586956521739, "grad_norm": 0.048041651215243814, "learning_rate": 7.017769506721186e-06, "loss": 0.9484, "step": 9138 }, { "epoch": 4.139039855072464, "grad_norm": 0.05140561456908414, "learning_rate": 7.0106155497068785e-06, "loss": 0.9127, "step": 9139 }, { "epoch": 4.1394927536231885, "grad_norm": 0.050395504665583304, "learning_rate": 7.003464890693603e-06, "loss": 0.955, "step": 9140 }, { "epoch": 4.139945652173913, "grad_norm": 0.0499613145588801, "learning_rate": 6.996317530396237e-06, "loss": 0.9165, "step": 9141 }, { "epoch": 4.140398550724638, "grad_norm": 0.047712451645442425, "learning_rate": 6.9891734695292975e-06, "loss": 0.9278, "step": 9142 }, { "epoch": 4.140851449275362, "grad_norm": 0.047822582313277365, "learning_rate": 6.982032708807005e-06, "loss": 0.9493, "step": 9143 }, { "epoch": 4.141304347826087, "grad_norm": 0.046216595382872705, "learning_rate": 6.974895248943223e-06, "loss": 0.8916, "step": 9144 }, { "epoch": 4.1417572463768115, "grad_norm": 0.04374956980496524, "learning_rate": 6.967761090651501e-06, "loss": 0.9159, "step": 9145 }, { "epoch": 4.142210144927536, "grad_norm": 0.04553168998449846, "learning_rate": 6.9606302346450335e-06, "loss": 0.9067, "step": 9146 }, { "epoch": 4.142663043478261, "grad_norm": 0.04429952191583516, "learning_rate": 6.953502681636726e-06, "loss": 0.9191, "step": 9147 }, { "epoch": 4.143115942028985, "grad_norm": 0.043495172967582374, "learning_rate": 6.946378432339114e-06, "loss": 0.9145, "step": 9148 }, { "epoch": 4.14356884057971, "grad_norm": 0.047744541342345584, "learning_rate": 6.939257487464437e-06, "loss": 0.9215, "step": 9149 }, { "epoch": 4.144021739130435, "grad_norm": 0.04406931627355592, "learning_rate": 6.932139847724566e-06, "loss": 0.9135, "step": 9150 }, { "epoch": 4.144474637681159, "grad_norm": 0.04753264117240782, "learning_rate": 6.925025513831087e-06, "loss": 0.916, "step": 9151 }, { "epoch": 4.144927536231884, "grad_norm": 0.04741597163544108, "learning_rate": 6.917914486495218e-06, "loss": 0.9085, "step": 9152 }, { "epoch": 4.145380434782608, "grad_norm": 0.04371509459026672, "learning_rate": 6.910806766427862e-06, "loss": 0.9225, "step": 9153 }, { "epoch": 4.145833333333333, "grad_norm": 0.0476220956142299, "learning_rate": 6.903702354339578e-06, "loss": 0.9211, "step": 9154 }, { "epoch": 4.146286231884058, "grad_norm": 0.045681305781454046, "learning_rate": 6.896601250940627e-06, "loss": 0.9138, "step": 9155 }, { "epoch": 4.146739130434782, "grad_norm": 0.043666557046422805, "learning_rate": 6.8895034569408915e-06, "loss": 0.8919, "step": 9156 }, { "epoch": 4.147192028985507, "grad_norm": 0.04331795504175423, "learning_rate": 6.882408973049979e-06, "loss": 0.9055, "step": 9157 }, { "epoch": 4.1476449275362315, "grad_norm": 0.04308041566923009, "learning_rate": 6.875317799977108e-06, "loss": 0.9278, "step": 9158 }, { "epoch": 4.148097826086956, "grad_norm": 0.045815439353294074, "learning_rate": 6.868229938431215e-06, "loss": 0.9243, "step": 9159 }, { "epoch": 4.148550724637682, "grad_norm": 0.0424835824033548, "learning_rate": 6.861145389120878e-06, "loss": 0.9285, "step": 9160 }, { "epoch": 4.149003623188406, "grad_norm": 0.0467403866229328, "learning_rate": 6.8540641527543495e-06, "loss": 0.9223, "step": 9161 }, { "epoch": 4.149456521739131, "grad_norm": 0.045015586689728035, "learning_rate": 6.846986230039543e-06, "loss": 0.9115, "step": 9162 }, { "epoch": 4.149909420289855, "grad_norm": 0.048038314346425665, "learning_rate": 6.83991162168407e-06, "loss": 0.9097, "step": 9163 }, { "epoch": 4.15036231884058, "grad_norm": 0.049834340072569726, "learning_rate": 6.832840328395165e-06, "loss": 0.9299, "step": 9164 }, { "epoch": 4.150815217391305, "grad_norm": 0.046372514511161235, "learning_rate": 6.8257723508797824e-06, "loss": 0.8972, "step": 9165 }, { "epoch": 4.151268115942029, "grad_norm": 0.05289022274735866, "learning_rate": 6.818707689844499e-06, "loss": 0.8985, "step": 9166 }, { "epoch": 4.151721014492754, "grad_norm": 0.046580233374344346, "learning_rate": 6.811646345995595e-06, "loss": 0.8949, "step": 9167 }, { "epoch": 4.1521739130434785, "grad_norm": 0.04978781992826108, "learning_rate": 6.804588320038998e-06, "loss": 0.9263, "step": 9168 }, { "epoch": 4.152626811594203, "grad_norm": 0.04894638775839725, "learning_rate": 6.797533612680309e-06, "loss": 0.9506, "step": 9169 }, { "epoch": 4.153079710144928, "grad_norm": 0.04710565739544265, "learning_rate": 6.7904822246247905e-06, "loss": 0.928, "step": 9170 }, { "epoch": 4.153532608695652, "grad_norm": 0.046522334923882844, "learning_rate": 6.783434156577398e-06, "loss": 0.9145, "step": 9171 }, { "epoch": 4.153985507246377, "grad_norm": 0.05037507214910648, "learning_rate": 6.776389409242718e-06, "loss": 0.9268, "step": 9172 }, { "epoch": 4.1544384057971016, "grad_norm": 0.04866551732272857, "learning_rate": 6.769347983325043e-06, "loss": 0.9384, "step": 9173 }, { "epoch": 4.154891304347826, "grad_norm": 0.042475827598386515, "learning_rate": 6.762309879528306e-06, "loss": 0.906, "step": 9174 }, { "epoch": 4.155344202898551, "grad_norm": 0.047381558612406866, "learning_rate": 6.755275098556127e-06, "loss": 0.9123, "step": 9175 }, { "epoch": 4.155797101449275, "grad_norm": 0.045894072900915704, "learning_rate": 6.7482436411117646e-06, "loss": 0.9083, "step": 9176 }, { "epoch": 4.15625, "grad_norm": 0.04292685142082182, "learning_rate": 6.7412155078981865e-06, "loss": 0.9202, "step": 9177 }, { "epoch": 4.156702898550725, "grad_norm": 0.045331484991103815, "learning_rate": 6.734190699618013e-06, "loss": 0.9219, "step": 9178 }, { "epoch": 4.157155797101449, "grad_norm": 0.04954067390960296, "learning_rate": 6.7271692169734855e-06, "loss": 0.916, "step": 9179 }, { "epoch": 4.157608695652174, "grad_norm": 0.046444595361296075, "learning_rate": 6.7201510606665865e-06, "loss": 0.9234, "step": 9180 }, { "epoch": 4.1580615942028984, "grad_norm": 0.04427186452689828, "learning_rate": 6.7131362313989135e-06, "loss": 0.9072, "step": 9181 }, { "epoch": 4.158514492753623, "grad_norm": 0.053113874863054036, "learning_rate": 6.706124729871768e-06, "loss": 0.8965, "step": 9182 }, { "epoch": 4.158967391304348, "grad_norm": 0.04462970892394902, "learning_rate": 6.6991165567860824e-06, "loss": 0.9263, "step": 9183 }, { "epoch": 4.159420289855072, "grad_norm": 0.0433310856502183, "learning_rate": 6.692111712842497e-06, "loss": 0.9249, "step": 9184 }, { "epoch": 4.159873188405797, "grad_norm": 0.048271555022017465, "learning_rate": 6.6851101987412695e-06, "loss": 0.9294, "step": 9185 }, { "epoch": 4.1603260869565215, "grad_norm": 0.04739344585277467, "learning_rate": 6.678112015182381e-06, "loss": 0.9318, "step": 9186 }, { "epoch": 4.160778985507246, "grad_norm": 0.050726457013325135, "learning_rate": 6.671117162865437e-06, "loss": 0.9175, "step": 9187 }, { "epoch": 4.161231884057971, "grad_norm": 0.042623764705300764, "learning_rate": 6.664125642489723e-06, "loss": 0.9171, "step": 9188 }, { "epoch": 4.161684782608695, "grad_norm": 0.05173257600860757, "learning_rate": 6.657137454754185e-06, "loss": 0.908, "step": 9189 }, { "epoch": 4.16213768115942, "grad_norm": 0.050903141396574564, "learning_rate": 6.650152600357463e-06, "loss": 0.9239, "step": 9190 }, { "epoch": 4.162590579710145, "grad_norm": 0.04390276353344801, "learning_rate": 6.643171079997821e-06, "loss": 0.9259, "step": 9191 }, { "epoch": 4.163043478260869, "grad_norm": 0.04759851030693755, "learning_rate": 6.636192894373232e-06, "loss": 0.9451, "step": 9192 }, { "epoch": 4.163496376811594, "grad_norm": 0.043975133680641494, "learning_rate": 6.629218044181303e-06, "loss": 0.9129, "step": 9193 }, { "epoch": 4.163949275362318, "grad_norm": 0.041624436982362935, "learning_rate": 6.622246530119336e-06, "loss": 0.9223, "step": 9194 }, { "epoch": 4.164402173913044, "grad_norm": 0.043558840517821565, "learning_rate": 6.615278352884274e-06, "loss": 0.9335, "step": 9195 }, { "epoch": 4.1648550724637685, "grad_norm": 0.046443831185338395, "learning_rate": 6.60831351317274e-06, "loss": 0.9164, "step": 9196 }, { "epoch": 4.165307971014493, "grad_norm": 0.04329129858312766, "learning_rate": 6.601352011681004e-06, "loss": 0.922, "step": 9197 }, { "epoch": 4.165760869565218, "grad_norm": 0.044259211468293716, "learning_rate": 6.594393849105043e-06, "loss": 0.9328, "step": 9198 }, { "epoch": 4.166213768115942, "grad_norm": 0.0430946017111116, "learning_rate": 6.587439026140456e-06, "loss": 0.9347, "step": 9199 }, { "epoch": 4.166666666666667, "grad_norm": 0.05011805782310666, "learning_rate": 6.58048754348255e-06, "loss": 0.9155, "step": 9200 }, { "epoch": 4.167119565217392, "grad_norm": 0.04423865154753742, "learning_rate": 6.5735394018262475e-06, "loss": 0.9283, "step": 9201 }, { "epoch": 4.167572463768116, "grad_norm": 0.04493618071633499, "learning_rate": 6.56659460186619e-06, "loss": 0.9195, "step": 9202 }, { "epoch": 4.168025362318841, "grad_norm": 0.04633072879724077, "learning_rate": 6.559653144296647e-06, "loss": 0.8964, "step": 9203 }, { "epoch": 4.168478260869565, "grad_norm": 0.044673208603626884, "learning_rate": 6.552715029811576e-06, "loss": 0.921, "step": 9204 }, { "epoch": 4.16893115942029, "grad_norm": 0.05021122523197682, "learning_rate": 6.545780259104573e-06, "loss": 0.9102, "step": 9205 }, { "epoch": 4.169384057971015, "grad_norm": 0.04681507183307578, "learning_rate": 6.538848832868936e-06, "loss": 0.932, "step": 9206 }, { "epoch": 4.169836956521739, "grad_norm": 0.053116545504054924, "learning_rate": 6.531920751797601e-06, "loss": 0.9327, "step": 9207 }, { "epoch": 4.170289855072464, "grad_norm": 0.044872396065566894, "learning_rate": 6.524996016583189e-06, "loss": 0.9031, "step": 9208 }, { "epoch": 4.1707427536231885, "grad_norm": 0.04430476199068899, "learning_rate": 6.518074627917963e-06, "loss": 0.9165, "step": 9209 }, { "epoch": 4.171195652173913, "grad_norm": 0.046900930809980645, "learning_rate": 6.511156586493883e-06, "loss": 0.9184, "step": 9210 }, { "epoch": 4.171648550724638, "grad_norm": 0.04858707113918245, "learning_rate": 6.504241893002548e-06, "loss": 0.9365, "step": 9211 }, { "epoch": 4.172101449275362, "grad_norm": 0.05243980642276721, "learning_rate": 6.497330548135225e-06, "loss": 0.9006, "step": 9212 }, { "epoch": 4.172554347826087, "grad_norm": 0.044686172153690876, "learning_rate": 6.490422552582854e-06, "loss": 0.9121, "step": 9213 }, { "epoch": 4.1730072463768115, "grad_norm": 0.04921746999867451, "learning_rate": 6.4835179070360435e-06, "loss": 0.918, "step": 9214 }, { "epoch": 4.173460144927536, "grad_norm": 0.050055271959626495, "learning_rate": 6.476616612185052e-06, "loss": 0.9346, "step": 9215 }, { "epoch": 4.173913043478261, "grad_norm": 0.04320195918586099, "learning_rate": 6.4697186687198285e-06, "loss": 0.9006, "step": 9216 }, { "epoch": 4.174365942028985, "grad_norm": 0.04392357030748295, "learning_rate": 6.462824077329957e-06, "loss": 0.9236, "step": 9217 }, { "epoch": 4.17481884057971, "grad_norm": 0.04526067015497013, "learning_rate": 6.455932838704715e-06, "loss": 0.9166, "step": 9218 }, { "epoch": 4.175271739130435, "grad_norm": 0.04765918143860958, "learning_rate": 6.449044953533024e-06, "loss": 0.919, "step": 9219 }, { "epoch": 4.175724637681159, "grad_norm": 0.044403729791124426, "learning_rate": 6.442160422503469e-06, "loss": 0.9316, "step": 9220 }, { "epoch": 4.176177536231884, "grad_norm": 0.05060186445996373, "learning_rate": 6.435279246304307e-06, "loss": 0.8999, "step": 9221 }, { "epoch": 4.176630434782608, "grad_norm": 0.047116806764574345, "learning_rate": 6.428401425623479e-06, "loss": 0.9226, "step": 9222 }, { "epoch": 4.177083333333333, "grad_norm": 0.04283950445136174, "learning_rate": 6.421526961148545e-06, "loss": 0.9141, "step": 9223 }, { "epoch": 4.177536231884058, "grad_norm": 0.04914630504588562, "learning_rate": 6.414655853566785e-06, "loss": 0.9201, "step": 9224 }, { "epoch": 4.177989130434782, "grad_norm": 0.04409491383653783, "learning_rate": 6.407788103565087e-06, "loss": 0.913, "step": 9225 }, { "epoch": 4.178442028985507, "grad_norm": 0.04666795629505375, "learning_rate": 6.4009237118300535e-06, "loss": 0.9333, "step": 9226 }, { "epoch": 4.1788949275362315, "grad_norm": 0.046670367482172744, "learning_rate": 6.39406267904791e-06, "loss": 0.9359, "step": 9227 }, { "epoch": 4.179347826086956, "grad_norm": 0.05524104919753754, "learning_rate": 6.387205005904582e-06, "loss": 0.9096, "step": 9228 }, { "epoch": 4.179800724637682, "grad_norm": 0.04815851307947499, "learning_rate": 6.380350693085637e-06, "loss": 0.9239, "step": 9229 }, { "epoch": 4.180253623188406, "grad_norm": 0.04649478256428627, "learning_rate": 6.373499741276305e-06, "loss": 0.9159, "step": 9230 }, { "epoch": 4.180706521739131, "grad_norm": 0.04538360148877471, "learning_rate": 6.3666521511614785e-06, "loss": 0.9205, "step": 9231 }, { "epoch": 4.181159420289855, "grad_norm": 0.04570112058294895, "learning_rate": 6.359807923425747e-06, "loss": 0.9467, "step": 9232 }, { "epoch": 4.18161231884058, "grad_norm": 0.05117605210801712, "learning_rate": 6.352967058753314e-06, "loss": 0.9131, "step": 9233 }, { "epoch": 4.182065217391305, "grad_norm": 0.047045196532057965, "learning_rate": 6.346129557828091e-06, "loss": 0.9137, "step": 9234 }, { "epoch": 4.182518115942029, "grad_norm": 0.04371951723121914, "learning_rate": 6.339295421333616e-06, "loss": 0.9166, "step": 9235 }, { "epoch": 4.182971014492754, "grad_norm": 0.05170909219038148, "learning_rate": 6.332464649953127e-06, "loss": 0.8896, "step": 9236 }, { "epoch": 4.1834239130434785, "grad_norm": 0.05296593246861269, "learning_rate": 6.325637244369498e-06, "loss": 0.9098, "step": 9237 }, { "epoch": 4.183876811594203, "grad_norm": 0.04442641377643161, "learning_rate": 6.318813205265275e-06, "loss": 0.9443, "step": 9238 }, { "epoch": 4.184329710144928, "grad_norm": 0.04778171001580719, "learning_rate": 6.311992533322655e-06, "loss": 0.9321, "step": 9239 }, { "epoch": 4.184782608695652, "grad_norm": 0.06031466738537187, "learning_rate": 6.305175229223541e-06, "loss": 0.921, "step": 9240 }, { "epoch": 4.185235507246377, "grad_norm": 0.052412884901566446, "learning_rate": 6.298361293649438e-06, "loss": 0.9415, "step": 9241 }, { "epoch": 4.1856884057971016, "grad_norm": 0.044277078022875, "learning_rate": 6.291550727281577e-06, "loss": 0.9243, "step": 9242 }, { "epoch": 4.186141304347826, "grad_norm": 0.049719975542785504, "learning_rate": 6.284743530800792e-06, "loss": 0.915, "step": 9243 }, { "epoch": 4.186594202898551, "grad_norm": 0.053104828123616575, "learning_rate": 6.277939704887633e-06, "loss": 0.9263, "step": 9244 }, { "epoch": 4.187047101449275, "grad_norm": 0.04901324573925122, "learning_rate": 6.271139250222278e-06, "loss": 0.9123, "step": 9245 }, { "epoch": 4.1875, "grad_norm": 0.04711725284760253, "learning_rate": 6.26434216748458e-06, "loss": 0.8974, "step": 9246 }, { "epoch": 4.187952898550725, "grad_norm": 0.05455203871479842, "learning_rate": 6.257548457354046e-06, "loss": 0.9115, "step": 9247 }, { "epoch": 4.188405797101449, "grad_norm": 0.04738956351057123, "learning_rate": 6.250758120509872e-06, "loss": 0.9144, "step": 9248 }, { "epoch": 4.188858695652174, "grad_norm": 0.04652576809097572, "learning_rate": 6.243971157630877e-06, "loss": 0.933, "step": 9249 }, { "epoch": 4.1893115942028984, "grad_norm": 0.05449774966398185, "learning_rate": 6.237187569395588e-06, "loss": 0.9053, "step": 9250 }, { "epoch": 4.189764492753623, "grad_norm": 0.049009039951334626, "learning_rate": 6.230407356482149e-06, "loss": 0.9353, "step": 9251 }, { "epoch": 4.190217391304348, "grad_norm": 0.04626937630205559, "learning_rate": 6.223630519568411e-06, "loss": 0.9204, "step": 9252 }, { "epoch": 4.190670289855072, "grad_norm": 0.04968597886680029, "learning_rate": 6.216857059331851e-06, "loss": 0.9381, "step": 9253 }, { "epoch": 4.191123188405797, "grad_norm": 0.0479770547871973, "learning_rate": 6.210086976449625e-06, "loss": 0.8998, "step": 9254 }, { "epoch": 4.1915760869565215, "grad_norm": 0.05538136014177518, "learning_rate": 6.203320271598539e-06, "loss": 0.9394, "step": 9255 }, { "epoch": 4.192028985507246, "grad_norm": 0.051632092264214724, "learning_rate": 6.19655694545509e-06, "loss": 0.9402, "step": 9256 }, { "epoch": 4.192481884057971, "grad_norm": 0.04793705857762218, "learning_rate": 6.189796998695401e-06, "loss": 0.9094, "step": 9257 }, { "epoch": 4.192934782608695, "grad_norm": 0.04583258846220978, "learning_rate": 6.1830404319952906e-06, "loss": 0.9069, "step": 9258 }, { "epoch": 4.19338768115942, "grad_norm": 0.05231005312776964, "learning_rate": 6.176287246030206e-06, "loss": 0.9077, "step": 9259 }, { "epoch": 4.193840579710145, "grad_norm": 0.05044474422080526, "learning_rate": 6.169537441475295e-06, "loss": 0.9017, "step": 9260 }, { "epoch": 4.194293478260869, "grad_norm": 0.05101027097396216, "learning_rate": 6.162791019005334e-06, "loss": 0.9292, "step": 9261 }, { "epoch": 4.194746376811594, "grad_norm": 0.047080132699624716, "learning_rate": 6.156047979294775e-06, "loss": 0.9247, "step": 9262 }, { "epoch": 4.195199275362318, "grad_norm": 0.04927965551772119, "learning_rate": 6.149308323017722e-06, "loss": 0.9185, "step": 9263 }, { "epoch": 4.195652173913044, "grad_norm": 0.051392306258627535, "learning_rate": 6.142572050847966e-06, "loss": 0.9213, "step": 9264 }, { "epoch": 4.1961050724637685, "grad_norm": 0.04256425086141542, "learning_rate": 6.135839163458923e-06, "loss": 0.9154, "step": 9265 }, { "epoch": 4.196557971014493, "grad_norm": 0.04614756105397516, "learning_rate": 6.129109661523709e-06, "loss": 0.9162, "step": 9266 }, { "epoch": 4.197010869565218, "grad_norm": 0.051188931506853545, "learning_rate": 6.122383545715069e-06, "loss": 0.9236, "step": 9267 }, { "epoch": 4.197463768115942, "grad_norm": 0.043184366947758526, "learning_rate": 6.115660816705435e-06, "loss": 0.9205, "step": 9268 }, { "epoch": 4.197916666666667, "grad_norm": 0.04668599710866952, "learning_rate": 6.108941475166879e-06, "loss": 0.8943, "step": 9269 }, { "epoch": 4.198369565217392, "grad_norm": 0.04797184743416011, "learning_rate": 6.102225521771154e-06, "loss": 0.9136, "step": 9270 }, { "epoch": 4.198822463768116, "grad_norm": 0.04615842368897833, "learning_rate": 6.095512957189664e-06, "loss": 0.9386, "step": 9271 }, { "epoch": 4.199275362318841, "grad_norm": 0.04886643544289137, "learning_rate": 6.088803782093466e-06, "loss": 0.925, "step": 9272 }, { "epoch": 4.199728260869565, "grad_norm": 0.04223613448169759, "learning_rate": 6.0820979971532825e-06, "loss": 0.9202, "step": 9273 }, { "epoch": 4.20018115942029, "grad_norm": 0.04748685692292357, "learning_rate": 6.075395603039518e-06, "loss": 0.9293, "step": 9274 }, { "epoch": 4.200634057971015, "grad_norm": 0.050404431011615174, "learning_rate": 6.068696600422206e-06, "loss": 0.9149, "step": 9275 }, { "epoch": 4.201086956521739, "grad_norm": 0.04418968654528176, "learning_rate": 6.062000989971073e-06, "loss": 0.9159, "step": 9276 }, { "epoch": 4.201539855072464, "grad_norm": 0.04511623468039159, "learning_rate": 6.055308772355468e-06, "loss": 0.921, "step": 9277 }, { "epoch": 4.2019927536231885, "grad_norm": 0.048029214705001114, "learning_rate": 6.048619948244448e-06, "loss": 0.8949, "step": 9278 }, { "epoch": 4.202445652173913, "grad_norm": 0.04485653291274694, "learning_rate": 6.041934518306689e-06, "loss": 0.9058, "step": 9279 }, { "epoch": 4.202898550724638, "grad_norm": 0.04806234311282053, "learning_rate": 6.035252483210547e-06, "loss": 0.9516, "step": 9280 }, { "epoch": 4.203351449275362, "grad_norm": 0.04658079482562603, "learning_rate": 6.028573843624031e-06, "loss": 0.9231, "step": 9281 }, { "epoch": 4.203804347826087, "grad_norm": 0.044278367152931354, "learning_rate": 6.021898600214826e-06, "loss": 0.9146, "step": 9282 }, { "epoch": 4.2042572463768115, "grad_norm": 0.04443911133539346, "learning_rate": 6.015226753650254e-06, "loss": 0.9282, "step": 9283 }, { "epoch": 4.204710144927536, "grad_norm": 0.04942098905607278, "learning_rate": 6.0085583045973224e-06, "loss": 0.9368, "step": 9284 }, { "epoch": 4.205163043478261, "grad_norm": 0.04435374277019379, "learning_rate": 6.001893253722677e-06, "loss": 0.9346, "step": 9285 }, { "epoch": 4.205615942028985, "grad_norm": 0.041835388924288265, "learning_rate": 5.995231601692641e-06, "loss": 0.8885, "step": 9286 }, { "epoch": 4.20606884057971, "grad_norm": 0.05081041482417246, "learning_rate": 5.988573349173195e-06, "loss": 0.9065, "step": 9287 }, { "epoch": 4.206521739130435, "grad_norm": 0.05047145146527746, "learning_rate": 5.981918496829959e-06, "loss": 0.9068, "step": 9288 }, { "epoch": 4.206974637681159, "grad_norm": 0.04510988839452639, "learning_rate": 5.97526704532823e-06, "loss": 0.93, "step": 9289 }, { "epoch": 4.207427536231884, "grad_norm": 0.044333332480897915, "learning_rate": 5.968618995332982e-06, "loss": 0.9263, "step": 9290 }, { "epoch": 4.207880434782608, "grad_norm": 0.048597553653060364, "learning_rate": 5.961974347508812e-06, "loss": 0.9299, "step": 9291 }, { "epoch": 4.208333333333333, "grad_norm": 0.044206720667914846, "learning_rate": 5.955333102520011e-06, "loss": 0.9278, "step": 9292 }, { "epoch": 4.208786231884058, "grad_norm": 0.05020211805127637, "learning_rate": 5.948695261030502e-06, "loss": 0.9291, "step": 9293 }, { "epoch": 4.209239130434782, "grad_norm": 0.04324723437662799, "learning_rate": 5.942060823703895e-06, "loss": 0.9332, "step": 9294 }, { "epoch": 4.209692028985507, "grad_norm": 0.04482941527415293, "learning_rate": 5.935429791203437e-06, "loss": 0.8977, "step": 9295 }, { "epoch": 4.2101449275362315, "grad_norm": 0.04338790693352158, "learning_rate": 5.92880216419204e-06, "loss": 0.9396, "step": 9296 }, { "epoch": 4.210597826086956, "grad_norm": 0.04574411273395491, "learning_rate": 5.922177943332275e-06, "loss": 0.9284, "step": 9297 }, { "epoch": 4.211050724637682, "grad_norm": 0.04477449272379397, "learning_rate": 5.915557129286389e-06, "loss": 0.9274, "step": 9298 }, { "epoch": 4.211503623188406, "grad_norm": 0.04481201762517809, "learning_rate": 5.908939722716258e-06, "loss": 0.9286, "step": 9299 }, { "epoch": 4.211956521739131, "grad_norm": 0.054591361385191765, "learning_rate": 5.9023257242834595e-06, "loss": 0.9199, "step": 9300 }, { "epoch": 4.212409420289855, "grad_norm": 0.04461160494901226, "learning_rate": 5.8957151346491756e-06, "loss": 0.9315, "step": 9301 }, { "epoch": 4.21286231884058, "grad_norm": 0.04598330782399452, "learning_rate": 5.8891079544743045e-06, "loss": 0.9127, "step": 9302 }, { "epoch": 4.213315217391305, "grad_norm": 0.04908173308729077, "learning_rate": 5.882504184419367e-06, "loss": 0.9373, "step": 9303 }, { "epoch": 4.213768115942029, "grad_norm": 0.046805620684586054, "learning_rate": 5.875903825144544e-06, "loss": 0.9344, "step": 9304 }, { "epoch": 4.214221014492754, "grad_norm": 0.046554865980275674, "learning_rate": 5.869306877309693e-06, "loss": 0.9352, "step": 9305 }, { "epoch": 4.2146739130434785, "grad_norm": 0.04793937738934069, "learning_rate": 5.862713341574311e-06, "loss": 0.9448, "step": 9306 }, { "epoch": 4.215126811594203, "grad_norm": 0.0434496191158703, "learning_rate": 5.856123218597578e-06, "loss": 0.9062, "step": 9307 }, { "epoch": 4.215579710144928, "grad_norm": 0.04601775846365246, "learning_rate": 5.849536509038305e-06, "loss": 0.9166, "step": 9308 }, { "epoch": 4.216032608695652, "grad_norm": 0.04306358465882486, "learning_rate": 5.8429532135549915e-06, "loss": 0.91, "step": 9309 }, { "epoch": 4.216485507246377, "grad_norm": 0.04477187088717955, "learning_rate": 5.83637333280576e-06, "loss": 0.9057, "step": 9310 }, { "epoch": 4.2169384057971016, "grad_norm": 0.048201880372018675, "learning_rate": 5.8297968674484364e-06, "loss": 0.9418, "step": 9311 }, { "epoch": 4.217391304347826, "grad_norm": 0.046045044876907575, "learning_rate": 5.823223818140458e-06, "loss": 0.9246, "step": 9312 }, { "epoch": 4.217844202898551, "grad_norm": 0.04323212417343363, "learning_rate": 5.816654185538957e-06, "loss": 0.9114, "step": 9313 }, { "epoch": 4.218297101449275, "grad_norm": 0.0441564293675109, "learning_rate": 5.810087970300711e-06, "loss": 0.9335, "step": 9314 }, { "epoch": 4.21875, "grad_norm": 0.041086550045336255, "learning_rate": 5.803525173082145e-06, "loss": 0.9304, "step": 9315 }, { "epoch": 4.219202898550725, "grad_norm": 0.048371930595306194, "learning_rate": 5.796965794539349e-06, "loss": 0.9289, "step": 9316 }, { "epoch": 4.219655797101449, "grad_norm": 0.0422314835175504, "learning_rate": 5.7904098353280904e-06, "loss": 0.9029, "step": 9317 }, { "epoch": 4.220108695652174, "grad_norm": 0.044500475990550604, "learning_rate": 5.783857296103761e-06, "loss": 0.9222, "step": 9318 }, { "epoch": 4.2205615942028984, "grad_norm": 0.04347286497016021, "learning_rate": 5.777308177521445e-06, "loss": 0.8982, "step": 9319 }, { "epoch": 4.221014492753623, "grad_norm": 0.04402688587401016, "learning_rate": 5.7707624802358565e-06, "loss": 0.9322, "step": 9320 }, { "epoch": 4.221467391304348, "grad_norm": 0.04289604748192675, "learning_rate": 5.764220204901389e-06, "loss": 0.9126, "step": 9321 }, { "epoch": 4.221920289855072, "grad_norm": 0.043170375995762, "learning_rate": 5.7576813521720824e-06, "loss": 0.9181, "step": 9322 }, { "epoch": 4.222373188405797, "grad_norm": 0.043403127460061794, "learning_rate": 5.751145922701629e-06, "loss": 0.9343, "step": 9323 }, { "epoch": 4.2228260869565215, "grad_norm": 0.04639980250949353, "learning_rate": 5.744613917143382e-06, "loss": 0.8995, "step": 9324 }, { "epoch": 4.223278985507246, "grad_norm": 0.04457812345897971, "learning_rate": 5.738085336150377e-06, "loss": 0.9275, "step": 9325 }, { "epoch": 4.223731884057971, "grad_norm": 0.04360624790231928, "learning_rate": 5.731560180375257e-06, "loss": 0.9253, "step": 9326 }, { "epoch": 4.224184782608695, "grad_norm": 0.04199726612729577, "learning_rate": 5.725038450470383e-06, "loss": 0.9309, "step": 9327 }, { "epoch": 4.22463768115942, "grad_norm": 0.04407079778889445, "learning_rate": 5.718520147087719e-06, "loss": 0.9243, "step": 9328 }, { "epoch": 4.225090579710145, "grad_norm": 0.04622027756581417, "learning_rate": 5.712005270878927e-06, "loss": 0.9252, "step": 9329 }, { "epoch": 4.225543478260869, "grad_norm": 0.04317509835290033, "learning_rate": 5.705493822495305e-06, "loss": 0.912, "step": 9330 }, { "epoch": 4.225996376811594, "grad_norm": 0.046463492333102846, "learning_rate": 5.698985802587804e-06, "loss": 0.9171, "step": 9331 }, { "epoch": 4.226449275362318, "grad_norm": 0.05036372668650682, "learning_rate": 5.692481211807046e-06, "loss": 0.923, "step": 9332 }, { "epoch": 4.226902173913044, "grad_norm": 0.043566678896951455, "learning_rate": 5.6859800508033105e-06, "loss": 0.9284, "step": 9333 }, { "epoch": 4.2273550724637685, "grad_norm": 0.0445527005874664, "learning_rate": 5.6794823202265125e-06, "loss": 0.936, "step": 9334 }, { "epoch": 4.227807971014493, "grad_norm": 0.05011101106107983, "learning_rate": 5.672988020726266e-06, "loss": 0.945, "step": 9335 }, { "epoch": 4.228260869565218, "grad_norm": 0.0466424653206864, "learning_rate": 5.666497152951791e-06, "loss": 0.9323, "step": 9336 }, { "epoch": 4.228713768115942, "grad_norm": 0.043444141835335945, "learning_rate": 5.660009717552015e-06, "loss": 0.9237, "step": 9337 }, { "epoch": 4.229166666666667, "grad_norm": 0.04889346478256003, "learning_rate": 5.653525715175483e-06, "loss": 0.9265, "step": 9338 }, { "epoch": 4.229619565217392, "grad_norm": 0.04847880219892926, "learning_rate": 5.647045146470409e-06, "loss": 0.8947, "step": 9339 }, { "epoch": 4.230072463768116, "grad_norm": 0.049217878837208275, "learning_rate": 5.640568012084662e-06, "loss": 0.9137, "step": 9340 }, { "epoch": 4.230525362318841, "grad_norm": 0.047350133557303616, "learning_rate": 5.634094312665785e-06, "loss": 0.8947, "step": 9341 }, { "epoch": 4.230978260869565, "grad_norm": 0.05264719682716076, "learning_rate": 5.6276240488609464e-06, "loss": 0.9089, "step": 9342 }, { "epoch": 4.23143115942029, "grad_norm": 0.05156132800489021, "learning_rate": 5.621157221317011e-06, "loss": 0.8983, "step": 9343 }, { "epoch": 4.231884057971015, "grad_norm": 0.04412145025609844, "learning_rate": 5.61469383068046e-06, "loss": 0.9305, "step": 9344 }, { "epoch": 4.232336956521739, "grad_norm": 0.051233276357322624, "learning_rate": 5.608233877597458e-06, "loss": 0.9126, "step": 9345 }, { "epoch": 4.232789855072464, "grad_norm": 0.04476707763705237, "learning_rate": 5.601777362713816e-06, "loss": 0.9382, "step": 9346 }, { "epoch": 4.2332427536231885, "grad_norm": 0.0509561345706236, "learning_rate": 5.595324286675001e-06, "loss": 0.9076, "step": 9347 }, { "epoch": 4.233695652173913, "grad_norm": 0.04493085495987364, "learning_rate": 5.588874650126128e-06, "loss": 0.9239, "step": 9348 }, { "epoch": 4.234148550724638, "grad_norm": 0.045750609274460015, "learning_rate": 5.582428453711992e-06, "loss": 0.9174, "step": 9349 }, { "epoch": 4.234601449275362, "grad_norm": 0.046311263190457426, "learning_rate": 5.575985698077016e-06, "loss": 0.9148, "step": 9350 }, { "epoch": 4.235054347826087, "grad_norm": 0.04807098068056031, "learning_rate": 5.569546383865305e-06, "loss": 0.9286, "step": 9351 }, { "epoch": 4.2355072463768115, "grad_norm": 0.04439594785375629, "learning_rate": 5.563110511720595e-06, "loss": 0.923, "step": 9352 }, { "epoch": 4.235960144927536, "grad_norm": 0.04515453219460295, "learning_rate": 5.556678082286304e-06, "loss": 0.9073, "step": 9353 }, { "epoch": 4.236413043478261, "grad_norm": 0.04597111297746932, "learning_rate": 5.5502490962054775e-06, "loss": 0.9192, "step": 9354 }, { "epoch": 4.236865942028985, "grad_norm": 0.04407469355075187, "learning_rate": 5.54382355412086e-06, "loss": 0.9212, "step": 9355 }, { "epoch": 4.23731884057971, "grad_norm": 0.048907217613122514, "learning_rate": 5.53740145667478e-06, "loss": 0.9175, "step": 9356 }, { "epoch": 4.237771739130435, "grad_norm": 0.04691313859897377, "learning_rate": 5.5309828045093e-06, "loss": 0.907, "step": 9357 }, { "epoch": 4.238224637681159, "grad_norm": 0.046101413727868724, "learning_rate": 5.524567598266078e-06, "loss": 0.9123, "step": 9358 }, { "epoch": 4.238677536231884, "grad_norm": 0.05224030810127048, "learning_rate": 5.518155838586472e-06, "loss": 0.9183, "step": 9359 }, { "epoch": 4.239130434782608, "grad_norm": 0.043156581760620885, "learning_rate": 5.511747526111459e-06, "loss": 0.9235, "step": 9360 }, { "epoch": 4.239583333333333, "grad_norm": 0.04468049453629877, "learning_rate": 5.5053426614817094e-06, "loss": 0.9058, "step": 9361 }, { "epoch": 4.240036231884058, "grad_norm": 0.04690148530420242, "learning_rate": 5.498941245337506e-06, "loss": 0.9409, "step": 9362 }, { "epoch": 4.240489130434782, "grad_norm": 0.04422030220582612, "learning_rate": 5.492543278318825e-06, "loss": 0.9361, "step": 9363 }, { "epoch": 4.240942028985507, "grad_norm": 0.04228062787375207, "learning_rate": 5.486148761065275e-06, "loss": 0.9217, "step": 9364 }, { "epoch": 4.2413949275362315, "grad_norm": 0.04460803676822155, "learning_rate": 5.479757694216128e-06, "loss": 0.9107, "step": 9365 }, { "epoch": 4.241847826086956, "grad_norm": 0.05014877910658825, "learning_rate": 5.4733700784102935e-06, "loss": 0.9151, "step": 9366 }, { "epoch": 4.242300724637682, "grad_norm": 0.04338211580685192, "learning_rate": 5.466985914286373e-06, "loss": 0.938, "step": 9367 }, { "epoch": 4.242753623188406, "grad_norm": 0.04374413262395208, "learning_rate": 5.460605202482585e-06, "loss": 0.9263, "step": 9368 }, { "epoch": 4.243206521739131, "grad_norm": 0.04610837663445453, "learning_rate": 5.454227943636836e-06, "loss": 0.908, "step": 9369 }, { "epoch": 4.243659420289855, "grad_norm": 0.041946674160695134, "learning_rate": 5.447854138386656e-06, "loss": 0.9352, "step": 9370 }, { "epoch": 4.24411231884058, "grad_norm": 0.04801897562268342, "learning_rate": 5.441483787369257e-06, "loss": 0.9343, "step": 9371 }, { "epoch": 4.244565217391305, "grad_norm": 0.04702917384405705, "learning_rate": 5.435116891221488e-06, "loss": 0.9276, "step": 9372 }, { "epoch": 4.245018115942029, "grad_norm": 0.04355102596344271, "learning_rate": 5.428753450579853e-06, "loss": 0.9259, "step": 9373 }, { "epoch": 4.245471014492754, "grad_norm": 0.046485251097366787, "learning_rate": 5.422393466080511e-06, "loss": 0.9342, "step": 9374 }, { "epoch": 4.2459239130434785, "grad_norm": 0.04425548467256047, "learning_rate": 5.416036938359295e-06, "loss": 0.9159, "step": 9375 }, { "epoch": 4.246376811594203, "grad_norm": 0.04236307503464299, "learning_rate": 5.409683868051661e-06, "loss": 0.9125, "step": 9376 }, { "epoch": 4.246829710144928, "grad_norm": 0.04486400214678031, "learning_rate": 5.403334255792754e-06, "loss": 0.9222, "step": 9377 }, { "epoch": 4.247282608695652, "grad_norm": 0.04966415861408436, "learning_rate": 5.396988102217333e-06, "loss": 0.9414, "step": 9378 }, { "epoch": 4.247735507246377, "grad_norm": 0.04422681744404439, "learning_rate": 5.390645407959856e-06, "loss": 0.9281, "step": 9379 }, { "epoch": 4.2481884057971016, "grad_norm": 0.044163131753056105, "learning_rate": 5.384306173654401e-06, "loss": 0.9333, "step": 9380 }, { "epoch": 4.248641304347826, "grad_norm": 0.04398910454642376, "learning_rate": 5.377970399934711e-06, "loss": 0.9057, "step": 9381 }, { "epoch": 4.249094202898551, "grad_norm": 0.04280036908599608, "learning_rate": 5.371638087434172e-06, "loss": 0.9107, "step": 9382 }, { "epoch": 4.249547101449275, "grad_norm": 0.043078529371104186, "learning_rate": 5.365309236785856e-06, "loss": 0.9298, "step": 9383 }, { "epoch": 4.25, "grad_norm": 0.045318241769989324, "learning_rate": 5.358983848622452e-06, "loss": 0.8881, "step": 9384 }, { "epoch": 4.250452898550725, "grad_norm": 0.04528426337551617, "learning_rate": 5.352661923576334e-06, "loss": 0.9237, "step": 9385 }, { "epoch": 4.250905797101449, "grad_norm": 0.04938539536849197, "learning_rate": 5.3463434622795e-06, "loss": 0.9466, "step": 9386 }, { "epoch": 4.251358695652174, "grad_norm": 0.040231366657333525, "learning_rate": 5.3400284653636295e-06, "loss": 0.9033, "step": 9387 }, { "epoch": 4.2518115942028984, "grad_norm": 0.0423008707321922, "learning_rate": 5.33371693346004e-06, "loss": 0.9133, "step": 9388 }, { "epoch": 4.252264492753623, "grad_norm": 0.04575006793034864, "learning_rate": 5.327408867199699e-06, "loss": 0.9167, "step": 9389 }, { "epoch": 4.252717391304348, "grad_norm": 0.04517096049851013, "learning_rate": 5.3211042672132305e-06, "loss": 0.9185, "step": 9390 }, { "epoch": 4.253170289855072, "grad_norm": 0.047119674425060924, "learning_rate": 5.31480313413093e-06, "loss": 0.9257, "step": 9391 }, { "epoch": 4.253623188405797, "grad_norm": 0.04401545522362487, "learning_rate": 5.308505468582712e-06, "loss": 0.9385, "step": 9392 }, { "epoch": 4.2540760869565215, "grad_norm": 0.045743355640440374, "learning_rate": 5.302211271198188e-06, "loss": 0.8985, "step": 9393 }, { "epoch": 4.254528985507246, "grad_norm": 0.05098703443877181, "learning_rate": 5.2959205426065744e-06, "loss": 0.9151, "step": 9394 }, { "epoch": 4.254981884057971, "grad_norm": 0.04366584770389839, "learning_rate": 5.28963328343679e-06, "loss": 0.9258, "step": 9395 }, { "epoch": 4.255434782608695, "grad_norm": 0.04355296399557841, "learning_rate": 5.283349494317369e-06, "loss": 0.902, "step": 9396 }, { "epoch": 4.25588768115942, "grad_norm": 0.046285238116038595, "learning_rate": 5.27706917587651e-06, "loss": 0.9012, "step": 9397 }, { "epoch": 4.256340579710145, "grad_norm": 0.04352544024723116, "learning_rate": 5.270792328742062e-06, "loss": 0.9045, "step": 9398 }, { "epoch": 4.256793478260869, "grad_norm": 0.04339887470298874, "learning_rate": 5.264518953541546e-06, "loss": 0.9122, "step": 9399 }, { "epoch": 4.257246376811594, "grad_norm": 0.04394281586779841, "learning_rate": 5.258249050902104e-06, "loss": 0.9049, "step": 9400 }, { "epoch": 4.257699275362318, "grad_norm": 0.05178529131567963, "learning_rate": 5.251982621450564e-06, "loss": 0.9092, "step": 9401 }, { "epoch": 4.258152173913043, "grad_norm": 0.04913839177813669, "learning_rate": 5.2457196658133804e-06, "loss": 0.9328, "step": 9402 }, { "epoch": 4.2586050724637685, "grad_norm": 0.04902798297075457, "learning_rate": 5.239460184616682e-06, "loss": 0.9322, "step": 9403 }, { "epoch": 4.259057971014493, "grad_norm": 0.05032250607751383, "learning_rate": 5.23320417848622e-06, "loss": 0.9097, "step": 9404 }, { "epoch": 4.259510869565218, "grad_norm": 0.04725966387134065, "learning_rate": 5.226951648047442e-06, "loss": 0.9201, "step": 9405 }, { "epoch": 4.259963768115942, "grad_norm": 0.043628652740775264, "learning_rate": 5.220702593925407e-06, "loss": 0.8922, "step": 9406 }, { "epoch": 4.260416666666667, "grad_norm": 0.048130220805046114, "learning_rate": 5.2144570167448475e-06, "loss": 0.9292, "step": 9407 }, { "epoch": 4.260869565217392, "grad_norm": 0.0483853515746185, "learning_rate": 5.208214917130132e-06, "loss": 0.9306, "step": 9408 }, { "epoch": 4.261322463768116, "grad_norm": 0.04488562452152101, "learning_rate": 5.201976295705318e-06, "loss": 0.9223, "step": 9409 }, { "epoch": 4.261775362318841, "grad_norm": 0.04749480658185855, "learning_rate": 5.19574115309406e-06, "loss": 0.933, "step": 9410 }, { "epoch": 4.262228260869565, "grad_norm": 0.040692209859185556, "learning_rate": 5.189509489919724e-06, "loss": 0.9004, "step": 9411 }, { "epoch": 4.26268115942029, "grad_norm": 0.0470281816085421, "learning_rate": 5.1832813068052765e-06, "loss": 0.9464, "step": 9412 }, { "epoch": 4.263134057971015, "grad_norm": 0.046161115468160825, "learning_rate": 5.177056604373376e-06, "loss": 0.9146, "step": 9413 }, { "epoch": 4.263586956521739, "grad_norm": 0.04420422010133239, "learning_rate": 5.170835383246311e-06, "loss": 0.9241, "step": 9414 }, { "epoch": 4.264039855072464, "grad_norm": 0.0482481437968206, "learning_rate": 5.164617644046024e-06, "loss": 0.9222, "step": 9415 }, { "epoch": 4.2644927536231885, "grad_norm": 0.0423450979040295, "learning_rate": 5.158403387394102e-06, "loss": 0.9122, "step": 9416 }, { "epoch": 4.264945652173913, "grad_norm": 0.04558510614574998, "learning_rate": 5.152192613911813e-06, "loss": 0.9277, "step": 9417 }, { "epoch": 4.265398550724638, "grad_norm": 0.04536710608363326, "learning_rate": 5.145985324220042e-06, "loss": 0.9266, "step": 9418 }, { "epoch": 4.265851449275362, "grad_norm": 0.0420466025254994, "learning_rate": 5.13978151893936e-06, "loss": 0.9458, "step": 9419 }, { "epoch": 4.266304347826087, "grad_norm": 0.0440593575352588, "learning_rate": 5.133581198689946e-06, "loss": 0.937, "step": 9420 }, { "epoch": 4.2667572463768115, "grad_norm": 0.042786855285977074, "learning_rate": 5.127384364091685e-06, "loss": 0.9247, "step": 9421 }, { "epoch": 4.267210144927536, "grad_norm": 0.043548876433993036, "learning_rate": 5.121191015764066e-06, "loss": 0.926, "step": 9422 }, { "epoch": 4.267663043478261, "grad_norm": 0.044680400896767884, "learning_rate": 5.1150011543262516e-06, "loss": 0.9159, "step": 9423 }, { "epoch": 4.268115942028985, "grad_norm": 0.04395642991746128, "learning_rate": 5.108814780397042e-06, "loss": 0.9345, "step": 9424 }, { "epoch": 4.26856884057971, "grad_norm": 0.04556301386879372, "learning_rate": 5.102631894594919e-06, "loss": 0.9403, "step": 9425 }, { "epoch": 4.269021739130435, "grad_norm": 0.05238535309108948, "learning_rate": 5.096452497537985e-06, "loss": 0.9312, "step": 9426 }, { "epoch": 4.269474637681159, "grad_norm": 0.043760324120003365, "learning_rate": 5.090276589843996e-06, "loss": 0.8972, "step": 9427 }, { "epoch": 4.269927536231884, "grad_norm": 0.04325278127783332, "learning_rate": 5.084104172130384e-06, "loss": 0.9111, "step": 9428 }, { "epoch": 4.270380434782608, "grad_norm": 0.05112047378574018, "learning_rate": 5.077935245014196e-06, "loss": 0.9142, "step": 9429 }, { "epoch": 4.270833333333333, "grad_norm": 0.04764708176179264, "learning_rate": 5.07176980911217e-06, "loss": 0.9124, "step": 9430 }, { "epoch": 4.271286231884058, "grad_norm": 0.04291129594884483, "learning_rate": 5.065607865040667e-06, "loss": 0.9052, "step": 9431 }, { "epoch": 4.271739130434782, "grad_norm": 0.049236144015026344, "learning_rate": 5.059449413415705e-06, "loss": 0.9104, "step": 9432 }, { "epoch": 4.272192028985507, "grad_norm": 0.049555103316848495, "learning_rate": 5.053294454852942e-06, "loss": 0.9051, "step": 9433 }, { "epoch": 4.2726449275362315, "grad_norm": 0.04970591507084452, "learning_rate": 5.047142989967717e-06, "loss": 0.9401, "step": 9434 }, { "epoch": 4.273097826086957, "grad_norm": 0.04477692991738972, "learning_rate": 5.040995019374988e-06, "loss": 0.9347, "step": 9435 }, { "epoch": 4.273550724637682, "grad_norm": 0.04833452072833305, "learning_rate": 5.034850543689391e-06, "loss": 0.9, "step": 9436 }, { "epoch": 4.274003623188406, "grad_norm": 0.04949105686281424, "learning_rate": 5.028709563525187e-06, "loss": 0.9121, "step": 9437 }, { "epoch": 4.274456521739131, "grad_norm": 0.042730244215808226, "learning_rate": 5.0225720794963105e-06, "loss": 0.9076, "step": 9438 }, { "epoch": 4.274909420289855, "grad_norm": 0.0479144087623217, "learning_rate": 5.0164380922163335e-06, "loss": 0.9104, "step": 9439 }, { "epoch": 4.27536231884058, "grad_norm": 0.050857140433632146, "learning_rate": 5.010307602298472e-06, "loss": 0.9125, "step": 9440 }, { "epoch": 4.275815217391305, "grad_norm": 0.04376748123197272, "learning_rate": 5.0041806103556e-06, "loss": 0.9435, "step": 9441 }, { "epoch": 4.276268115942029, "grad_norm": 0.04539133653777298, "learning_rate": 4.998057117000255e-06, "loss": 0.9286, "step": 9442 }, { "epoch": 4.276721014492754, "grad_norm": 0.04801715246093088, "learning_rate": 4.991937122844599e-06, "loss": 0.9157, "step": 9443 }, { "epoch": 4.2771739130434785, "grad_norm": 0.04444313471626092, "learning_rate": 4.985820628500473e-06, "loss": 0.9388, "step": 9444 }, { "epoch": 4.277626811594203, "grad_norm": 0.057592297420331916, "learning_rate": 4.979707634579329e-06, "loss": 0.9094, "step": 9445 }, { "epoch": 4.278079710144928, "grad_norm": 0.04774576096909576, "learning_rate": 4.97359814169232e-06, "loss": 0.9247, "step": 9446 }, { "epoch": 4.278532608695652, "grad_norm": 0.05282909359051347, "learning_rate": 4.9674921504502e-06, "loss": 0.9138, "step": 9447 }, { "epoch": 4.278985507246377, "grad_norm": 0.05022952353464201, "learning_rate": 4.9613896614634136e-06, "loss": 0.9046, "step": 9448 }, { "epoch": 4.2794384057971016, "grad_norm": 0.045893152416133384, "learning_rate": 4.955290675342022e-06, "loss": 0.9212, "step": 9449 }, { "epoch": 4.279891304347826, "grad_norm": 0.04513075349209318, "learning_rate": 4.9491951926957574e-06, "loss": 0.9329, "step": 9450 }, { "epoch": 4.280344202898551, "grad_norm": 0.0519551446797922, "learning_rate": 4.943103214133986e-06, "loss": 0.9142, "step": 9451 }, { "epoch": 4.280797101449275, "grad_norm": 0.04590820912156594, "learning_rate": 4.937014740265742e-06, "loss": 0.9109, "step": 9452 }, { "epoch": 4.28125, "grad_norm": 0.044909357308458416, "learning_rate": 4.930929771699693e-06, "loss": 0.9268, "step": 9453 }, { "epoch": 4.281702898550725, "grad_norm": 0.04658027795578857, "learning_rate": 4.92484830904417e-06, "loss": 0.9163, "step": 9454 }, { "epoch": 4.282155797101449, "grad_norm": 0.040795832615909626, "learning_rate": 4.918770352907136e-06, "loss": 0.9221, "step": 9455 }, { "epoch": 4.282608695652174, "grad_norm": 0.04279260615814601, "learning_rate": 4.91269590389623e-06, "loss": 0.941, "step": 9456 }, { "epoch": 4.2830615942028984, "grad_norm": 0.04318028164307463, "learning_rate": 4.9066249626187155e-06, "loss": 0.8998, "step": 9457 }, { "epoch": 4.283514492753623, "grad_norm": 0.03975212544297382, "learning_rate": 4.900557529681513e-06, "loss": 0.9382, "step": 9458 }, { "epoch": 4.283967391304348, "grad_norm": 0.04172077508790216, "learning_rate": 4.894493605691186e-06, "loss": 0.921, "step": 9459 }, { "epoch": 4.284420289855072, "grad_norm": 0.04373461709798942, "learning_rate": 4.888433191253974e-06, "loss": 0.9343, "step": 9460 }, { "epoch": 4.284873188405797, "grad_norm": 0.0444388303151466, "learning_rate": 4.882376286975721e-06, "loss": 0.9155, "step": 9461 }, { "epoch": 4.2853260869565215, "grad_norm": 0.04307768434452821, "learning_rate": 4.876322893461974e-06, "loss": 0.9297, "step": 9462 }, { "epoch": 4.285778985507246, "grad_norm": 0.04175202296685514, "learning_rate": 4.87027301131787e-06, "loss": 0.9064, "step": 9463 }, { "epoch": 4.286231884057971, "grad_norm": 0.04314139855603146, "learning_rate": 4.864226641148256e-06, "loss": 0.9016, "step": 9464 }, { "epoch": 4.286684782608695, "grad_norm": 0.04369930803959031, "learning_rate": 4.858183783557575e-06, "loss": 0.9114, "step": 9465 }, { "epoch": 4.28713768115942, "grad_norm": 0.04578384813622058, "learning_rate": 4.852144439149955e-06, "loss": 0.928, "step": 9466 }, { "epoch": 4.287590579710145, "grad_norm": 0.04316484567361699, "learning_rate": 4.846108608529139e-06, "loss": 0.9192, "step": 9467 }, { "epoch": 4.288043478260869, "grad_norm": 0.0423092078679909, "learning_rate": 4.8400762922985635e-06, "loss": 0.9327, "step": 9468 }, { "epoch": 4.288496376811594, "grad_norm": 0.04606820258915135, "learning_rate": 4.834047491061267e-06, "loss": 0.9015, "step": 9469 }, { "epoch": 4.288949275362318, "grad_norm": 0.046816196072131895, "learning_rate": 4.828022205419975e-06, "loss": 0.9089, "step": 9470 }, { "epoch": 4.289402173913043, "grad_norm": 0.04442139298656876, "learning_rate": 4.82200043597703e-06, "loss": 0.9036, "step": 9471 }, { "epoch": 4.2898550724637685, "grad_norm": 0.043244291065335984, "learning_rate": 4.815982183334456e-06, "loss": 0.9195, "step": 9472 }, { "epoch": 4.290307971014493, "grad_norm": 0.04445989565656179, "learning_rate": 4.8099674480938955e-06, "loss": 0.9249, "step": 9473 }, { "epoch": 4.290760869565218, "grad_norm": 0.04491612329940572, "learning_rate": 4.80395623085665e-06, "loss": 0.9255, "step": 9474 }, { "epoch": 4.291213768115942, "grad_norm": 0.04231714511893, "learning_rate": 4.797948532223666e-06, "loss": 0.9145, "step": 9475 }, { "epoch": 4.291666666666667, "grad_norm": 0.04516527073222936, "learning_rate": 4.791944352795561e-06, "loss": 0.9416, "step": 9476 }, { "epoch": 4.292119565217392, "grad_norm": 0.05024666466687317, "learning_rate": 4.785943693172561e-06, "loss": 0.9224, "step": 9477 }, { "epoch": 4.292572463768116, "grad_norm": 0.04514607825421716, "learning_rate": 4.779946553954578e-06, "loss": 0.9296, "step": 9478 }, { "epoch": 4.293025362318841, "grad_norm": 0.041611207474021555, "learning_rate": 4.773952935741144e-06, "loss": 0.9361, "step": 9479 }, { "epoch": 4.293478260869565, "grad_norm": 0.04683218076649312, "learning_rate": 4.7679628391314615e-06, "loss": 0.9337, "step": 9480 }, { "epoch": 4.29393115942029, "grad_norm": 0.04914599672767355, "learning_rate": 4.761976264724366e-06, "loss": 0.9142, "step": 9481 }, { "epoch": 4.294384057971015, "grad_norm": 0.046887805108910816, "learning_rate": 4.7559932131183395e-06, "loss": 0.9037, "step": 9482 }, { "epoch": 4.294836956521739, "grad_norm": 0.04287594989980111, "learning_rate": 4.750013684911512e-06, "loss": 0.931, "step": 9483 }, { "epoch": 4.295289855072464, "grad_norm": 0.04645263563419294, "learning_rate": 4.744037680701686e-06, "loss": 0.9094, "step": 9484 }, { "epoch": 4.2957427536231885, "grad_norm": 0.0497276576779348, "learning_rate": 4.73806520108627e-06, "loss": 0.9358, "step": 9485 }, { "epoch": 4.296195652173913, "grad_norm": 0.04097772199049459, "learning_rate": 4.732096246662368e-06, "loss": 0.9189, "step": 9486 }, { "epoch": 4.296648550724638, "grad_norm": 0.04356447093287494, "learning_rate": 4.7261308180266774e-06, "loss": 0.9242, "step": 9487 }, { "epoch": 4.297101449275362, "grad_norm": 0.05045546098942286, "learning_rate": 4.7201689157755935e-06, "loss": 0.9221, "step": 9488 }, { "epoch": 4.297554347826087, "grad_norm": 0.04967136007464626, "learning_rate": 4.71421054050512e-06, "loss": 0.9064, "step": 9489 }, { "epoch": 4.2980072463768115, "grad_norm": 0.046531030430631365, "learning_rate": 4.7082556928109435e-06, "loss": 0.9125, "step": 9490 }, { "epoch": 4.298460144927536, "grad_norm": 0.05234316958758019, "learning_rate": 4.702304373288367e-06, "loss": 0.9278, "step": 9491 }, { "epoch": 4.298913043478261, "grad_norm": 0.04527319559636484, "learning_rate": 4.6963565825323576e-06, "loss": 0.9159, "step": 9492 }, { "epoch": 4.299365942028985, "grad_norm": 0.04296601742735339, "learning_rate": 4.690412321137521e-06, "loss": 0.9324, "step": 9493 }, { "epoch": 4.29981884057971, "grad_norm": 0.04999545586980462, "learning_rate": 4.684471589698119e-06, "loss": 0.9138, "step": 9494 }, { "epoch": 4.300271739130435, "grad_norm": 0.04865969907413367, "learning_rate": 4.678534388808049e-06, "loss": 0.9345, "step": 9495 }, { "epoch": 4.300724637681159, "grad_norm": 0.041703817484399684, "learning_rate": 4.672600719060878e-06, "loss": 0.9241, "step": 9496 }, { "epoch": 4.301177536231884, "grad_norm": 0.04722024667550518, "learning_rate": 4.666670581049784e-06, "loss": 0.9174, "step": 9497 }, { "epoch": 4.301630434782608, "grad_norm": 0.049155887737429, "learning_rate": 4.66074397536763e-06, "loss": 0.9076, "step": 9498 }, { "epoch": 4.302083333333333, "grad_norm": 0.04933187975395464, "learning_rate": 4.654820902606898e-06, "loss": 0.9081, "step": 9499 }, { "epoch": 4.302536231884058, "grad_norm": 0.04327143522693924, "learning_rate": 4.648901363359737e-06, "loss": 0.9253, "step": 9500 }, { "epoch": 4.302989130434782, "grad_norm": 0.04659499090688357, "learning_rate": 4.642985358217913e-06, "loss": 0.9238, "step": 9501 }, { "epoch": 4.303442028985507, "grad_norm": 0.04588128340503536, "learning_rate": 4.637072887772877e-06, "loss": 0.9466, "step": 9502 }, { "epoch": 4.3038949275362315, "grad_norm": 0.0446961657191757, "learning_rate": 4.631163952615696e-06, "loss": 0.8906, "step": 9503 }, { "epoch": 4.304347826086957, "grad_norm": 0.046723969883586194, "learning_rate": 4.625258553337108e-06, "loss": 0.9328, "step": 9504 }, { "epoch": 4.304800724637682, "grad_norm": 0.043549959288206, "learning_rate": 4.61935669052747e-06, "loss": 0.9284, "step": 9505 }, { "epoch": 4.305253623188406, "grad_norm": 0.04637282100685613, "learning_rate": 4.613458364776819e-06, "loss": 0.9024, "step": 9506 }, { "epoch": 4.305706521739131, "grad_norm": 0.0483319430197605, "learning_rate": 4.60756357667481e-06, "loss": 0.8938, "step": 9507 }, { "epoch": 4.306159420289855, "grad_norm": 0.046980886349045146, "learning_rate": 4.6016723268107555e-06, "loss": 0.9088, "step": 9508 }, { "epoch": 4.30661231884058, "grad_norm": 0.03990319382855315, "learning_rate": 4.595784615773599e-06, "loss": 0.921, "step": 9509 }, { "epoch": 4.307065217391305, "grad_norm": 0.04672219261243195, "learning_rate": 4.589900444151973e-06, "loss": 0.9116, "step": 9510 }, { "epoch": 4.307518115942029, "grad_norm": 0.04716396085637045, "learning_rate": 4.5840198125341e-06, "loss": 0.9064, "step": 9511 }, { "epoch": 4.307971014492754, "grad_norm": 0.043546113608978476, "learning_rate": 4.578142721507894e-06, "loss": 0.9123, "step": 9512 }, { "epoch": 4.3084239130434785, "grad_norm": 0.04360434517354263, "learning_rate": 4.572269171660883e-06, "loss": 0.9281, "step": 9513 }, { "epoch": 4.308876811594203, "grad_norm": 0.04456581784180386, "learning_rate": 4.566399163580273e-06, "loss": 0.9236, "step": 9514 }, { "epoch": 4.309329710144928, "grad_norm": 0.0427399894089927, "learning_rate": 4.560532697852886e-06, "loss": 0.9129, "step": 9515 }, { "epoch": 4.309782608695652, "grad_norm": 0.04455492919968616, "learning_rate": 4.554669775065206e-06, "loss": 0.937, "step": 9516 }, { "epoch": 4.310235507246377, "grad_norm": 0.043139458300152096, "learning_rate": 4.548810395803345e-06, "loss": 0.8801, "step": 9517 }, { "epoch": 4.3106884057971016, "grad_norm": 0.04311994946701077, "learning_rate": 4.542954560653097e-06, "loss": 0.9322, "step": 9518 }, { "epoch": 4.311141304347826, "grad_norm": 0.04331975080882622, "learning_rate": 4.53710227019986e-06, "loss": 0.938, "step": 9519 }, { "epoch": 4.311594202898551, "grad_norm": 0.04339432387889823, "learning_rate": 4.5312535250287136e-06, "loss": 0.9302, "step": 9520 }, { "epoch": 4.312047101449275, "grad_norm": 0.04375960474184896, "learning_rate": 4.525408325724346e-06, "loss": 0.8993, "step": 9521 }, { "epoch": 4.3125, "grad_norm": 0.04387837826752962, "learning_rate": 4.519566672871132e-06, "loss": 0.9184, "step": 9522 }, { "epoch": 4.312952898550725, "grad_norm": 0.04770518183323944, "learning_rate": 4.513728567053064e-06, "loss": 0.9284, "step": 9523 }, { "epoch": 4.313405797101449, "grad_norm": 0.04231840940585805, "learning_rate": 4.507894008853781e-06, "loss": 0.9211, "step": 9524 }, { "epoch": 4.313858695652174, "grad_norm": 0.044949889192802954, "learning_rate": 4.502062998856569e-06, "loss": 0.9139, "step": 9525 }, { "epoch": 4.3143115942028984, "grad_norm": 0.04621032290111854, "learning_rate": 4.496235537644378e-06, "loss": 0.9327, "step": 9526 }, { "epoch": 4.314764492753623, "grad_norm": 0.04265645619505876, "learning_rate": 4.490411625799769e-06, "loss": 0.9293, "step": 9527 }, { "epoch": 4.315217391304348, "grad_norm": 0.045853354097714995, "learning_rate": 4.484591263904992e-06, "loss": 0.9114, "step": 9528 }, { "epoch": 4.315670289855072, "grad_norm": 0.048794394168911216, "learning_rate": 4.478774452541896e-06, "loss": 0.9117, "step": 9529 }, { "epoch": 4.316123188405797, "grad_norm": 0.04134016008878836, "learning_rate": 4.472961192292018e-06, "loss": 0.8966, "step": 9530 }, { "epoch": 4.3165760869565215, "grad_norm": 0.04671638293119816, "learning_rate": 4.467151483736496e-06, "loss": 0.9298, "step": 9531 }, { "epoch": 4.317028985507246, "grad_norm": 0.042724179317113226, "learning_rate": 4.461345327456159e-06, "loss": 0.9179, "step": 9532 }, { "epoch": 4.317481884057971, "grad_norm": 0.04244747467127619, "learning_rate": 4.455542724031436e-06, "loss": 0.9178, "step": 9533 }, { "epoch": 4.317934782608695, "grad_norm": 0.04094020383714698, "learning_rate": 4.449743674042437e-06, "loss": 0.9013, "step": 9534 }, { "epoch": 4.31838768115942, "grad_norm": 0.042663363052542196, "learning_rate": 4.443948178068888e-06, "loss": 0.9373, "step": 9535 }, { "epoch": 4.318840579710145, "grad_norm": 0.04278112230281241, "learning_rate": 4.438156236690194e-06, "loss": 0.9147, "step": 9536 }, { "epoch": 4.319293478260869, "grad_norm": 0.041864460235444136, "learning_rate": 4.432367850485362e-06, "loss": 0.9256, "step": 9537 }, { "epoch": 4.319746376811594, "grad_norm": 0.04208499866400806, "learning_rate": 4.426583020033088e-06, "loss": 0.915, "step": 9538 }, { "epoch": 4.320199275362318, "grad_norm": 0.04773899420727591, "learning_rate": 4.420801745911672e-06, "loss": 0.9072, "step": 9539 }, { "epoch": 4.320652173913043, "grad_norm": 0.046571078911755596, "learning_rate": 4.415024028699093e-06, "loss": 0.9103, "step": 9540 }, { "epoch": 4.3211050724637685, "grad_norm": 0.041896080477392905, "learning_rate": 4.409249868972954e-06, "loss": 0.9058, "step": 9541 }, { "epoch": 4.321557971014493, "grad_norm": 0.04068831832160625, "learning_rate": 4.4034792673104974e-06, "loss": 0.9247, "step": 9542 }, { "epoch": 4.322010869565218, "grad_norm": 0.042295672878351336, "learning_rate": 4.397712224288624e-06, "loss": 0.9206, "step": 9543 }, { "epoch": 4.322463768115942, "grad_norm": 0.04416288069479975, "learning_rate": 4.39194874048388e-06, "loss": 0.9313, "step": 9544 }, { "epoch": 4.322916666666667, "grad_norm": 0.04340436411695944, "learning_rate": 4.386188816472441e-06, "loss": 0.9002, "step": 9545 }, { "epoch": 4.323369565217392, "grad_norm": 0.04205797356760906, "learning_rate": 4.380432452830148e-06, "loss": 0.9288, "step": 9546 }, { "epoch": 4.323822463768116, "grad_norm": 0.04556621411684935, "learning_rate": 4.374679650132461e-06, "loss": 0.9306, "step": 9547 }, { "epoch": 4.324275362318841, "grad_norm": 0.04391060455148062, "learning_rate": 4.36893040895451e-06, "loss": 0.9131, "step": 9548 }, { "epoch": 4.324728260869565, "grad_norm": 0.042719971483731696, "learning_rate": 4.363184729871046e-06, "loss": 0.9368, "step": 9549 }, { "epoch": 4.32518115942029, "grad_norm": 0.04527415658004791, "learning_rate": 4.357442613456475e-06, "loss": 0.9154, "step": 9550 }, { "epoch": 4.325634057971015, "grad_norm": 0.04686003136235214, "learning_rate": 4.35170406028485e-06, "loss": 0.9398, "step": 9551 }, { "epoch": 4.326086956521739, "grad_norm": 0.043054201978436145, "learning_rate": 4.345969070929848e-06, "loss": 0.9031, "step": 9552 }, { "epoch": 4.326539855072464, "grad_norm": 0.04075281145583729, "learning_rate": 4.340237645964829e-06, "loss": 0.9336, "step": 9553 }, { "epoch": 4.3269927536231885, "grad_norm": 0.04354796654836161, "learning_rate": 4.334509785962749e-06, "loss": 0.9143, "step": 9554 }, { "epoch": 4.327445652173913, "grad_norm": 0.042770179662169845, "learning_rate": 4.32878549149625e-06, "loss": 0.8898, "step": 9555 }, { "epoch": 4.327898550724638, "grad_norm": 0.04214151421793354, "learning_rate": 4.323064763137587e-06, "loss": 0.9118, "step": 9556 }, { "epoch": 4.328351449275362, "grad_norm": 0.041852423696810435, "learning_rate": 4.317347601458677e-06, "loss": 0.929, "step": 9557 }, { "epoch": 4.328804347826087, "grad_norm": 0.04509461306989833, "learning_rate": 4.311634007031078e-06, "loss": 0.9194, "step": 9558 }, { "epoch": 4.3292572463768115, "grad_norm": 0.04377497847952206, "learning_rate": 4.3059239804259744e-06, "loss": 0.9137, "step": 9559 }, { "epoch": 4.329710144927536, "grad_norm": 0.04187074554750641, "learning_rate": 4.300217522214208e-06, "loss": 0.9156, "step": 9560 }, { "epoch": 4.330163043478261, "grad_norm": 0.04288747938989028, "learning_rate": 4.2945146329662756e-06, "loss": 0.9009, "step": 9561 }, { "epoch": 4.330615942028985, "grad_norm": 0.04034365231093875, "learning_rate": 4.288815313252284e-06, "loss": 0.9174, "step": 9562 }, { "epoch": 4.33106884057971, "grad_norm": 0.04113047426221731, "learning_rate": 4.283119563642024e-06, "loss": 0.9193, "step": 9563 }, { "epoch": 4.331521739130435, "grad_norm": 0.04140519966755668, "learning_rate": 4.277427384704895e-06, "loss": 0.9359, "step": 9564 }, { "epoch": 4.331974637681159, "grad_norm": 0.04085975129650327, "learning_rate": 4.271738777009962e-06, "loss": 0.9047, "step": 9565 }, { "epoch": 4.332427536231884, "grad_norm": 0.04453709369008602, "learning_rate": 4.266053741125919e-06, "loss": 0.9249, "step": 9566 }, { "epoch": 4.332880434782608, "grad_norm": 0.0424551336940926, "learning_rate": 4.260372277621114e-06, "loss": 0.9189, "step": 9567 }, { "epoch": 4.333333333333333, "grad_norm": 0.04093842426358286, "learning_rate": 4.254694387063514e-06, "loss": 0.9225, "step": 9568 }, { "epoch": 4.333786231884058, "grad_norm": 0.04187826452768484, "learning_rate": 4.249020070020766e-06, "loss": 0.9219, "step": 9569 }, { "epoch": 4.334239130434782, "grad_norm": 0.04142509192638549, "learning_rate": 4.243349327060129e-06, "loss": 0.9127, "step": 9570 }, { "epoch": 4.334692028985507, "grad_norm": 0.042514751271189324, "learning_rate": 4.237682158748531e-06, "loss": 0.903, "step": 9571 }, { "epoch": 4.3351449275362315, "grad_norm": 0.04101207396353202, "learning_rate": 4.232018565652505e-06, "loss": 0.9242, "step": 9572 }, { "epoch": 4.335597826086957, "grad_norm": 0.04382741984593741, "learning_rate": 4.226358548338274e-06, "loss": 0.9423, "step": 9573 }, { "epoch": 4.336050724637682, "grad_norm": 0.03969709507459621, "learning_rate": 4.2207021073716656e-06, "loss": 0.9215, "step": 9574 }, { "epoch": 4.336503623188406, "grad_norm": 0.04111351506476282, "learning_rate": 4.215049243318165e-06, "loss": 0.9137, "step": 9575 }, { "epoch": 4.336956521739131, "grad_norm": 0.043163553666620495, "learning_rate": 4.20939995674289e-06, "loss": 0.9066, "step": 9576 }, { "epoch": 4.337409420289855, "grad_norm": 0.04611103378208283, "learning_rate": 4.203754248210623e-06, "loss": 0.9154, "step": 9577 }, { "epoch": 4.33786231884058, "grad_norm": 0.04101265818260614, "learning_rate": 4.1981121182857624e-06, "loss": 0.8995, "step": 9578 }, { "epoch": 4.338315217391305, "grad_norm": 0.04170013266706157, "learning_rate": 4.192473567532372e-06, "loss": 0.918, "step": 9579 }, { "epoch": 4.338768115942029, "grad_norm": 0.04032797436730328, "learning_rate": 4.186838596514133e-06, "loss": 0.9342, "step": 9580 }, { "epoch": 4.339221014492754, "grad_norm": 0.045540967670757646, "learning_rate": 4.181207205794397e-06, "loss": 0.9022, "step": 9581 }, { "epoch": 4.3396739130434785, "grad_norm": 0.04141845680048078, "learning_rate": 4.175579395936131e-06, "loss": 0.9249, "step": 9582 }, { "epoch": 4.340126811594203, "grad_norm": 0.04248638955430333, "learning_rate": 4.169955167501969e-06, "loss": 0.9369, "step": 9583 }, { "epoch": 4.340579710144928, "grad_norm": 0.04469207376893668, "learning_rate": 4.164334521054167e-06, "loss": 0.9163, "step": 9584 }, { "epoch": 4.341032608695652, "grad_norm": 0.041776566168536665, "learning_rate": 4.158717457154628e-06, "loss": 0.9187, "step": 9585 }, { "epoch": 4.341485507246377, "grad_norm": 0.04056309237546835, "learning_rate": 4.153103976364894e-06, "loss": 0.9186, "step": 9586 }, { "epoch": 4.3419384057971016, "grad_norm": 0.04485189210250384, "learning_rate": 4.147494079246164e-06, "loss": 0.9286, "step": 9587 }, { "epoch": 4.342391304347826, "grad_norm": 0.04376474241257088, "learning_rate": 4.1418877663592605e-06, "loss": 0.9198, "step": 9588 }, { "epoch": 4.342844202898551, "grad_norm": 0.044915862379441035, "learning_rate": 4.136285038264665e-06, "loss": 0.9469, "step": 9589 }, { "epoch": 4.343297101449275, "grad_norm": 0.04599863690539296, "learning_rate": 4.130685895522479e-06, "loss": 0.9339, "step": 9590 }, { "epoch": 4.34375, "grad_norm": 0.04306018342940915, "learning_rate": 4.12509033869247e-06, "loss": 0.9141, "step": 9591 }, { "epoch": 4.344202898550725, "grad_norm": 0.04576955397661237, "learning_rate": 4.119498368334029e-06, "loss": 0.9245, "step": 9592 }, { "epoch": 4.344655797101449, "grad_norm": 0.04662048344476137, "learning_rate": 4.11390998500619e-06, "loss": 0.9189, "step": 9593 }, { "epoch": 4.345108695652174, "grad_norm": 0.04649504564881133, "learning_rate": 4.108325189267634e-06, "loss": 0.9084, "step": 9594 }, { "epoch": 4.3455615942028984, "grad_norm": 0.04189986808176013, "learning_rate": 4.102743981676689e-06, "loss": 0.9295, "step": 9595 }, { "epoch": 4.346014492753623, "grad_norm": 0.04365745528033013, "learning_rate": 4.0971663627913025e-06, "loss": 0.9138, "step": 9596 }, { "epoch": 4.346467391304348, "grad_norm": 0.049976702604087035, "learning_rate": 4.091592333169101e-06, "loss": 0.9281, "step": 9597 }, { "epoch": 4.346920289855072, "grad_norm": 0.04489140897131066, "learning_rate": 4.086021893367305e-06, "loss": 0.9198, "step": 9598 }, { "epoch": 4.347373188405797, "grad_norm": 0.043543911477880226, "learning_rate": 4.0804550439428195e-06, "loss": 0.9267, "step": 9599 }, { "epoch": 4.3478260869565215, "grad_norm": 0.04633643721307111, "learning_rate": 4.07489178545216e-06, "loss": 0.9124, "step": 9600 }, { "epoch": 4.348278985507246, "grad_norm": 0.04880135585974273, "learning_rate": 4.069332118451499e-06, "loss": 0.9225, "step": 9601 }, { "epoch": 4.348731884057971, "grad_norm": 0.044107552486430564, "learning_rate": 4.063776043496632e-06, "loss": 0.9139, "step": 9602 }, { "epoch": 4.349184782608695, "grad_norm": 0.045107088219106195, "learning_rate": 4.058223561143031e-06, "loss": 0.9163, "step": 9603 }, { "epoch": 4.34963768115942, "grad_norm": 0.049869115233605404, "learning_rate": 4.052674671945767e-06, "loss": 0.9259, "step": 9604 }, { "epoch": 4.350090579710145, "grad_norm": 0.04517336876788795, "learning_rate": 4.047129376459586e-06, "loss": 0.9246, "step": 9605 }, { "epoch": 4.350543478260869, "grad_norm": 0.041044370783719766, "learning_rate": 4.041587675238847e-06, "loss": 0.9031, "step": 9606 }, { "epoch": 4.350996376811594, "grad_norm": 0.04207860883561241, "learning_rate": 4.036049568837577e-06, "loss": 0.9416, "step": 9607 }, { "epoch": 4.351449275362318, "grad_norm": 0.04558202141350657, "learning_rate": 4.030515057809421e-06, "loss": 0.9301, "step": 9608 }, { "epoch": 4.351902173913043, "grad_norm": 0.04563451371883578, "learning_rate": 4.0249841427076756e-06, "loss": 0.9385, "step": 9609 }, { "epoch": 4.3523550724637685, "grad_norm": 0.04549977476321962, "learning_rate": 4.019456824085262e-06, "loss": 0.9207, "step": 9610 }, { "epoch": 4.352807971014493, "grad_norm": 0.04848947232403014, "learning_rate": 4.013933102494774e-06, "loss": 0.9108, "step": 9611 }, { "epoch": 4.353260869565218, "grad_norm": 0.04431151884157984, "learning_rate": 4.008412978488409e-06, "loss": 0.9335, "step": 9612 }, { "epoch": 4.353713768115942, "grad_norm": 0.04745815970499913, "learning_rate": 4.002896452618044e-06, "loss": 0.9218, "step": 9613 }, { "epoch": 4.354166666666667, "grad_norm": 0.04957123662983453, "learning_rate": 3.997383525435154e-06, "loss": 0.914, "step": 9614 }, { "epoch": 4.354619565217392, "grad_norm": 0.046714139352103244, "learning_rate": 3.991874197490888e-06, "loss": 0.9144, "step": 9615 }, { "epoch": 4.355072463768116, "grad_norm": 0.03917437971736447, "learning_rate": 3.986368469336022e-06, "loss": 0.9211, "step": 9616 }, { "epoch": 4.355525362318841, "grad_norm": 0.045034626064850036, "learning_rate": 3.980866341520968e-06, "loss": 0.9425, "step": 9617 }, { "epoch": 4.355978260869565, "grad_norm": 0.04513629359083036, "learning_rate": 3.975367814595776e-06, "loss": 0.9185, "step": 9618 }, { "epoch": 4.35643115942029, "grad_norm": 0.04709967434483765, "learning_rate": 3.969872889110153e-06, "loss": 0.9104, "step": 9619 }, { "epoch": 4.356884057971015, "grad_norm": 0.04239154682417617, "learning_rate": 3.9643815656134285e-06, "loss": 0.9099, "step": 9620 }, { "epoch": 4.357336956521739, "grad_norm": 0.04095177107324452, "learning_rate": 3.95889384465459e-06, "loss": 0.9201, "step": 9621 }, { "epoch": 4.357789855072464, "grad_norm": 0.044038081572952324, "learning_rate": 3.953409726782238e-06, "loss": 0.9201, "step": 9622 }, { "epoch": 4.3582427536231885, "grad_norm": 0.04407665831054526, "learning_rate": 3.947929212544641e-06, "loss": 0.8999, "step": 9623 }, { "epoch": 4.358695652173913, "grad_norm": 0.04936165905247283, "learning_rate": 3.942452302489686e-06, "loss": 0.9067, "step": 9624 }, { "epoch": 4.359148550724638, "grad_norm": 0.04166794201179558, "learning_rate": 3.936978997164915e-06, "loss": 0.9245, "step": 9625 }, { "epoch": 4.359601449275362, "grad_norm": 0.04567338728278904, "learning_rate": 3.931509297117506e-06, "loss": 0.9019, "step": 9626 }, { "epoch": 4.360054347826087, "grad_norm": 0.04893660087935092, "learning_rate": 3.926043202894265e-06, "loss": 0.9268, "step": 9627 }, { "epoch": 4.3605072463768115, "grad_norm": 0.042869108713543866, "learning_rate": 3.920580715041644e-06, "loss": 0.9185, "step": 9628 }, { "epoch": 4.360960144927536, "grad_norm": 0.0420621093903959, "learning_rate": 3.915121834105748e-06, "loss": 0.9286, "step": 9629 }, { "epoch": 4.361413043478261, "grad_norm": 0.044180313064583386, "learning_rate": 3.909666560632292e-06, "loss": 0.9337, "step": 9630 }, { "epoch": 4.361865942028985, "grad_norm": 0.04402989211065452, "learning_rate": 3.904214895166672e-06, "loss": 0.9405, "step": 9631 }, { "epoch": 4.36231884057971, "grad_norm": 0.04243907032478462, "learning_rate": 3.89876683825388e-06, "loss": 0.9112, "step": 9632 }, { "epoch": 4.362771739130435, "grad_norm": 0.041747478710506936, "learning_rate": 3.893322390438581e-06, "loss": 0.9239, "step": 9633 }, { "epoch": 4.363224637681159, "grad_norm": 0.04160403647939244, "learning_rate": 3.887881552265058e-06, "loss": 0.9333, "step": 9634 }, { "epoch": 4.363677536231884, "grad_norm": 0.04324654523653411, "learning_rate": 3.882444324277241e-06, "loss": 0.9372, "step": 9635 }, { "epoch": 4.364130434782608, "grad_norm": 0.041766033913830075, "learning_rate": 3.877010707018691e-06, "loss": 0.9254, "step": 9636 }, { "epoch": 4.364583333333333, "grad_norm": 0.04135439980288356, "learning_rate": 3.871580701032631e-06, "loss": 0.9368, "step": 9637 }, { "epoch": 4.365036231884058, "grad_norm": 0.04329211476674673, "learning_rate": 3.866154306861889e-06, "loss": 0.9279, "step": 9638 }, { "epoch": 4.365489130434782, "grad_norm": 0.04443346839107116, "learning_rate": 3.860731525048969e-06, "loss": 0.9171, "step": 9639 }, { "epoch": 4.365942028985507, "grad_norm": 0.04376372741218375, "learning_rate": 3.855312356135978e-06, "loss": 0.8898, "step": 9640 }, { "epoch": 4.3663949275362315, "grad_norm": 0.04056645329993594, "learning_rate": 3.849896800664694e-06, "loss": 0.9227, "step": 9641 }, { "epoch": 4.366847826086957, "grad_norm": 0.07612969503916056, "learning_rate": 3.844484859176514e-06, "loss": 0.9289, "step": 9642 }, { "epoch": 4.367300724637682, "grad_norm": 0.04091400415448403, "learning_rate": 3.839076532212476e-06, "loss": 0.9455, "step": 9643 }, { "epoch": 4.367753623188406, "grad_norm": 0.04499337263145185, "learning_rate": 3.83367182031325e-06, "loss": 0.9311, "step": 9644 }, { "epoch": 4.368206521739131, "grad_norm": 0.04413453130559542, "learning_rate": 3.828270724019167e-06, "loss": 0.9181, "step": 9645 }, { "epoch": 4.368659420289855, "grad_norm": 0.04271508210217735, "learning_rate": 3.8228732438701755e-06, "loss": 0.9218, "step": 9646 }, { "epoch": 4.36911231884058, "grad_norm": 0.041380720534006175, "learning_rate": 3.8174793804058775e-06, "loss": 0.9206, "step": 9647 }, { "epoch": 4.369565217391305, "grad_norm": 0.040405381687363344, "learning_rate": 3.8120891341654955e-06, "loss": 0.9273, "step": 9648 }, { "epoch": 4.370018115942029, "grad_norm": 0.04360734360556548, "learning_rate": 3.8067025056879182e-06, "loss": 0.9179, "step": 9649 }, { "epoch": 4.370471014492754, "grad_norm": 0.046988856943589945, "learning_rate": 3.8013194955116397e-06, "loss": 0.9178, "step": 9650 }, { "epoch": 4.3709239130434785, "grad_norm": 0.04267648031387622, "learning_rate": 3.7959401041748155e-06, "loss": 0.9003, "step": 9651 }, { "epoch": 4.371376811594203, "grad_norm": 0.04468521603036157, "learning_rate": 3.790564332215225e-06, "loss": 0.9237, "step": 9652 }, { "epoch": 4.371829710144928, "grad_norm": 0.04797177149200086, "learning_rate": 3.785192180170301e-06, "loss": 0.925, "step": 9653 }, { "epoch": 4.372282608695652, "grad_norm": 0.04508040280256609, "learning_rate": 3.7798236485770925e-06, "loss": 0.9208, "step": 9654 }, { "epoch": 4.372735507246377, "grad_norm": 0.04794776630907515, "learning_rate": 3.774458737972322e-06, "loss": 0.9341, "step": 9655 }, { "epoch": 4.3731884057971016, "grad_norm": 0.04141957486555887, "learning_rate": 3.7690974488923027e-06, "loss": 0.9248, "step": 9656 }, { "epoch": 4.373641304347826, "grad_norm": 0.041207478885498854, "learning_rate": 3.763739781873037e-06, "loss": 0.9137, "step": 9657 }, { "epoch": 4.374094202898551, "grad_norm": 0.04550958656553675, "learning_rate": 3.758385737450123e-06, "loss": 0.9181, "step": 9658 }, { "epoch": 4.374547101449275, "grad_norm": 0.04592695376664755, "learning_rate": 3.7530353161588175e-06, "loss": 0.9137, "step": 9659 }, { "epoch": 4.375, "grad_norm": 0.04467367951619694, "learning_rate": 3.747688518534003e-06, "loss": 0.9294, "step": 9660 }, { "epoch": 4.375452898550725, "grad_norm": 0.042509019714187815, "learning_rate": 3.742345345110221e-06, "loss": 0.9367, "step": 9661 }, { "epoch": 4.375905797101449, "grad_norm": 0.04229372027092109, "learning_rate": 3.7370057964216223e-06, "loss": 0.9046, "step": 9662 }, { "epoch": 4.376358695652174, "grad_norm": 0.04311076779404793, "learning_rate": 3.731669873002024e-06, "loss": 0.9221, "step": 9663 }, { "epoch": 4.3768115942028984, "grad_norm": 0.045261993022962715, "learning_rate": 3.7263375753848573e-06, "loss": 0.9167, "step": 9664 }, { "epoch": 4.377264492753623, "grad_norm": 0.04820646245724673, "learning_rate": 3.7210089041032073e-06, "loss": 0.9217, "step": 9665 }, { "epoch": 4.377717391304348, "grad_norm": 0.045402639763739545, "learning_rate": 3.7156838596897803e-06, "loss": 0.9206, "step": 9666 }, { "epoch": 4.378170289855072, "grad_norm": 0.043620037804805195, "learning_rate": 3.7103624426769424e-06, "loss": 0.9305, "step": 9667 }, { "epoch": 4.378623188405797, "grad_norm": 0.04742727792149184, "learning_rate": 3.7050446535966765e-06, "loss": 0.9373, "step": 9668 }, { "epoch": 4.3790760869565215, "grad_norm": 0.043105422845421015, "learning_rate": 3.699730492980611e-06, "loss": 0.9541, "step": 9669 }, { "epoch": 4.379528985507246, "grad_norm": 0.042828203797060274, "learning_rate": 3.694419961360005e-06, "loss": 0.9185, "step": 9670 }, { "epoch": 4.379981884057971, "grad_norm": 0.04533477887125144, "learning_rate": 3.689113059265772e-06, "loss": 0.9137, "step": 9671 }, { "epoch": 4.380434782608695, "grad_norm": 0.04418940473181836, "learning_rate": 3.6838097872284382e-06, "loss": 0.8923, "step": 9672 }, { "epoch": 4.38088768115942, "grad_norm": 0.04316387931827334, "learning_rate": 3.678510145778198e-06, "loss": 0.9357, "step": 9673 }, { "epoch": 4.381340579710145, "grad_norm": 0.041476857601395856, "learning_rate": 3.6732141354448536e-06, "loss": 0.9053, "step": 9674 }, { "epoch": 4.381793478260869, "grad_norm": 0.042434726123934195, "learning_rate": 3.6679217567578486e-06, "loss": 0.9173, "step": 9675 }, { "epoch": 4.382246376811594, "grad_norm": 0.0442843724619016, "learning_rate": 3.662633010246288e-06, "loss": 0.9143, "step": 9676 }, { "epoch": 4.382699275362318, "grad_norm": 0.043057271995002394, "learning_rate": 3.657347896438883e-06, "loss": 0.9142, "step": 9677 }, { "epoch": 4.383152173913043, "grad_norm": 0.04269475707224895, "learning_rate": 3.652066415864002e-06, "loss": 0.9117, "step": 9678 }, { "epoch": 4.3836050724637685, "grad_norm": 0.04107937408999397, "learning_rate": 3.646788569049631e-06, "loss": 0.9132, "step": 9679 }, { "epoch": 4.384057971014493, "grad_norm": 0.042378259683658256, "learning_rate": 3.641514356523419e-06, "loss": 0.8882, "step": 9680 }, { "epoch": 4.384510869565218, "grad_norm": 0.04431446049267093, "learning_rate": 3.6362437788126292e-06, "loss": 0.9348, "step": 9681 }, { "epoch": 4.384963768115942, "grad_norm": 0.04176276346802351, "learning_rate": 3.6309768364441776e-06, "loss": 0.9164, "step": 9682 }, { "epoch": 4.385416666666667, "grad_norm": 0.040969874038813865, "learning_rate": 3.6257135299445943e-06, "loss": 0.9105, "step": 9683 }, { "epoch": 4.385869565217392, "grad_norm": 0.042158774964811865, "learning_rate": 3.620453859840076e-06, "loss": 0.9131, "step": 9684 }, { "epoch": 4.386322463768116, "grad_norm": 0.04334882269356182, "learning_rate": 3.615197826656438e-06, "loss": 0.9223, "step": 9685 }, { "epoch": 4.386775362318841, "grad_norm": 0.04540313481313408, "learning_rate": 3.6099454309191262e-06, "loss": 0.9289, "step": 9686 }, { "epoch": 4.387228260869565, "grad_norm": 0.03923510737549273, "learning_rate": 3.6046966731532275e-06, "loss": 0.9159, "step": 9687 }, { "epoch": 4.38768115942029, "grad_norm": 0.04138860010776083, "learning_rate": 3.599451553883482e-06, "loss": 0.9328, "step": 9688 }, { "epoch": 4.388134057971015, "grad_norm": 0.04436237528966916, "learning_rate": 3.5942100736342433e-06, "loss": 0.9314, "step": 9689 }, { "epoch": 4.388586956521739, "grad_norm": 0.041183743283340776, "learning_rate": 3.5889722329295152e-06, "loss": 0.9372, "step": 9690 }, { "epoch": 4.389039855072464, "grad_norm": 0.04133670901671538, "learning_rate": 3.5837380322929273e-06, "loss": 0.9231, "step": 9691 }, { "epoch": 4.3894927536231885, "grad_norm": 0.04209214700258309, "learning_rate": 3.578507472247763e-06, "loss": 0.9019, "step": 9692 }, { "epoch": 4.389945652173913, "grad_norm": 0.04031640615066562, "learning_rate": 3.57328055331692e-06, "loss": 0.9113, "step": 9693 }, { "epoch": 4.390398550724638, "grad_norm": 0.04815005404683151, "learning_rate": 3.5680572760229446e-06, "loss": 0.9221, "step": 9694 }, { "epoch": 4.390851449275362, "grad_norm": 0.042379414283246285, "learning_rate": 3.5628376408880062e-06, "loss": 0.8923, "step": 9695 }, { "epoch": 4.391304347826087, "grad_norm": 0.04392253793690465, "learning_rate": 3.5576216484339356e-06, "loss": 0.9048, "step": 9696 }, { "epoch": 4.3917572463768115, "grad_norm": 0.04151082919697397, "learning_rate": 3.552409299182169e-06, "loss": 0.945, "step": 9697 }, { "epoch": 4.392210144927536, "grad_norm": 0.04532031327448144, "learning_rate": 3.54720059365381e-06, "loss": 0.9311, "step": 9698 }, { "epoch": 4.392663043478261, "grad_norm": 0.044507410099288394, "learning_rate": 3.541995532369562e-06, "loss": 0.9079, "step": 9699 }, { "epoch": 4.393115942028985, "grad_norm": 0.042024346597855526, "learning_rate": 3.536794115849804e-06, "loss": 0.9117, "step": 9700 }, { "epoch": 4.39356884057971, "grad_norm": 0.04243964816275726, "learning_rate": 3.531596344614516e-06, "loss": 0.9221, "step": 9701 }, { "epoch": 4.394021739130435, "grad_norm": 0.040773669388735265, "learning_rate": 3.5264022191833313e-06, "loss": 0.9182, "step": 9702 }, { "epoch": 4.394474637681159, "grad_norm": 0.03909535650479815, "learning_rate": 3.5212117400755097e-06, "loss": 0.8995, "step": 9703 }, { "epoch": 4.394927536231884, "grad_norm": 0.04471025670162706, "learning_rate": 3.5160249078099606e-06, "loss": 0.9109, "step": 9704 }, { "epoch": 4.395380434782608, "grad_norm": 0.04054302164522414, "learning_rate": 3.510841722905207e-06, "loss": 0.9383, "step": 9705 }, { "epoch": 4.395833333333333, "grad_norm": 0.042943364488866896, "learning_rate": 3.5056621858794393e-06, "loss": 0.9379, "step": 9706 }, { "epoch": 4.396286231884058, "grad_norm": 0.04280118029226939, "learning_rate": 3.5004862972504426e-06, "loss": 0.9051, "step": 9707 }, { "epoch": 4.396739130434782, "grad_norm": 0.04259121445184882, "learning_rate": 3.495314057535679e-06, "loss": 0.9197, "step": 9708 }, { "epoch": 4.397192028985507, "grad_norm": 0.04154023753396887, "learning_rate": 3.4901454672522063e-06, "loss": 0.9283, "step": 9709 }, { "epoch": 4.3976449275362315, "grad_norm": 0.041375743838625155, "learning_rate": 3.4849805269167615e-06, "loss": 0.9251, "step": 9710 }, { "epoch": 4.398097826086957, "grad_norm": 0.045225741806755225, "learning_rate": 3.479819237045661e-06, "loss": 0.9162, "step": 9711 }, { "epoch": 4.398550724637682, "grad_norm": 0.07738133741935924, "learning_rate": 3.4746615981549094e-06, "loss": 0.9353, "step": 9712 }, { "epoch": 4.399003623188406, "grad_norm": 0.04211542480160807, "learning_rate": 3.4695076107601124e-06, "loss": 0.9121, "step": 9713 }, { "epoch": 4.399456521739131, "grad_norm": 0.04199431113750718, "learning_rate": 3.464357275376533e-06, "loss": 0.9224, "step": 9714 }, { "epoch": 4.399909420289855, "grad_norm": 0.043532035290331283, "learning_rate": 3.4592105925190446e-06, "loss": 0.9323, "step": 9715 }, { "epoch": 4.40036231884058, "grad_norm": 0.04209147296216725, "learning_rate": 3.4540675627021856e-06, "loss": 0.8928, "step": 9716 }, { "epoch": 4.400815217391305, "grad_norm": 0.04385002633703962, "learning_rate": 3.448928186440097e-06, "loss": 0.9461, "step": 9717 }, { "epoch": 4.401268115942029, "grad_norm": 0.04140748068195046, "learning_rate": 3.4437924642465846e-06, "loss": 0.9084, "step": 9718 }, { "epoch": 4.401721014492754, "grad_norm": 0.04892909099484071, "learning_rate": 3.438660396635065e-06, "loss": 0.9188, "step": 9719 }, { "epoch": 4.4021739130434785, "grad_norm": 0.046066872993294074, "learning_rate": 3.4335319841186077e-06, "loss": 0.9143, "step": 9720 }, { "epoch": 4.402626811594203, "grad_norm": 0.044097723203537414, "learning_rate": 3.428407227209891e-06, "loss": 0.9359, "step": 9721 }, { "epoch": 4.403079710144928, "grad_norm": 0.04505693967848357, "learning_rate": 3.4232861264212657e-06, "loss": 0.9182, "step": 9722 }, { "epoch": 4.403532608695652, "grad_norm": 0.0425015559248569, "learning_rate": 3.4181686822646777e-06, "loss": 0.9447, "step": 9723 }, { "epoch": 4.403985507246377, "grad_norm": 0.04389412945501769, "learning_rate": 3.413054895251744e-06, "loss": 0.9402, "step": 9724 }, { "epoch": 4.4044384057971016, "grad_norm": 0.04551422724841657, "learning_rate": 3.4079447658936825e-06, "loss": 0.9225, "step": 9725 }, { "epoch": 4.404891304347826, "grad_norm": 0.04251662752476926, "learning_rate": 3.402838294701374e-06, "loss": 0.8944, "step": 9726 }, { "epoch": 4.405344202898551, "grad_norm": 0.04553734938698583, "learning_rate": 3.397735482185316e-06, "loss": 0.9213, "step": 9727 }, { "epoch": 4.405797101449275, "grad_norm": 0.04197439554637481, "learning_rate": 3.392636328855643e-06, "loss": 0.9344, "step": 9728 }, { "epoch": 4.40625, "grad_norm": 0.03989603418786107, "learning_rate": 3.3875408352221164e-06, "loss": 0.9161, "step": 9729 }, { "epoch": 4.406702898550725, "grad_norm": 0.04282111815630622, "learning_rate": 3.3824490017941592e-06, "loss": 0.9254, "step": 9730 }, { "epoch": 4.407155797101449, "grad_norm": 0.046512658266823326, "learning_rate": 3.3773608290807914e-06, "loss": 0.9114, "step": 9731 }, { "epoch": 4.407608695652174, "grad_norm": 0.04043271022618968, "learning_rate": 3.3722763175907037e-06, "loss": 0.9288, "step": 9732 }, { "epoch": 4.4080615942028984, "grad_norm": 0.042300468957859896, "learning_rate": 3.3671954678321873e-06, "loss": 0.9157, "step": 9733 }, { "epoch": 4.408514492753623, "grad_norm": 0.04284726458475782, "learning_rate": 3.3621182803131957e-06, "loss": 0.9099, "step": 9734 }, { "epoch": 4.408967391304348, "grad_norm": 0.043077259416286255, "learning_rate": 3.3570447555412964e-06, "loss": 0.9328, "step": 9735 }, { "epoch": 4.409420289855072, "grad_norm": 0.04128663493008231, "learning_rate": 3.3519748940237017e-06, "loss": 0.9236, "step": 9736 }, { "epoch": 4.409873188405797, "grad_norm": 0.04134438135077055, "learning_rate": 3.3469086962672413e-06, "loss": 0.9242, "step": 9737 }, { "epoch": 4.4103260869565215, "grad_norm": 0.04026251994156679, "learning_rate": 3.3418461627784083e-06, "loss": 0.8902, "step": 9738 }, { "epoch": 4.410778985507246, "grad_norm": 0.04338325017258158, "learning_rate": 3.3367872940632952e-06, "loss": 0.9144, "step": 9739 }, { "epoch": 4.411231884057971, "grad_norm": 0.04064835180154265, "learning_rate": 3.331732090627666e-06, "loss": 0.9243, "step": 9740 }, { "epoch": 4.411684782608695, "grad_norm": 0.03883472730065705, "learning_rate": 3.326680552976873e-06, "loss": 0.9393, "step": 9741 }, { "epoch": 4.41213768115942, "grad_norm": 0.04399946783621468, "learning_rate": 3.3216326816159517e-06, "loss": 0.9312, "step": 9742 }, { "epoch": 4.412590579710145, "grad_norm": 0.041026507829469705, "learning_rate": 3.316588477049529e-06, "loss": 0.9223, "step": 9743 }, { "epoch": 4.413043478260869, "grad_norm": 0.040768126176698, "learning_rate": 3.311547939781887e-06, "loss": 0.9267, "step": 9744 }, { "epoch": 4.413496376811594, "grad_norm": 0.044413056684727156, "learning_rate": 3.3065110703169247e-06, "loss": 0.9532, "step": 9745 }, { "epoch": 4.413949275362318, "grad_norm": 0.03944552330865929, "learning_rate": 3.3014778691582073e-06, "loss": 0.913, "step": 9746 }, { "epoch": 4.414402173913043, "grad_norm": 0.0402256024218657, "learning_rate": 3.2964483368088973e-06, "loss": 0.9224, "step": 9747 }, { "epoch": 4.4148550724637685, "grad_norm": 0.04270441276189197, "learning_rate": 3.2914224737718105e-06, "loss": 0.9264, "step": 9748 }, { "epoch": 4.415307971014493, "grad_norm": 0.048788428970004225, "learning_rate": 3.2864002805493845e-06, "loss": 0.9192, "step": 9749 }, { "epoch": 4.415760869565218, "grad_norm": 0.04040546167765701, "learning_rate": 3.2813817576437068e-06, "loss": 0.9277, "step": 9750 }, { "epoch": 4.416213768115942, "grad_norm": 0.04037207848601015, "learning_rate": 3.2763669055564784e-06, "loss": 0.9388, "step": 9751 }, { "epoch": 4.416666666666667, "grad_norm": 0.043137215560445585, "learning_rate": 3.2713557247890447e-06, "loss": 0.8914, "step": 9752 }, { "epoch": 4.417119565217392, "grad_norm": 0.0458155002046612, "learning_rate": 3.2663482158423743e-06, "loss": 0.9212, "step": 9753 }, { "epoch": 4.417572463768116, "grad_norm": 0.04145880487976854, "learning_rate": 3.261344379217093e-06, "loss": 0.9157, "step": 9754 }, { "epoch": 4.418025362318841, "grad_norm": 0.04512199801497495, "learning_rate": 3.256344215413418e-06, "loss": 0.9197, "step": 9755 }, { "epoch": 4.418478260869565, "grad_norm": 0.04487479957687217, "learning_rate": 3.2513477249312488e-06, "loss": 0.923, "step": 9756 }, { "epoch": 4.41893115942029, "grad_norm": 0.04511186803476127, "learning_rate": 3.246354908270073e-06, "loss": 0.8909, "step": 9757 }, { "epoch": 4.419384057971015, "grad_norm": 0.0413298809065793, "learning_rate": 3.2413657659290433e-06, "loss": 0.9145, "step": 9758 }, { "epoch": 4.419836956521739, "grad_norm": 0.042561765412695844, "learning_rate": 3.2363802984069205e-06, "loss": 0.9234, "step": 9759 }, { "epoch": 4.420289855072464, "grad_norm": 0.046646106090679006, "learning_rate": 3.2313985062021236e-06, "loss": 0.9257, "step": 9760 }, { "epoch": 4.4207427536231885, "grad_norm": 0.04423449122463453, "learning_rate": 3.226420389812681e-06, "loss": 0.9045, "step": 9761 }, { "epoch": 4.421195652173913, "grad_norm": 0.04464817951139176, "learning_rate": 3.221445949736266e-06, "loss": 0.8919, "step": 9762 }, { "epoch": 4.421648550724638, "grad_norm": 0.04153027655400501, "learning_rate": 3.2164751864701736e-06, "loss": 0.9238, "step": 9763 }, { "epoch": 4.422101449275362, "grad_norm": 0.04511928719905976, "learning_rate": 3.2115081005113536e-06, "loss": 0.9188, "step": 9764 }, { "epoch": 4.422554347826087, "grad_norm": 0.04558760797574036, "learning_rate": 3.206544692356359e-06, "loss": 0.9146, "step": 9765 }, { "epoch": 4.4230072463768115, "grad_norm": 0.04265939976192522, "learning_rate": 3.2015849625013985e-06, "loss": 0.9374, "step": 9766 }, { "epoch": 4.423460144927536, "grad_norm": 0.04685555232022986, "learning_rate": 3.1966289114422964e-06, "loss": 0.9488, "step": 9767 }, { "epoch": 4.423913043478261, "grad_norm": 0.04268551907109459, "learning_rate": 3.1916765396745288e-06, "loss": 0.9203, "step": 9768 }, { "epoch": 4.424365942028985, "grad_norm": 0.04599048124067127, "learning_rate": 3.1867278476931874e-06, "loss": 0.9366, "step": 9769 }, { "epoch": 4.42481884057971, "grad_norm": 0.04150415400640876, "learning_rate": 3.181782835992997e-06, "loss": 0.9065, "step": 9770 }, { "epoch": 4.425271739130435, "grad_norm": 0.04255178732358897, "learning_rate": 3.176841505068313e-06, "loss": 0.9164, "step": 9771 }, { "epoch": 4.425724637681159, "grad_norm": 0.047573551870315915, "learning_rate": 3.171903855413141e-06, "loss": 0.932, "step": 9772 }, { "epoch": 4.426177536231884, "grad_norm": 0.04151621515872737, "learning_rate": 3.1669698875210943e-06, "loss": 0.9408, "step": 9773 }, { "epoch": 4.426630434782608, "grad_norm": 0.04250762499336011, "learning_rate": 3.1620396018854403e-06, "loss": 0.9101, "step": 9774 }, { "epoch": 4.427083333333333, "grad_norm": 0.04709477314157629, "learning_rate": 3.157112998999057e-06, "loss": 0.9294, "step": 9775 }, { "epoch": 4.427536231884058, "grad_norm": 0.041279271203225984, "learning_rate": 3.152190079354478e-06, "loss": 0.9305, "step": 9776 }, { "epoch": 4.427989130434782, "grad_norm": 0.03942813371104809, "learning_rate": 3.1472708434438437e-06, "loss": 0.9163, "step": 9777 }, { "epoch": 4.428442028985507, "grad_norm": 0.0393480511162431, "learning_rate": 3.1423552917589427e-06, "loss": 0.9042, "step": 9778 }, { "epoch": 4.4288949275362315, "grad_norm": 0.04357711166484602, "learning_rate": 3.1374434247911823e-06, "loss": 0.933, "step": 9779 }, { "epoch": 4.429347826086957, "grad_norm": 0.04368265418130268, "learning_rate": 3.1325352430316224e-06, "loss": 0.919, "step": 9780 }, { "epoch": 4.429800724637682, "grad_norm": 0.04371548778485909, "learning_rate": 3.1276307469709287e-06, "loss": 0.9227, "step": 9781 }, { "epoch": 4.430253623188406, "grad_norm": 0.046466059314412465, "learning_rate": 3.1227299370994246e-06, "loss": 0.9331, "step": 9782 }, { "epoch": 4.430706521739131, "grad_norm": 0.04101362281938477, "learning_rate": 3.1178328139070422e-06, "loss": 0.9269, "step": 9783 }, { "epoch": 4.431159420289855, "grad_norm": 0.04148865828628967, "learning_rate": 3.1129393778833638e-06, "loss": 0.9302, "step": 9784 }, { "epoch": 4.43161231884058, "grad_norm": 0.04274029044928363, "learning_rate": 3.108049629517589e-06, "loss": 0.9157, "step": 9785 }, { "epoch": 4.432065217391305, "grad_norm": 0.04353634081158974, "learning_rate": 3.103163569298553e-06, "loss": 0.9283, "step": 9786 }, { "epoch": 4.432518115942029, "grad_norm": 0.04425950271813747, "learning_rate": 3.098281197714714e-06, "loss": 0.9357, "step": 9787 }, { "epoch": 4.432971014492754, "grad_norm": 0.04342248024820919, "learning_rate": 3.0934025152541847e-06, "loss": 0.9194, "step": 9788 }, { "epoch": 4.4334239130434785, "grad_norm": 0.04086446505869579, "learning_rate": 3.088527522404685e-06, "loss": 0.9342, "step": 9789 }, { "epoch": 4.433876811594203, "grad_norm": 0.04065672495325063, "learning_rate": 3.08365621965359e-06, "loss": 0.9265, "step": 9790 }, { "epoch": 4.434329710144928, "grad_norm": 0.04034405289899904, "learning_rate": 3.07878860748787e-06, "loss": 0.8966, "step": 9791 }, { "epoch": 4.434782608695652, "grad_norm": 0.043904690508280705, "learning_rate": 3.0739246863941717e-06, "loss": 0.908, "step": 9792 }, { "epoch": 4.435235507246377, "grad_norm": 0.042163126536590406, "learning_rate": 3.0690644568587324e-06, "loss": 0.9262, "step": 9793 }, { "epoch": 4.4356884057971016, "grad_norm": 0.04226474508142952, "learning_rate": 3.0642079193674436e-06, "loss": 0.9414, "step": 9794 }, { "epoch": 4.436141304347826, "grad_norm": 0.0441395353684039, "learning_rate": 3.05935507440581e-06, "loss": 0.9282, "step": 9795 }, { "epoch": 4.436594202898551, "grad_norm": 0.04366865887498598, "learning_rate": 3.0545059224589947e-06, "loss": 0.9293, "step": 9796 }, { "epoch": 4.437047101449275, "grad_norm": 0.0434594954464609, "learning_rate": 3.049660464011761e-06, "loss": 0.9104, "step": 9797 }, { "epoch": 4.4375, "grad_norm": 0.04289892100336461, "learning_rate": 3.0448186995485307e-06, "loss": 0.9114, "step": 9798 }, { "epoch": 4.437952898550725, "grad_norm": 0.04214970290986748, "learning_rate": 3.039980629553334e-06, "loss": 0.904, "step": 9799 }, { "epoch": 4.438405797101449, "grad_norm": 0.042021237266610736, "learning_rate": 3.0351462545098374e-06, "loss": 0.9286, "step": 9800 }, { "epoch": 4.438858695652174, "grad_norm": 0.04314591662823096, "learning_rate": 3.0303155749013523e-06, "loss": 0.918, "step": 9801 }, { "epoch": 4.4393115942028984, "grad_norm": 0.043658179610311175, "learning_rate": 3.0254885912107946e-06, "loss": 0.9158, "step": 9802 }, { "epoch": 4.439764492753623, "grad_norm": 0.04238983485618411, "learning_rate": 3.0206653039207424e-06, "loss": 0.9122, "step": 9803 }, { "epoch": 4.440217391304348, "grad_norm": 0.042270038986285924, "learning_rate": 3.0158457135133745e-06, "loss": 0.9198, "step": 9804 }, { "epoch": 4.440670289855072, "grad_norm": 0.04574148926982373, "learning_rate": 3.011029820470519e-06, "loss": 0.9313, "step": 9805 }, { "epoch": 4.441123188405797, "grad_norm": 0.04433723764296025, "learning_rate": 3.006217625273622e-06, "loss": 0.9277, "step": 9806 }, { "epoch": 4.4415760869565215, "grad_norm": 0.04358339132705394, "learning_rate": 3.0014091284037737e-06, "loss": 0.9097, "step": 9807 }, { "epoch": 4.442028985507246, "grad_norm": 0.047460667056028874, "learning_rate": 2.996604330341679e-06, "loss": 0.9229, "step": 9808 }, { "epoch": 4.442481884057971, "grad_norm": 0.0446254513867483, "learning_rate": 2.9918032315676957e-06, "loss": 0.9068, "step": 9809 }, { "epoch": 4.442934782608695, "grad_norm": 0.04141191020075635, "learning_rate": 2.987005832561778e-06, "loss": 0.9192, "step": 9810 }, { "epoch": 4.44338768115942, "grad_norm": 0.0411918960545854, "learning_rate": 2.9822121338035457e-06, "loss": 0.9008, "step": 9811 }, { "epoch": 4.443840579710145, "grad_norm": 0.043868163433695075, "learning_rate": 2.9774221357722298e-06, "loss": 0.9122, "step": 9812 }, { "epoch": 4.444293478260869, "grad_norm": 0.051053413147772564, "learning_rate": 2.972635838946687e-06, "loss": 0.9051, "step": 9813 }, { "epoch": 4.444746376811594, "grad_norm": 0.04353682544923963, "learning_rate": 2.9678532438054096e-06, "loss": 0.9067, "step": 9814 }, { "epoch": 4.445199275362318, "grad_norm": 0.04060985093983426, "learning_rate": 2.9630743508265313e-06, "loss": 0.9206, "step": 9815 }, { "epoch": 4.445652173913043, "grad_norm": 0.040459277381352854, "learning_rate": 2.95829916048779e-06, "loss": 0.9115, "step": 9816 }, { "epoch": 4.4461050724637685, "grad_norm": 0.04169761663486849, "learning_rate": 2.9535276732665897e-06, "loss": 0.919, "step": 9817 }, { "epoch": 4.446557971014493, "grad_norm": 0.0450314330748827, "learning_rate": 2.948759889639927e-06, "loss": 0.9301, "step": 9818 }, { "epoch": 4.447010869565218, "grad_norm": 0.045858824434387574, "learning_rate": 2.9439958100844567e-06, "loss": 0.9338, "step": 9819 }, { "epoch": 4.447463768115942, "grad_norm": 0.04218409907018321, "learning_rate": 2.939235435076446e-06, "loss": 0.933, "step": 9820 }, { "epoch": 4.447916666666667, "grad_norm": 0.04249905761275686, "learning_rate": 2.934478765091795e-06, "loss": 0.9264, "step": 9821 }, { "epoch": 4.448369565217392, "grad_norm": 0.04365392293492062, "learning_rate": 2.9297258006060335e-06, "loss": 0.92, "step": 9822 }, { "epoch": 4.448822463768116, "grad_norm": 0.04401870510966664, "learning_rate": 2.9249765420943287e-06, "loss": 0.9377, "step": 9823 }, { "epoch": 4.449275362318841, "grad_norm": 0.042243961523434716, "learning_rate": 2.920230990031465e-06, "loss": 0.895, "step": 9824 }, { "epoch": 4.449728260869565, "grad_norm": 0.041367658873755084, "learning_rate": 2.9154891448918764e-06, "loss": 0.9345, "step": 9825 }, { "epoch": 4.45018115942029, "grad_norm": 0.04166415925102393, "learning_rate": 2.9107510071495928e-06, "loss": 0.9212, "step": 9826 }, { "epoch": 4.450634057971015, "grad_norm": 0.04377393034008896, "learning_rate": 2.9060165772783144e-06, "loss": 0.9372, "step": 9827 }, { "epoch": 4.451086956521739, "grad_norm": 0.04176545928336405, "learning_rate": 2.9012858557513344e-06, "loss": 0.9094, "step": 9828 }, { "epoch": 4.451539855072464, "grad_norm": 0.04278180196822356, "learning_rate": 2.8965588430415993e-06, "loss": 0.9202, "step": 9829 }, { "epoch": 4.4519927536231885, "grad_norm": 0.040777672087953576, "learning_rate": 2.8918355396216636e-06, "loss": 0.9204, "step": 9830 }, { "epoch": 4.452445652173913, "grad_norm": 0.042500848675750734, "learning_rate": 2.887115945963741e-06, "loss": 0.9175, "step": 9831 }, { "epoch": 4.452898550724638, "grad_norm": 0.04372515813805701, "learning_rate": 2.882400062539641e-06, "loss": 0.9284, "step": 9832 }, { "epoch": 4.453351449275362, "grad_norm": 0.04138802114032788, "learning_rate": 2.877687889820826e-06, "loss": 0.9109, "step": 9833 }, { "epoch": 4.453804347826087, "grad_norm": 0.0428091951880904, "learning_rate": 2.8729794282783776e-06, "loss": 0.9257, "step": 9834 }, { "epoch": 4.4542572463768115, "grad_norm": 0.03973321047522969, "learning_rate": 2.8682746783830118e-06, "loss": 0.9074, "step": 9835 }, { "epoch": 4.454710144927536, "grad_norm": 0.04335918372120389, "learning_rate": 2.8635736406050683e-06, "loss": 0.9261, "step": 9836 }, { "epoch": 4.455163043478261, "grad_norm": 0.0416995397184332, "learning_rate": 2.858876315414514e-06, "loss": 0.9151, "step": 9837 }, { "epoch": 4.455615942028985, "grad_norm": 0.04190476373378722, "learning_rate": 2.8541827032809453e-06, "loss": 0.9357, "step": 9838 }, { "epoch": 4.45606884057971, "grad_norm": 0.03969968848253955, "learning_rate": 2.849492804673597e-06, "loss": 0.9327, "step": 9839 }, { "epoch": 4.456521739130435, "grad_norm": 0.04343329652766427, "learning_rate": 2.84480662006132e-06, "loss": 0.9195, "step": 9840 }, { "epoch": 4.456974637681159, "grad_norm": 0.03959305494281249, "learning_rate": 2.8401241499126064e-06, "loss": 0.9215, "step": 9841 }, { "epoch": 4.457427536231884, "grad_norm": 0.039731352489869906, "learning_rate": 2.835445394695562e-06, "loss": 0.9213, "step": 9842 }, { "epoch": 4.457880434782608, "grad_norm": 0.038092647662898244, "learning_rate": 2.8307703548779406e-06, "loss": 0.9247, "step": 9843 }, { "epoch": 4.458333333333333, "grad_norm": 0.042551582659275784, "learning_rate": 2.8260990309270987e-06, "loss": 0.9253, "step": 9844 }, { "epoch": 4.458786231884058, "grad_norm": 0.042533818180032434, "learning_rate": 2.8214314233100526e-06, "loss": 0.8966, "step": 9845 }, { "epoch": 4.459239130434782, "grad_norm": 0.041247504256054446, "learning_rate": 2.8167675324934207e-06, "loss": 0.9391, "step": 9846 }, { "epoch": 4.459692028985507, "grad_norm": 0.03875843344102132, "learning_rate": 2.8121073589434566e-06, "loss": 0.908, "step": 9847 }, { "epoch": 4.4601449275362315, "grad_norm": 0.04204553979866974, "learning_rate": 2.8074509031260454e-06, "loss": 0.9179, "step": 9848 }, { "epoch": 4.460597826086957, "grad_norm": 0.041810391795283224, "learning_rate": 2.8027981655067126e-06, "loss": 0.9193, "step": 9849 }, { "epoch": 4.461050724637682, "grad_norm": 0.04170156856787928, "learning_rate": 2.798149146550584e-06, "loss": 0.9215, "step": 9850 }, { "epoch": 4.461503623188406, "grad_norm": 0.04078194558405382, "learning_rate": 2.793503846722443e-06, "loss": 0.9131, "step": 9851 }, { "epoch": 4.461956521739131, "grad_norm": 0.041968125695390886, "learning_rate": 2.7888622664866736e-06, "loss": 0.9348, "step": 9852 }, { "epoch": 4.462409420289855, "grad_norm": 0.041449717485013395, "learning_rate": 2.7842244063073143e-06, "loss": 0.9056, "step": 9853 }, { "epoch": 4.46286231884058, "grad_norm": 0.03999527284308288, "learning_rate": 2.779590266648016e-06, "loss": 0.8986, "step": 9854 }, { "epoch": 4.463315217391305, "grad_norm": 0.04111621302788705, "learning_rate": 2.7749598479720608e-06, "loss": 0.9268, "step": 9855 }, { "epoch": 4.463768115942029, "grad_norm": 0.0421244728074743, "learning_rate": 2.7703331507423504e-06, "loss": 0.9198, "step": 9856 }, { "epoch": 4.464221014492754, "grad_norm": 0.040197879983201665, "learning_rate": 2.7657101754214343e-06, "loss": 0.9017, "step": 9857 }, { "epoch": 4.4646739130434785, "grad_norm": 0.04021088094292391, "learning_rate": 2.761090922471472e-06, "loss": 0.9341, "step": 9858 }, { "epoch": 4.465126811594203, "grad_norm": 0.04267126391368, "learning_rate": 2.7564753923542676e-06, "loss": 0.923, "step": 9859 }, { "epoch": 4.465579710144928, "grad_norm": 0.04105946039785552, "learning_rate": 2.751863585531225e-06, "loss": 0.9238, "step": 9860 }, { "epoch": 4.466032608695652, "grad_norm": 0.040983132152707616, "learning_rate": 2.747255502463415e-06, "loss": 0.9124, "step": 9861 }, { "epoch": 4.466485507246377, "grad_norm": 0.042725938542501146, "learning_rate": 2.742651143611501e-06, "loss": 0.9137, "step": 9862 }, { "epoch": 4.4669384057971016, "grad_norm": 0.039999007519930144, "learning_rate": 2.7380505094357947e-06, "loss": 0.9051, "step": 9863 }, { "epoch": 4.467391304347826, "grad_norm": 0.042968544742328404, "learning_rate": 2.7334536003962166e-06, "loss": 0.9278, "step": 9864 }, { "epoch": 4.467844202898551, "grad_norm": 0.0406921655998602, "learning_rate": 2.7288604169523456e-06, "loss": 0.8958, "step": 9865 }, { "epoch": 4.468297101449275, "grad_norm": 0.041092951749336026, "learning_rate": 2.7242709595633486e-06, "loss": 0.9289, "step": 9866 }, { "epoch": 4.46875, "grad_norm": 0.042873401905290924, "learning_rate": 2.7196852286880624e-06, "loss": 0.9316, "step": 9867 }, { "epoch": 4.469202898550725, "grad_norm": 0.03918320599771702, "learning_rate": 2.715103224784912e-06, "loss": 0.9212, "step": 9868 }, { "epoch": 4.469655797101449, "grad_norm": 0.0389088118239587, "learning_rate": 2.7105249483119834e-06, "loss": 0.9105, "step": 9869 }, { "epoch": 4.470108695652174, "grad_norm": 0.040181287987936055, "learning_rate": 2.705950399726964e-06, "loss": 0.9015, "step": 9870 }, { "epoch": 4.4705615942028984, "grad_norm": 0.03950325857516948, "learning_rate": 2.7013795794871776e-06, "loss": 0.9211, "step": 9871 }, { "epoch": 4.471014492753623, "grad_norm": 0.04145550560779898, "learning_rate": 2.6968124880495783e-06, "loss": 0.9171, "step": 9872 }, { "epoch": 4.471467391304348, "grad_norm": 0.03987085727227028, "learning_rate": 2.6922491258707516e-06, "loss": 0.9285, "step": 9873 }, { "epoch": 4.471920289855072, "grad_norm": 0.04249211721141813, "learning_rate": 2.687689493406889e-06, "loss": 0.9175, "step": 9874 }, { "epoch": 4.472373188405797, "grad_norm": 0.04248701091961805, "learning_rate": 2.683133591113847e-06, "loss": 0.8986, "step": 9875 }, { "epoch": 4.4728260869565215, "grad_norm": 0.04006516560938218, "learning_rate": 2.6785814194470616e-06, "loss": 0.9323, "step": 9876 }, { "epoch": 4.473278985507246, "grad_norm": 0.039637355485257014, "learning_rate": 2.67403297886164e-06, "loss": 0.919, "step": 9877 }, { "epoch": 4.473731884057971, "grad_norm": 0.041591921904462516, "learning_rate": 2.6694882698122903e-06, "loss": 0.9354, "step": 9878 }, { "epoch": 4.474184782608695, "grad_norm": 0.045454524116139795, "learning_rate": 2.664947292753355e-06, "loss": 0.9378, "step": 9879 }, { "epoch": 4.47463768115942, "grad_norm": 0.043001014983790276, "learning_rate": 2.6604100481387952e-06, "loss": 0.9291, "step": 9880 }, { "epoch": 4.475090579710145, "grad_norm": 0.041060455686060836, "learning_rate": 2.6558765364222174e-06, "loss": 0.922, "step": 9881 }, { "epoch": 4.475543478260869, "grad_norm": 0.04187767629808704, "learning_rate": 2.651346758056832e-06, "loss": 0.926, "step": 9882 }, { "epoch": 4.475996376811594, "grad_norm": 0.044154936601525474, "learning_rate": 2.6468207134955036e-06, "loss": 0.9173, "step": 9883 }, { "epoch": 4.476449275362318, "grad_norm": 0.04126010996380109, "learning_rate": 2.6422984031906886e-06, "loss": 0.9403, "step": 9884 }, { "epoch": 4.476902173913043, "grad_norm": 0.04202635688355087, "learning_rate": 2.637779827594513e-06, "loss": 0.9123, "step": 9885 }, { "epoch": 4.4773550724637685, "grad_norm": 0.043674608737448384, "learning_rate": 2.6332649871586836e-06, "loss": 0.9002, "step": 9886 }, { "epoch": 4.477807971014493, "grad_norm": 0.044277403010846844, "learning_rate": 2.6287538823345805e-06, "loss": 0.9412, "step": 9887 }, { "epoch": 4.478260869565218, "grad_norm": 0.04252857795674011, "learning_rate": 2.6242465135731587e-06, "loss": 0.9158, "step": 9888 }, { "epoch": 4.478713768115942, "grad_norm": 0.04202813757580597, "learning_rate": 2.6197428813250448e-06, "loss": 0.9017, "step": 9889 }, { "epoch": 4.479166666666667, "grad_norm": 0.04845825863548128, "learning_rate": 2.6152429860404647e-06, "loss": 0.9199, "step": 9890 }, { "epoch": 4.479619565217392, "grad_norm": 0.04837554807868598, "learning_rate": 2.6107468281692904e-06, "loss": 0.9236, "step": 9891 }, { "epoch": 4.480072463768116, "grad_norm": 0.03957270501287891, "learning_rate": 2.606254408160993e-06, "loss": 0.9182, "step": 9892 }, { "epoch": 4.480525362318841, "grad_norm": 0.039514158570672676, "learning_rate": 2.6017657264647113e-06, "loss": 0.929, "step": 9893 }, { "epoch": 4.480978260869565, "grad_norm": 0.04161096349538299, "learning_rate": 2.5972807835291613e-06, "loss": 0.9246, "step": 9894 }, { "epoch": 4.48143115942029, "grad_norm": 0.04822785124695438, "learning_rate": 2.5927995798027317e-06, "loss": 0.907, "step": 9895 }, { "epoch": 4.481884057971015, "grad_norm": 0.046886890215785446, "learning_rate": 2.5883221157334015e-06, "loss": 0.915, "step": 9896 }, { "epoch": 4.482336956521739, "grad_norm": 0.04242848786231923, "learning_rate": 2.583848391768795e-06, "loss": 0.9213, "step": 9897 }, { "epoch": 4.482789855072464, "grad_norm": 0.05422536351671409, "learning_rate": 2.57937840835615e-06, "loss": 0.922, "step": 9898 }, { "epoch": 4.4832427536231885, "grad_norm": 0.043803976942330555, "learning_rate": 2.574912165942345e-06, "loss": 0.9266, "step": 9899 }, { "epoch": 4.483695652173913, "grad_norm": 0.04754573663728057, "learning_rate": 2.5704496649738754e-06, "loss": 0.9161, "step": 9900 }, { "epoch": 4.484148550724638, "grad_norm": 0.03964607886020123, "learning_rate": 2.565990905896869e-06, "loss": 0.9215, "step": 9901 }, { "epoch": 4.484601449275362, "grad_norm": 0.03872765239842927, "learning_rate": 2.561535889157063e-06, "loss": 0.8937, "step": 9902 }, { "epoch": 4.485054347826087, "grad_norm": 0.04334531789319006, "learning_rate": 2.5570846151998517e-06, "loss": 0.9196, "step": 9903 }, { "epoch": 4.4855072463768115, "grad_norm": 0.0394928163389771, "learning_rate": 2.552637084470222e-06, "loss": 0.9305, "step": 9904 }, { "epoch": 4.485960144927536, "grad_norm": 0.03987251052387697, "learning_rate": 2.5481932974128e-06, "loss": 0.9233, "step": 9905 }, { "epoch": 4.486413043478261, "grad_norm": 0.10382357354503892, "learning_rate": 2.543753254471839e-06, "loss": 0.9164, "step": 9906 }, { "epoch": 4.486865942028985, "grad_norm": 0.04076407908562161, "learning_rate": 2.5393169560912247e-06, "loss": 0.9184, "step": 9907 }, { "epoch": 4.48731884057971, "grad_norm": 0.03857684790770713, "learning_rate": 2.5348844027144503e-06, "loss": 0.9094, "step": 9908 }, { "epoch": 4.487771739130435, "grad_norm": 0.038478250731550714, "learning_rate": 2.53045559478466e-06, "loss": 0.92, "step": 9909 }, { "epoch": 4.488224637681159, "grad_norm": 0.039896567354685386, "learning_rate": 2.526030532744588e-06, "loss": 0.9267, "step": 9910 }, { "epoch": 4.488677536231884, "grad_norm": 0.04251453445991466, "learning_rate": 2.521609217036636e-06, "loss": 0.9224, "step": 9911 }, { "epoch": 4.489130434782608, "grad_norm": 0.03912877530721284, "learning_rate": 2.5171916481027968e-06, "loss": 0.941, "step": 9912 }, { "epoch": 4.489583333333333, "grad_norm": 0.03854623698256135, "learning_rate": 2.512777826384709e-06, "loss": 0.9136, "step": 9913 }, { "epoch": 4.490036231884058, "grad_norm": 0.041939981747623555, "learning_rate": 2.5083677523236194e-06, "loss": 0.9109, "step": 9914 }, { "epoch": 4.490489130434782, "grad_norm": 0.03974441412263593, "learning_rate": 2.5039614263604194e-06, "loss": 0.8998, "step": 9915 }, { "epoch": 4.490942028985507, "grad_norm": 0.040008557389792955, "learning_rate": 2.499558848935606e-06, "loss": 0.9375, "step": 9916 }, { "epoch": 4.4913949275362315, "grad_norm": 0.04164404313589985, "learning_rate": 2.495160020489329e-06, "loss": 0.9234, "step": 9917 }, { "epoch": 4.491847826086957, "grad_norm": 0.04018160794493344, "learning_rate": 2.4907649414613256e-06, "loss": 0.9364, "step": 9918 }, { "epoch": 4.492300724637682, "grad_norm": 0.03862458298452444, "learning_rate": 2.486373612291e-06, "loss": 0.9168, "step": 9919 }, { "epoch": 4.492753623188406, "grad_norm": 0.03944647653410964, "learning_rate": 2.4819860334173473e-06, "loss": 0.916, "step": 9920 }, { "epoch": 4.493206521739131, "grad_norm": 0.03800992956475972, "learning_rate": 2.477602205279004e-06, "loss": 0.9208, "step": 9921 }, { "epoch": 4.493659420289855, "grad_norm": 0.04125790941402979, "learning_rate": 2.473222128314223e-06, "loss": 0.9181, "step": 9922 }, { "epoch": 4.49411231884058, "grad_norm": 0.044484914221132606, "learning_rate": 2.4688458029608952e-06, "loss": 0.9179, "step": 9923 }, { "epoch": 4.494565217391305, "grad_norm": 0.04077973706078421, "learning_rate": 2.4644732296565277e-06, "loss": 0.9365, "step": 9924 }, { "epoch": 4.495018115942029, "grad_norm": 0.039283662467312644, "learning_rate": 2.4601044088382465e-06, "loss": 0.9074, "step": 9925 }, { "epoch": 4.495471014492754, "grad_norm": 0.03851454090078672, "learning_rate": 2.4557393409428176e-06, "loss": 0.9015, "step": 9926 }, { "epoch": 4.4959239130434785, "grad_norm": 0.03888278803560064, "learning_rate": 2.4513780264066165e-06, "loss": 0.9234, "step": 9927 }, { "epoch": 4.496376811594203, "grad_norm": 0.04299570818466077, "learning_rate": 2.4470204656656638e-06, "loss": 0.928, "step": 9928 }, { "epoch": 4.496829710144928, "grad_norm": 0.03963598262730123, "learning_rate": 2.4426666591555794e-06, "loss": 0.9159, "step": 9929 }, { "epoch": 4.497282608695652, "grad_norm": 0.04292913776187536, "learning_rate": 2.4383166073116236e-06, "loss": 0.926, "step": 9930 }, { "epoch": 4.497735507246377, "grad_norm": 0.042233671203013405, "learning_rate": 2.4339703105686717e-06, "loss": 0.9247, "step": 9931 }, { "epoch": 4.4981884057971016, "grad_norm": 0.03994335043410694, "learning_rate": 2.4296277693612424e-06, "loss": 0.9154, "step": 9932 }, { "epoch": 4.498641304347826, "grad_norm": 0.038639712988380086, "learning_rate": 2.4252889841234594e-06, "loss": 0.9224, "step": 9933 }, { "epoch": 4.499094202898551, "grad_norm": 0.04586895625628346, "learning_rate": 2.420953955289078e-06, "loss": 0.9129, "step": 9934 }, { "epoch": 4.499547101449275, "grad_norm": 0.03946061914759013, "learning_rate": 2.4166226832914764e-06, "loss": 0.9128, "step": 9935 }, { "epoch": 4.5, "grad_norm": 0.03860505180239921, "learning_rate": 2.4122951685636674e-06, "loss": 0.9183, "step": 9936 }, { "epoch": 4.500452898550725, "grad_norm": 0.041315531060507164, "learning_rate": 2.407971411538266e-06, "loss": 0.9248, "step": 9937 }, { "epoch": 4.500905797101449, "grad_norm": 0.04069143599012861, "learning_rate": 2.4036514126475384e-06, "loss": 0.9376, "step": 9938 }, { "epoch": 4.501358695652174, "grad_norm": 0.040078736544416425, "learning_rate": 2.3993351723233584e-06, "loss": 0.8974, "step": 9939 }, { "epoch": 4.5018115942028984, "grad_norm": 0.043242441423197796, "learning_rate": 2.39502269099722e-06, "loss": 0.9223, "step": 9940 }, { "epoch": 4.502264492753623, "grad_norm": 0.04172431383971911, "learning_rate": 2.390713969100249e-06, "loss": 0.9056, "step": 9941 }, { "epoch": 4.502717391304348, "grad_norm": 0.0399570950901782, "learning_rate": 2.386409007063204e-06, "loss": 0.9225, "step": 9942 }, { "epoch": 4.503170289855072, "grad_norm": 0.0422506602500687, "learning_rate": 2.3821078053164514e-06, "loss": 0.9359, "step": 9943 }, { "epoch": 4.503623188405797, "grad_norm": 0.04661537557243387, "learning_rate": 2.3778103642899985e-06, "loss": 0.9219, "step": 9944 }, { "epoch": 4.5040760869565215, "grad_norm": 0.03974577454317304, "learning_rate": 2.3735166844134526e-06, "loss": 0.9084, "step": 9945 }, { "epoch": 4.504528985507246, "grad_norm": 0.04127644686278609, "learning_rate": 2.369226766116075e-06, "loss": 0.9364, "step": 9946 }, { "epoch": 4.504981884057971, "grad_norm": 0.040646547493695596, "learning_rate": 2.3649406098267315e-06, "loss": 0.9205, "step": 9947 }, { "epoch": 4.505434782608695, "grad_norm": 0.03982775580739411, "learning_rate": 2.36065821597391e-06, "loss": 0.9163, "step": 9948 }, { "epoch": 4.50588768115942, "grad_norm": 0.04447225554308911, "learning_rate": 2.356379584985722e-06, "loss": 0.9356, "step": 9949 }, { "epoch": 4.506340579710145, "grad_norm": 0.042619012001301596, "learning_rate": 2.3521047172899313e-06, "loss": 0.932, "step": 9950 }, { "epoch": 4.506793478260869, "grad_norm": 0.04308472878288739, "learning_rate": 2.3478336133138812e-06, "loss": 0.9298, "step": 9951 }, { "epoch": 4.507246376811594, "grad_norm": 0.04119656692823584, "learning_rate": 2.343566273484572e-06, "loss": 0.9086, "step": 9952 }, { "epoch": 4.507699275362318, "grad_norm": 0.04164256820448924, "learning_rate": 2.339302698228614e-06, "loss": 0.9135, "step": 9953 }, { "epoch": 4.508152173913043, "grad_norm": 0.04077041833787913, "learning_rate": 2.3350428879722476e-06, "loss": 0.8948, "step": 9954 }, { "epoch": 4.508605072463768, "grad_norm": 0.04259158276069315, "learning_rate": 2.3307868431413283e-06, "loss": 0.9077, "step": 9955 }, { "epoch": 4.509057971014493, "grad_norm": 0.04139514637651785, "learning_rate": 2.326534564161338e-06, "loss": 0.9276, "step": 9956 }, { "epoch": 4.509510869565218, "grad_norm": 0.04041156954312102, "learning_rate": 2.322286051457385e-06, "loss": 0.9073, "step": 9957 }, { "epoch": 4.509963768115942, "grad_norm": 0.03802457092791362, "learning_rate": 2.318041305454202e-06, "loss": 0.9329, "step": 9958 }, { "epoch": 4.510416666666667, "grad_norm": 0.03886092664455706, "learning_rate": 2.313800326576141e-06, "loss": 0.9123, "step": 9959 }, { "epoch": 4.510869565217392, "grad_norm": 0.04004194820631621, "learning_rate": 2.309563115247184e-06, "loss": 0.9267, "step": 9960 }, { "epoch": 4.511322463768116, "grad_norm": 0.03849719726367972, "learning_rate": 2.3053296718909215e-06, "loss": 0.9208, "step": 9961 }, { "epoch": 4.511775362318841, "grad_norm": 0.039319305451400735, "learning_rate": 2.3010999969305913e-06, "loss": 0.9127, "step": 9962 }, { "epoch": 4.512228260869565, "grad_norm": 0.04019482117012664, "learning_rate": 2.2968740907890296e-06, "loss": 0.9209, "step": 9963 }, { "epoch": 4.51268115942029, "grad_norm": 0.037103554038032734, "learning_rate": 2.292651953888716e-06, "loss": 0.9127, "step": 9964 }, { "epoch": 4.513134057971015, "grad_norm": 0.04127691119285168, "learning_rate": 2.2884335866517305e-06, "loss": 0.9137, "step": 9965 }, { "epoch": 4.513586956521739, "grad_norm": 0.04104544587036869, "learning_rate": 2.284218989499802e-06, "loss": 0.9174, "step": 9966 }, { "epoch": 4.514039855072464, "grad_norm": 0.03874631570017181, "learning_rate": 2.2800081628542657e-06, "loss": 0.9349, "step": 9967 }, { "epoch": 4.5144927536231885, "grad_norm": 0.03982492986337119, "learning_rate": 2.275801107136086e-06, "loss": 0.9184, "step": 9968 }, { "epoch": 4.514945652173913, "grad_norm": 0.042783283995076596, "learning_rate": 2.2715978227658473e-06, "loss": 0.924, "step": 9969 }, { "epoch": 4.515398550724638, "grad_norm": 0.04160116347270392, "learning_rate": 2.267398310163764e-06, "loss": 0.9296, "step": 9970 }, { "epoch": 4.515851449275362, "grad_norm": 0.041567761968614815, "learning_rate": 2.263202569749665e-06, "loss": 0.9276, "step": 9971 }, { "epoch": 4.516304347826087, "grad_norm": 0.04404725674994601, "learning_rate": 2.2590106019430057e-06, "loss": 0.9364, "step": 9972 }, { "epoch": 4.5167572463768115, "grad_norm": 0.041990666811118425, "learning_rate": 2.2548224071628557e-06, "loss": 0.9245, "step": 9973 }, { "epoch": 4.517210144927536, "grad_norm": 0.04053862125857819, "learning_rate": 2.2506379858279283e-06, "loss": 0.9116, "step": 9974 }, { "epoch": 4.517663043478261, "grad_norm": 0.042212200024142534, "learning_rate": 2.246457338356529e-06, "loss": 0.9342, "step": 9975 }, { "epoch": 4.518115942028985, "grad_norm": 0.03980878235011469, "learning_rate": 2.2422804651666265e-06, "loss": 0.9224, "step": 9976 }, { "epoch": 4.51856884057971, "grad_norm": 0.04005352048074143, "learning_rate": 2.23810736667577e-06, "loss": 0.9131, "step": 9977 }, { "epoch": 4.519021739130435, "grad_norm": 0.038630225787884935, "learning_rate": 2.2339380433011693e-06, "loss": 0.9151, "step": 9978 }, { "epoch": 4.519474637681159, "grad_norm": 0.03915594923791138, "learning_rate": 2.229772495459619e-06, "loss": 0.9254, "step": 9979 }, { "epoch": 4.519927536231884, "grad_norm": 0.03953149538262559, "learning_rate": 2.2256107235675684e-06, "loss": 0.9125, "step": 9980 }, { "epoch": 4.520380434782608, "grad_norm": 0.041174932544344234, "learning_rate": 2.2214527280410712e-06, "loss": 0.9297, "step": 9981 }, { "epoch": 4.520833333333333, "grad_norm": 0.04269083631748789, "learning_rate": 2.217298509295813e-06, "loss": 0.9263, "step": 9982 }, { "epoch": 4.521286231884058, "grad_norm": 0.03791267989190939, "learning_rate": 2.2131480677470886e-06, "loss": 0.9324, "step": 9983 }, { "epoch": 4.521739130434782, "grad_norm": 0.039321148710168084, "learning_rate": 2.2090014038098364e-06, "loss": 0.9236, "step": 9984 }, { "epoch": 4.522192028985507, "grad_norm": 0.039119662163051995, "learning_rate": 2.204858517898596e-06, "loss": 0.906, "step": 9985 }, { "epoch": 4.522644927536232, "grad_norm": 0.04010122867827893, "learning_rate": 2.200719410427543e-06, "loss": 0.9181, "step": 9986 }, { "epoch": 4.523097826086957, "grad_norm": 0.04040553786649055, "learning_rate": 2.1965840818104623e-06, "loss": 0.9022, "step": 9987 }, { "epoch": 4.523550724637682, "grad_norm": 0.0390770091092123, "learning_rate": 2.1924525324607825e-06, "loss": 0.9273, "step": 9988 }, { "epoch": 4.524003623188406, "grad_norm": 0.03957498188826107, "learning_rate": 2.1883247627915338e-06, "loss": 0.9303, "step": 9989 }, { "epoch": 4.524456521739131, "grad_norm": 0.04016880604255566, "learning_rate": 2.184200773215377e-06, "loss": 0.9387, "step": 9990 }, { "epoch": 4.524909420289855, "grad_norm": 0.03932108501435434, "learning_rate": 2.1800805641445867e-06, "loss": 0.9354, "step": 9991 }, { "epoch": 4.52536231884058, "grad_norm": 0.039991578538734154, "learning_rate": 2.1759641359910777e-06, "loss": 0.9334, "step": 9992 }, { "epoch": 4.525815217391305, "grad_norm": 0.04018247827025461, "learning_rate": 2.17185148916637e-06, "loss": 0.9256, "step": 9993 }, { "epoch": 4.526268115942029, "grad_norm": 0.04005297529107022, "learning_rate": 2.1677426240816146e-06, "loss": 0.9309, "step": 9994 }, { "epoch": 4.526721014492754, "grad_norm": 0.040300438136042616, "learning_rate": 2.1636375411475765e-06, "loss": 0.9101, "step": 9995 }, { "epoch": 4.5271739130434785, "grad_norm": 0.03838608409132758, "learning_rate": 2.159536240774651e-06, "loss": 0.9255, "step": 9996 }, { "epoch": 4.527626811594203, "grad_norm": 0.04187151832708662, "learning_rate": 2.155438723372858e-06, "loss": 0.932, "step": 9997 }, { "epoch": 4.528079710144928, "grad_norm": 0.041552569240840914, "learning_rate": 2.1513449893518203e-06, "loss": 0.9291, "step": 9998 }, { "epoch": 4.528532608695652, "grad_norm": 0.03892975709636427, "learning_rate": 2.147255039120797e-06, "loss": 0.9156, "step": 9999 }, { "epoch": 4.528985507246377, "grad_norm": 0.03891506677628064, "learning_rate": 2.1431688730886747e-06, "loss": 0.9257, "step": 10000 }, { "epoch": 4.5294384057971016, "grad_norm": 0.03847140065556041, "learning_rate": 2.139086491663944e-06, "loss": 0.9247, "step": 10001 }, { "epoch": 4.529891304347826, "grad_norm": 0.03824412043762994, "learning_rate": 2.135007895254737e-06, "loss": 0.9178, "step": 10002 }, { "epoch": 4.530344202898551, "grad_norm": 0.038116699775768245, "learning_rate": 2.1309330842687893e-06, "loss": 0.9259, "step": 10003 }, { "epoch": 4.530797101449275, "grad_norm": 0.04185816675424923, "learning_rate": 2.126862059113477e-06, "loss": 0.945, "step": 10004 }, { "epoch": 4.53125, "grad_norm": 0.036764690322696976, "learning_rate": 2.122794820195777e-06, "loss": 0.9071, "step": 10005 }, { "epoch": 4.531702898550725, "grad_norm": 0.03906260595756779, "learning_rate": 2.118731367922302e-06, "loss": 0.9225, "step": 10006 }, { "epoch": 4.532155797101449, "grad_norm": 0.04027024129668935, "learning_rate": 2.1146717026992735e-06, "loss": 0.9291, "step": 10007 }, { "epoch": 4.532608695652174, "grad_norm": 0.039788076393509446, "learning_rate": 2.1106158249325536e-06, "loss": 0.927, "step": 10008 }, { "epoch": 4.5330615942028984, "grad_norm": 0.03927032178941205, "learning_rate": 2.1065637350276045e-06, "loss": 0.9343, "step": 10009 }, { "epoch": 4.533514492753623, "grad_norm": 0.03952177651030603, "learning_rate": 2.1025154333895337e-06, "loss": 0.9264, "step": 10010 }, { "epoch": 4.533967391304348, "grad_norm": 0.03841812348503569, "learning_rate": 2.0984709204230437e-06, "loss": 0.9178, "step": 10011 }, { "epoch": 4.534420289855072, "grad_norm": 0.04036127249325122, "learning_rate": 2.094430196532482e-06, "loss": 0.9185, "step": 10012 }, { "epoch": 4.534873188405797, "grad_norm": 0.03933489624083629, "learning_rate": 2.0903932621217972e-06, "loss": 0.9194, "step": 10013 }, { "epoch": 4.5353260869565215, "grad_norm": 0.0403198715764858, "learning_rate": 2.0863601175945723e-06, "loss": 0.9327, "step": 10014 }, { "epoch": 4.535778985507246, "grad_norm": 0.0379697917935643, "learning_rate": 2.082330763354001e-06, "loss": 0.9112, "step": 10015 }, { "epoch": 4.536231884057971, "grad_norm": 0.03966376524814871, "learning_rate": 2.078305199802917e-06, "loss": 0.9398, "step": 10016 }, { "epoch": 4.536684782608695, "grad_norm": 0.04071682320797205, "learning_rate": 2.074283427343744e-06, "loss": 0.9061, "step": 10017 }, { "epoch": 4.53713768115942, "grad_norm": 0.03861968553081656, "learning_rate": 2.070265446378561e-06, "loss": 0.9284, "step": 10018 }, { "epoch": 4.537590579710145, "grad_norm": 0.041008442198445245, "learning_rate": 2.0662512573090422e-06, "loss": 0.9099, "step": 10019 }, { "epoch": 4.538043478260869, "grad_norm": 0.04027147051584781, "learning_rate": 2.0622408605365017e-06, "loss": 0.8913, "step": 10020 }, { "epoch": 4.538496376811594, "grad_norm": 0.039649220849133075, "learning_rate": 2.058234256461855e-06, "loss": 0.925, "step": 10021 }, { "epoch": 4.538949275362318, "grad_norm": 0.03808942898920331, "learning_rate": 2.0542314454856614e-06, "loss": 0.9029, "step": 10022 }, { "epoch": 4.539402173913043, "grad_norm": 0.038874274867873024, "learning_rate": 2.0502324280080766e-06, "loss": 0.9198, "step": 10023 }, { "epoch": 4.539855072463768, "grad_norm": 0.03916515156690464, "learning_rate": 2.0462372044288916e-06, "loss": 0.9268, "step": 10024 }, { "epoch": 4.540307971014493, "grad_norm": 0.04246397515632974, "learning_rate": 2.0422457751475157e-06, "loss": 0.9511, "step": 10025 }, { "epoch": 4.540760869565218, "grad_norm": 0.042874244534622505, "learning_rate": 2.0382581405629807e-06, "loss": 0.895, "step": 10026 }, { "epoch": 4.541213768115942, "grad_norm": 0.039243876439055955, "learning_rate": 2.0342743010739286e-06, "loss": 0.8884, "step": 10027 }, { "epoch": 4.541666666666667, "grad_norm": 0.0402145289412902, "learning_rate": 2.0302942570786446e-06, "loss": 0.9312, "step": 10028 }, { "epoch": 4.542119565217392, "grad_norm": 0.04280223046967751, "learning_rate": 2.0263180089750057e-06, "loss": 0.932, "step": 10029 }, { "epoch": 4.542572463768116, "grad_norm": 0.043087932624790035, "learning_rate": 2.0223455571605344e-06, "loss": 0.925, "step": 10030 }, { "epoch": 4.543025362318841, "grad_norm": 0.04342595263812623, "learning_rate": 2.0183769020323574e-06, "loss": 0.9217, "step": 10031 }, { "epoch": 4.543478260869565, "grad_norm": 0.04050142799016908, "learning_rate": 2.014412043987233e-06, "loss": 0.9376, "step": 10032 }, { "epoch": 4.54393115942029, "grad_norm": 0.03937358924249398, "learning_rate": 2.010450983421519e-06, "loss": 0.9156, "step": 10033 }, { "epoch": 4.544384057971015, "grad_norm": 0.03998830761005779, "learning_rate": 2.0064937207312286e-06, "loss": 0.9261, "step": 10034 }, { "epoch": 4.544836956521739, "grad_norm": 0.04012501412726133, "learning_rate": 2.00254025631196e-06, "loss": 0.9085, "step": 10035 }, { "epoch": 4.545289855072464, "grad_norm": 0.04462027261138263, "learning_rate": 1.998590590558962e-06, "loss": 0.9246, "step": 10036 }, { "epoch": 4.5457427536231885, "grad_norm": 0.0434366137125565, "learning_rate": 1.9946447238670742e-06, "loss": 0.9003, "step": 10037 }, { "epoch": 4.546195652173913, "grad_norm": 0.04306043417813072, "learning_rate": 1.990702656630781e-06, "loss": 0.9249, "step": 10038 }, { "epoch": 4.546648550724638, "grad_norm": 0.043428645406297095, "learning_rate": 1.9867643892441755e-06, "loss": 0.9331, "step": 10039 }, { "epoch": 4.547101449275362, "grad_norm": 0.040127832740021435, "learning_rate": 1.982829922100975e-06, "loss": 0.9337, "step": 10040 }, { "epoch": 4.547554347826087, "grad_norm": 0.041747536179214506, "learning_rate": 1.9788992555944996e-06, "loss": 0.945, "step": 10041 }, { "epoch": 4.5480072463768115, "grad_norm": 0.03809671989725341, "learning_rate": 1.974972390117724e-06, "loss": 0.914, "step": 10042 }, { "epoch": 4.548460144927536, "grad_norm": 0.04134005223229415, "learning_rate": 1.971049326063206e-06, "loss": 0.9128, "step": 10043 }, { "epoch": 4.548913043478261, "grad_norm": 0.04092179956153421, "learning_rate": 1.9671300638231594e-06, "loss": 0.9459, "step": 10044 }, { "epoch": 4.549365942028985, "grad_norm": 0.03907146948757855, "learning_rate": 1.9632146037893786e-06, "loss": 0.9234, "step": 10045 }, { "epoch": 4.54981884057971, "grad_norm": 0.04134852482118254, "learning_rate": 1.9593029463533143e-06, "loss": 0.9323, "step": 10046 }, { "epoch": 4.550271739130435, "grad_norm": 0.038806124025650784, "learning_rate": 1.9553950919060182e-06, "loss": 0.918, "step": 10047 }, { "epoch": 4.550724637681159, "grad_norm": 0.04142002150708427, "learning_rate": 1.951491040838156e-06, "loss": 0.925, "step": 10048 }, { "epoch": 4.551177536231884, "grad_norm": 0.04235655509931912, "learning_rate": 1.947590793540033e-06, "loss": 0.924, "step": 10049 }, { "epoch": 4.551630434782608, "grad_norm": 0.046575296592519064, "learning_rate": 1.9436943504015505e-06, "loss": 0.9198, "step": 10050 }, { "epoch": 4.552083333333333, "grad_norm": 0.03920919757381803, "learning_rate": 1.9398017118122546e-06, "loss": 0.8853, "step": 10051 }, { "epoch": 4.552536231884058, "grad_norm": 0.037844375063589474, "learning_rate": 1.935912878161288e-06, "loss": 0.91, "step": 10052 }, { "epoch": 4.552989130434782, "grad_norm": 0.040692811373865115, "learning_rate": 1.9320278498374323e-06, "loss": 0.9088, "step": 10053 }, { "epoch": 4.553442028985507, "grad_norm": 0.04123033431446222, "learning_rate": 1.9281466272290704e-06, "loss": 0.9295, "step": 10054 }, { "epoch": 4.553894927536232, "grad_norm": 0.041350101264751304, "learning_rate": 1.9242692107242257e-06, "loss": 0.938, "step": 10055 }, { "epoch": 4.554347826086957, "grad_norm": 0.04120819684540487, "learning_rate": 1.9203956007105293e-06, "loss": 0.8968, "step": 10056 }, { "epoch": 4.554800724637682, "grad_norm": 0.0418903564799635, "learning_rate": 1.9165257975752193e-06, "loss": 0.9344, "step": 10057 }, { "epoch": 4.555253623188406, "grad_norm": 0.038049778941613405, "learning_rate": 1.9126598017051766e-06, "loss": 0.8856, "step": 10058 }, { "epoch": 4.555706521739131, "grad_norm": 0.03988228344526852, "learning_rate": 1.908797613486888e-06, "loss": 0.9182, "step": 10059 }, { "epoch": 4.556159420289855, "grad_norm": 0.04854223029573165, "learning_rate": 1.9049392333064577e-06, "loss": 0.9338, "step": 10060 }, { "epoch": 4.55661231884058, "grad_norm": 0.041057787846662766, "learning_rate": 1.901084661549626e-06, "loss": 0.9089, "step": 10061 }, { "epoch": 4.557065217391305, "grad_norm": 0.04227854290873366, "learning_rate": 1.8972338986017247e-06, "loss": 0.9168, "step": 10062 }, { "epoch": 4.557518115942029, "grad_norm": 0.03867527277355586, "learning_rate": 1.8933869448477394e-06, "loss": 0.9163, "step": 10063 }, { "epoch": 4.557971014492754, "grad_norm": 0.03941609255806152, "learning_rate": 1.8895438006722377e-06, "loss": 0.9016, "step": 10064 }, { "epoch": 4.5584239130434785, "grad_norm": 0.041022308922345055, "learning_rate": 1.885704466459446e-06, "loss": 0.9216, "step": 10065 }, { "epoch": 4.558876811594203, "grad_norm": 0.038535466800140326, "learning_rate": 1.881868942593168e-06, "loss": 0.9197, "step": 10066 }, { "epoch": 4.559329710144928, "grad_norm": 0.03900607142856148, "learning_rate": 1.8780372294568573e-06, "loss": 0.9114, "step": 10067 }, { "epoch": 4.559782608695652, "grad_norm": 0.04114228006217492, "learning_rate": 1.874209327433567e-06, "loss": 0.8939, "step": 10068 }, { "epoch": 4.560235507246377, "grad_norm": 0.04124740537546044, "learning_rate": 1.8703852369059961e-06, "loss": 0.8964, "step": 10069 }, { "epoch": 4.5606884057971016, "grad_norm": 0.038702547997074274, "learning_rate": 1.8665649582564248e-06, "loss": 0.8886, "step": 10070 }, { "epoch": 4.561141304347826, "grad_norm": 0.039212220552496055, "learning_rate": 1.862748491866788e-06, "loss": 0.9197, "step": 10071 }, { "epoch": 4.561594202898551, "grad_norm": 0.03872091075273513, "learning_rate": 1.8589358381186163e-06, "loss": 0.9276, "step": 10072 }, { "epoch": 4.562047101449275, "grad_norm": 0.03797993268093969, "learning_rate": 1.8551269973930707e-06, "loss": 0.9468, "step": 10073 }, { "epoch": 4.5625, "grad_norm": 0.04258556321559123, "learning_rate": 1.8513219700709272e-06, "loss": 0.9241, "step": 10074 }, { "epoch": 4.562952898550725, "grad_norm": 0.03813238663821889, "learning_rate": 1.8475207565325747e-06, "loss": 0.9225, "step": 10075 }, { "epoch": 4.563405797101449, "grad_norm": 0.039304005800618644, "learning_rate": 1.8437233571580294e-06, "loss": 0.9227, "step": 10076 }, { "epoch": 4.563858695652174, "grad_norm": 0.03857261496743153, "learning_rate": 1.839929772326925e-06, "loss": 0.9301, "step": 10077 }, { "epoch": 4.5643115942028984, "grad_norm": 0.040569321575253725, "learning_rate": 1.836140002418505e-06, "loss": 0.909, "step": 10078 }, { "epoch": 4.564764492753623, "grad_norm": 0.04164530734051734, "learning_rate": 1.832354047811653e-06, "loss": 0.9232, "step": 10079 }, { "epoch": 4.565217391304348, "grad_norm": 0.0386632092343813, "learning_rate": 1.8285719088848442e-06, "loss": 0.9241, "step": 10080 }, { "epoch": 4.565670289855072, "grad_norm": 0.03784731054950305, "learning_rate": 1.8247935860161937e-06, "loss": 0.9374, "step": 10081 }, { "epoch": 4.566123188405797, "grad_norm": 0.04017617852330147, "learning_rate": 1.8210190795834216e-06, "loss": 0.9362, "step": 10082 }, { "epoch": 4.5665760869565215, "grad_norm": 0.0400351760717525, "learning_rate": 1.8172483899638704e-06, "loss": 0.9143, "step": 10083 }, { "epoch": 4.567028985507246, "grad_norm": 0.03770859409968299, "learning_rate": 1.8134815175345011e-06, "loss": 0.9085, "step": 10084 }, { "epoch": 4.567481884057971, "grad_norm": 0.03757607754556694, "learning_rate": 1.8097184626719011e-06, "loss": 0.9087, "step": 10085 }, { "epoch": 4.567934782608695, "grad_norm": 0.03918755227106339, "learning_rate": 1.8059592257522584e-06, "loss": 0.9182, "step": 10086 }, { "epoch": 4.56838768115942, "grad_norm": 0.03984132939179244, "learning_rate": 1.802203807151397e-06, "loss": 0.8935, "step": 10087 }, { "epoch": 4.568840579710145, "grad_norm": 0.0382602306987963, "learning_rate": 1.7984522072447497e-06, "loss": 0.9293, "step": 10088 }, { "epoch": 4.569293478260869, "grad_norm": 0.038850269185031205, "learning_rate": 1.7947044264073766e-06, "loss": 0.9105, "step": 10089 }, { "epoch": 4.569746376811594, "grad_norm": 0.03893689253046081, "learning_rate": 1.7909604650139423e-06, "loss": 0.9196, "step": 10090 }, { "epoch": 4.570199275362318, "grad_norm": 0.04006123035898107, "learning_rate": 1.7872203234387342e-06, "loss": 0.9522, "step": 10091 }, { "epoch": 4.570652173913043, "grad_norm": 0.03925945686941771, "learning_rate": 1.783484002055662e-06, "loss": 0.9106, "step": 10092 }, { "epoch": 4.571105072463768, "grad_norm": 0.03819710175942643, "learning_rate": 1.7797515012382583e-06, "loss": 0.902, "step": 10093 }, { "epoch": 4.571557971014493, "grad_norm": 0.03921340926179909, "learning_rate": 1.7760228213596598e-06, "loss": 0.9329, "step": 10094 }, { "epoch": 4.572010869565218, "grad_norm": 0.03984037615039146, "learning_rate": 1.7722979627926307e-06, "loss": 0.8971, "step": 10095 }, { "epoch": 4.572463768115942, "grad_norm": 0.0385930778577031, "learning_rate": 1.7685769259095486e-06, "loss": 0.9349, "step": 10096 }, { "epoch": 4.572916666666667, "grad_norm": 0.038250096212064474, "learning_rate": 1.7648597110824183e-06, "loss": 0.9096, "step": 10097 }, { "epoch": 4.573369565217392, "grad_norm": 0.039480055062009675, "learning_rate": 1.7611463186828537e-06, "loss": 0.9301, "step": 10098 }, { "epoch": 4.573822463768116, "grad_norm": 0.04080988405095852, "learning_rate": 1.7574367490820865e-06, "loss": 0.8867, "step": 10099 }, { "epoch": 4.574275362318841, "grad_norm": 0.04069758613827322, "learning_rate": 1.753731002650958e-06, "loss": 0.9185, "step": 10100 }, { "epoch": 4.574728260869565, "grad_norm": 0.0394028674843086, "learning_rate": 1.7500290797599585e-06, "loss": 0.9054, "step": 10101 }, { "epoch": 4.57518115942029, "grad_norm": 0.037659198751147524, "learning_rate": 1.7463309807791562e-06, "loss": 0.9349, "step": 10102 }, { "epoch": 4.575634057971015, "grad_norm": 0.0408109795836978, "learning_rate": 1.742636706078269e-06, "loss": 0.9279, "step": 10103 }, { "epoch": 4.576086956521739, "grad_norm": 0.038987014613643844, "learning_rate": 1.7389462560266102e-06, "loss": 0.9075, "step": 10104 }, { "epoch": 4.576539855072464, "grad_norm": 0.03948265225163065, "learning_rate": 1.7352596309931292e-06, "loss": 0.9317, "step": 10105 }, { "epoch": 4.5769927536231885, "grad_norm": 0.037119800154330904, "learning_rate": 1.7315768313463755e-06, "loss": 0.9188, "step": 10106 }, { "epoch": 4.577445652173913, "grad_norm": 0.037250383795911494, "learning_rate": 1.7278978574545301e-06, "loss": 0.91, "step": 10107 }, { "epoch": 4.577898550724638, "grad_norm": 0.03988767386106684, "learning_rate": 1.724222709685379e-06, "loss": 0.9183, "step": 10108 }, { "epoch": 4.578351449275362, "grad_norm": 0.038633328544248675, "learning_rate": 1.7205513884063396e-06, "loss": 0.9213, "step": 10109 }, { "epoch": 4.578804347826087, "grad_norm": 0.04034314869280561, "learning_rate": 1.7168838939844335e-06, "loss": 0.9188, "step": 10110 }, { "epoch": 4.5792572463768115, "grad_norm": 0.03883044574360337, "learning_rate": 1.713220226786314e-06, "loss": 0.9251, "step": 10111 }, { "epoch": 4.579710144927536, "grad_norm": 0.03783094398495944, "learning_rate": 1.709560387178235e-06, "loss": 0.9164, "step": 10112 }, { "epoch": 4.580163043478261, "grad_norm": 0.039689300502585775, "learning_rate": 1.7059043755260817e-06, "loss": 0.922, "step": 10113 }, { "epoch": 4.580615942028985, "grad_norm": 0.03832688826460731, "learning_rate": 1.7022521921953482e-06, "loss": 0.9351, "step": 10114 }, { "epoch": 4.58106884057971, "grad_norm": 0.03838561147247905, "learning_rate": 1.698603837551156e-06, "loss": 0.9046, "step": 10115 }, { "epoch": 4.581521739130435, "grad_norm": 0.04073989654121995, "learning_rate": 1.6949593119582352e-06, "loss": 0.9322, "step": 10116 }, { "epoch": 4.581974637681159, "grad_norm": 0.03821529148559216, "learning_rate": 1.6913186157809257e-06, "loss": 0.9342, "step": 10117 }, { "epoch": 4.582427536231884, "grad_norm": 0.03896933924026428, "learning_rate": 1.6876817493831987e-06, "loss": 0.9328, "step": 10118 }, { "epoch": 4.582880434782608, "grad_norm": 0.03680883797470526, "learning_rate": 1.6840487131286388e-06, "loss": 0.925, "step": 10119 }, { "epoch": 4.583333333333333, "grad_norm": 0.037664709161110616, "learning_rate": 1.6804195073804442e-06, "loss": 0.9283, "step": 10120 }, { "epoch": 4.583786231884058, "grad_norm": 0.03885291113201161, "learning_rate": 1.6767941325014402e-06, "loss": 0.9087, "step": 10121 }, { "epoch": 4.584239130434782, "grad_norm": 0.040000111446501264, "learning_rate": 1.6731725888540485e-06, "loss": 0.9362, "step": 10122 }, { "epoch": 4.584692028985507, "grad_norm": 0.04281525560636885, "learning_rate": 1.66955487680033e-06, "loss": 0.9092, "step": 10123 }, { "epoch": 4.585144927536232, "grad_norm": 0.04004029769756653, "learning_rate": 1.6659409967019513e-06, "loss": 0.9196, "step": 10124 }, { "epoch": 4.585597826086957, "grad_norm": 0.039137180178173706, "learning_rate": 1.6623309489201967e-06, "loss": 0.9183, "step": 10125 }, { "epoch": 4.586050724637682, "grad_norm": 0.03873086730664337, "learning_rate": 1.658724733815964e-06, "loss": 0.9202, "step": 10126 }, { "epoch": 4.586503623188406, "grad_norm": 0.04077479996899807, "learning_rate": 1.6551223517497827e-06, "loss": 0.9289, "step": 10127 }, { "epoch": 4.586956521739131, "grad_norm": 0.03994480721539332, "learning_rate": 1.651523803081778e-06, "loss": 0.9094, "step": 10128 }, { "epoch": 4.587409420289855, "grad_norm": 0.03956511277050616, "learning_rate": 1.6479290881717114e-06, "loss": 0.9163, "step": 10129 }, { "epoch": 4.58786231884058, "grad_norm": 0.04191316870191311, "learning_rate": 1.6443382073789393e-06, "loss": 0.9114, "step": 10130 }, { "epoch": 4.588315217391305, "grad_norm": 0.03944492409928112, "learning_rate": 1.6407511610624638e-06, "loss": 0.9194, "step": 10131 }, { "epoch": 4.588768115942029, "grad_norm": 0.04004334554423078, "learning_rate": 1.6371679495808823e-06, "loss": 0.9293, "step": 10132 }, { "epoch": 4.589221014492754, "grad_norm": 0.040845430423470754, "learning_rate": 1.6335885732924107e-06, "loss": 0.9099, "step": 10133 }, { "epoch": 4.5896739130434785, "grad_norm": 0.03991618679910894, "learning_rate": 1.6300130325548825e-06, "loss": 0.9133, "step": 10134 }, { "epoch": 4.590126811594203, "grad_norm": 0.03987147936876144, "learning_rate": 1.626441327725754e-06, "loss": 0.9109, "step": 10135 }, { "epoch": 4.590579710144928, "grad_norm": 0.04517875658039588, "learning_rate": 1.6228734591620953e-06, "loss": 0.9084, "step": 10136 }, { "epoch": 4.591032608695652, "grad_norm": 0.04073846711255332, "learning_rate": 1.61930942722059e-06, "loss": 0.9132, "step": 10137 }, { "epoch": 4.591485507246377, "grad_norm": 0.03864909990708976, "learning_rate": 1.6157492322575396e-06, "loss": 0.9272, "step": 10138 }, { "epoch": 4.5919384057971016, "grad_norm": 0.03907944377562256, "learning_rate": 1.6121928746288684e-06, "loss": 0.9203, "step": 10139 }, { "epoch": 4.592391304347826, "grad_norm": 0.03911841687506254, "learning_rate": 1.6086403546901053e-06, "loss": 0.903, "step": 10140 }, { "epoch": 4.592844202898551, "grad_norm": 0.039188265652666456, "learning_rate": 1.6050916727964017e-06, "loss": 0.9182, "step": 10141 }, { "epoch": 4.593297101449275, "grad_norm": 0.04017983679339389, "learning_rate": 1.6015468293025227e-06, "loss": 0.922, "step": 10142 }, { "epoch": 4.59375, "grad_norm": 0.041938618966357026, "learning_rate": 1.598005824562856e-06, "loss": 0.9138, "step": 10143 }, { "epoch": 4.594202898550725, "grad_norm": 0.039497930360927566, "learning_rate": 1.5944686589313984e-06, "loss": 0.907, "step": 10144 }, { "epoch": 4.594655797101449, "grad_norm": 0.0410312740193904, "learning_rate": 1.5909353327617693e-06, "loss": 0.9157, "step": 10145 }, { "epoch": 4.595108695652174, "grad_norm": 0.03802950715751411, "learning_rate": 1.5874058464071974e-06, "loss": 0.9256, "step": 10146 }, { "epoch": 4.5955615942028984, "grad_norm": 0.039235736617378135, "learning_rate": 1.5838802002205377e-06, "loss": 0.9365, "step": 10147 }, { "epoch": 4.596014492753623, "grad_norm": 0.04015324883987756, "learning_rate": 1.5803583945542511e-06, "loss": 0.9231, "step": 10148 }, { "epoch": 4.596467391304348, "grad_norm": 0.039202042188348864, "learning_rate": 1.5768404297604112e-06, "loss": 0.9142, "step": 10149 }, { "epoch": 4.596920289855072, "grad_norm": 0.038389168443228795, "learning_rate": 1.5733263061907234e-06, "loss": 0.93, "step": 10150 }, { "epoch": 4.597373188405797, "grad_norm": 0.04233709448773613, "learning_rate": 1.5698160241964978e-06, "loss": 0.9325, "step": 10151 }, { "epoch": 4.5978260869565215, "grad_norm": 0.03818662703021069, "learning_rate": 1.5663095841286535e-06, "loss": 0.9127, "step": 10152 }, { "epoch": 4.598278985507246, "grad_norm": 0.039899071100948424, "learning_rate": 1.5628069863377548e-06, "loss": 0.9332, "step": 10153 }, { "epoch": 4.598731884057971, "grad_norm": 0.03772738009147555, "learning_rate": 1.5593082311739439e-06, "loss": 0.9003, "step": 10154 }, { "epoch": 4.599184782608695, "grad_norm": 0.03819785630168048, "learning_rate": 1.5558133189870073e-06, "loss": 0.9214, "step": 10155 }, { "epoch": 4.59963768115942, "grad_norm": 0.03815688297270171, "learning_rate": 1.552322250126328e-06, "loss": 0.9367, "step": 10156 }, { "epoch": 4.600090579710145, "grad_norm": 0.04023249786437164, "learning_rate": 1.5488350249409245e-06, "loss": 0.9177, "step": 10157 }, { "epoch": 4.600543478260869, "grad_norm": 0.04495686902871401, "learning_rate": 1.5453516437794159e-06, "loss": 0.918, "step": 10158 }, { "epoch": 4.600996376811594, "grad_norm": 0.04190549463319152, "learning_rate": 1.5418721069900388e-06, "loss": 0.9312, "step": 10159 }, { "epoch": 4.601449275362318, "grad_norm": 0.03733384216847763, "learning_rate": 1.5383964149206442e-06, "loss": 0.9045, "step": 10160 }, { "epoch": 4.601902173913043, "grad_norm": 0.03926444515136123, "learning_rate": 1.5349245679187141e-06, "loss": 0.9076, "step": 10161 }, { "epoch": 4.602355072463768, "grad_norm": 0.038694930673867985, "learning_rate": 1.5314565663313219e-06, "loss": 0.9312, "step": 10162 }, { "epoch": 4.602807971014493, "grad_norm": 0.03696295130728079, "learning_rate": 1.5279924105051813e-06, "loss": 0.9428, "step": 10163 }, { "epoch": 4.603260869565218, "grad_norm": 0.03972004933270338, "learning_rate": 1.5245321007865977e-06, "loss": 0.913, "step": 10164 }, { "epoch": 4.603713768115942, "grad_norm": 0.04028449122971261, "learning_rate": 1.5210756375215118e-06, "loss": 0.9254, "step": 10165 }, { "epoch": 4.604166666666667, "grad_norm": 0.039688631444191456, "learning_rate": 1.5176230210554744e-06, "loss": 0.9353, "step": 10166 }, { "epoch": 4.604619565217392, "grad_norm": 0.04142393568030116, "learning_rate": 1.51417425173364e-06, "loss": 0.9379, "step": 10167 }, { "epoch": 4.605072463768116, "grad_norm": 0.03922592523972006, "learning_rate": 1.5107293299007907e-06, "loss": 0.9354, "step": 10168 }, { "epoch": 4.605525362318841, "grad_norm": 0.036491438672229336, "learning_rate": 1.5072882559013225e-06, "loss": 0.9, "step": 10169 }, { "epoch": 4.605978260869565, "grad_norm": 0.03895976222074788, "learning_rate": 1.503851030079244e-06, "loss": 0.9362, "step": 10170 }, { "epoch": 4.60643115942029, "grad_norm": 0.03837550022094289, "learning_rate": 1.5004176527781833e-06, "loss": 0.915, "step": 10171 }, { "epoch": 4.606884057971015, "grad_norm": 0.03959085439483514, "learning_rate": 1.4969881243413764e-06, "loss": 0.9115, "step": 10172 }, { "epoch": 4.607336956521739, "grad_norm": 0.04120497254591398, "learning_rate": 1.4935624451116737e-06, "loss": 0.9214, "step": 10173 }, { "epoch": 4.607789855072464, "grad_norm": 0.0392768277973522, "learning_rate": 1.4901406154315612e-06, "loss": 0.9171, "step": 10174 }, { "epoch": 4.6082427536231885, "grad_norm": 0.042417725988372, "learning_rate": 1.4867226356431118e-06, "loss": 0.9029, "step": 10175 }, { "epoch": 4.608695652173913, "grad_norm": 0.03955342690955391, "learning_rate": 1.4833085060880349e-06, "loss": 0.9207, "step": 10176 }, { "epoch": 4.609148550724638, "grad_norm": 0.03780989552340065, "learning_rate": 1.4798982271076302e-06, "loss": 0.916, "step": 10177 }, { "epoch": 4.609601449275362, "grad_norm": 0.03972584244984162, "learning_rate": 1.4764917990428518e-06, "loss": 0.8999, "step": 10178 }, { "epoch": 4.610054347826087, "grad_norm": 0.04090776457514047, "learning_rate": 1.473089222234223e-06, "loss": 0.9324, "step": 10179 }, { "epoch": 4.6105072463768115, "grad_norm": 0.04079406388788661, "learning_rate": 1.469690497021925e-06, "loss": 0.916, "step": 10180 }, { "epoch": 4.610960144927536, "grad_norm": 0.04044914899017874, "learning_rate": 1.4662956237457172e-06, "loss": 0.9279, "step": 10181 }, { "epoch": 4.611413043478261, "grad_norm": 0.039963651122685996, "learning_rate": 1.4629046027450034e-06, "loss": 0.9069, "step": 10182 }, { "epoch": 4.611865942028985, "grad_norm": 0.03990125908543132, "learning_rate": 1.4595174343587838e-06, "loss": 0.9229, "step": 10183 }, { "epoch": 4.61231884057971, "grad_norm": 0.039985043212129506, "learning_rate": 1.4561341189256762e-06, "loss": 0.9179, "step": 10184 }, { "epoch": 4.612771739130435, "grad_norm": 0.03761769806659468, "learning_rate": 1.4527546567839169e-06, "loss": 0.9177, "step": 10185 }, { "epoch": 4.613224637681159, "grad_norm": 0.04055842605757917, "learning_rate": 1.4493790482713644e-06, "loss": 0.9216, "step": 10186 }, { "epoch": 4.613677536231884, "grad_norm": 0.03907531214458134, "learning_rate": 1.4460072937254688e-06, "loss": 0.9412, "step": 10187 }, { "epoch": 4.614130434782608, "grad_norm": 0.04050410793548243, "learning_rate": 1.4426393934833249e-06, "loss": 0.9396, "step": 10188 }, { "epoch": 4.614583333333333, "grad_norm": 0.039602041617499625, "learning_rate": 1.4392753478816145e-06, "loss": 0.9115, "step": 10189 }, { "epoch": 4.615036231884058, "grad_norm": 0.03809160982841719, "learning_rate": 1.4359151572566598e-06, "loss": 0.9178, "step": 10190 }, { "epoch": 4.615489130434782, "grad_norm": 0.039658733672373526, "learning_rate": 1.4325588219443743e-06, "loss": 0.9358, "step": 10191 }, { "epoch": 4.615942028985507, "grad_norm": 0.03809588942579177, "learning_rate": 1.4292063422802982e-06, "loss": 0.9008, "step": 10192 }, { "epoch": 4.616394927536232, "grad_norm": 0.038277551604054605, "learning_rate": 1.4258577185995815e-06, "loss": 0.925, "step": 10193 }, { "epoch": 4.616847826086957, "grad_norm": 0.04194334564701172, "learning_rate": 1.4225129512369962e-06, "loss": 0.9194, "step": 10194 }, { "epoch": 4.617300724637682, "grad_norm": 0.04123917005721856, "learning_rate": 1.4191720405269194e-06, "loss": 0.9252, "step": 10195 }, { "epoch": 4.617753623188406, "grad_norm": 0.03942146770599655, "learning_rate": 1.4158349868033594e-06, "loss": 0.9098, "step": 10196 }, { "epoch": 4.618206521739131, "grad_norm": 0.0370208483646501, "learning_rate": 1.4125017903999073e-06, "loss": 0.9153, "step": 10197 }, { "epoch": 4.618659420289855, "grad_norm": 0.038057016035453604, "learning_rate": 1.409172451649803e-06, "loss": 0.9295, "step": 10198 }, { "epoch": 4.61911231884058, "grad_norm": 0.03833902187640348, "learning_rate": 1.405846970885878e-06, "loss": 0.9233, "step": 10199 }, { "epoch": 4.619565217391305, "grad_norm": 0.037853454752192435, "learning_rate": 1.4025253484405954e-06, "loss": 0.9428, "step": 10200 }, { "epoch": 4.620018115942029, "grad_norm": 0.04014698532991687, "learning_rate": 1.399207584646014e-06, "loss": 0.9391, "step": 10201 }, { "epoch": 4.620471014492754, "grad_norm": 0.03758562740633703, "learning_rate": 1.395893679833824e-06, "loss": 0.9312, "step": 10202 }, { "epoch": 4.6209239130434785, "grad_norm": 0.04214882591648239, "learning_rate": 1.3925836343353073e-06, "loss": 0.9115, "step": 10203 }, { "epoch": 4.621376811594203, "grad_norm": 0.03934383196219543, "learning_rate": 1.3892774484813853e-06, "loss": 0.9275, "step": 10204 }, { "epoch": 4.621829710144928, "grad_norm": 0.038591659117707545, "learning_rate": 1.3859751226025808e-06, "loss": 0.9218, "step": 10205 }, { "epoch": 4.622282608695652, "grad_norm": 0.038096252036588035, "learning_rate": 1.382676657029034e-06, "loss": 0.918, "step": 10206 }, { "epoch": 4.622735507246377, "grad_norm": 0.03901097618896555, "learning_rate": 1.3793820520904944e-06, "loss": 0.9286, "step": 10207 }, { "epoch": 4.6231884057971016, "grad_norm": 0.04133568060313472, "learning_rate": 1.3760913081163341e-06, "loss": 0.9361, "step": 10208 }, { "epoch": 4.623641304347826, "grad_norm": 0.04008120930460799, "learning_rate": 1.37280442543553e-06, "loss": 0.9192, "step": 10209 }, { "epoch": 4.624094202898551, "grad_norm": 0.03618010873850004, "learning_rate": 1.369521404376677e-06, "loss": 0.9183, "step": 10210 }, { "epoch": 4.624547101449275, "grad_norm": 0.03696711602236812, "learning_rate": 1.3662422452679792e-06, "loss": 0.9056, "step": 10211 }, { "epoch": 4.625, "grad_norm": 0.03989330975323007, "learning_rate": 1.3629669484372722e-06, "loss": 0.9109, "step": 10212 }, { "epoch": 4.625452898550725, "grad_norm": 0.039028252153018354, "learning_rate": 1.3596955142119784e-06, "loss": 0.9315, "step": 10213 }, { "epoch": 4.625905797101449, "grad_norm": 0.03964773711620325, "learning_rate": 1.3564279429191607e-06, "loss": 0.9404, "step": 10214 }, { "epoch": 4.626358695652174, "grad_norm": 0.04032362628033614, "learning_rate": 1.353164234885469e-06, "loss": 0.9182, "step": 10215 }, { "epoch": 4.6268115942028984, "grad_norm": 0.039639672278429405, "learning_rate": 1.3499043904371978e-06, "loss": 0.9477, "step": 10216 }, { "epoch": 4.627264492753623, "grad_norm": 0.036912284802768223, "learning_rate": 1.346648409900233e-06, "loss": 0.9149, "step": 10217 }, { "epoch": 4.627717391304348, "grad_norm": 0.038838366060545304, "learning_rate": 1.3433962936000745e-06, "loss": 0.908, "step": 10218 }, { "epoch": 4.628170289855072, "grad_norm": 0.038824830887532934, "learning_rate": 1.340148041861844e-06, "loss": 0.9484, "step": 10219 }, { "epoch": 4.628623188405797, "grad_norm": 0.03769125694405492, "learning_rate": 1.3369036550102776e-06, "loss": 0.912, "step": 10220 }, { "epoch": 4.6290760869565215, "grad_norm": 0.04067004284450722, "learning_rate": 1.333663133369716e-06, "loss": 0.9247, "step": 10221 }, { "epoch": 4.629528985507246, "grad_norm": 0.039372658800980405, "learning_rate": 1.3304264772641263e-06, "loss": 0.919, "step": 10222 }, { "epoch": 4.629981884057971, "grad_norm": 0.038075209263883074, "learning_rate": 1.3271936870170764e-06, "loss": 0.9158, "step": 10223 }, { "epoch": 4.630434782608695, "grad_norm": 0.037588138804240195, "learning_rate": 1.3239647629517572e-06, "loss": 0.9191, "step": 10224 }, { "epoch": 4.63088768115942, "grad_norm": 0.03735847769365327, "learning_rate": 1.320739705390972e-06, "loss": 0.9097, "step": 10225 }, { "epoch": 4.631340579710145, "grad_norm": 0.03779811659809753, "learning_rate": 1.3175185146571302e-06, "loss": 0.9016, "step": 10226 }, { "epoch": 4.631793478260869, "grad_norm": 0.03979219642755163, "learning_rate": 1.3143011910722536e-06, "loss": 0.9165, "step": 10227 }, { "epoch": 4.632246376811594, "grad_norm": 0.0399254616216245, "learning_rate": 1.311087734957992e-06, "loss": 0.9144, "step": 10228 }, { "epoch": 4.632699275362318, "grad_norm": 0.03598043617408016, "learning_rate": 1.3078781466355995e-06, "loss": 0.926, "step": 10229 }, { "epoch": 4.633152173913043, "grad_norm": 0.0373405224234102, "learning_rate": 1.3046724264259435e-06, "loss": 0.9269, "step": 10230 }, { "epoch": 4.633605072463768, "grad_norm": 0.039373256810510834, "learning_rate": 1.3014705746494972e-06, "loss": 0.9448, "step": 10231 }, { "epoch": 4.634057971014493, "grad_norm": 0.03910670784556459, "learning_rate": 1.2982725916263638e-06, "loss": 0.9228, "step": 10232 }, { "epoch": 4.634510869565218, "grad_norm": 0.03992338547326075, "learning_rate": 1.295078477676248e-06, "loss": 0.8992, "step": 10233 }, { "epoch": 4.634963768115942, "grad_norm": 0.040848320643000556, "learning_rate": 1.2918882331184724e-06, "loss": 0.9272, "step": 10234 }, { "epoch": 4.635416666666667, "grad_norm": 0.04010139880200475, "learning_rate": 1.2887018582719634e-06, "loss": 0.9223, "step": 10235 }, { "epoch": 4.635869565217392, "grad_norm": 0.03804299217612856, "learning_rate": 1.2855193534552756e-06, "loss": 0.9085, "step": 10236 }, { "epoch": 4.636322463768116, "grad_norm": 0.038539461537823456, "learning_rate": 1.2823407189865633e-06, "loss": 0.9302, "step": 10237 }, { "epoch": 4.636775362318841, "grad_norm": 0.03724076263838944, "learning_rate": 1.2791659551836078e-06, "loss": 0.919, "step": 10238 }, { "epoch": 4.637228260869565, "grad_norm": 0.03903091459404328, "learning_rate": 1.2759950623637863e-06, "loss": 0.9059, "step": 10239 }, { "epoch": 4.63768115942029, "grad_norm": 0.04000786191889384, "learning_rate": 1.272828040844103e-06, "loss": 0.9302, "step": 10240 }, { "epoch": 4.638134057971015, "grad_norm": 0.03748812301536488, "learning_rate": 1.269664890941167e-06, "loss": 0.9324, "step": 10241 }, { "epoch": 4.638586956521739, "grad_norm": 0.03856698497709144, "learning_rate": 1.2665056129712138e-06, "loss": 0.9259, "step": 10242 }, { "epoch": 4.639039855072464, "grad_norm": 0.03984485264942209, "learning_rate": 1.263350207250067e-06, "loss": 0.9526, "step": 10243 }, { "epoch": 4.6394927536231885, "grad_norm": 0.03805593619260021, "learning_rate": 1.260198674093185e-06, "loss": 0.9118, "step": 10244 }, { "epoch": 4.639945652173913, "grad_norm": 0.039142971033805256, "learning_rate": 1.2570510138156223e-06, "loss": 0.9305, "step": 10245 }, { "epoch": 4.640398550724638, "grad_norm": 0.041560444823355386, "learning_rate": 1.253907226732074e-06, "loss": 0.9152, "step": 10246 }, { "epoch": 4.640851449275362, "grad_norm": 0.03770422118302599, "learning_rate": 1.2507673131568087e-06, "loss": 0.916, "step": 10247 }, { "epoch": 4.641304347826087, "grad_norm": 0.040557241983340674, "learning_rate": 1.2476312734037488e-06, "loss": 0.9144, "step": 10248 }, { "epoch": 4.6417572463768115, "grad_norm": 0.038902374811915684, "learning_rate": 1.24449910778639e-06, "loss": 0.9129, "step": 10249 }, { "epoch": 4.642210144927536, "grad_norm": 0.038528129325305804, "learning_rate": 1.241370816617873e-06, "loss": 0.9082, "step": 10250 }, { "epoch": 4.642663043478261, "grad_norm": 0.04201687691473573, "learning_rate": 1.238246400210934e-06, "loss": 0.9101, "step": 10251 }, { "epoch": 4.643115942028985, "grad_norm": 0.039103871699918996, "learning_rate": 1.235125858877928e-06, "loss": 0.9122, "step": 10252 }, { "epoch": 4.64356884057971, "grad_norm": 0.041253950474280286, "learning_rate": 1.2320091929308142e-06, "loss": 0.9216, "step": 10253 }, { "epoch": 4.644021739130435, "grad_norm": 0.038856104017758276, "learning_rate": 1.2288964026811744e-06, "loss": 0.8916, "step": 10254 }, { "epoch": 4.644474637681159, "grad_norm": 0.04147500892361515, "learning_rate": 1.2257874884402e-06, "loss": 0.9257, "step": 10255 }, { "epoch": 4.644927536231884, "grad_norm": 0.037796127367400796, "learning_rate": 1.222682450518695e-06, "loss": 0.9381, "step": 10256 }, { "epoch": 4.645380434782608, "grad_norm": 0.03983640145895968, "learning_rate": 1.2195812892270698e-06, "loss": 0.9247, "step": 10257 }, { "epoch": 4.645833333333333, "grad_norm": 0.03786978227365415, "learning_rate": 1.2164840048753602e-06, "loss": 0.9048, "step": 10258 }, { "epoch": 4.646286231884058, "grad_norm": 0.038511056944260394, "learning_rate": 1.2133905977731986e-06, "loss": 0.9253, "step": 10259 }, { "epoch": 4.646739130434782, "grad_norm": 0.03964350318519202, "learning_rate": 1.2103010682298489e-06, "loss": 0.9092, "step": 10260 }, { "epoch": 4.647192028985507, "grad_norm": 0.03882918849341601, "learning_rate": 1.2072154165541572e-06, "loss": 0.921, "step": 10261 }, { "epoch": 4.647644927536232, "grad_norm": 0.03889601854240105, "learning_rate": 1.2041336430546235e-06, "loss": 0.9346, "step": 10262 }, { "epoch": 4.648097826086957, "grad_norm": 0.03878239197865718, "learning_rate": 1.201055748039317e-06, "loss": 0.92, "step": 10263 }, { "epoch": 4.648550724637682, "grad_norm": 0.03801772081083895, "learning_rate": 1.1979817318159559e-06, "loss": 0.9111, "step": 10264 }, { "epoch": 4.649003623188406, "grad_norm": 0.03947951025684038, "learning_rate": 1.1949115946918454e-06, "loss": 0.9199, "step": 10265 }, { "epoch": 4.649456521739131, "grad_norm": 0.03835236431231093, "learning_rate": 1.191845336973918e-06, "loss": 0.9306, "step": 10266 }, { "epoch": 4.649909420289855, "grad_norm": 0.04249819178580581, "learning_rate": 1.1887829589687105e-06, "loss": 0.9048, "step": 10267 }, { "epoch": 4.65036231884058, "grad_norm": 0.03861495563252067, "learning_rate": 1.1857244609823693e-06, "loss": 0.942, "step": 10268 }, { "epoch": 4.650815217391305, "grad_norm": 0.03899803141872258, "learning_rate": 1.1826698433206584e-06, "loss": 0.9408, "step": 10269 }, { "epoch": 4.651268115942029, "grad_norm": 0.03752668267125149, "learning_rate": 1.1796191062889606e-06, "loss": 0.9093, "step": 10270 }, { "epoch": 4.651721014492754, "grad_norm": 0.036871992290843016, "learning_rate": 1.1765722501922494e-06, "loss": 0.9055, "step": 10271 }, { "epoch": 4.6521739130434785, "grad_norm": 0.04012378631517065, "learning_rate": 1.1735292753351346e-06, "loss": 0.9394, "step": 10272 }, { "epoch": 4.652626811594203, "grad_norm": 0.038864658693254776, "learning_rate": 1.1704901820218217e-06, "loss": 0.9327, "step": 10273 }, { "epoch": 4.653079710144928, "grad_norm": 0.03770269593083958, "learning_rate": 1.167454970556139e-06, "loss": 0.9068, "step": 10274 }, { "epoch": 4.653532608695652, "grad_norm": 0.04020901043604558, "learning_rate": 1.1644236412415189e-06, "loss": 0.9231, "step": 10275 }, { "epoch": 4.653985507246377, "grad_norm": 0.0368476706102729, "learning_rate": 1.161396194381008e-06, "loss": 0.903, "step": 10276 }, { "epoch": 4.6544384057971016, "grad_norm": 0.04011967940886335, "learning_rate": 1.1583726302772625e-06, "loss": 0.9031, "step": 10277 }, { "epoch": 4.654891304347826, "grad_norm": 0.03762421564209233, "learning_rate": 1.1553529492325555e-06, "loss": 0.9347, "step": 10278 }, { "epoch": 4.655344202898551, "grad_norm": 0.04077460643366113, "learning_rate": 1.1523371515487657e-06, "loss": 0.8987, "step": 10279 }, { "epoch": 4.655797101449275, "grad_norm": 0.03882807117517786, "learning_rate": 1.1493252375273944e-06, "loss": 0.9051, "step": 10280 }, { "epoch": 4.65625, "grad_norm": 0.04150752600395751, "learning_rate": 1.1463172074695428e-06, "loss": 0.916, "step": 10281 }, { "epoch": 4.656702898550725, "grad_norm": 0.038948141503903445, "learning_rate": 1.1433130616759304e-06, "loss": 0.9115, "step": 10282 }, { "epoch": 4.657155797101449, "grad_norm": 0.042059036975461636, "learning_rate": 1.140312800446881e-06, "loss": 0.8881, "step": 10283 }, { "epoch": 4.657608695652174, "grad_norm": 0.03684820907905384, "learning_rate": 1.137316424082342e-06, "loss": 0.9066, "step": 10284 }, { "epoch": 4.6580615942028984, "grad_norm": 0.03647916480165372, "learning_rate": 1.13432393288186e-06, "loss": 0.9177, "step": 10285 }, { "epoch": 4.658514492753623, "grad_norm": 0.04213130783640793, "learning_rate": 1.1313353271446093e-06, "loss": 0.9215, "step": 10286 }, { "epoch": 4.658967391304348, "grad_norm": 0.03930070811651734, "learning_rate": 1.1283506071693506e-06, "loss": 0.9146, "step": 10287 }, { "epoch": 4.659420289855072, "grad_norm": 0.03952006575897464, "learning_rate": 1.1253697732544855e-06, "loss": 0.9099, "step": 10288 }, { "epoch": 4.659873188405797, "grad_norm": 0.04124796263134606, "learning_rate": 1.1223928256979977e-06, "loss": 0.9234, "step": 10289 }, { "epoch": 4.6603260869565215, "grad_norm": 0.03788434998603067, "learning_rate": 1.1194197647975158e-06, "loss": 0.9225, "step": 10290 }, { "epoch": 4.660778985507246, "grad_norm": 0.038280380653193335, "learning_rate": 1.1164505908502465e-06, "loss": 0.9016, "step": 10291 }, { "epoch": 4.661231884057971, "grad_norm": 0.037883213956948616, "learning_rate": 1.113485304153028e-06, "loss": 0.9394, "step": 10292 }, { "epoch": 4.661684782608695, "grad_norm": 0.03761346110123443, "learning_rate": 1.1105239050023075e-06, "loss": 0.9274, "step": 10293 }, { "epoch": 4.66213768115942, "grad_norm": 0.040851892764719296, "learning_rate": 1.1075663936941373e-06, "loss": 0.9376, "step": 10294 }, { "epoch": 4.662590579710145, "grad_norm": 0.037294979754297686, "learning_rate": 1.1046127705241871e-06, "loss": 0.9396, "step": 10295 }, { "epoch": 4.663043478260869, "grad_norm": 0.03916162528651294, "learning_rate": 1.1016630357877278e-06, "loss": 0.9184, "step": 10296 }, { "epoch": 4.663496376811594, "grad_norm": 0.038465779957549746, "learning_rate": 1.0987171897796567e-06, "loss": 0.9167, "step": 10297 }, { "epoch": 4.663949275362318, "grad_norm": 0.04076730618771455, "learning_rate": 1.095775232794467e-06, "loss": 0.9311, "step": 10298 }, { "epoch": 4.664402173913043, "grad_norm": 0.04112189851476086, "learning_rate": 1.0928371651262837e-06, "loss": 0.9111, "step": 10299 }, { "epoch": 4.664855072463768, "grad_norm": 0.03851725310984283, "learning_rate": 1.0899029870688183e-06, "loss": 0.9105, "step": 10300 }, { "epoch": 4.665307971014493, "grad_norm": 0.03836257432336832, "learning_rate": 1.086972698915414e-06, "loss": 0.8829, "step": 10301 }, { "epoch": 4.665760869565218, "grad_norm": 0.03662752715905604, "learning_rate": 1.084046300959014e-06, "loss": 0.9069, "step": 10302 }, { "epoch": 4.666213768115942, "grad_norm": 0.03855108886317635, "learning_rate": 1.081123793492167e-06, "loss": 0.9246, "step": 10303 }, { "epoch": 4.666666666666667, "grad_norm": 0.03785748883405847, "learning_rate": 1.0782051768070477e-06, "loss": 0.9294, "step": 10304 }, { "epoch": 4.667119565217392, "grad_norm": 0.03961373545883071, "learning_rate": 1.0752904511954364e-06, "loss": 0.9282, "step": 10305 }, { "epoch": 4.667572463768116, "grad_norm": 0.03678734984713123, "learning_rate": 1.0723796169487178e-06, "loss": 0.9484, "step": 10306 }, { "epoch": 4.668025362318841, "grad_norm": 0.03696995566416789, "learning_rate": 1.0694726743578944e-06, "loss": 0.9075, "step": 10307 }, { "epoch": 4.668478260869565, "grad_norm": 0.03734997145131694, "learning_rate": 1.0665696237135782e-06, "loss": 0.9193, "step": 10308 }, { "epoch": 4.66893115942029, "grad_norm": 0.038841417103290754, "learning_rate": 1.0636704653059949e-06, "loss": 0.9366, "step": 10309 }, { "epoch": 4.669384057971015, "grad_norm": 0.04169600913543081, "learning_rate": 1.0607751994249749e-06, "loss": 0.9232, "step": 10310 }, { "epoch": 4.669836956521739, "grad_norm": 0.03982592967837628, "learning_rate": 1.0578838263599622e-06, "loss": 0.9294, "step": 10311 }, { "epoch": 4.670289855072464, "grad_norm": 0.03689127378705932, "learning_rate": 1.0549963464000102e-06, "loss": 0.9229, "step": 10312 }, { "epoch": 4.6707427536231885, "grad_norm": 0.03671445066871851, "learning_rate": 1.0521127598337899e-06, "loss": 0.9305, "step": 10313 }, { "epoch": 4.671195652173913, "grad_norm": 0.03814423088772221, "learning_rate": 1.0492330669495688e-06, "loss": 0.9269, "step": 10314 }, { "epoch": 4.671648550724638, "grad_norm": 0.03826333753959214, "learning_rate": 1.0463572680352495e-06, "loss": 0.9235, "step": 10315 }, { "epoch": 4.672101449275362, "grad_norm": 0.03854327598293935, "learning_rate": 1.0434853633783137e-06, "loss": 0.8927, "step": 10316 }, { "epoch": 4.672554347826087, "grad_norm": 0.038526459120470945, "learning_rate": 1.040617353265887e-06, "loss": 0.9166, "step": 10317 }, { "epoch": 4.6730072463768115, "grad_norm": 0.037220308702188, "learning_rate": 1.0377532379846734e-06, "loss": 0.9232, "step": 10318 }, { "epoch": 4.673460144927536, "grad_norm": 0.038351804758694485, "learning_rate": 1.0348930178210125e-06, "loss": 0.9074, "step": 10319 }, { "epoch": 4.673913043478261, "grad_norm": 0.03830323409828969, "learning_rate": 1.0320366930608405e-06, "loss": 0.9046, "step": 10320 }, { "epoch": 4.674365942028985, "grad_norm": 0.038342355065544106, "learning_rate": 1.029184263989711e-06, "loss": 0.9281, "step": 10321 }, { "epoch": 4.67481884057971, "grad_norm": 0.03884326616895421, "learning_rate": 1.0263357308927823e-06, "loss": 0.9111, "step": 10322 }, { "epoch": 4.675271739130435, "grad_norm": 0.03831882433267317, "learning_rate": 1.023491094054836e-06, "loss": 0.9096, "step": 10323 }, { "epoch": 4.675724637681159, "grad_norm": 0.04616520095859584, "learning_rate": 1.0206503537602396e-06, "loss": 0.927, "step": 10324 }, { "epoch": 4.676177536231884, "grad_norm": 0.03905674280906037, "learning_rate": 1.017813510293002e-06, "loss": 0.9404, "step": 10325 }, { "epoch": 4.676630434782608, "grad_norm": 0.03985044293701003, "learning_rate": 1.0149805639367226e-06, "loss": 0.9226, "step": 10326 }, { "epoch": 4.677083333333333, "grad_norm": 0.03846346463515557, "learning_rate": 1.0121515149746108e-06, "loss": 0.9146, "step": 10327 }, { "epoch": 4.677536231884058, "grad_norm": 0.03735427493063335, "learning_rate": 1.009326363689489e-06, "loss": 0.9126, "step": 10328 }, { "epoch": 4.677989130434782, "grad_norm": 0.03870682063015068, "learning_rate": 1.0065051103638025e-06, "loss": 0.9215, "step": 10329 }, { "epoch": 4.678442028985507, "grad_norm": 0.038120348401327865, "learning_rate": 1.003687755279583e-06, "loss": 0.9017, "step": 10330 }, { "epoch": 4.678894927536232, "grad_norm": 0.03934956746307867, "learning_rate": 1.000874298718504e-06, "loss": 0.9421, "step": 10331 }, { "epoch": 4.679347826086957, "grad_norm": 0.04038721023213418, "learning_rate": 9.980647409618106e-07, "loss": 0.9323, "step": 10332 }, { "epoch": 4.679800724637682, "grad_norm": 0.03885688609713341, "learning_rate": 9.95259082290394e-07, "loss": 0.9277, "step": 10333 }, { "epoch": 4.680253623188406, "grad_norm": 0.0377724871145879, "learning_rate": 9.924573229847368e-07, "loss": 0.9465, "step": 10334 }, { "epoch": 4.680706521739131, "grad_norm": 0.038491795827122796, "learning_rate": 9.896594633249301e-07, "loss": 0.9157, "step": 10335 }, { "epoch": 4.681159420289855, "grad_norm": 0.03679832249344336, "learning_rate": 9.868655035906883e-07, "loss": 0.9074, "step": 10336 }, { "epoch": 4.68161231884058, "grad_norm": 0.04024426019590344, "learning_rate": 9.840754440613253e-07, "loss": 0.9261, "step": 10337 }, { "epoch": 4.682065217391305, "grad_norm": 0.03766798401449295, "learning_rate": 9.812892850157563e-07, "loss": 0.9099, "step": 10338 }, { "epoch": 4.682518115942029, "grad_norm": 0.03701911059463839, "learning_rate": 9.785070267325358e-07, "loss": 0.925, "step": 10339 }, { "epoch": 4.682971014492754, "grad_norm": 0.037896739545924325, "learning_rate": 9.757286694898016e-07, "loss": 0.9066, "step": 10340 }, { "epoch": 4.6834239130434785, "grad_norm": 0.036639647405761505, "learning_rate": 9.729542135653092e-07, "loss": 0.9184, "step": 10341 }, { "epoch": 4.683876811594203, "grad_norm": 0.03698358628222451, "learning_rate": 9.701836592364278e-07, "loss": 0.9023, "step": 10342 }, { "epoch": 4.684329710144928, "grad_norm": 0.037587032611088714, "learning_rate": 9.674170067801359e-07, "loss": 0.9211, "step": 10343 }, { "epoch": 4.684782608695652, "grad_norm": 0.03654446583146772, "learning_rate": 9.64654256473021e-07, "loss": 0.9123, "step": 10344 }, { "epoch": 4.685235507246377, "grad_norm": 0.03754035618086566, "learning_rate": 9.618954085912757e-07, "loss": 0.9408, "step": 10345 }, { "epoch": 4.6856884057971016, "grad_norm": 0.03718528180577924, "learning_rate": 9.591404634107015e-07, "loss": 0.9016, "step": 10346 }, { "epoch": 4.686141304347826, "grad_norm": 0.03816593668354792, "learning_rate": 9.56389421206727e-07, "loss": 0.9399, "step": 10347 }, { "epoch": 4.686594202898551, "grad_norm": 0.039281320329493934, "learning_rate": 9.536422822543678e-07, "loss": 0.9279, "step": 10348 }, { "epoch": 4.687047101449275, "grad_norm": 0.03973118050212915, "learning_rate": 9.508990468282664e-07, "loss": 0.9124, "step": 10349 }, { "epoch": 4.6875, "grad_norm": 0.04089781775362704, "learning_rate": 9.481597152026656e-07, "loss": 0.9052, "step": 10350 }, { "epoch": 4.687952898550725, "grad_norm": 0.03834299936850109, "learning_rate": 9.454242876514219e-07, "loss": 0.933, "step": 10351 }, { "epoch": 4.688405797101449, "grad_norm": 0.03641088207584505, "learning_rate": 9.42692764448001e-07, "loss": 0.9195, "step": 10352 }, { "epoch": 4.688858695652174, "grad_norm": 0.03699452850875072, "learning_rate": 9.399651458654735e-07, "loss": 0.9238, "step": 10353 }, { "epoch": 4.6893115942028984, "grad_norm": 0.03776040710451125, "learning_rate": 9.372414321765233e-07, "loss": 0.9508, "step": 10354 }, { "epoch": 4.689764492753623, "grad_norm": 0.03684863510004585, "learning_rate": 9.345216236534527e-07, "loss": 0.9221, "step": 10355 }, { "epoch": 4.690217391304348, "grad_norm": 0.04014912464939504, "learning_rate": 9.318057205681552e-07, "loss": 0.9308, "step": 10356 }, { "epoch": 4.690670289855072, "grad_norm": 0.03849945821732011, "learning_rate": 9.290937231921516e-07, "loss": 0.9163, "step": 10357 }, { "epoch": 4.691123188405797, "grad_norm": 0.03877882471396457, "learning_rate": 9.26385631796558e-07, "loss": 0.9119, "step": 10358 }, { "epoch": 4.6915760869565215, "grad_norm": 0.040126106112901676, "learning_rate": 9.236814466521138e-07, "loss": 0.9246, "step": 10359 }, { "epoch": 4.692028985507246, "grad_norm": 0.03841756248611169, "learning_rate": 9.209811680291581e-07, "loss": 0.931, "step": 10360 }, { "epoch": 4.692481884057971, "grad_norm": 0.038500733565449974, "learning_rate": 9.182847961976438e-07, "loss": 0.9195, "step": 10361 }, { "epoch": 4.692934782608695, "grad_norm": 0.04055765888240192, "learning_rate": 9.155923314271197e-07, "loss": 0.9286, "step": 10362 }, { "epoch": 4.69338768115942, "grad_norm": 0.03650198754182788, "learning_rate": 9.129037739867707e-07, "loss": 0.9074, "step": 10363 }, { "epoch": 4.693840579710145, "grad_norm": 0.03775466697718136, "learning_rate": 9.102191241453684e-07, "loss": 0.9314, "step": 10364 }, { "epoch": 4.694293478260869, "grad_norm": 0.03866490669598154, "learning_rate": 9.075383821713069e-07, "loss": 0.9236, "step": 10365 }, { "epoch": 4.694746376811594, "grad_norm": 0.04122223499322153, "learning_rate": 9.048615483325762e-07, "loss": 0.9151, "step": 10366 }, { "epoch": 4.695199275362318, "grad_norm": 0.03952993289165594, "learning_rate": 9.021886228967891e-07, "loss": 0.924, "step": 10367 }, { "epoch": 4.695652173913043, "grad_norm": 0.03825546294996581, "learning_rate": 8.995196061311628e-07, "loss": 0.9196, "step": 10368 }, { "epoch": 4.696105072463768, "grad_norm": 0.03828461327399491, "learning_rate": 8.96854498302524e-07, "loss": 0.9335, "step": 10369 }, { "epoch": 4.696557971014493, "grad_norm": 0.03714033951724087, "learning_rate": 8.941932996772995e-07, "loss": 0.9146, "step": 10370 }, { "epoch": 4.697010869565218, "grad_norm": 0.040635387106769276, "learning_rate": 8.915360105215431e-07, "loss": 0.9253, "step": 10371 }, { "epoch": 4.697463768115942, "grad_norm": 0.03742229101884284, "learning_rate": 8.888826311009002e-07, "loss": 0.9065, "step": 10372 }, { "epoch": 4.697916666666667, "grad_norm": 0.037437051232585984, "learning_rate": 8.862331616806385e-07, "loss": 0.9362, "step": 10373 }, { "epoch": 4.698369565217392, "grad_norm": 0.03902516441076413, "learning_rate": 8.835876025256307e-07, "loss": 0.9466, "step": 10374 }, { "epoch": 4.698822463768116, "grad_norm": 0.038729788651111124, "learning_rate": 8.809459539003584e-07, "loss": 0.9195, "step": 10375 }, { "epoch": 4.699275362318841, "grad_norm": 0.04000411361481296, "learning_rate": 8.783082160688994e-07, "loss": 0.9227, "step": 10376 }, { "epoch": 4.699728260869565, "grad_norm": 0.03998058485092, "learning_rate": 8.756743892949715e-07, "loss": 0.9152, "step": 10377 }, { "epoch": 4.70018115942029, "grad_norm": 0.038286201660181894, "learning_rate": 8.730444738418708e-07, "loss": 0.9173, "step": 10378 }, { "epoch": 4.700634057971015, "grad_norm": 0.04289369457653661, "learning_rate": 8.70418469972516e-07, "loss": 0.9321, "step": 10379 }, { "epoch": 4.701086956521739, "grad_norm": 0.03977993962368787, "learning_rate": 8.677963779494259e-07, "loss": 0.9199, "step": 10380 }, { "epoch": 4.701539855072464, "grad_norm": 0.03669391393690091, "learning_rate": 8.651781980347507e-07, "loss": 0.9113, "step": 10381 }, { "epoch": 4.7019927536231885, "grad_norm": 0.037178802946567516, "learning_rate": 8.625639304902189e-07, "loss": 0.9188, "step": 10382 }, { "epoch": 4.702445652173913, "grad_norm": 0.03922938205658775, "learning_rate": 8.599535755771948e-07, "loss": 0.9031, "step": 10383 }, { "epoch": 4.702898550724638, "grad_norm": 0.03861378165956963, "learning_rate": 8.573471335566297e-07, "loss": 0.9223, "step": 10384 }, { "epoch": 4.703351449275362, "grad_norm": 0.03846711987632096, "learning_rate": 8.547446046891017e-07, "loss": 0.9232, "step": 10385 }, { "epoch": 4.703804347826087, "grad_norm": 0.037983048574318785, "learning_rate": 8.521459892347894e-07, "loss": 0.9057, "step": 10386 }, { "epoch": 4.7042572463768115, "grad_norm": 0.038007999435476975, "learning_rate": 8.495512874534761e-07, "loss": 0.9138, "step": 10387 }, { "epoch": 4.704710144927536, "grad_norm": 0.03664525215942689, "learning_rate": 8.469604996045544e-07, "loss": 0.9336, "step": 10388 }, { "epoch": 4.705163043478261, "grad_norm": 0.039188754642871165, "learning_rate": 8.443736259470392e-07, "loss": 0.9341, "step": 10389 }, { "epoch": 4.705615942028985, "grad_norm": 0.03775564386101192, "learning_rate": 8.41790666739537e-07, "loss": 0.9259, "step": 10390 }, { "epoch": 4.70606884057971, "grad_norm": 0.039477017244129116, "learning_rate": 8.392116222402768e-07, "loss": 0.9116, "step": 10391 }, { "epoch": 4.706521739130435, "grad_norm": 0.042033734810031954, "learning_rate": 8.366364927070836e-07, "loss": 0.8959, "step": 10392 }, { "epoch": 4.706974637681159, "grad_norm": 0.03911944825995978, "learning_rate": 8.340652783974046e-07, "loss": 0.9393, "step": 10393 }, { "epoch": 4.707427536231884, "grad_norm": 0.04074023441665327, "learning_rate": 8.314979795682787e-07, "loss": 0.9163, "step": 10394 }, { "epoch": 4.707880434782608, "grad_norm": 0.04380004669692229, "learning_rate": 8.289345964763718e-07, "loss": 0.9477, "step": 10395 }, { "epoch": 4.708333333333333, "grad_norm": 0.037315139190655496, "learning_rate": 8.263751293779409e-07, "loss": 0.9339, "step": 10396 }, { "epoch": 4.708786231884058, "grad_norm": 0.0393712928268729, "learning_rate": 8.238195785288661e-07, "loss": 0.9252, "step": 10397 }, { "epoch": 4.709239130434782, "grad_norm": 0.03762901868151051, "learning_rate": 8.212679441846272e-07, "loss": 0.9295, "step": 10398 }, { "epoch": 4.709692028985507, "grad_norm": 0.03652622204636126, "learning_rate": 8.187202266003181e-07, "loss": 0.917, "step": 10399 }, { "epoch": 4.710144927536232, "grad_norm": 0.040032104349505146, "learning_rate": 8.161764260306327e-07, "loss": 0.9203, "step": 10400 }, { "epoch": 4.710597826086957, "grad_norm": 0.03931299689973503, "learning_rate": 8.136365427298831e-07, "loss": 0.9196, "step": 10401 }, { "epoch": 4.711050724637682, "grad_norm": 0.03895746224557188, "learning_rate": 8.111005769519864e-07, "loss": 0.9305, "step": 10402 }, { "epoch": 4.711503623188406, "grad_norm": 0.04081890498320308, "learning_rate": 8.085685289504641e-07, "loss": 0.9364, "step": 10403 }, { "epoch": 4.711956521739131, "grad_norm": 0.0400485217021274, "learning_rate": 8.060403989784516e-07, "loss": 0.9201, "step": 10404 }, { "epoch": 4.712409420289855, "grad_norm": 0.03800837855418398, "learning_rate": 8.035161872886887e-07, "loss": 0.929, "step": 10405 }, { "epoch": 4.71286231884058, "grad_norm": 0.038845286130375245, "learning_rate": 8.009958941335205e-07, "loss": 0.9526, "step": 10406 }, { "epoch": 4.713315217391305, "grad_norm": 0.03839494492207934, "learning_rate": 7.984795197649098e-07, "loss": 0.9349, "step": 10407 }, { "epoch": 4.713768115942029, "grad_norm": 0.039422140045337964, "learning_rate": 7.959670644344242e-07, "loss": 0.9488, "step": 10408 }, { "epoch": 4.714221014492754, "grad_norm": 0.03802743375086967, "learning_rate": 7.934585283932317e-07, "loss": 0.9191, "step": 10409 }, { "epoch": 4.7146739130434785, "grad_norm": 0.03918310991927449, "learning_rate": 7.909539118921228e-07, "loss": 0.9146, "step": 10410 }, { "epoch": 4.715126811594203, "grad_norm": 0.03789804393249812, "learning_rate": 7.884532151814839e-07, "loss": 0.8973, "step": 10411 }, { "epoch": 4.715579710144928, "grad_norm": 0.03799543471699956, "learning_rate": 7.859564385113105e-07, "loss": 0.9232, "step": 10412 }, { "epoch": 4.716032608695652, "grad_norm": 0.03697411528305496, "learning_rate": 7.83463582131212e-07, "loss": 0.9248, "step": 10413 }, { "epoch": 4.716485507246377, "grad_norm": 0.03876300778861716, "learning_rate": 7.809746462904022e-07, "loss": 0.9268, "step": 10414 }, { "epoch": 4.7169384057971016, "grad_norm": 0.03825937380304353, "learning_rate": 7.784896312377044e-07, "loss": 0.9077, "step": 10415 }, { "epoch": 4.717391304347826, "grad_norm": 0.038352031448123665, "learning_rate": 7.760085372215509e-07, "loss": 0.9276, "step": 10416 }, { "epoch": 4.717844202898551, "grad_norm": 0.0398074193899418, "learning_rate": 7.735313644899789e-07, "loss": 0.9323, "step": 10417 }, { "epoch": 4.718297101449275, "grad_norm": 0.03787078208362362, "learning_rate": 7.710581132906391e-07, "loss": 0.9065, "step": 10418 }, { "epoch": 4.71875, "grad_norm": 0.03884982635228142, "learning_rate": 7.685887838707828e-07, "loss": 0.923, "step": 10419 }, { "epoch": 4.719202898550725, "grad_norm": 0.037071226223985336, "learning_rate": 7.661233764772746e-07, "loss": 0.9207, "step": 10420 }, { "epoch": 4.719655797101449, "grad_norm": 0.03903232168679493, "learning_rate": 7.636618913565796e-07, "loss": 0.9092, "step": 10421 }, { "epoch": 4.720108695652174, "grad_norm": 0.03954616161179176, "learning_rate": 7.612043287547855e-07, "loss": 0.9053, "step": 10422 }, { "epoch": 4.7205615942028984, "grad_norm": 0.03712587374558307, "learning_rate": 7.587506889175666e-07, "loss": 0.9219, "step": 10423 }, { "epoch": 4.721014492753623, "grad_norm": 0.038144418822263736, "learning_rate": 7.563009720902292e-07, "loss": 0.9285, "step": 10424 }, { "epoch": 4.721467391304348, "grad_norm": 0.04184136075995, "learning_rate": 7.538551785176707e-07, "loss": 0.9158, "step": 10425 }, { "epoch": 4.721920289855072, "grad_norm": 0.03996738238875558, "learning_rate": 7.514133084443975e-07, "loss": 0.9288, "step": 10426 }, { "epoch": 4.722373188405797, "grad_norm": 0.03582951455113826, "learning_rate": 7.489753621145346e-07, "loss": 0.8997, "step": 10427 }, { "epoch": 4.7228260869565215, "grad_norm": 0.03894738873291933, "learning_rate": 7.465413397718024e-07, "loss": 0.9361, "step": 10428 }, { "epoch": 4.723278985507246, "grad_norm": 0.037319259652520656, "learning_rate": 7.441112416595353e-07, "loss": 0.9047, "step": 10429 }, { "epoch": 4.723731884057971, "grad_norm": 0.03731343259458829, "learning_rate": 7.416850680206722e-07, "loss": 0.9017, "step": 10430 }, { "epoch": 4.724184782608695, "grad_norm": 0.03892196864862223, "learning_rate": 7.392628190977613e-07, "loss": 0.9128, "step": 10431 }, { "epoch": 4.72463768115942, "grad_norm": 0.04000104167267482, "learning_rate": 7.368444951329645e-07, "loss": 0.9192, "step": 10432 }, { "epoch": 4.725090579710145, "grad_norm": 0.03810983270274422, "learning_rate": 7.344300963680395e-07, "loss": 0.9221, "step": 10433 }, { "epoch": 4.725543478260869, "grad_norm": 0.036722704900095555, "learning_rate": 7.320196230443577e-07, "loss": 0.9281, "step": 10434 }, { "epoch": 4.725996376811594, "grad_norm": 0.03915931048257273, "learning_rate": 7.296130754028997e-07, "loss": 0.9072, "step": 10435 }, { "epoch": 4.726449275362318, "grad_norm": 0.03812460571180458, "learning_rate": 7.27210453684255e-07, "loss": 0.9129, "step": 10436 }, { "epoch": 4.726902173913043, "grad_norm": 0.03847592243397702, "learning_rate": 7.24811758128614e-07, "loss": 0.9256, "step": 10437 }, { "epoch": 4.727355072463768, "grad_norm": 0.03833429333964274, "learning_rate": 7.224169889757804e-07, "loss": 0.9433, "step": 10438 }, { "epoch": 4.727807971014493, "grad_norm": 0.04004253991221881, "learning_rate": 7.200261464651537e-07, "loss": 0.9109, "step": 10439 }, { "epoch": 4.728260869565218, "grad_norm": 0.040112114158250604, "learning_rate": 7.176392308357649e-07, "loss": 0.9251, "step": 10440 }, { "epoch": 4.728713768115942, "grad_norm": 0.039552225370209144, "learning_rate": 7.152562423262232e-07, "loss": 0.9226, "step": 10441 }, { "epoch": 4.729166666666667, "grad_norm": 0.03662072381950908, "learning_rate": 7.128771811747737e-07, "loss": 0.9316, "step": 10442 }, { "epoch": 4.729619565217392, "grad_norm": 0.03786466370859477, "learning_rate": 7.105020476192437e-07, "loss": 0.9358, "step": 10443 }, { "epoch": 4.730072463768116, "grad_norm": 0.0364629946881178, "learning_rate": 7.081308418970878e-07, "loss": 0.9364, "step": 10444 }, { "epoch": 4.730525362318841, "grad_norm": 0.03637584172608129, "learning_rate": 7.057635642453519e-07, "loss": 0.9465, "step": 10445 }, { "epoch": 4.730978260869565, "grad_norm": 0.04087298543676972, "learning_rate": 7.034002149007002e-07, "loss": 0.932, "step": 10446 }, { "epoch": 4.73143115942029, "grad_norm": 0.040979004270500496, "learning_rate": 7.010407940993969e-07, "loss": 0.9379, "step": 10447 }, { "epoch": 4.731884057971015, "grad_norm": 0.03790772457972925, "learning_rate": 6.986853020773243e-07, "loss": 0.9241, "step": 10448 }, { "epoch": 4.732336956521739, "grad_norm": 0.03896568059767878, "learning_rate": 6.963337390699565e-07, "loss": 0.9271, "step": 10449 }, { "epoch": 4.732789855072464, "grad_norm": 0.0385099540595427, "learning_rate": 6.939861053123897e-07, "loss": 0.9283, "step": 10450 }, { "epoch": 4.7332427536231885, "grad_norm": 0.03735090092717992, "learning_rate": 6.916424010393164e-07, "loss": 0.9268, "step": 10451 }, { "epoch": 4.733695652173913, "grad_norm": 0.03826670138367364, "learning_rate": 6.893026264850466e-07, "loss": 0.9241, "step": 10452 }, { "epoch": 4.734148550724638, "grad_norm": 0.038748887432527426, "learning_rate": 6.869667818834824e-07, "loss": 0.9119, "step": 10453 }, { "epoch": 4.734601449275362, "grad_norm": 0.03894919125157793, "learning_rate": 6.846348674681436e-07, "loss": 0.9103, "step": 10454 }, { "epoch": 4.735054347826087, "grad_norm": 0.039613436250340694, "learning_rate": 6.823068834721591e-07, "loss": 0.9233, "step": 10455 }, { "epoch": 4.7355072463768115, "grad_norm": 0.03957054202704786, "learning_rate": 6.799828301282629e-07, "loss": 0.9133, "step": 10456 }, { "epoch": 4.735960144927536, "grad_norm": 0.03777489085325715, "learning_rate": 6.776627076687847e-07, "loss": 0.9148, "step": 10457 }, { "epoch": 4.736413043478261, "grad_norm": 0.04174778311576777, "learning_rate": 6.753465163256812e-07, "loss": 0.9148, "step": 10458 }, { "epoch": 4.736865942028985, "grad_norm": 0.03805887306187193, "learning_rate": 6.730342563305003e-07, "loss": 0.914, "step": 10459 }, { "epoch": 4.73731884057971, "grad_norm": 0.036256109594394834, "learning_rate": 6.707259279144085e-07, "loss": 0.915, "step": 10460 }, { "epoch": 4.737771739130435, "grad_norm": 0.03610814453005923, "learning_rate": 6.684215313081632e-07, "loss": 0.9359, "step": 10461 }, { "epoch": 4.738224637681159, "grad_norm": 0.040041250137334616, "learning_rate": 6.661210667421491e-07, "loss": 0.9379, "step": 10462 }, { "epoch": 4.738677536231884, "grad_norm": 0.04078801628320759, "learning_rate": 6.638245344463335e-07, "loss": 0.9292, "step": 10463 }, { "epoch": 4.739130434782608, "grad_norm": 0.03692399336652508, "learning_rate": 6.61531934650319e-07, "loss": 0.916, "step": 10464 }, { "epoch": 4.739583333333333, "grad_norm": 0.03869395709583493, "learning_rate": 6.592432675832916e-07, "loss": 0.9142, "step": 10465 }, { "epoch": 4.740036231884058, "grad_norm": 0.03754822867144402, "learning_rate": 6.569585334740546e-07, "loss": 0.9155, "step": 10466 }, { "epoch": 4.740489130434782, "grad_norm": 0.03730820659322543, "learning_rate": 6.546777325510167e-07, "loss": 0.9177, "step": 10467 }, { "epoch": 4.740942028985507, "grad_norm": 0.036622217656815934, "learning_rate": 6.524008650421998e-07, "loss": 0.9426, "step": 10468 }, { "epoch": 4.741394927536232, "grad_norm": 0.03766589988493923, "learning_rate": 6.501279311752129e-07, "loss": 0.9295, "step": 10469 }, { "epoch": 4.741847826086957, "grad_norm": 0.03680266937633646, "learning_rate": 6.478589311772965e-07, "loss": 0.9147, "step": 10470 }, { "epoch": 4.742300724637682, "grad_norm": 0.03875785376871532, "learning_rate": 6.455938652752824e-07, "loss": 0.909, "step": 10471 }, { "epoch": 4.742753623188406, "grad_norm": 0.0379741542297821, "learning_rate": 6.433327336956119e-07, "loss": 0.9364, "step": 10472 }, { "epoch": 4.743206521739131, "grad_norm": 0.04206595486341994, "learning_rate": 6.410755366643307e-07, "loss": 0.9232, "step": 10473 }, { "epoch": 4.743659420289855, "grad_norm": 0.036579513368038864, "learning_rate": 6.388222744070982e-07, "loss": 0.9315, "step": 10474 }, { "epoch": 4.74411231884058, "grad_norm": 0.038578481618590704, "learning_rate": 6.365729471491743e-07, "loss": 0.9212, "step": 10475 }, { "epoch": 4.744565217391305, "grad_norm": 0.035867236301177344, "learning_rate": 6.343275551154326e-07, "loss": 0.9201, "step": 10476 }, { "epoch": 4.745018115942029, "grad_norm": 0.037690195738803445, "learning_rate": 6.320860985303467e-07, "loss": 0.9074, "step": 10477 }, { "epoch": 4.745471014492754, "grad_norm": 0.039251820520941984, "learning_rate": 6.298485776179953e-07, "loss": 0.9269, "step": 10478 }, { "epoch": 4.7459239130434785, "grad_norm": 0.03898591637453071, "learning_rate": 6.276149926020703e-07, "loss": 0.9292, "step": 10479 }, { "epoch": 4.746376811594203, "grad_norm": 0.03707297060053359, "learning_rate": 6.253853437058644e-07, "loss": 0.9048, "step": 10480 }, { "epoch": 4.746829710144928, "grad_norm": 0.03802917744968732, "learning_rate": 6.231596311522792e-07, "loss": 0.8975, "step": 10481 }, { "epoch": 4.747282608695652, "grad_norm": 0.04163878061168961, "learning_rate": 6.209378551638256e-07, "loss": 0.9149, "step": 10482 }, { "epoch": 4.747735507246377, "grad_norm": 0.03880790565465085, "learning_rate": 6.18720015962615e-07, "loss": 0.9223, "step": 10483 }, { "epoch": 4.7481884057971016, "grad_norm": 0.03696914829842618, "learning_rate": 6.165061137703677e-07, "loss": 0.9155, "step": 10484 }, { "epoch": 4.748641304347826, "grad_norm": 0.0376912783819811, "learning_rate": 6.142961488084131e-07, "loss": 0.9215, "step": 10485 }, { "epoch": 4.749094202898551, "grad_norm": 0.03821658164815072, "learning_rate": 6.12090121297686e-07, "loss": 0.929, "step": 10486 }, { "epoch": 4.749547101449275, "grad_norm": 0.036133942173413244, "learning_rate": 6.098880314587296e-07, "loss": 0.9184, "step": 10487 }, { "epoch": 4.75, "grad_norm": 0.03707045493767377, "learning_rate": 6.076898795116792e-07, "loss": 0.9, "step": 10488 }, { "epoch": 4.750452898550725, "grad_norm": 0.037320238371788245, "learning_rate": 6.054956656762967e-07, "loss": 0.9079, "step": 10489 }, { "epoch": 4.750905797101449, "grad_norm": 0.03866063759126723, "learning_rate": 6.033053901719355e-07, "loss": 0.9196, "step": 10490 }, { "epoch": 4.751358695652174, "grad_norm": 0.03856555240659706, "learning_rate": 6.01119053217567e-07, "loss": 0.9132, "step": 10491 }, { "epoch": 4.7518115942028984, "grad_norm": 0.03980579009228084, "learning_rate": 5.989366550317588e-07, "loss": 0.9404, "step": 10492 }, { "epoch": 4.752264492753623, "grad_norm": 0.036377984620127486, "learning_rate": 5.967581958326873e-07, "loss": 0.9334, "step": 10493 }, { "epoch": 4.752717391304348, "grad_norm": 0.03901002323445454, "learning_rate": 5.945836758381429e-07, "loss": 0.9321, "step": 10494 }, { "epoch": 4.753170289855072, "grad_norm": 0.03747337746677525, "learning_rate": 5.924130952655071e-07, "loss": 0.9242, "step": 10495 }, { "epoch": 4.753623188405797, "grad_norm": 0.03742746309010339, "learning_rate": 5.902464543317843e-07, "loss": 0.9281, "step": 10496 }, { "epoch": 4.7540760869565215, "grad_norm": 0.037382512486775626, "learning_rate": 5.8808375325357e-07, "loss": 0.9233, "step": 10497 }, { "epoch": 4.754528985507246, "grad_norm": 0.03771794895953296, "learning_rate": 5.859249922470823e-07, "loss": 0.9185, "step": 10498 }, { "epoch": 4.754981884057971, "grad_norm": 0.037167515396453195, "learning_rate": 5.837701715281263e-07, "loss": 0.9221, "step": 10499 }, { "epoch": 4.755434782608695, "grad_norm": 0.0379284562886249, "learning_rate": 5.816192913121255e-07, "loss": 0.9004, "step": 10500 }, { "epoch": 4.75588768115942, "grad_norm": 0.03607437695294167, "learning_rate": 5.794723518141077e-07, "loss": 0.9146, "step": 10501 }, { "epoch": 4.756340579710145, "grad_norm": 0.04042331392755136, "learning_rate": 5.773293532487101e-07, "loss": 0.8895, "step": 10502 }, { "epoch": 4.756793478260869, "grad_norm": 0.03745922065632566, "learning_rate": 5.751902958301658e-07, "loss": 0.9072, "step": 10503 }, { "epoch": 4.757246376811594, "grad_norm": 0.038041742539511904, "learning_rate": 5.730551797723216e-07, "loss": 0.9415, "step": 10504 }, { "epoch": 4.757699275362318, "grad_norm": 0.03681288868023213, "learning_rate": 5.709240052886289e-07, "loss": 0.9155, "step": 10505 }, { "epoch": 4.758152173913043, "grad_norm": 0.036300826019734285, "learning_rate": 5.687967725921439e-07, "loss": 0.8942, "step": 10506 }, { "epoch": 4.758605072463768, "grad_norm": 0.03694768027655317, "learning_rate": 5.666734818955233e-07, "loss": 0.9052, "step": 10507 }, { "epoch": 4.759057971014493, "grad_norm": 0.03751356183246157, "learning_rate": 5.645541334110505e-07, "loss": 0.9264, "step": 10508 }, { "epoch": 4.759510869565218, "grad_norm": 0.036891903747225126, "learning_rate": 5.624387273505871e-07, "loss": 0.9207, "step": 10509 }, { "epoch": 4.759963768115942, "grad_norm": 0.03840641864959044, "learning_rate": 5.603272639256219e-07, "loss": 0.9377, "step": 10510 }, { "epoch": 4.760416666666667, "grad_norm": 0.038671309893277425, "learning_rate": 5.582197433472348e-07, "loss": 0.9153, "step": 10511 }, { "epoch": 4.760869565217392, "grad_norm": 0.03687470915139375, "learning_rate": 5.561161658261238e-07, "loss": 0.9221, "step": 10512 }, { "epoch": 4.761322463768116, "grad_norm": 0.03804692140792699, "learning_rate": 5.540165315725832e-07, "loss": 0.9175, "step": 10513 }, { "epoch": 4.761775362318841, "grad_norm": 0.040087223488528306, "learning_rate": 5.51920840796516e-07, "loss": 0.9305, "step": 10514 }, { "epoch": 4.762228260869565, "grad_norm": 0.04090383323787189, "learning_rate": 5.498290937074303e-07, "loss": 0.9441, "step": 10515 }, { "epoch": 4.76268115942029, "grad_norm": 0.03819044051820345, "learning_rate": 5.477412905144519e-07, "loss": 0.934, "step": 10516 }, { "epoch": 4.763134057971015, "grad_norm": 0.0370757817743727, "learning_rate": 5.456574314262853e-07, "loss": 0.9368, "step": 10517 }, { "epoch": 4.763586956521739, "grad_norm": 0.03907134508063529, "learning_rate": 5.435775166512747e-07, "loss": 0.9131, "step": 10518 }, { "epoch": 4.764039855072464, "grad_norm": 0.03654580292488658, "learning_rate": 5.415015463973383e-07, "loss": 0.9034, "step": 10519 }, { "epoch": 4.7644927536231885, "grad_norm": 0.0383810717551107, "learning_rate": 5.394295208720213e-07, "loss": 0.8958, "step": 10520 }, { "epoch": 4.764945652173913, "grad_norm": 0.03896458304098573, "learning_rate": 5.373614402824645e-07, "loss": 0.9142, "step": 10521 }, { "epoch": 4.765398550724638, "grad_norm": 0.038534010298380494, "learning_rate": 5.352973048354226e-07, "loss": 0.9274, "step": 10522 }, { "epoch": 4.765851449275362, "grad_norm": 0.03774236838046834, "learning_rate": 5.332371147372417e-07, "loss": 0.9204, "step": 10523 }, { "epoch": 4.766304347826087, "grad_norm": 0.037512780777222754, "learning_rate": 5.311808701938859e-07, "loss": 0.9244, "step": 10524 }, { "epoch": 4.7667572463768115, "grad_norm": 0.039033782723494906, "learning_rate": 5.29128571410924e-07, "loss": 0.9387, "step": 10525 }, { "epoch": 4.767210144927536, "grad_norm": 0.0385615812738586, "learning_rate": 5.270802185935298e-07, "loss": 0.9203, "step": 10526 }, { "epoch": 4.767663043478261, "grad_norm": 0.038118423686049556, "learning_rate": 5.250358119464682e-07, "loss": 0.9156, "step": 10527 }, { "epoch": 4.768115942028985, "grad_norm": 0.03839932825245445, "learning_rate": 5.229953516741359e-07, "loss": 0.8868, "step": 10528 }, { "epoch": 4.76856884057971, "grad_norm": 0.03646718824830232, "learning_rate": 5.209588379805164e-07, "loss": 0.9098, "step": 10529 }, { "epoch": 4.769021739130435, "grad_norm": 0.03789407497500323, "learning_rate": 5.189262710691977e-07, "loss": 0.902, "step": 10530 }, { "epoch": 4.769474637681159, "grad_norm": 0.037722472630242074, "learning_rate": 5.168976511433865e-07, "loss": 0.9266, "step": 10531 }, { "epoch": 4.769927536231884, "grad_norm": 0.037307980029041496, "learning_rate": 5.148729784058804e-07, "loss": 0.9245, "step": 10532 }, { "epoch": 4.770380434782608, "grad_norm": 0.03613587088318933, "learning_rate": 5.128522530590907e-07, "loss": 0.9265, "step": 10533 }, { "epoch": 4.770833333333333, "grad_norm": 0.037577111903033006, "learning_rate": 5.108354753050381e-07, "loss": 0.9097, "step": 10534 }, { "epoch": 4.771286231884058, "grad_norm": 0.036173019147930995, "learning_rate": 5.088226453453349e-07, "loss": 0.9121, "step": 10535 }, { "epoch": 4.771739130434782, "grad_norm": 0.036986113700422305, "learning_rate": 5.068137633812153e-07, "loss": 0.9305, "step": 10536 }, { "epoch": 4.772192028985507, "grad_norm": 0.037656498878770255, "learning_rate": 5.048088296135056e-07, "loss": 0.9189, "step": 10537 }, { "epoch": 4.772644927536232, "grad_norm": 0.03732561538670619, "learning_rate": 5.028078442426409e-07, "loss": 0.9271, "step": 10538 }, { "epoch": 4.773097826086957, "grad_norm": 0.03701145304513952, "learning_rate": 5.008108074686657e-07, "loss": 0.9207, "step": 10539 }, { "epoch": 4.773550724637682, "grad_norm": 0.036623803634482116, "learning_rate": 4.988177194912247e-07, "loss": 0.9012, "step": 10540 }, { "epoch": 4.774003623188406, "grad_norm": 0.03651756492997742, "learning_rate": 4.968285805095674e-07, "loss": 0.917, "step": 10541 }, { "epoch": 4.774456521739131, "grad_norm": 0.037413134672758946, "learning_rate": 4.948433907225614e-07, "loss": 0.9245, "step": 10542 }, { "epoch": 4.774909420289855, "grad_norm": 0.03752720224833184, "learning_rate": 4.928621503286613e-07, "loss": 0.918, "step": 10543 }, { "epoch": 4.77536231884058, "grad_norm": 0.037744107446389764, "learning_rate": 4.908848595259352e-07, "loss": 0.9334, "step": 10544 }, { "epoch": 4.775815217391305, "grad_norm": 0.03565071928107841, "learning_rate": 4.889115185120608e-07, "loss": 0.9198, "step": 10545 }, { "epoch": 4.776268115942029, "grad_norm": 0.03746016999560616, "learning_rate": 4.869421274843112e-07, "loss": 0.9368, "step": 10546 }, { "epoch": 4.776721014492754, "grad_norm": 0.034642329404877915, "learning_rate": 4.849766866395689e-07, "loss": 0.9115, "step": 10547 }, { "epoch": 4.7771739130434785, "grad_norm": 0.03814369253569078, "learning_rate": 4.830151961743257e-07, "loss": 0.9222, "step": 10548 }, { "epoch": 4.777626811594203, "grad_norm": 0.03759657547383634, "learning_rate": 4.810576562846737e-07, "loss": 0.9453, "step": 10549 }, { "epoch": 4.778079710144928, "grad_norm": 0.03746968973771417, "learning_rate": 4.79104067166305e-07, "loss": 0.9179, "step": 10550 }, { "epoch": 4.778532608695652, "grad_norm": 0.03832616031737046, "learning_rate": 4.771544290145347e-07, "loss": 0.931, "step": 10551 }, { "epoch": 4.778985507246377, "grad_norm": 0.0375463095101844, "learning_rate": 4.752087420242646e-07, "loss": 0.9339, "step": 10552 }, { "epoch": 4.7794384057971016, "grad_norm": 0.03670098190405818, "learning_rate": 4.732670063900102e-07, "loss": 0.9018, "step": 10553 }, { "epoch": 4.779891304347826, "grad_norm": 0.03695346008155152, "learning_rate": 4.71329222305883e-07, "loss": 0.912, "step": 10554 }, { "epoch": 4.780344202898551, "grad_norm": 0.03653024586542537, "learning_rate": 4.693953899656167e-07, "loss": 0.9322, "step": 10555 }, { "epoch": 4.780797101449275, "grad_norm": 0.03648344693105936, "learning_rate": 4.6746550956253687e-07, "loss": 0.9172, "step": 10556 }, { "epoch": 4.78125, "grad_norm": 0.03742312669139994, "learning_rate": 4.6553958128957355e-07, "loss": 0.9314, "step": 10557 }, { "epoch": 4.781702898550725, "grad_norm": 0.03848194582798898, "learning_rate": 4.636176053392616e-07, "loss": 0.9397, "step": 10558 }, { "epoch": 4.782155797101449, "grad_norm": 0.037502425076108376, "learning_rate": 4.6169958190375395e-07, "loss": 0.9091, "step": 10559 }, { "epoch": 4.782608695652174, "grad_norm": 0.03922188370515544, "learning_rate": 4.5978551117479063e-07, "loss": 0.9079, "step": 10560 }, { "epoch": 4.7830615942028984, "grad_norm": 0.038876573372804715, "learning_rate": 4.578753933437252e-07, "loss": 0.9142, "step": 10561 }, { "epoch": 4.783514492753623, "grad_norm": 0.038243674851867415, "learning_rate": 4.5596922860151603e-07, "loss": 0.9162, "step": 10562 }, { "epoch": 4.783967391304348, "grad_norm": 0.03696397236173819, "learning_rate": 4.540670171387307e-07, "loss": 0.9364, "step": 10563 }, { "epoch": 4.784420289855072, "grad_norm": 0.03563984177102542, "learning_rate": 4.5216875914552815e-07, "loss": 0.907, "step": 10564 }, { "epoch": 4.784873188405797, "grad_norm": 0.03549754455522038, "learning_rate": 4.502744548116855e-07, "loss": 0.9102, "step": 10565 }, { "epoch": 4.7853260869565215, "grad_norm": 0.03643519856911507, "learning_rate": 4.4838410432657573e-07, "loss": 0.8996, "step": 10566 }, { "epoch": 4.785778985507246, "grad_norm": 0.03597775292597927, "learning_rate": 4.46497707879181e-07, "loss": 0.9237, "step": 10567 }, { "epoch": 4.786231884057971, "grad_norm": 0.03628553702283573, "learning_rate": 4.446152656580882e-07, "loss": 0.8933, "step": 10568 }, { "epoch": 4.786684782608695, "grad_norm": 0.036673903100135775, "learning_rate": 4.427367778514935e-07, "loss": 0.9018, "step": 10569 }, { "epoch": 4.78713768115942, "grad_norm": 0.036858362634298925, "learning_rate": 4.4086224464717995e-07, "loss": 0.9402, "step": 10570 }, { "epoch": 4.787590579710145, "grad_norm": 0.03822680472670304, "learning_rate": 4.389916662325577e-07, "loss": 0.9158, "step": 10571 }, { "epoch": 4.788043478260869, "grad_norm": 0.038889052352307295, "learning_rate": 4.3712504279463275e-07, "loss": 0.9194, "step": 10572 }, { "epoch": 4.788496376811594, "grad_norm": 0.03849877866688539, "learning_rate": 4.3526237452000684e-07, "loss": 0.9335, "step": 10573 }, { "epoch": 4.788949275362318, "grad_norm": 0.03723543277658097, "learning_rate": 4.3340366159489553e-07, "loss": 0.9358, "step": 10574 }, { "epoch": 4.789402173913043, "grad_norm": 0.038249830940915074, "learning_rate": 4.315489042051191e-07, "loss": 0.9296, "step": 10575 }, { "epoch": 4.789855072463768, "grad_norm": 0.03744934414699527, "learning_rate": 4.29698102536098e-07, "loss": 0.9236, "step": 10576 }, { "epoch": 4.790307971014493, "grad_norm": 0.037525573721482935, "learning_rate": 4.2785125677286654e-07, "loss": 0.9211, "step": 10577 }, { "epoch": 4.790760869565218, "grad_norm": 0.03836842479634056, "learning_rate": 4.260083671000459e-07, "loss": 0.9114, "step": 10578 }, { "epoch": 4.791213768115942, "grad_norm": 0.03767939658043592, "learning_rate": 4.241694337018887e-07, "loss": 0.9155, "step": 10579 }, { "epoch": 4.791666666666667, "grad_norm": 0.03626079352824451, "learning_rate": 4.223344567622212e-07, "loss": 0.9195, "step": 10580 }, { "epoch": 4.792119565217392, "grad_norm": 0.041217414747548566, "learning_rate": 4.205034364644966e-07, "loss": 0.9262, "step": 10581 }, { "epoch": 4.792572463768116, "grad_norm": 0.03953139950925628, "learning_rate": 4.1867637299175536e-07, "loss": 0.9212, "step": 10582 }, { "epoch": 4.793025362318841, "grad_norm": 0.037654312730191174, "learning_rate": 4.168532665266689e-07, "loss": 0.9257, "step": 10583 }, { "epoch": 4.793478260869565, "grad_norm": 0.038550586664363146, "learning_rate": 4.1503411725147824e-07, "loss": 0.9092, "step": 10584 }, { "epoch": 4.79393115942029, "grad_norm": 0.03760022424000417, "learning_rate": 4.132189253480601e-07, "loss": 0.9254, "step": 10585 }, { "epoch": 4.794384057971015, "grad_norm": 0.037402023411180874, "learning_rate": 4.114076909978737e-07, "loss": 0.9035, "step": 10586 }, { "epoch": 4.794836956521739, "grad_norm": 0.03774190009214437, "learning_rate": 4.096004143819965e-07, "loss": 0.9102, "step": 10587 }, { "epoch": 4.795289855072464, "grad_norm": 0.0373478966398757, "learning_rate": 4.0779709568109727e-07, "loss": 0.9192, "step": 10588 }, { "epoch": 4.7957427536231885, "grad_norm": 0.044697208433611406, "learning_rate": 4.0599773507546735e-07, "loss": 0.9276, "step": 10589 }, { "epoch": 4.796195652173913, "grad_norm": 0.03843669712594281, "learning_rate": 4.0420233274498064e-07, "loss": 0.9127, "step": 10590 }, { "epoch": 4.796648550724638, "grad_norm": 0.03788088713346357, "learning_rate": 4.02410888869138e-07, "loss": 0.915, "step": 10591 }, { "epoch": 4.797101449275362, "grad_norm": 0.03863782637265877, "learning_rate": 4.006234036270229e-07, "loss": 0.9193, "step": 10592 }, { "epoch": 4.797554347826087, "grad_norm": 0.038192031488405845, "learning_rate": 3.9883987719733673e-07, "loss": 0.9101, "step": 10593 }, { "epoch": 4.7980072463768115, "grad_norm": 0.037991905184982376, "learning_rate": 3.9706030975837697e-07, "loss": 0.9337, "step": 10594 }, { "epoch": 4.798460144927536, "grad_norm": 0.03617978523738002, "learning_rate": 3.9528470148806344e-07, "loss": 0.9056, "step": 10595 }, { "epoch": 4.798913043478261, "grad_norm": 0.03758733416365461, "learning_rate": 3.935130525638897e-07, "loss": 0.9056, "step": 10596 }, { "epoch": 4.799365942028985, "grad_norm": 0.03902693776236551, "learning_rate": 3.917453631629853e-07, "loss": 0.9223, "step": 10597 }, { "epoch": 4.79981884057971, "grad_norm": 0.03965462388767912, "learning_rate": 3.899816334620532e-07, "loss": 0.9167, "step": 10598 }, { "epoch": 4.800271739130435, "grad_norm": 0.03735395824880587, "learning_rate": 3.882218636374324e-07, "loss": 0.9058, "step": 10599 }, { "epoch": 4.800724637681159, "grad_norm": 0.03774124432824577, "learning_rate": 3.8646605386503554e-07, "loss": 0.9075, "step": 10600 }, { "epoch": 4.801177536231884, "grad_norm": 0.037226414976384375, "learning_rate": 3.8471420432040664e-07, "loss": 0.9206, "step": 10601 }, { "epoch": 4.801630434782608, "grad_norm": 0.03686300564050065, "learning_rate": 3.829663151786678e-07, "loss": 0.9303, "step": 10602 }, { "epoch": 4.802083333333333, "grad_norm": 0.03733541909064235, "learning_rate": 3.8122238661456814e-07, "loss": 0.9085, "step": 10603 }, { "epoch": 4.802536231884058, "grad_norm": 0.03564647990794748, "learning_rate": 3.794824188024482e-07, "loss": 0.9246, "step": 10604 }, { "epoch": 4.802989130434782, "grad_norm": 0.03863835578075057, "learning_rate": 3.7774641191625773e-07, "loss": 0.9297, "step": 10605 }, { "epoch": 4.803442028985507, "grad_norm": 0.037608856860735854, "learning_rate": 3.7601436612954236e-07, "loss": 0.9282, "step": 10606 }, { "epoch": 4.803894927536232, "grad_norm": 0.03757640954393161, "learning_rate": 3.7428628161546135e-07, "loss": 0.9217, "step": 10607 }, { "epoch": 4.804347826086957, "grad_norm": 0.03635967702860455, "learning_rate": 3.725621585467698e-07, "loss": 0.9193, "step": 10608 }, { "epoch": 4.804800724637682, "grad_norm": 0.038679453269976335, "learning_rate": 3.708419970958366e-07, "loss": 0.9148, "step": 10609 }, { "epoch": 4.805253623188406, "grad_norm": 0.037621827754976585, "learning_rate": 3.691257974346263e-07, "loss": 0.9161, "step": 10610 }, { "epoch": 4.805706521739131, "grad_norm": 0.03700060516544409, "learning_rate": 3.674135597347128e-07, "loss": 0.9382, "step": 10611 }, { "epoch": 4.806159420289855, "grad_norm": 0.03893563231132247, "learning_rate": 3.6570528416726593e-07, "loss": 0.9382, "step": 10612 }, { "epoch": 4.80661231884058, "grad_norm": 0.03760366783364334, "learning_rate": 3.6400097090306894e-07, "loss": 0.9219, "step": 10613 }, { "epoch": 4.807065217391305, "grad_norm": 0.03695888214129618, "learning_rate": 3.6230062011250565e-07, "loss": 0.9115, "step": 10614 }, { "epoch": 4.807518115942029, "grad_norm": 0.035687351144600345, "learning_rate": 3.6060423196556005e-07, "loss": 0.9207, "step": 10615 }, { "epoch": 4.807971014492754, "grad_norm": 0.039935794560013935, "learning_rate": 3.589118066318209e-07, "loss": 0.9288, "step": 10616 }, { "epoch": 4.8084239130434785, "grad_norm": 0.03599024980338347, "learning_rate": 3.572233442804862e-07, "loss": 0.9118, "step": 10617 }, { "epoch": 4.808876811594203, "grad_norm": 0.038383526873859146, "learning_rate": 3.555388450803499e-07, "loss": 0.9273, "step": 10618 }, { "epoch": 4.809329710144928, "grad_norm": 0.03787326677763803, "learning_rate": 3.538583091998238e-07, "loss": 0.9345, "step": 10619 }, { "epoch": 4.809782608695652, "grad_norm": 0.038507357650960625, "learning_rate": 3.521817368069025e-07, "loss": 0.9297, "step": 10620 }, { "epoch": 4.810235507246377, "grad_norm": 0.03676410972960563, "learning_rate": 3.50509128069203e-07, "loss": 0.9177, "step": 10621 }, { "epoch": 4.8106884057971016, "grad_norm": 0.03652905420310821, "learning_rate": 3.4884048315393384e-07, "loss": 0.911, "step": 10622 }, { "epoch": 4.811141304347826, "grad_norm": 0.03678236444713294, "learning_rate": 3.47175802227917e-07, "loss": 0.8894, "step": 10623 }, { "epoch": 4.811594202898551, "grad_norm": 0.039944424148308004, "learning_rate": 3.455150854575662e-07, "loss": 0.9091, "step": 10624 }, { "epoch": 4.812047101449275, "grad_norm": 0.03812302794504689, "learning_rate": 3.4385833300891735e-07, "loss": 0.9252, "step": 10625 }, { "epoch": 4.8125, "grad_norm": 0.21241809884971602, "learning_rate": 3.4220554504758475e-07, "loss": 0.925, "step": 10626 }, { "epoch": 4.812952898550725, "grad_norm": 0.0361380158527046, "learning_rate": 3.405567217388095e-07, "loss": 0.9035, "step": 10627 }, { "epoch": 4.813405797101449, "grad_norm": 0.038353163881177155, "learning_rate": 3.3891186324741974e-07, "loss": 0.9255, "step": 10628 }, { "epoch": 4.813858695652174, "grad_norm": 0.039399800733277084, "learning_rate": 3.372709697378662e-07, "loss": 0.913, "step": 10629 }, { "epoch": 4.8143115942028984, "grad_norm": 0.03720392726860619, "learning_rate": 3.356340413741821e-07, "loss": 0.9222, "step": 10630 }, { "epoch": 4.814764492753623, "grad_norm": 0.03674621740590935, "learning_rate": 3.3400107832001873e-07, "loss": 0.9041, "step": 10631 }, { "epoch": 4.815217391304348, "grad_norm": 0.03774411442296342, "learning_rate": 3.3237208073862324e-07, "loss": 0.9261, "step": 10632 }, { "epoch": 4.815670289855072, "grad_norm": 0.03767165559631646, "learning_rate": 3.3074704879284767e-07, "loss": 0.9203, "step": 10633 }, { "epoch": 4.816123188405797, "grad_norm": 0.0390275739704688, "learning_rate": 3.291259826451487e-07, "loss": 0.9046, "step": 10634 }, { "epoch": 4.8165760869565215, "grad_norm": 0.03803787696048321, "learning_rate": 3.2750888245759226e-07, "loss": 0.9303, "step": 10635 }, { "epoch": 4.817028985507246, "grad_norm": 0.036499736656232386, "learning_rate": 3.2589574839184015e-07, "loss": 0.9106, "step": 10636 }, { "epoch": 4.817481884057971, "grad_norm": 0.03658731439290878, "learning_rate": 3.2428658060915887e-07, "loss": 0.9001, "step": 10637 }, { "epoch": 4.817934782608695, "grad_norm": 0.03614334819295043, "learning_rate": 3.2268137927041977e-07, "loss": 0.9078, "step": 10638 }, { "epoch": 4.81838768115942, "grad_norm": 0.03794136565742263, "learning_rate": 3.2108014453609446e-07, "loss": 0.9242, "step": 10639 }, { "epoch": 4.818840579710145, "grad_norm": 0.037525735917955765, "learning_rate": 3.1948287656626387e-07, "loss": 0.9444, "step": 10640 }, { "epoch": 4.819293478260869, "grad_norm": 0.03775727269042126, "learning_rate": 3.178895755206135e-07, "loss": 0.9394, "step": 10641 }, { "epoch": 4.819746376811594, "grad_norm": 0.037017393073495095, "learning_rate": 3.163002415584204e-07, "loss": 0.9232, "step": 10642 }, { "epoch": 4.820199275362318, "grad_norm": 0.037048138918030056, "learning_rate": 3.147148748385753e-07, "loss": 0.9234, "step": 10643 }, { "epoch": 4.820652173913043, "grad_norm": 0.03645618341525069, "learning_rate": 3.131334755195736e-07, "loss": 0.9307, "step": 10644 }, { "epoch": 4.821105072463768, "grad_norm": 0.03711519080865067, "learning_rate": 3.1155604375950666e-07, "loss": 0.9305, "step": 10645 }, { "epoch": 4.821557971014493, "grad_norm": 0.03776965034436461, "learning_rate": 3.0998257971607494e-07, "loss": 0.9437, "step": 10646 }, { "epoch": 4.822010869565218, "grad_norm": 0.03699184631163096, "learning_rate": 3.0841308354657486e-07, "loss": 0.9214, "step": 10647 }, { "epoch": 4.822463768115942, "grad_norm": 0.035320539374940105, "learning_rate": 3.068475554079209e-07, "loss": 0.9198, "step": 10648 }, { "epoch": 4.822916666666667, "grad_norm": 0.03641742663063575, "learning_rate": 3.0528599545661453e-07, "loss": 0.9164, "step": 10649 }, { "epoch": 4.823369565217392, "grad_norm": 0.03701973309344248, "learning_rate": 3.0372840384876646e-07, "loss": 0.9012, "step": 10650 }, { "epoch": 4.823822463768116, "grad_norm": 0.036509065143747924, "learning_rate": 3.0217478074009656e-07, "loss": 0.9356, "step": 10651 }, { "epoch": 4.824275362318841, "grad_norm": 0.03847344546701371, "learning_rate": 3.0062512628592054e-07, "loss": 0.9443, "step": 10652 }, { "epoch": 4.824728260869565, "grad_norm": 0.03690861204801993, "learning_rate": 2.9907944064115903e-07, "loss": 0.9059, "step": 10653 }, { "epoch": 4.82518115942029, "grad_norm": 0.03929295338247468, "learning_rate": 2.9753772396033276e-07, "loss": 0.9183, "step": 10654 }, { "epoch": 4.825634057971015, "grad_norm": 0.036939502976125825, "learning_rate": 2.959999763975807e-07, "loss": 0.9145, "step": 10655 }, { "epoch": 4.826086956521739, "grad_norm": 0.03593418237839516, "learning_rate": 2.9446619810662434e-07, "loss": 0.9116, "step": 10656 }, { "epoch": 4.826539855072464, "grad_norm": 0.03725431397855533, "learning_rate": 2.9293638924080327e-07, "loss": 0.9284, "step": 10657 }, { "epoch": 4.8269927536231885, "grad_norm": 0.03888246759505578, "learning_rate": 2.9141054995305285e-07, "loss": 0.9121, "step": 10658 }, { "epoch": 4.827445652173913, "grad_norm": 0.03655353113817438, "learning_rate": 2.898886803959089e-07, "loss": 0.9317, "step": 10659 }, { "epoch": 4.827898550724638, "grad_norm": 0.03645491037346689, "learning_rate": 2.8837078072152083e-07, "loss": 0.939, "step": 10660 }, { "epoch": 4.828351449275362, "grad_norm": 0.03688812871149905, "learning_rate": 2.868568510816383e-07, "loss": 0.9119, "step": 10661 }, { "epoch": 4.828804347826087, "grad_norm": 0.03707620352418911, "learning_rate": 2.8534689162760255e-07, "loss": 0.9229, "step": 10662 }, { "epoch": 4.8292572463768115, "grad_norm": 0.03592102744087389, "learning_rate": 2.8384090251037724e-07, "loss": 0.9063, "step": 10663 }, { "epoch": 4.829710144927536, "grad_norm": 0.038611366631588734, "learning_rate": 2.823388838805086e-07, "loss": 0.9149, "step": 10664 }, { "epoch": 4.830163043478261, "grad_norm": 0.036463907402544266, "learning_rate": 2.8084083588816093e-07, "loss": 0.9342, "step": 10665 }, { "epoch": 4.830615942028985, "grad_norm": 0.035447954958863444, "learning_rate": 2.7934675868309e-07, "loss": 0.9169, "step": 10666 }, { "epoch": 4.83106884057971, "grad_norm": 0.03692587237372806, "learning_rate": 2.778566524146742e-07, "loss": 0.9166, "step": 10667 }, { "epoch": 4.831521739130435, "grad_norm": 0.03767039682975681, "learning_rate": 2.763705172318698e-07, "loss": 0.9127, "step": 10668 }, { "epoch": 4.831974637681159, "grad_norm": 0.03624624189441471, "learning_rate": 2.7488835328325137e-07, "loss": 0.9156, "step": 10669 }, { "epoch": 4.832427536231884, "grad_norm": 0.040673693143369646, "learning_rate": 2.734101607169981e-07, "loss": 0.9373, "step": 10670 }, { "epoch": 4.832880434782608, "grad_norm": 0.0365753323485456, "learning_rate": 2.7193593968088074e-07, "loss": 0.9409, "step": 10671 }, { "epoch": 4.833333333333333, "grad_norm": 0.037694458462568, "learning_rate": 2.704656903222791e-07, "loss": 0.9171, "step": 10672 }, { "epoch": 4.833786231884058, "grad_norm": 0.03583324556701352, "learning_rate": 2.6899941278818673e-07, "loss": 0.9362, "step": 10673 }, { "epoch": 4.834239130434782, "grad_norm": 0.03666831871948922, "learning_rate": 2.6753710722517536e-07, "loss": 0.9059, "step": 10674 }, { "epoch": 4.834692028985507, "grad_norm": 0.037211479083181676, "learning_rate": 2.660787737794435e-07, "loss": 0.9002, "step": 10675 }, { "epoch": 4.835144927536232, "grad_norm": 0.039003129622124384, "learning_rate": 2.6462441259678115e-07, "loss": 0.9414, "step": 10676 }, { "epoch": 4.835597826086957, "grad_norm": 0.03835796214507823, "learning_rate": 2.631740238225788e-07, "loss": 0.9157, "step": 10677 }, { "epoch": 4.836050724637682, "grad_norm": 0.03673516075493141, "learning_rate": 2.6172760760184047e-07, "loss": 0.9199, "step": 10678 }, { "epoch": 4.836503623188406, "grad_norm": 0.038002069957811004, "learning_rate": 2.602851640791659e-07, "loss": 0.9149, "step": 10679 }, { "epoch": 4.836956521739131, "grad_norm": 0.03623288680056421, "learning_rate": 2.5884669339875546e-07, "loss": 0.9371, "step": 10680 }, { "epoch": 4.837409420289855, "grad_norm": 0.036065008070738536, "learning_rate": 2.57412195704414e-07, "loss": 0.9104, "step": 10681 }, { "epoch": 4.83786231884058, "grad_norm": 0.03730117570549245, "learning_rate": 2.5598167113955574e-07, "loss": 0.9272, "step": 10682 }, { "epoch": 4.838315217391305, "grad_norm": 0.03844536442287483, "learning_rate": 2.5455511984718626e-07, "loss": 0.9446, "step": 10683 }, { "epoch": 4.838768115942029, "grad_norm": 0.03679364046089589, "learning_rate": 2.531325419699249e-07, "loss": 0.9125, "step": 10684 }, { "epoch": 4.839221014492754, "grad_norm": 0.03585818771469278, "learning_rate": 2.517139376499866e-07, "loss": 0.9194, "step": 10685 }, { "epoch": 4.8396739130434785, "grad_norm": 0.037669330058218896, "learning_rate": 2.5029930702919594e-07, "loss": 0.9259, "step": 10686 }, { "epoch": 4.840126811594203, "grad_norm": 0.03726262519606968, "learning_rate": 2.488886502489685e-07, "loss": 0.921, "step": 10687 }, { "epoch": 4.840579710144928, "grad_norm": 0.03768006493094889, "learning_rate": 2.474819674503381e-07, "loss": 0.9279, "step": 10688 }, { "epoch": 4.841032608695652, "grad_norm": 0.0387296118683848, "learning_rate": 2.4607925877392135e-07, "loss": 0.9336, "step": 10689 }, { "epoch": 4.841485507246377, "grad_norm": 0.0372729670098015, "learning_rate": 2.4468052435996146e-07, "loss": 0.9148, "step": 10690 }, { "epoch": 4.8419384057971016, "grad_norm": 0.03643892028327749, "learning_rate": 2.4328576434828887e-07, "loss": 0.9346, "step": 10691 }, { "epoch": 4.842391304347826, "grad_norm": 0.03825804858742245, "learning_rate": 2.4189497887833866e-07, "loss": 0.9119, "step": 10692 }, { "epoch": 4.842844202898551, "grad_norm": 0.037549135822549415, "learning_rate": 2.4050816808914635e-07, "loss": 0.9147, "step": 10693 }, { "epoch": 4.843297101449275, "grad_norm": 0.08459504118382516, "learning_rate": 2.3912533211935653e-07, "loss": 0.9364, "step": 10694 }, { "epoch": 4.84375, "grad_norm": 0.037006914994833495, "learning_rate": 2.3774647110721415e-07, "loss": 0.8955, "step": 10695 }, { "epoch": 4.844202898550725, "grad_norm": 0.037078913284902024, "learning_rate": 2.3637158519056902e-07, "loss": 0.9053, "step": 10696 }, { "epoch": 4.844655797101449, "grad_norm": 0.03689373213142292, "learning_rate": 2.3500067450686225e-07, "loss": 0.9268, "step": 10697 }, { "epoch": 4.845108695652174, "grad_norm": 0.03745252035936597, "learning_rate": 2.3363373919315757e-07, "loss": 0.9171, "step": 10698 }, { "epoch": 4.8455615942028984, "grad_norm": 0.03724257105465131, "learning_rate": 2.3227077938609678e-07, "loss": 0.9237, "step": 10699 }, { "epoch": 4.846014492753623, "grad_norm": 0.037481350156715464, "learning_rate": 2.3091179522194862e-07, "loss": 0.9153, "step": 10700 }, { "epoch": 4.846467391304348, "grad_norm": 0.036024300568956114, "learning_rate": 2.2955678683656445e-07, "loss": 0.9058, "step": 10701 }, { "epoch": 4.846920289855072, "grad_norm": 0.03641705668551761, "learning_rate": 2.282057543654137e-07, "loss": 0.9246, "step": 10702 }, { "epoch": 4.847373188405797, "grad_norm": 0.037796077555917997, "learning_rate": 2.2685869794355274e-07, "loss": 0.9359, "step": 10703 }, { "epoch": 4.8478260869565215, "grad_norm": 0.035877929693568435, "learning_rate": 2.2551561770565611e-07, "loss": 0.9186, "step": 10704 }, { "epoch": 4.848278985507246, "grad_norm": 0.03646119608010473, "learning_rate": 2.241765137859897e-07, "loss": 0.9395, "step": 10705 }, { "epoch": 4.848731884057971, "grad_norm": 0.038892653433395395, "learning_rate": 2.2284138631842867e-07, "loss": 0.9382, "step": 10706 }, { "epoch": 4.849184782608695, "grad_norm": 0.03789821182234211, "learning_rate": 2.2151023543644845e-07, "loss": 0.915, "step": 10707 }, { "epoch": 4.84963768115942, "grad_norm": 0.03771634808875971, "learning_rate": 2.201830612731204e-07, "loss": 0.9318, "step": 10708 }, { "epoch": 4.850090579710145, "grad_norm": 0.0360870154105985, "learning_rate": 2.1885986396112947e-07, "loss": 0.9174, "step": 10709 }, { "epoch": 4.850543478260869, "grad_norm": 0.03681833364790615, "learning_rate": 2.1754064363276096e-07, "loss": 0.9176, "step": 10710 }, { "epoch": 4.850996376811594, "grad_norm": 0.03670762821246909, "learning_rate": 2.1622540041988716e-07, "loss": 0.9156, "step": 10711 }, { "epoch": 4.851449275362318, "grad_norm": 0.03689508186864757, "learning_rate": 2.1491413445401177e-07, "loss": 0.9315, "step": 10712 }, { "epoch": 4.851902173913043, "grad_norm": 0.03648211886543346, "learning_rate": 2.1360684586620772e-07, "loss": 0.9134, "step": 10713 }, { "epoch": 4.852355072463768, "grad_norm": 0.037374781879552114, "learning_rate": 2.1230353478717935e-07, "loss": 0.9329, "step": 10714 }, { "epoch": 4.852807971014493, "grad_norm": 0.0370264478743678, "learning_rate": 2.1100420134721356e-07, "loss": 0.8972, "step": 10715 }, { "epoch": 4.853260869565218, "grad_norm": 0.03628548042101432, "learning_rate": 2.0970884567620642e-07, "loss": 0.9055, "step": 10716 }, { "epoch": 4.853713768115942, "grad_norm": 0.038520734084308283, "learning_rate": 2.0841746790366325e-07, "loss": 0.9249, "step": 10717 }, { "epoch": 4.854166666666667, "grad_norm": 0.036728672547270066, "learning_rate": 2.0713006815868075e-07, "loss": 0.9153, "step": 10718 }, { "epoch": 4.854619565217392, "grad_norm": 0.035933133031097714, "learning_rate": 2.0584664656995602e-07, "loss": 0.912, "step": 10719 }, { "epoch": 4.855072463768116, "grad_norm": 0.03627944119441385, "learning_rate": 2.0456720326580858e-07, "loss": 0.9045, "step": 10720 }, { "epoch": 4.855525362318841, "grad_norm": 0.03649680952267741, "learning_rate": 2.032917383741362e-07, "loss": 0.9306, "step": 10721 }, { "epoch": 4.855978260869565, "grad_norm": 0.03667588344458692, "learning_rate": 2.0202025202245457e-07, "loss": 0.9473, "step": 10722 }, { "epoch": 4.85643115942029, "grad_norm": 0.035204013763236244, "learning_rate": 2.0075274433787096e-07, "loss": 0.917, "step": 10723 }, { "epoch": 4.856884057971015, "grad_norm": 0.03573230953278012, "learning_rate": 1.9948921544710177e-07, "loss": 0.9085, "step": 10724 }, { "epoch": 4.857336956521739, "grad_norm": 0.03647043647727198, "learning_rate": 1.9822966547646815e-07, "loss": 0.9422, "step": 10725 }, { "epoch": 4.857789855072464, "grad_norm": 0.03645819247500158, "learning_rate": 1.9697409455188278e-07, "loss": 0.8929, "step": 10726 }, { "epoch": 4.8582427536231885, "grad_norm": 0.03562232645614927, "learning_rate": 1.9572250279887185e-07, "loss": 0.9166, "step": 10727 }, { "epoch": 4.858695652173913, "grad_norm": 0.03837807128689944, "learning_rate": 1.9447489034256196e-07, "loss": 0.9227, "step": 10728 }, { "epoch": 4.859148550724638, "grad_norm": 0.038207217041792516, "learning_rate": 1.9323125730766666e-07, "loss": 0.9191, "step": 10729 }, { "epoch": 4.859601449275362, "grad_norm": 0.0362470332636922, "learning_rate": 1.9199160381852656e-07, "loss": 0.9146, "step": 10730 }, { "epoch": 4.860054347826087, "grad_norm": 0.03536225689927718, "learning_rate": 1.907559299990691e-07, "loss": 0.8898, "step": 10731 }, { "epoch": 4.8605072463768115, "grad_norm": 0.03713709928690301, "learning_rate": 1.895242359728222e-07, "loss": 0.917, "step": 10732 }, { "epoch": 4.860960144927536, "grad_norm": 0.0381478229978641, "learning_rate": 1.882965218629229e-07, "loss": 0.9093, "step": 10733 }, { "epoch": 4.861413043478261, "grad_norm": 0.03768195596847281, "learning_rate": 1.8707278779210415e-07, "loss": 0.9215, "step": 10734 }, { "epoch": 4.861865942028985, "grad_norm": 0.03830390276794107, "learning_rate": 1.8585303388271247e-07, "loss": 0.9077, "step": 10735 }, { "epoch": 4.86231884057971, "grad_norm": 0.0375136085683048, "learning_rate": 1.846372602566815e-07, "loss": 0.9277, "step": 10736 }, { "epoch": 4.862771739130435, "grad_norm": 0.036113957226585176, "learning_rate": 1.8342546703555396e-07, "loss": 0.8966, "step": 10737 }, { "epoch": 4.863224637681159, "grad_norm": 0.03739932084369057, "learning_rate": 1.822176543404819e-07, "loss": 0.9182, "step": 10738 }, { "epoch": 4.863677536231884, "grad_norm": 0.037743159842745656, "learning_rate": 1.810138222922042e-07, "loss": 0.9296, "step": 10739 }, { "epoch": 4.864130434782608, "grad_norm": 0.03635998543258008, "learning_rate": 1.7981397101107355e-07, "loss": 0.9074, "step": 10740 }, { "epoch": 4.864583333333333, "grad_norm": 0.03606719767789758, "learning_rate": 1.7861810061704287e-07, "loss": 0.9201, "step": 10741 }, { "epoch": 4.865036231884058, "grad_norm": 0.036069108420441885, "learning_rate": 1.774262112296654e-07, "loss": 0.9182, "step": 10742 }, { "epoch": 4.865489130434782, "grad_norm": 0.036383422823734524, "learning_rate": 1.7623830296808587e-07, "loss": 0.931, "step": 10743 }, { "epoch": 4.865942028985507, "grad_norm": 0.03717282672459077, "learning_rate": 1.750543759510759e-07, "loss": 0.9203, "step": 10744 }, { "epoch": 4.866394927536232, "grad_norm": 0.037068751053657936, "learning_rate": 1.7387443029698525e-07, "loss": 0.9031, "step": 10745 }, { "epoch": 4.866847826086957, "grad_norm": 0.03611644218621632, "learning_rate": 1.7269846612378182e-07, "loss": 0.9052, "step": 10746 }, { "epoch": 4.867300724637682, "grad_norm": 0.03608454151178297, "learning_rate": 1.7152648354902047e-07, "loss": 0.92, "step": 10747 }, { "epoch": 4.867753623188406, "grad_norm": 0.0376231005418955, "learning_rate": 1.7035848268986966e-07, "loss": 0.9159, "step": 10748 }, { "epoch": 4.868206521739131, "grad_norm": 0.03624790975879652, "learning_rate": 1.6919446366309822e-07, "loss": 0.9192, "step": 10749 }, { "epoch": 4.868659420289855, "grad_norm": 0.03964404938717008, "learning_rate": 1.680344265850753e-07, "loss": 0.9047, "step": 10750 }, { "epoch": 4.86911231884058, "grad_norm": 0.03729107311580959, "learning_rate": 1.668783715717659e-07, "loss": 0.9175, "step": 10751 }, { "epoch": 4.869565217391305, "grad_norm": 0.037267699094383676, "learning_rate": 1.6572629873874869e-07, "loss": 0.8909, "step": 10752 }, { "epoch": 4.870018115942029, "grad_norm": 0.03606721485327864, "learning_rate": 1.645782082011982e-07, "loss": 0.8849, "step": 10753 }, { "epoch": 4.870471014492754, "grad_norm": 0.035838036799518615, "learning_rate": 1.6343410007388483e-07, "loss": 0.9174, "step": 10754 }, { "epoch": 4.8709239130434785, "grad_norm": 0.03681913363732336, "learning_rate": 1.6229397447119267e-07, "loss": 0.934, "step": 10755 }, { "epoch": 4.871376811594203, "grad_norm": 0.03819345888922696, "learning_rate": 1.6115783150710162e-07, "loss": 0.9144, "step": 10756 }, { "epoch": 4.871829710144928, "grad_norm": 0.036479621342397815, "learning_rate": 1.6002567129519197e-07, "loss": 0.9205, "step": 10757 }, { "epoch": 4.872282608695652, "grad_norm": 0.036038552571371764, "learning_rate": 1.5889749394864872e-07, "loss": 0.9049, "step": 10758 }, { "epoch": 4.872735507246377, "grad_norm": 0.03693949029824824, "learning_rate": 1.577732995802528e-07, "loss": 0.922, "step": 10759 }, { "epoch": 4.8731884057971016, "grad_norm": 0.03728075997729463, "learning_rate": 1.5665308830239867e-07, "loss": 0.9114, "step": 10760 }, { "epoch": 4.873641304347826, "grad_norm": 0.039655439040737615, "learning_rate": 1.5553686022707237e-07, "loss": 0.9049, "step": 10761 }, { "epoch": 4.874094202898551, "grad_norm": 0.03762251215401235, "learning_rate": 1.5442461546586461e-07, "loss": 0.9171, "step": 10762 }, { "epoch": 4.874547101449275, "grad_norm": 0.036349179146887345, "learning_rate": 1.533163541299665e-07, "loss": 0.9446, "step": 10763 }, { "epoch": 4.875, "grad_norm": 0.03660250248287623, "learning_rate": 1.522120763301782e-07, "loss": 0.9227, "step": 10764 }, { "epoch": 4.875452898550725, "grad_norm": 0.03551746277300124, "learning_rate": 1.511117821768915e-07, "loss": 0.9268, "step": 10765 }, { "epoch": 4.875905797101449, "grad_norm": 0.03633434160042215, "learning_rate": 1.5001547178010723e-07, "loss": 0.9186, "step": 10766 }, { "epoch": 4.876358695652174, "grad_norm": 0.03923592285918481, "learning_rate": 1.489231452494222e-07, "loss": 0.9014, "step": 10767 }, { "epoch": 4.8768115942028984, "grad_norm": 0.036600758525110945, "learning_rate": 1.4783480269404238e-07, "loss": 0.9094, "step": 10768 }, { "epoch": 4.877264492753623, "grad_norm": 0.037054808940100194, "learning_rate": 1.4675044422276517e-07, "loss": 0.9104, "step": 10769 }, { "epoch": 4.877717391304348, "grad_norm": 0.03875299060093917, "learning_rate": 1.4567006994400167e-07, "loss": 0.9181, "step": 10770 }, { "epoch": 4.878170289855072, "grad_norm": 0.03897359307522935, "learning_rate": 1.4459367996575436e-07, "loss": 0.9194, "step": 10771 }, { "epoch": 4.878623188405797, "grad_norm": 0.03807883071344497, "learning_rate": 1.4352127439563045e-07, "loss": 0.9085, "step": 10772 }, { "epoch": 4.8790760869565215, "grad_norm": 0.038108053505660516, "learning_rate": 1.4245285334084646e-07, "loss": 0.9042, "step": 10773 }, { "epoch": 4.879528985507246, "grad_norm": 0.0370603924858772, "learning_rate": 1.4138841690821027e-07, "loss": 0.9109, "step": 10774 }, { "epoch": 4.879981884057971, "grad_norm": 0.0373544210234529, "learning_rate": 1.403279652041345e-07, "loss": 0.9154, "step": 10775 }, { "epoch": 4.880434782608695, "grad_norm": 0.03722847906153618, "learning_rate": 1.392714983346366e-07, "loss": 0.9241, "step": 10776 }, { "epoch": 4.88088768115942, "grad_norm": 0.036036843056177555, "learning_rate": 1.3821901640532986e-07, "loss": 0.9177, "step": 10777 }, { "epoch": 4.881340579710145, "grad_norm": 0.0347149759698179, "learning_rate": 1.371705195214368e-07, "loss": 0.9132, "step": 10778 }, { "epoch": 4.881793478260869, "grad_norm": 0.03595472261573151, "learning_rate": 1.361260077877713e-07, "loss": 0.8933, "step": 10779 }, { "epoch": 4.882246376811594, "grad_norm": 0.03748894057648203, "learning_rate": 1.3508548130876097e-07, "loss": 0.9284, "step": 10780 }, { "epoch": 4.882699275362318, "grad_norm": 0.03699364674993241, "learning_rate": 1.3404894018842484e-07, "loss": 0.9123, "step": 10781 }, { "epoch": 4.883152173913043, "grad_norm": 0.036319973007556364, "learning_rate": 1.3301638453039112e-07, "loss": 0.9273, "step": 10782 }, { "epoch": 4.883605072463768, "grad_norm": 0.0361223562478319, "learning_rate": 1.319878144378839e-07, "loss": 0.9387, "step": 10783 }, { "epoch": 4.884057971014493, "grad_norm": 0.03622030911888137, "learning_rate": 1.30963230013732e-07, "loss": 0.9428, "step": 10784 }, { "epoch": 4.884510869565218, "grad_norm": 0.03535034858534817, "learning_rate": 1.2994263136036024e-07, "loss": 0.909, "step": 10785 }, { "epoch": 4.884963768115942, "grad_norm": 0.03723150474257502, "learning_rate": 1.289260185798069e-07, "loss": 0.907, "step": 10786 }, { "epoch": 4.885416666666667, "grad_norm": 0.037311880369006595, "learning_rate": 1.2791339177369745e-07, "loss": 0.946, "step": 10787 }, { "epoch": 4.885869565217392, "grad_norm": 0.03655588322549183, "learning_rate": 1.2690475104327082e-07, "loss": 0.9196, "step": 10788 }, { "epoch": 4.886322463768116, "grad_norm": 0.03690908145241759, "learning_rate": 1.2590009648935752e-07, "loss": 0.9094, "step": 10789 }, { "epoch": 4.886775362318841, "grad_norm": 0.03770452268204135, "learning_rate": 1.2489942821240164e-07, "loss": 0.9291, "step": 10790 }, { "epoch": 4.887228260869565, "grad_norm": 0.0370686118490859, "learning_rate": 1.2390274631243872e-07, "loss": 0.9157, "step": 10791 }, { "epoch": 4.88768115942029, "grad_norm": 0.03631369543232735, "learning_rate": 1.2291005088910457e-07, "loss": 0.9147, "step": 10792 }, { "epoch": 4.888134057971015, "grad_norm": 0.037734999772805075, "learning_rate": 1.219213420416443e-07, "loss": 0.9241, "step": 10793 }, { "epoch": 4.888586956521739, "grad_norm": 0.03709735326980153, "learning_rate": 1.2093661986889883e-07, "loss": 0.9337, "step": 10794 }, { "epoch": 4.889039855072464, "grad_norm": 0.03861402280111023, "learning_rate": 1.1995588446931382e-07, "loss": 0.9146, "step": 10795 }, { "epoch": 4.8894927536231885, "grad_norm": 0.03678185433832854, "learning_rate": 1.1897913594093091e-07, "loss": 0.9236, "step": 10796 }, { "epoch": 4.889945652173913, "grad_norm": 0.0368888204348242, "learning_rate": 1.1800637438140528e-07, "loss": 0.9321, "step": 10797 }, { "epoch": 4.890398550724638, "grad_norm": 0.0368432872105831, "learning_rate": 1.1703759988797914e-07, "loss": 0.9156, "step": 10798 }, { "epoch": 4.890851449275362, "grad_norm": 0.035858801982942216, "learning_rate": 1.1607281255750391e-07, "loss": 0.9045, "step": 10799 }, { "epoch": 4.891304347826087, "grad_norm": 0.03660738852643562, "learning_rate": 1.1511201248643134e-07, "loss": 0.9105, "step": 10800 }, { "epoch": 4.8917572463768115, "grad_norm": 0.03647880842411636, "learning_rate": 1.1415519977081347e-07, "loss": 0.923, "step": 10801 }, { "epoch": 4.892210144927536, "grad_norm": 0.03716120453087953, "learning_rate": 1.132023745063071e-07, "loss": 0.9329, "step": 10802 }, { "epoch": 4.892663043478261, "grad_norm": 0.036890174802952175, "learning_rate": 1.1225353678816497e-07, "loss": 0.9292, "step": 10803 }, { "epoch": 4.893115942028985, "grad_norm": 0.038175432315067016, "learning_rate": 1.1130868671124006e-07, "loss": 0.9302, "step": 10804 }, { "epoch": 4.89356884057971, "grad_norm": 0.037899261985000085, "learning_rate": 1.1036782436999905e-07, "loss": 0.9209, "step": 10805 }, { "epoch": 4.894021739130435, "grad_norm": 0.03580538473175293, "learning_rate": 1.0943094985849556e-07, "loss": 0.9218, "step": 10806 }, { "epoch": 4.894474637681159, "grad_norm": 0.03677781703807567, "learning_rate": 1.0849806327039248e-07, "loss": 0.926, "step": 10807 }, { "epoch": 4.894927536231884, "grad_norm": 0.036497225323038035, "learning_rate": 1.0756916469894851e-07, "loss": 0.9394, "step": 10808 }, { "epoch": 4.895380434782608, "grad_norm": 0.036989916795988696, "learning_rate": 1.0664425423703606e-07, "loss": 0.9223, "step": 10809 }, { "epoch": 4.895833333333333, "grad_norm": 0.03670847261805673, "learning_rate": 1.0572333197711005e-07, "loss": 0.9181, "step": 10810 }, { "epoch": 4.896286231884058, "grad_norm": 0.03842164616587709, "learning_rate": 1.0480639801124348e-07, "loss": 0.9337, "step": 10811 }, { "epoch": 4.896739130434782, "grad_norm": 0.03645522461364307, "learning_rate": 1.0389345243109639e-07, "loss": 0.936, "step": 10812 }, { "epoch": 4.897192028985507, "grad_norm": 0.03564294781489119, "learning_rate": 1.0298449532794241e-07, "loss": 0.9178, "step": 10813 }, { "epoch": 4.897644927536232, "grad_norm": 0.03490815087224096, "learning_rate": 1.020795267926511e-07, "loss": 0.9161, "step": 10814 }, { "epoch": 4.898097826086957, "grad_norm": 0.03684657339597248, "learning_rate": 1.0117854691569672e-07, "loss": 0.9326, "step": 10815 }, { "epoch": 4.898550724637682, "grad_norm": 0.038122070701352144, "learning_rate": 1.0028155578714504e-07, "loss": 0.9329, "step": 10816 }, { "epoch": 4.899003623188406, "grad_norm": 0.05182549710445685, "learning_rate": 9.938855349667542e-08, "loss": 0.921, "step": 10817 }, { "epoch": 4.899456521739131, "grad_norm": 0.03698632333836683, "learning_rate": 9.849954013355423e-08, "loss": 0.9075, "step": 10818 }, { "epoch": 4.899909420289855, "grad_norm": 0.03569272157168366, "learning_rate": 9.761451578666592e-08, "loss": 0.9138, "step": 10819 }, { "epoch": 4.90036231884058, "grad_norm": 0.03819089424701537, "learning_rate": 9.673348054448639e-08, "loss": 0.9214, "step": 10820 }, { "epoch": 4.900815217391305, "grad_norm": 0.038125562559550026, "learning_rate": 9.58564344950963e-08, "loss": 0.9226, "step": 10821 }, { "epoch": 4.901268115942029, "grad_norm": 0.03590824353691461, "learning_rate": 9.498337772616773e-08, "loss": 0.912, "step": 10822 }, { "epoch": 4.901721014492754, "grad_norm": 0.036573398990437905, "learning_rate": 9.411431032498642e-08, "loss": 0.9202, "step": 10823 }, { "epoch": 4.9021739130434785, "grad_norm": 0.03911252927933686, "learning_rate": 9.324923237842953e-08, "loss": 0.9089, "step": 10824 }, { "epoch": 4.902626811594203, "grad_norm": 0.037190012315308024, "learning_rate": 9.238814397299234e-08, "loss": 0.9267, "step": 10825 }, { "epoch": 4.903079710144928, "grad_norm": 0.03759984519178244, "learning_rate": 9.15310451947482e-08, "loss": 0.9229, "step": 10826 }, { "epoch": 4.903532608695652, "grad_norm": 0.037221235439281, "learning_rate": 9.067793612938414e-08, "loss": 0.9363, "step": 10827 }, { "epoch": 4.903985507246377, "grad_norm": 0.036751473693627225, "learning_rate": 8.982881686219191e-08, "loss": 0.9246, "step": 10828 }, { "epoch": 4.9044384057971016, "grad_norm": 0.03605668923147379, "learning_rate": 8.898368747805474e-08, "loss": 0.8873, "step": 10829 }, { "epoch": 4.904891304347826, "grad_norm": 0.03720811433167872, "learning_rate": 8.81425480614606e-08, "loss": 0.9438, "step": 10830 }, { "epoch": 4.905344202898551, "grad_norm": 0.037369327756503284, "learning_rate": 8.730539869650224e-08, "loss": 0.9311, "step": 10831 }, { "epoch": 4.905797101449275, "grad_norm": 0.03642930109931737, "learning_rate": 8.647223946687266e-08, "loss": 0.909, "step": 10832 }, { "epoch": 4.90625, "grad_norm": 0.03795954928962299, "learning_rate": 8.564307045586085e-08, "loss": 0.9416, "step": 10833 }, { "epoch": 4.906702898550725, "grad_norm": 0.03655699763718464, "learning_rate": 8.481789174636046e-08, "loss": 0.9272, "step": 10834 }, { "epoch": 4.907155797101449, "grad_norm": 0.03698227216120044, "learning_rate": 8.399670342086553e-08, "loss": 0.9208, "step": 10835 }, { "epoch": 4.907608695652174, "grad_norm": 0.03807655146340165, "learning_rate": 8.31795055614748e-08, "loss": 0.9164, "step": 10836 }, { "epoch": 4.9080615942028984, "grad_norm": 0.03776259318645283, "learning_rate": 8.236629824987851e-08, "loss": 0.9407, "step": 10837 }, { "epoch": 4.908514492753623, "grad_norm": 0.03754490379279583, "learning_rate": 8.155708156738052e-08, "loss": 0.9265, "step": 10838 }, { "epoch": 4.908967391304348, "grad_norm": 0.036898098249178175, "learning_rate": 8.075185559488053e-08, "loss": 0.9187, "step": 10839 }, { "epoch": 4.909420289855072, "grad_norm": 0.03673247761268371, "learning_rate": 7.995062041286972e-08, "loss": 0.9063, "step": 10840 }, { "epoch": 4.909873188405797, "grad_norm": 0.0364444378493812, "learning_rate": 7.915337610145735e-08, "loss": 0.9142, "step": 10841 }, { "epoch": 4.9103260869565215, "grad_norm": 0.035641068628892984, "learning_rate": 7.83601227403441e-08, "loss": 0.9124, "step": 10842 }, { "epoch": 4.910778985507246, "grad_norm": 0.03589274899212376, "learning_rate": 7.757086040882655e-08, "loss": 0.9195, "step": 10843 }, { "epoch": 4.911231884057971, "grad_norm": 0.036380801156025766, "learning_rate": 7.678558918581492e-08, "loss": 0.9185, "step": 10844 }, { "epoch": 4.911684782608695, "grad_norm": 0.036249716186028985, "learning_rate": 7.600430914981083e-08, "loss": 0.9001, "step": 10845 }, { "epoch": 4.91213768115942, "grad_norm": 0.035277276412465516, "learning_rate": 7.522702037892071e-08, "loss": 0.8891, "step": 10846 }, { "epoch": 4.912590579710145, "grad_norm": 0.03768413774642421, "learning_rate": 7.44537229508513e-08, "loss": 0.9181, "step": 10847 }, { "epoch": 4.913043478260869, "grad_norm": 0.03697670658757093, "learning_rate": 7.368441694290963e-08, "loss": 0.9045, "step": 10848 }, { "epoch": 4.913496376811594, "grad_norm": 0.036916357433043966, "learning_rate": 7.291910243200751e-08, "loss": 0.9045, "step": 10849 }, { "epoch": 4.913949275362318, "grad_norm": 0.03630808004999342, "learning_rate": 7.215777949465264e-08, "loss": 0.9112, "step": 10850 }, { "epoch": 4.914402173913043, "grad_norm": 0.03631755143524964, "learning_rate": 7.140044820695302e-08, "loss": 0.905, "step": 10851 }, { "epoch": 4.914855072463768, "grad_norm": 0.03698047917521259, "learning_rate": 7.064710864462587e-08, "loss": 0.9381, "step": 10852 }, { "epoch": 4.915307971014493, "grad_norm": 0.0362328136352571, "learning_rate": 6.989776088297983e-08, "loss": 0.9201, "step": 10853 }, { "epoch": 4.915760869565218, "grad_norm": 0.03721837632677431, "learning_rate": 6.915240499692832e-08, "loss": 0.9236, "step": 10854 }, { "epoch": 4.916213768115942, "grad_norm": 0.037368753804804575, "learning_rate": 6.84110410609895e-08, "loss": 0.9035, "step": 10855 }, { "epoch": 4.916666666666667, "grad_norm": 0.03600824192056641, "learning_rate": 6.767366914927298e-08, "loss": 0.9172, "step": 10856 }, { "epoch": 4.917119565217392, "grad_norm": 0.03754154401005763, "learning_rate": 6.694028933549757e-08, "loss": 0.9202, "step": 10857 }, { "epoch": 4.917572463768116, "grad_norm": 0.03664110704856859, "learning_rate": 6.621090169298238e-08, "loss": 0.9471, "step": 10858 }, { "epoch": 4.918025362318841, "grad_norm": 0.0395995382983369, "learning_rate": 6.548550629464245e-08, "loss": 0.9242, "step": 10859 }, { "epoch": 4.918478260869565, "grad_norm": 0.0362837916037849, "learning_rate": 6.476410321300197e-08, "loss": 0.9212, "step": 10860 }, { "epoch": 4.91893115942029, "grad_norm": 0.03716490426196905, "learning_rate": 6.404669252017214e-08, "loss": 0.9098, "step": 10861 }, { "epoch": 4.919384057971015, "grad_norm": 0.036792504931065007, "learning_rate": 6.333327428788227e-08, "loss": 0.9229, "step": 10862 }, { "epoch": 4.919836956521739, "grad_norm": 0.03691459246538742, "learning_rate": 6.262384858744863e-08, "loss": 0.9289, "step": 10863 }, { "epoch": 4.920289855072464, "grad_norm": 0.03703246402971473, "learning_rate": 6.191841548979227e-08, "loss": 0.9182, "step": 10864 }, { "epoch": 4.9207427536231885, "grad_norm": 0.03591474874277584, "learning_rate": 6.121697506544345e-08, "loss": 0.9287, "step": 10865 }, { "epoch": 4.921195652173913, "grad_norm": 0.037037486741681336, "learning_rate": 6.051952738451939e-08, "loss": 0.9057, "step": 10866 }, { "epoch": 4.921648550724638, "grad_norm": 0.03634718226326351, "learning_rate": 5.982607251675098e-08, "loss": 0.9331, "step": 10867 }, { "epoch": 4.922101449275362, "grad_norm": 0.03671698620393521, "learning_rate": 5.913661053145614e-08, "loss": 0.9154, "step": 10868 }, { "epoch": 4.922554347826087, "grad_norm": 0.03648642198272883, "learning_rate": 5.8451141497570804e-08, "loss": 0.9135, "step": 10869 }, { "epoch": 4.9230072463768115, "grad_norm": 0.03675864914980606, "learning_rate": 5.776966548361351e-08, "loss": 0.9389, "step": 10870 }, { "epoch": 4.923460144927536, "grad_norm": 0.03655751801336593, "learning_rate": 5.709218255771643e-08, "loss": 0.914, "step": 10871 }, { "epoch": 4.923913043478261, "grad_norm": 0.03607808537019597, "learning_rate": 5.641869278761203e-08, "loss": 0.9306, "step": 10872 }, { "epoch": 4.924365942028985, "grad_norm": 0.036733253954195084, "learning_rate": 5.574919624062425e-08, "loss": 0.9015, "step": 10873 }, { "epoch": 4.92481884057971, "grad_norm": 0.03584629893374109, "learning_rate": 5.508369298369065e-08, "loss": 0.8874, "step": 10874 }, { "epoch": 4.925271739130435, "grad_norm": 0.03610989048935422, "learning_rate": 5.44221830833358e-08, "loss": 0.9099, "step": 10875 }, { "epoch": 4.925724637681159, "grad_norm": 0.0356389842559711, "learning_rate": 5.376466660569346e-08, "loss": 0.9248, "step": 10876 }, { "epoch": 4.926177536231884, "grad_norm": 0.036824326607498, "learning_rate": 5.311114361649772e-08, "loss": 0.9181, "step": 10877 }, { "epoch": 4.926630434782608, "grad_norm": 0.03601718695325375, "learning_rate": 5.246161418108298e-08, "loss": 0.9148, "step": 10878 }, { "epoch": 4.927083333333333, "grad_norm": 0.03593045067602935, "learning_rate": 5.1816078364383956e-08, "loss": 0.9343, "step": 10879 }, { "epoch": 4.927536231884058, "grad_norm": 0.03600292489591814, "learning_rate": 5.117453623093127e-08, "loss": 0.9304, "step": 10880 }, { "epoch": 4.927989130434782, "grad_norm": 0.03696227040418064, "learning_rate": 5.053698784486916e-08, "loss": 0.9173, "step": 10881 }, { "epoch": 4.928442028985507, "grad_norm": 0.03643242754793031, "learning_rate": 4.9903433269928856e-08, "loss": 0.9366, "step": 10882 }, { "epoch": 4.928894927536232, "grad_norm": 0.03528394006386958, "learning_rate": 4.927387256944638e-08, "loss": 0.9242, "step": 10883 }, { "epoch": 4.929347826086957, "grad_norm": 0.03718387147801549, "learning_rate": 4.864830580636248e-08, "loss": 0.9163, "step": 10884 }, { "epoch": 4.929800724637682, "grad_norm": 0.03534234608802333, "learning_rate": 4.8026733043218256e-08, "loss": 0.946, "step": 10885 }, { "epoch": 4.930253623188406, "grad_norm": 0.03663137002997746, "learning_rate": 4.7409154342146215e-08, "loss": 0.9333, "step": 10886 }, { "epoch": 4.930706521739131, "grad_norm": 0.038254440308456705, "learning_rate": 4.6795569764896966e-08, "loss": 0.9189, "step": 10887 }, { "epoch": 4.931159420289855, "grad_norm": 0.03698386411014669, "learning_rate": 4.618597937280367e-08, "loss": 0.9296, "step": 10888 }, { "epoch": 4.93161231884058, "grad_norm": 0.035663085784991624, "learning_rate": 4.558038322680869e-08, "loss": 0.9417, "step": 10889 }, { "epoch": 4.932065217391305, "grad_norm": 0.036804392505159704, "learning_rate": 4.497878138745915e-08, "loss": 0.9079, "step": 10890 }, { "epoch": 4.932518115942029, "grad_norm": 0.03573959305718368, "learning_rate": 4.4381173914893605e-08, "loss": 0.9048, "step": 10891 }, { "epoch": 4.932971014492754, "grad_norm": 0.037349328765344456, "learning_rate": 4.378756086885538e-08, "loss": 0.9331, "step": 10892 }, { "epoch": 4.9334239130434785, "grad_norm": 0.03593904833651632, "learning_rate": 4.3197942308692564e-08, "loss": 0.9263, "step": 10893 }, { "epoch": 4.933876811594203, "grad_norm": 0.03654756863952798, "learning_rate": 4.261231829334467e-08, "loss": 0.944, "step": 10894 }, { "epoch": 4.934329710144928, "grad_norm": 0.0372188368930612, "learning_rate": 4.2030688881364854e-08, "loss": 0.9105, "step": 10895 }, { "epoch": 4.934782608695652, "grad_norm": 0.03553109684810498, "learning_rate": 4.145305413089773e-08, "loss": 0.8809, "step": 10896 }, { "epoch": 4.935235507246377, "grad_norm": 0.03601700766798287, "learning_rate": 4.0879414099688205e-08, "loss": 0.9162, "step": 10897 }, { "epoch": 4.9356884057971016, "grad_norm": 0.037379798912688826, "learning_rate": 4.030976884508153e-08, "loss": 0.9081, "step": 10898 }, { "epoch": 4.936141304347826, "grad_norm": 0.035636381121941, "learning_rate": 3.97441184240277e-08, "loss": 0.9086, "step": 10899 }, { "epoch": 4.936594202898551, "grad_norm": 0.035813931507800024, "learning_rate": 3.918246289308147e-08, "loss": 0.9187, "step": 10900 }, { "epoch": 4.937047101449275, "grad_norm": 0.03552561030639323, "learning_rate": 3.862480230838461e-08, "loss": 0.9078, "step": 10901 }, { "epoch": 4.9375, "grad_norm": 0.03631426631844661, "learning_rate": 3.8071136725688074e-08, "loss": 0.9322, "step": 10902 }, { "epoch": 4.937952898550725, "grad_norm": 0.036202515637506455, "learning_rate": 3.752146620035202e-08, "loss": 0.9393, "step": 10903 }, { "epoch": 4.938405797101449, "grad_norm": 0.03649261374701106, "learning_rate": 3.697579078731473e-08, "loss": 0.9208, "step": 10904 }, { "epoch": 4.938858695652174, "grad_norm": 0.036558545150337324, "learning_rate": 3.643411054113699e-08, "loss": 0.912, "step": 10905 }, { "epoch": 4.9393115942028984, "grad_norm": 0.03743328856296528, "learning_rate": 3.589642551596662e-08, "loss": 0.9161, "step": 10906 }, { "epoch": 4.939764492753623, "grad_norm": 0.03637770524812691, "learning_rate": 3.5362735765565036e-08, "loss": 0.9293, "step": 10907 }, { "epoch": 4.940217391304348, "grad_norm": 0.035921695147366, "learning_rate": 3.483304134327181e-08, "loss": 0.9263, "step": 10908 }, { "epoch": 4.940670289855072, "grad_norm": 0.036130986648022445, "learning_rate": 3.43073423020579e-08, "loss": 0.9528, "step": 10909 }, { "epoch": 4.941123188405797, "grad_norm": 0.037054549536922436, "learning_rate": 3.37856386944635e-08, "loss": 0.9414, "step": 10910 }, { "epoch": 4.9415760869565215, "grad_norm": 0.03809216705771078, "learning_rate": 3.326793057265576e-08, "loss": 0.9301, "step": 10911 }, { "epoch": 4.942028985507246, "grad_norm": 0.0351791454025937, "learning_rate": 3.2754217988384406e-08, "loss": 0.9356, "step": 10912 }, { "epoch": 4.942481884057971, "grad_norm": 0.03571827136966171, "learning_rate": 3.2244500993003916e-08, "loss": 0.9219, "step": 10913 }, { "epoch": 4.942934782608695, "grad_norm": 0.03800473568543704, "learning_rate": 3.173877963747796e-08, "loss": 0.9087, "step": 10914 }, { "epoch": 4.94338768115942, "grad_norm": 0.036238547646024515, "learning_rate": 3.1237053972361653e-08, "loss": 0.9272, "step": 10915 }, { "epoch": 4.943840579710145, "grad_norm": 0.03630842433697802, "learning_rate": 3.0739324047810436e-08, "loss": 0.9104, "step": 10916 }, { "epoch": 4.944293478260869, "grad_norm": 0.03773231758305298, "learning_rate": 3.024558991358895e-08, "loss": 0.9369, "step": 10917 }, { "epoch": 4.944746376811594, "grad_norm": 0.03705491162147226, "learning_rate": 2.9755851619053256e-08, "loss": 0.9296, "step": 10918 }, { "epoch": 4.945199275362318, "grad_norm": 0.03719957939556092, "learning_rate": 2.9270109213159757e-08, "loss": 0.915, "step": 10919 }, { "epoch": 4.945652173913043, "grad_norm": 0.036859979655173285, "learning_rate": 2.8788362744478492e-08, "loss": 0.9058, "step": 10920 }, { "epoch": 4.946105072463768, "grad_norm": 0.035711399810633294, "learning_rate": 2.8310612261157612e-08, "loss": 0.9082, "step": 10921 }, { "epoch": 4.946557971014493, "grad_norm": 0.035975828836340595, "learning_rate": 2.7836857810972227e-08, "loss": 0.9136, "step": 10922 }, { "epoch": 4.947010869565218, "grad_norm": 0.03611141900842947, "learning_rate": 2.736709944127114e-08, "loss": 0.9301, "step": 10923 }, { "epoch": 4.947463768115942, "grad_norm": 0.036982117942070306, "learning_rate": 2.6901337199025656e-08, "loss": 0.9151, "step": 10924 }, { "epoch": 4.947916666666667, "grad_norm": 0.03721934681306233, "learning_rate": 2.6439571130798536e-08, "loss": 0.9213, "step": 10925 }, { "epoch": 4.948369565217392, "grad_norm": 0.03590111081329061, "learning_rate": 2.598180128274841e-08, "loss": 0.8975, "step": 10926 }, { "epoch": 4.948822463768116, "grad_norm": 0.037599447359856174, "learning_rate": 2.552802770064311e-08, "loss": 0.9047, "step": 10927 }, { "epoch": 4.949275362318841, "grad_norm": 0.03530796774083084, "learning_rate": 2.5078250429846352e-08, "loss": 0.898, "step": 10928 }, { "epoch": 4.949728260869565, "grad_norm": 0.03602667888274943, "learning_rate": 2.4632469515317725e-08, "loss": 0.9349, "step": 10929 }, { "epoch": 4.95018115942029, "grad_norm": 0.03701653191379181, "learning_rate": 2.4190685001634906e-08, "loss": 0.9173, "step": 10930 }, { "epoch": 4.950634057971015, "grad_norm": 0.0372193105419159, "learning_rate": 2.3752896932949244e-08, "loss": 0.9215, "step": 10931 }, { "epoch": 4.951086956521739, "grad_norm": 0.035999744899003844, "learning_rate": 2.331910535303905e-08, "loss": 0.9096, "step": 10932 }, { "epoch": 4.951539855072464, "grad_norm": 0.03537886079186417, "learning_rate": 2.28893103052652e-08, "loss": 0.9056, "step": 10933 }, { "epoch": 4.9519927536231885, "grad_norm": 0.03716611309878116, "learning_rate": 2.246351183259776e-08, "loss": 0.9094, "step": 10934 }, { "epoch": 4.952445652173913, "grad_norm": 0.03585976794138235, "learning_rate": 2.2041709977598248e-08, "loss": 0.9194, "step": 10935 }, { "epoch": 4.952898550724638, "grad_norm": 0.03670903255898562, "learning_rate": 2.1623904782446246e-08, "loss": 0.9214, "step": 10936 }, { "epoch": 4.953351449275362, "grad_norm": 0.035305942765110926, "learning_rate": 2.1210096288899473e-08, "loss": 0.8926, "step": 10937 }, { "epoch": 4.953804347826087, "grad_norm": 0.03599135796814854, "learning_rate": 2.0800284538329274e-08, "loss": 0.9278, "step": 10938 }, { "epoch": 4.9542572463768115, "grad_norm": 0.03755731756167451, "learning_rate": 2.039446957170732e-08, "loss": 0.94, "step": 10939 }, { "epoch": 4.954710144927536, "grad_norm": 0.03614080778846184, "learning_rate": 1.9992651429610042e-08, "loss": 0.9425, "step": 10940 }, { "epoch": 4.955163043478261, "grad_norm": 0.03714889388849555, "learning_rate": 1.959483015219199e-08, "loss": 0.8945, "step": 10941 }, { "epoch": 4.955615942028985, "grad_norm": 0.037272107371116255, "learning_rate": 1.9201005779239113e-08, "loss": 0.9489, "step": 10942 }, { "epoch": 4.95606884057971, "grad_norm": 0.0360808966539375, "learning_rate": 1.8811178350115476e-08, "loss": 0.9159, "step": 10943 }, { "epoch": 4.956521739130435, "grad_norm": 0.037359887546082535, "learning_rate": 1.8425347903794355e-08, "loss": 0.9079, "step": 10944 }, { "epoch": 4.956974637681159, "grad_norm": 0.038162241715599385, "learning_rate": 1.8043514478849334e-08, "loss": 0.9299, "step": 10945 }, { "epoch": 4.957427536231884, "grad_norm": 0.03532947552474045, "learning_rate": 1.766567811344544e-08, "loss": 0.8992, "step": 10946 }, { "epoch": 4.957880434782608, "grad_norm": 0.03615101243087767, "learning_rate": 1.7291838845365783e-08, "loss": 0.9061, "step": 10947 }, { "epoch": 4.958333333333333, "grad_norm": 0.036515260814991554, "learning_rate": 1.6921996711976028e-08, "loss": 0.9388, "step": 10948 }, { "epoch": 4.958786231884058, "grad_norm": 0.03705515988077403, "learning_rate": 1.655615175025549e-08, "loss": 0.9062, "step": 10949 }, { "epoch": 4.959239130434782, "grad_norm": 0.03682936996519072, "learning_rate": 1.6194303996774906e-08, "loss": 0.9254, "step": 10950 }, { "epoch": 4.959692028985507, "grad_norm": 0.038029226190317206, "learning_rate": 1.583645348770979e-08, "loss": 0.9315, "step": 10951 }, { "epoch": 4.960144927536232, "grad_norm": 0.037482508248899, "learning_rate": 1.5482600258831526e-08, "loss": 0.9157, "step": 10952 }, { "epoch": 4.960597826086957, "grad_norm": 0.0383656007882811, "learning_rate": 1.5132744345525142e-08, "loss": 0.9134, "step": 10953 }, { "epoch": 4.961050724637682, "grad_norm": 0.03973615150426157, "learning_rate": 1.478688578275378e-08, "loss": 0.9046, "step": 10954 }, { "epoch": 4.961503623188406, "grad_norm": 0.03585241657076289, "learning_rate": 1.4445024605103108e-08, "loss": 0.9341, "step": 10955 }, { "epoch": 4.961956521739131, "grad_norm": 0.03518954184272152, "learning_rate": 1.4107160846741353e-08, "loss": 0.9149, "step": 10956 }, { "epoch": 4.962409420289855, "grad_norm": 0.035700630538783376, "learning_rate": 1.3773294541454818e-08, "loss": 0.9072, "step": 10957 }, { "epoch": 4.96286231884058, "grad_norm": 0.03520808994361503, "learning_rate": 1.3443425722612368e-08, "loss": 0.9056, "step": 10958 }, { "epoch": 4.963315217391305, "grad_norm": 0.037027564450019, "learning_rate": 1.311755442319651e-08, "loss": 0.9164, "step": 10959 }, { "epoch": 4.963768115942029, "grad_norm": 0.03647960644301926, "learning_rate": 1.2795680675781186e-08, "loss": 0.9204, "step": 10960 }, { "epoch": 4.964221014492754, "grad_norm": 0.03529050920510702, "learning_rate": 1.247780451254954e-08, "loss": 0.8945, "step": 10961 }, { "epoch": 4.9646739130434785, "grad_norm": 0.03628699365098604, "learning_rate": 1.2163925965276157e-08, "loss": 0.9143, "step": 10962 }, { "epoch": 4.965126811594203, "grad_norm": 0.03688021503602977, "learning_rate": 1.185404506534038e-08, "loss": 0.9273, "step": 10963 }, { "epoch": 4.965579710144928, "grad_norm": 0.03824672948881226, "learning_rate": 1.1548161843726313e-08, "loss": 0.9129, "step": 10964 }, { "epoch": 4.966032608695652, "grad_norm": 0.037259399940699094, "learning_rate": 1.1246276331005058e-08, "loss": 0.9344, "step": 10965 }, { "epoch": 4.966485507246377, "grad_norm": 0.03662157138800928, "learning_rate": 1.0948388557365797e-08, "loss": 0.9363, "step": 10966 }, { "epoch": 4.9669384057971016, "grad_norm": 0.03529545269995018, "learning_rate": 1.0654498552580272e-08, "loss": 0.8953, "step": 10967 }, { "epoch": 4.967391304347826, "grad_norm": 0.03682335585491427, "learning_rate": 1.0364606346033867e-08, "loss": 0.9261, "step": 10968 }, { "epoch": 4.967844202898551, "grad_norm": 0.03653935037378184, "learning_rate": 1.0078711966707844e-08, "loss": 0.9326, "step": 10969 }, { "epoch": 4.968297101449275, "grad_norm": 0.03600400537742652, "learning_rate": 9.796815443179342e-09, "loss": 0.9233, "step": 10970 }, { "epoch": 4.96875, "grad_norm": 0.036434968195986885, "learning_rate": 9.518916803634703e-09, "loss": 0.9068, "step": 10971 }, { "epoch": 4.969202898550725, "grad_norm": 0.03749206686429881, "learning_rate": 9.245016075856151e-09, "loss": 0.9317, "step": 10972 }, { "epoch": 4.969655797101449, "grad_norm": 0.03665995041215544, "learning_rate": 8.975113287221782e-09, "loss": 0.9253, "step": 10973 }, { "epoch": 4.970108695652174, "grad_norm": 0.03546105825300887, "learning_rate": 8.7092084647189e-09, "loss": 0.934, "step": 10974 }, { "epoch": 4.9705615942028984, "grad_norm": 0.037001806339453776, "learning_rate": 8.447301634926242e-09, "loss": 0.9238, "step": 10975 }, { "epoch": 4.971014492753623, "grad_norm": 0.03562028238590576, "learning_rate": 8.18939282403175e-09, "loss": 0.892, "step": 10976 }, { "epoch": 4.971467391304348, "grad_norm": 0.0367786928810114, "learning_rate": 7.935482057814802e-09, "loss": 0.9098, "step": 10977 }, { "epoch": 4.971920289855072, "grad_norm": 0.035974348017705673, "learning_rate": 7.68556936165954e-09, "loss": 0.9042, "step": 10978 }, { "epoch": 4.972373188405797, "grad_norm": 0.03784973186270394, "learning_rate": 7.439654760550419e-09, "loss": 0.9058, "step": 10979 }, { "epoch": 4.9728260869565215, "grad_norm": 0.03899672017828897, "learning_rate": 7.197738279072219e-09, "loss": 0.9119, "step": 10980 }, { "epoch": 4.973278985507246, "grad_norm": 0.03696067475900119, "learning_rate": 6.95981994141448e-09, "loss": 0.9052, "step": 10981 }, { "epoch": 4.973731884057971, "grad_norm": 0.037367918835549165, "learning_rate": 6.725899771353739e-09, "loss": 0.9292, "step": 10982 }, { "epoch": 4.974184782608695, "grad_norm": 0.03764392329021098, "learning_rate": 6.495977792280173e-09, "loss": 0.9049, "step": 10983 }, { "epoch": 4.97463768115942, "grad_norm": 0.03647054536905502, "learning_rate": 6.27005402717984e-09, "loss": 0.9037, "step": 10984 }, { "epoch": 4.975090579710145, "grad_norm": 0.036348467752537535, "learning_rate": 6.048128498639116e-09, "loss": 0.9175, "step": 10985 }, { "epoch": 4.975543478260869, "grad_norm": 0.036067656007712195, "learning_rate": 5.8302012288402596e-09, "loss": 0.9077, "step": 10986 }, { "epoch": 4.975996376811594, "grad_norm": 0.03738351157306031, "learning_rate": 5.616272239570286e-09, "loss": 0.9291, "step": 10987 }, { "epoch": 4.976449275362318, "grad_norm": 0.03472141571375057, "learning_rate": 5.406341552225414e-09, "loss": 0.9148, "step": 10988 }, { "epoch": 4.976902173913043, "grad_norm": 0.03621289904792055, "learning_rate": 5.200409187779976e-09, "loss": 0.9209, "step": 10989 }, { "epoch": 4.977355072463768, "grad_norm": 0.03732364155152455, "learning_rate": 4.9984751668263885e-09, "loss": 0.9085, "step": 10990 }, { "epoch": 4.977807971014493, "grad_norm": 0.03599918440439071, "learning_rate": 4.800539509552948e-09, "loss": 0.918, "step": 10991 }, { "epoch": 4.978260869565218, "grad_norm": 0.03615136874231791, "learning_rate": 4.606602235752711e-09, "loss": 0.9244, "step": 10992 }, { "epoch": 4.978713768115942, "grad_norm": 0.03674624583251946, "learning_rate": 4.416663364801288e-09, "loss": 0.9289, "step": 10993 }, { "epoch": 4.979166666666667, "grad_norm": 0.03759915451248596, "learning_rate": 4.230722915701257e-09, "loss": 0.9358, "step": 10994 }, { "epoch": 4.979619565217392, "grad_norm": 0.03687657600553294, "learning_rate": 4.0487809070288705e-09, "loss": 0.9172, "step": 10995 }, { "epoch": 4.980072463768116, "grad_norm": 0.03649453066046883, "learning_rate": 3.870837356982904e-09, "loss": 0.8865, "step": 10996 }, { "epoch": 4.980525362318841, "grad_norm": 0.0363192167149325, "learning_rate": 3.6968922833446885e-09, "loss": 0.9087, "step": 10997 }, { "epoch": 4.980978260869565, "grad_norm": 0.03764081480149489, "learning_rate": 3.526945703509199e-09, "loss": 0.9177, "step": 10998 }, { "epoch": 4.98143115942029, "grad_norm": 0.03696950582704788, "learning_rate": 3.360997634467289e-09, "loss": 0.9352, "step": 10999 }, { "epoch": 4.981884057971015, "grad_norm": 0.03662913332489977, "learning_rate": 3.19904809280569e-09, "loss": 0.9275, "step": 11000 }, { "epoch": 4.982336956521739, "grad_norm": 0.03466218507583154, "learning_rate": 3.041097094711454e-09, "loss": 0.9268, "step": 11001 }, { "epoch": 4.982789855072464, "grad_norm": 0.036882136880295784, "learning_rate": 2.8871446559852744e-09, "loss": 0.9062, "step": 11002 }, { "epoch": 4.9832427536231885, "grad_norm": 0.036159244920915226, "learning_rate": 2.7371907920104022e-09, "loss": 0.9327, "step": 11003 }, { "epoch": 4.983695652173913, "grad_norm": 0.03647521241077318, "learning_rate": 2.591235517779289e-09, "loss": 0.9419, "step": 11004 }, { "epoch": 4.984148550724638, "grad_norm": 0.03532422218190119, "learning_rate": 2.4492788478802654e-09, "loss": 0.9024, "step": 11005 }, { "epoch": 4.984601449275362, "grad_norm": 0.03520997891513662, "learning_rate": 2.311320796510863e-09, "loss": 0.9054, "step": 11006 }, { "epoch": 4.985054347826087, "grad_norm": 0.03615649449499151, "learning_rate": 2.177361377460052e-09, "loss": 0.9198, "step": 11007 }, { "epoch": 4.9855072463768115, "grad_norm": 0.03679487909592031, "learning_rate": 2.0474006041215634e-09, "loss": 0.9249, "step": 11008 }, { "epoch": 4.985960144927536, "grad_norm": 0.0384872839008101, "learning_rate": 1.9214384894894468e-09, "loss": 0.9117, "step": 11009 }, { "epoch": 4.986413043478261, "grad_norm": 0.03668027526135608, "learning_rate": 1.7994750461491905e-09, "loss": 0.9201, "step": 11010 }, { "epoch": 4.986865942028985, "grad_norm": 0.03751766646647131, "learning_rate": 1.6815102862999255e-09, "loss": 0.9251, "step": 11011 }, { "epoch": 4.98731884057971, "grad_norm": 0.03559450478226427, "learning_rate": 1.5675442217322201e-09, "loss": 0.9173, "step": 11012 }, { "epoch": 4.987771739130435, "grad_norm": 0.036743326337558654, "learning_rate": 1.4575768638369626e-09, "loss": 0.9316, "step": 11013 }, { "epoch": 4.988224637681159, "grad_norm": 0.03585810526640892, "learning_rate": 1.3516082236142424e-09, "loss": 0.9211, "step": 11014 }, { "epoch": 4.988677536231884, "grad_norm": 0.03688202045365322, "learning_rate": 1.2496383116555877e-09, "loss": 0.9227, "step": 11015 }, { "epoch": 4.989130434782608, "grad_norm": 0.03721839825598312, "learning_rate": 1.1516671381528454e-09, "loss": 0.9136, "step": 11016 }, { "epoch": 4.989583333333333, "grad_norm": 0.03610891229156521, "learning_rate": 1.057694712902624e-09, "loss": 0.9192, "step": 11017 }, { "epoch": 4.990036231884058, "grad_norm": 0.03604845989128548, "learning_rate": 9.677210452974095e-10, "loss": 0.918, "step": 11018 }, { "epoch": 4.990489130434782, "grad_norm": 0.037672576921710055, "learning_rate": 8.81746144330009e-10, "loss": 0.9154, "step": 11019 }, { "epoch": 4.990942028985507, "grad_norm": 0.036063310473133395, "learning_rate": 7.997700186024304e-10, "loss": 0.9195, "step": 11020 }, { "epoch": 4.991394927536232, "grad_norm": 0.03663980130853759, "learning_rate": 7.217926763036787e-10, "loss": 0.9384, "step": 11021 }, { "epoch": 4.991847826086957, "grad_norm": 0.03679177241149953, "learning_rate": 6.478141252319603e-10, "loss": 0.9301, "step": 11022 }, { "epoch": 4.992300724637682, "grad_norm": 0.036109891128034664, "learning_rate": 5.778343727858016e-10, "loss": 0.9202, "step": 11023 }, { "epoch": 4.992753623188406, "grad_norm": 0.03881630944065768, "learning_rate": 5.118534259551667e-10, "loss": 0.9255, "step": 11024 }, { "epoch": 4.993206521739131, "grad_norm": 0.035917876367757014, "learning_rate": 4.4987129133922116e-10, "loss": 0.9301, "step": 11025 }, { "epoch": 4.993659420289855, "grad_norm": 0.0367186527372499, "learning_rate": 3.9188797513300956e-10, "loss": 0.9337, "step": 11026 }, { "epoch": 4.99411231884058, "grad_norm": 0.03651971618077037, "learning_rate": 3.37903483136337e-10, "loss": 0.9054, "step": 11027 }, { "epoch": 4.994565217391305, "grad_norm": 0.03797331435766102, "learning_rate": 2.879178207448874e-10, "loss": 0.9164, "step": 11028 }, { "epoch": 4.995018115942029, "grad_norm": 0.036371882086587586, "learning_rate": 2.4193099295466426e-10, "loss": 0.9215, "step": 11029 }, { "epoch": 4.995471014492754, "grad_norm": 0.036969405754608185, "learning_rate": 1.9994300436643187e-10, "loss": 0.9175, "step": 11030 }, { "epoch": 4.9959239130434785, "grad_norm": 0.03604914189152817, "learning_rate": 1.6195385917239238e-10, "loss": 0.915, "step": 11031 }, { "epoch": 4.996376811594203, "grad_norm": 0.035555764072428414, "learning_rate": 1.2796356117394936e-10, "loss": 0.9015, "step": 11032 }, { "epoch": 4.996829710144928, "grad_norm": 0.036172989436399586, "learning_rate": 9.797211376394445e-11, "loss": 0.9137, "step": 11033 }, { "epoch": 4.997282608695652, "grad_norm": 0.0375073504017792, "learning_rate": 7.197951994886154e-11, "loss": 0.9171, "step": 11034 }, { "epoch": 4.997735507246377, "grad_norm": 0.03694954039734701, "learning_rate": 4.9985782322181654e-11, "loss": 0.9157, "step": 11035 }, { "epoch": 4.9981884057971016, "grad_norm": 0.0357163903473403, "learning_rate": 3.199090308658725e-11, "loss": 0.9099, "step": 11036 }, { "epoch": 4.998641304347826, "grad_norm": 0.036739401068547214, "learning_rate": 1.7994884036198757e-11, "loss": 0.9148, "step": 11037 }, { "epoch": 4.999094202898551, "grad_norm": 0.0362585105726279, "learning_rate": 7.997726569897168e-12, "loss": 0.9147, "step": 11038 }, { "epoch": 4.999547101449275, "grad_norm": 0.03549708032179242, "learning_rate": 1.999431691324105e-12, "loss": 0.9088, "step": 11039 }, { "epoch": 5.0, "grad_norm": 0.036632343406672395, "learning_rate": 0.0, "loss": 0.9244, "step": 11040 }, { "epoch": 5.0, "step": 11040, "total_flos": 8.051311333343232e+16, "train_loss": 0.36911470478643543, "train_runtime": 49525.0547, "train_samples_per_second": 114.1, "train_steps_per_second": 0.223 } ], "logging_steps": 1, "max_steps": 11040, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8.051311333343232e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }