{ "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 11040, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0004528985507246377, "grad_norm": 3.39732863417477, "learning_rate": 7.246376811594204e-08, "loss": 1.0567, "step": 1 }, { "epoch": 0.0009057971014492754, "grad_norm": 3.4298865509240826, "learning_rate": 1.4492753623188408e-07, "loss": 1.0662, "step": 2 }, { "epoch": 0.001358695652173913, "grad_norm": 3.4594259500675837, "learning_rate": 2.173913043478261e-07, "loss": 1.0765, "step": 3 }, { "epoch": 0.0018115942028985507, "grad_norm": 3.4037689939630478, "learning_rate": 2.8985507246376816e-07, "loss": 1.1008, "step": 4 }, { "epoch": 0.0022644927536231885, "grad_norm": 3.3652998107506225, "learning_rate": 3.623188405797102e-07, "loss": 1.0974, "step": 5 }, { "epoch": 0.002717391304347826, "grad_norm": 3.394000729068578, "learning_rate": 4.347826086956522e-07, "loss": 1.0683, "step": 6 }, { "epoch": 0.003170289855072464, "grad_norm": 3.4333410623520835, "learning_rate": 5.072463768115942e-07, "loss": 1.0642, "step": 7 }, { "epoch": 0.0036231884057971015, "grad_norm": 3.3999472925732843, "learning_rate": 5.797101449275363e-07, "loss": 1.0621, "step": 8 }, { "epoch": 0.004076086956521739, "grad_norm": 3.3361347250165436, "learning_rate": 6.521739130434783e-07, "loss": 1.0643, "step": 9 }, { "epoch": 0.004528985507246377, "grad_norm": 3.334299147880356, "learning_rate": 7.246376811594204e-07, "loss": 1.0531, "step": 10 }, { "epoch": 0.004981884057971014, "grad_norm": 3.402778194741388, "learning_rate": 7.971014492753623e-07, "loss": 1.0763, "step": 11 }, { "epoch": 0.005434782608695652, "grad_norm": 3.344480485773468, "learning_rate": 8.695652173913044e-07, "loss": 1.0632, "step": 12 }, { "epoch": 0.00588768115942029, "grad_norm": 3.3554867096020735, "learning_rate": 9.420289855072465e-07, "loss": 1.0842, "step": 13 }, { "epoch": 0.006340579710144928, "grad_norm": 3.319819119248795, "learning_rate": 1.0144927536231885e-06, "loss": 1.0548, "step": 14 }, { "epoch": 0.006793478260869565, "grad_norm": 3.256406383424863, "learning_rate": 1.0869565217391306e-06, "loss": 1.0832, "step": 15 }, { "epoch": 0.007246376811594203, "grad_norm": 3.227685173399398, "learning_rate": 1.1594202898550726e-06, "loss": 1.0596, "step": 16 }, { "epoch": 0.007699275362318841, "grad_norm": 2.9312036646028923, "learning_rate": 1.2318840579710147e-06, "loss": 1.0465, "step": 17 }, { "epoch": 0.008152173913043478, "grad_norm": 2.9333750577200672, "learning_rate": 1.3043478260869566e-06, "loss": 1.0656, "step": 18 }, { "epoch": 0.008605072463768116, "grad_norm": 2.9276742804426785, "learning_rate": 1.3768115942028987e-06, "loss": 1.0592, "step": 19 }, { "epoch": 0.009057971014492754, "grad_norm": 2.865792595428422, "learning_rate": 1.4492753623188408e-06, "loss": 1.0414, "step": 20 }, { "epoch": 0.009510869565217392, "grad_norm": 2.8379285206611784, "learning_rate": 1.521739130434783e-06, "loss": 1.0212, "step": 21 }, { "epoch": 0.009963768115942028, "grad_norm": 2.591331240019748, "learning_rate": 1.5942028985507246e-06, "loss": 1.0225, "step": 22 }, { "epoch": 0.010416666666666666, "grad_norm": 2.0684790481268625, "learning_rate": 1.6666666666666667e-06, "loss": 1.0485, "step": 23 }, { "epoch": 0.010869565217391304, "grad_norm": 2.0101408800033616, "learning_rate": 1.7391304347826088e-06, "loss": 1.0338, "step": 24 }, { "epoch": 0.011322463768115942, "grad_norm": 1.90086291060798, "learning_rate": 1.8115942028985508e-06, "loss": 0.9983, "step": 25 }, { "epoch": 0.01177536231884058, "grad_norm": 1.8750108858737726, "learning_rate": 1.884057971014493e-06, "loss": 1.0196, "step": 26 }, { "epoch": 0.012228260869565218, "grad_norm": 1.7785814618064852, "learning_rate": 1.956521739130435e-06, "loss": 1.0375, "step": 27 }, { "epoch": 0.012681159420289856, "grad_norm": 1.670433962778606, "learning_rate": 2.028985507246377e-06, "loss": 1.0146, "step": 28 }, { "epoch": 0.013134057971014492, "grad_norm": 1.6152781513901113, "learning_rate": 2.101449275362319e-06, "loss": 1.0357, "step": 29 }, { "epoch": 0.01358695652173913, "grad_norm": 1.1952303155059478, "learning_rate": 2.173913043478261e-06, "loss": 0.9971, "step": 30 }, { "epoch": 0.014039855072463768, "grad_norm": 0.8936200033447007, "learning_rate": 2.246376811594203e-06, "loss": 1.0104, "step": 31 }, { "epoch": 0.014492753623188406, "grad_norm": 1.064194707904204, "learning_rate": 2.3188405797101453e-06, "loss": 1.0091, "step": 32 }, { "epoch": 0.014945652173913044, "grad_norm": 1.3025370577199196, "learning_rate": 2.391304347826087e-06, "loss": 0.9958, "step": 33 }, { "epoch": 0.015398550724637682, "grad_norm": 1.4627081141417213, "learning_rate": 2.4637681159420295e-06, "loss": 1.0025, "step": 34 }, { "epoch": 0.01585144927536232, "grad_norm": 1.435915634419328, "learning_rate": 2.5362318840579714e-06, "loss": 1.0093, "step": 35 }, { "epoch": 0.016304347826086956, "grad_norm": 1.4249792395063776, "learning_rate": 2.6086956521739132e-06, "loss": 1.0261, "step": 36 }, { "epoch": 0.016757246376811596, "grad_norm": 1.2343448554298366, "learning_rate": 2.6811594202898555e-06, "loss": 0.9911, "step": 37 }, { "epoch": 0.017210144927536232, "grad_norm": 1.0886715549593102, "learning_rate": 2.7536231884057974e-06, "loss": 0.9814, "step": 38 }, { "epoch": 0.017663043478260868, "grad_norm": 0.9585035844236229, "learning_rate": 2.8260869565217393e-06, "loss": 1.0076, "step": 39 }, { "epoch": 0.018115942028985508, "grad_norm": 0.847092209488619, "learning_rate": 2.8985507246376816e-06, "loss": 1.0659, "step": 40 }, { "epoch": 0.018568840579710144, "grad_norm": 0.6968108600363109, "learning_rate": 2.9710144927536235e-06, "loss": 0.9996, "step": 41 }, { "epoch": 0.019021739130434784, "grad_norm": 0.769396324460943, "learning_rate": 3.043478260869566e-06, "loss": 1.0106, "step": 42 }, { "epoch": 0.01947463768115942, "grad_norm": 0.6881782296141182, "learning_rate": 3.1159420289855073e-06, "loss": 0.9993, "step": 43 }, { "epoch": 0.019927536231884056, "grad_norm": 0.6328226044524058, "learning_rate": 3.188405797101449e-06, "loss": 0.9956, "step": 44 }, { "epoch": 0.020380434782608696, "grad_norm": 0.5545708716636935, "learning_rate": 3.2608695652173914e-06, "loss": 0.9849, "step": 45 }, { "epoch": 0.020833333333333332, "grad_norm": 0.542590152120146, "learning_rate": 3.3333333333333333e-06, "loss": 0.998, "step": 46 }, { "epoch": 0.021286231884057972, "grad_norm": 0.5556127327659357, "learning_rate": 3.4057971014492756e-06, "loss": 0.9936, "step": 47 }, { "epoch": 0.021739130434782608, "grad_norm": 0.57943604411544, "learning_rate": 3.4782608695652175e-06, "loss": 0.9844, "step": 48 }, { "epoch": 0.022192028985507248, "grad_norm": 0.5836297722677296, "learning_rate": 3.55072463768116e-06, "loss": 0.9847, "step": 49 }, { "epoch": 0.022644927536231884, "grad_norm": 0.5212660887922798, "learning_rate": 3.6231884057971017e-06, "loss": 0.9785, "step": 50 }, { "epoch": 0.02309782608695652, "grad_norm": 0.45717568569875167, "learning_rate": 3.6956521739130436e-06, "loss": 0.9758, "step": 51 }, { "epoch": 0.02355072463768116, "grad_norm": 0.3828640545853302, "learning_rate": 3.768115942028986e-06, "loss": 0.9658, "step": 52 }, { "epoch": 0.024003623188405796, "grad_norm": 0.31778317210211665, "learning_rate": 3.840579710144928e-06, "loss": 0.9731, "step": 53 }, { "epoch": 0.024456521739130436, "grad_norm": 0.31356281027449867, "learning_rate": 3.91304347826087e-06, "loss": 0.9501, "step": 54 }, { "epoch": 0.024909420289855072, "grad_norm": 0.3338255524212708, "learning_rate": 3.9855072463768115e-06, "loss": 0.9658, "step": 55 }, { "epoch": 0.025362318840579712, "grad_norm": 0.31679158546355374, "learning_rate": 4.057971014492754e-06, "loss": 0.9655, "step": 56 }, { "epoch": 0.025815217391304348, "grad_norm": 0.3253970981161815, "learning_rate": 4.130434782608696e-06, "loss": 0.9893, "step": 57 }, { "epoch": 0.026268115942028984, "grad_norm": 0.2879383914483881, "learning_rate": 4.202898550724638e-06, "loss": 0.9863, "step": 58 }, { "epoch": 0.026721014492753624, "grad_norm": 0.27972834732368723, "learning_rate": 4.27536231884058e-06, "loss": 0.9632, "step": 59 }, { "epoch": 0.02717391304347826, "grad_norm": 0.2670216441640564, "learning_rate": 4.347826086956522e-06, "loss": 0.9755, "step": 60 }, { "epoch": 0.0276268115942029, "grad_norm": 0.2932448421424406, "learning_rate": 4.4202898550724645e-06, "loss": 0.9746, "step": 61 }, { "epoch": 0.028079710144927536, "grad_norm": 0.24589338497145874, "learning_rate": 4.492753623188406e-06, "loss": 0.9602, "step": 62 }, { "epoch": 0.028532608695652172, "grad_norm": 0.21003656020400616, "learning_rate": 4.565217391304348e-06, "loss": 0.9804, "step": 63 }, { "epoch": 0.028985507246376812, "grad_norm": 0.20126649121075427, "learning_rate": 4.637681159420291e-06, "loss": 0.9793, "step": 64 }, { "epoch": 0.029438405797101448, "grad_norm": 0.20518175912590234, "learning_rate": 4.710144927536232e-06, "loss": 0.9592, "step": 65 }, { "epoch": 0.029891304347826088, "grad_norm": 0.21327185547746552, "learning_rate": 4.782608695652174e-06, "loss": 0.9311, "step": 66 }, { "epoch": 0.030344202898550724, "grad_norm": 0.2221462748065046, "learning_rate": 4.855072463768117e-06, "loss": 0.9523, "step": 67 }, { "epoch": 0.030797101449275364, "grad_norm": 0.2191499516537148, "learning_rate": 4.927536231884059e-06, "loss": 0.9497, "step": 68 }, { "epoch": 0.03125, "grad_norm": 0.1514793992580924, "learning_rate": 5e-06, "loss": 0.9841, "step": 69 }, { "epoch": 0.03170289855072464, "grad_norm": 0.1501770309505511, "learning_rate": 5.072463768115943e-06, "loss": 0.9777, "step": 70 }, { "epoch": 0.03215579710144927, "grad_norm": 0.1444659990017339, "learning_rate": 5.144927536231884e-06, "loss": 0.9684, "step": 71 }, { "epoch": 0.03260869565217391, "grad_norm": 0.18955082986730495, "learning_rate": 5.2173913043478265e-06, "loss": 0.9414, "step": 72 }, { "epoch": 0.03306159420289855, "grad_norm": 0.1881818084292502, "learning_rate": 5.289855072463769e-06, "loss": 0.9529, "step": 73 }, { "epoch": 0.03351449275362319, "grad_norm": 0.15384471104623437, "learning_rate": 5.362318840579711e-06, "loss": 0.9718, "step": 74 }, { "epoch": 0.033967391304347824, "grad_norm": 0.13357280475161476, "learning_rate": 5.4347826086956525e-06, "loss": 0.9526, "step": 75 }, { "epoch": 0.034420289855072464, "grad_norm": 0.13194252313850047, "learning_rate": 5.507246376811595e-06, "loss": 0.9612, "step": 76 }, { "epoch": 0.034873188405797104, "grad_norm": 0.14546061057259882, "learning_rate": 5.579710144927537e-06, "loss": 0.964, "step": 77 }, { "epoch": 0.035326086956521736, "grad_norm": 0.16591011926582636, "learning_rate": 5.652173913043479e-06, "loss": 0.9867, "step": 78 }, { "epoch": 0.035778985507246376, "grad_norm": 0.11277758978336146, "learning_rate": 5.724637681159421e-06, "loss": 0.9365, "step": 79 }, { "epoch": 0.036231884057971016, "grad_norm": 0.10663177691162679, "learning_rate": 5.797101449275363e-06, "loss": 0.9566, "step": 80 }, { "epoch": 0.036684782608695655, "grad_norm": 0.11047622370716603, "learning_rate": 5.8695652173913055e-06, "loss": 0.9651, "step": 81 }, { "epoch": 0.03713768115942029, "grad_norm": 0.12055160403530289, "learning_rate": 5.942028985507247e-06, "loss": 0.9702, "step": 82 }, { "epoch": 0.03759057971014493, "grad_norm": 0.12103895236173973, "learning_rate": 6.014492753623189e-06, "loss": 0.944, "step": 83 }, { "epoch": 0.03804347826086957, "grad_norm": 0.11894037037854498, "learning_rate": 6.086956521739132e-06, "loss": 0.9573, "step": 84 }, { "epoch": 0.0384963768115942, "grad_norm": 0.09631570715857289, "learning_rate": 6.159420289855072e-06, "loss": 0.9672, "step": 85 }, { "epoch": 0.03894927536231884, "grad_norm": 0.20635537158651998, "learning_rate": 6.2318840579710145e-06, "loss": 0.9663, "step": 86 }, { "epoch": 0.03940217391304348, "grad_norm": 0.10828001047712539, "learning_rate": 6.304347826086958e-06, "loss": 0.9587, "step": 87 }, { "epoch": 0.03985507246376811, "grad_norm": 0.11838314526073247, "learning_rate": 6.376811594202898e-06, "loss": 0.9506, "step": 88 }, { "epoch": 0.04030797101449275, "grad_norm": 0.09944063801394436, "learning_rate": 6.449275362318841e-06, "loss": 0.972, "step": 89 }, { "epoch": 0.04076086956521739, "grad_norm": 0.10627971711153358, "learning_rate": 6.521739130434783e-06, "loss": 0.9662, "step": 90 }, { "epoch": 0.04121376811594203, "grad_norm": 0.10282676977850515, "learning_rate": 6.594202898550725e-06, "loss": 0.9645, "step": 91 }, { "epoch": 0.041666666666666664, "grad_norm": 0.12352524623271946, "learning_rate": 6.666666666666667e-06, "loss": 0.9756, "step": 92 }, { "epoch": 0.042119565217391304, "grad_norm": 0.6331610618090454, "learning_rate": 6.739130434782609e-06, "loss": 0.9773, "step": 93 }, { "epoch": 0.042572463768115944, "grad_norm": 0.10931423506515056, "learning_rate": 6.811594202898551e-06, "loss": 0.9705, "step": 94 }, { "epoch": 0.043025362318840576, "grad_norm": 0.10668623775497925, "learning_rate": 6.884057971014493e-06, "loss": 0.967, "step": 95 }, { "epoch": 0.043478260869565216, "grad_norm": 0.11697200632330902, "learning_rate": 6.956521739130435e-06, "loss": 0.9547, "step": 96 }, { "epoch": 0.043931159420289856, "grad_norm": 0.09528060110211414, "learning_rate": 7.028985507246377e-06, "loss": 0.9628, "step": 97 }, { "epoch": 0.044384057971014496, "grad_norm": 0.08857124043337387, "learning_rate": 7.10144927536232e-06, "loss": 0.9566, "step": 98 }, { "epoch": 0.04483695652173913, "grad_norm": 0.11274091154755868, "learning_rate": 7.173913043478261e-06, "loss": 0.9567, "step": 99 }, { "epoch": 0.04528985507246377, "grad_norm": 0.11667678765798527, "learning_rate": 7.246376811594203e-06, "loss": 0.9535, "step": 100 }, { "epoch": 0.04574275362318841, "grad_norm": 0.1008997179402615, "learning_rate": 7.318840579710146e-06, "loss": 0.9705, "step": 101 }, { "epoch": 0.04619565217391304, "grad_norm": 0.08984329216352704, "learning_rate": 7.391304347826087e-06, "loss": 0.952, "step": 102 }, { "epoch": 0.04664855072463768, "grad_norm": 0.08864789497574734, "learning_rate": 7.4637681159420295e-06, "loss": 0.9543, "step": 103 }, { "epoch": 0.04710144927536232, "grad_norm": 0.10272763353257724, "learning_rate": 7.536231884057972e-06, "loss": 0.9461, "step": 104 }, { "epoch": 0.04755434782608696, "grad_norm": 0.08988841173984777, "learning_rate": 7.608695652173914e-06, "loss": 0.9458, "step": 105 }, { "epoch": 0.04800724637681159, "grad_norm": 0.09471781708118013, "learning_rate": 7.681159420289856e-06, "loss": 0.9586, "step": 106 }, { "epoch": 0.04846014492753623, "grad_norm": 0.09720446365782988, "learning_rate": 7.753623188405797e-06, "loss": 0.9569, "step": 107 }, { "epoch": 0.04891304347826087, "grad_norm": 0.09737416336990094, "learning_rate": 7.82608695652174e-06, "loss": 0.9714, "step": 108 }, { "epoch": 0.049365942028985504, "grad_norm": 0.08603808550102579, "learning_rate": 7.898550724637682e-06, "loss": 0.9599, "step": 109 }, { "epoch": 0.049818840579710144, "grad_norm": 0.09819605436125065, "learning_rate": 7.971014492753623e-06, "loss": 0.9475, "step": 110 }, { "epoch": 0.050271739130434784, "grad_norm": 0.08854766784957023, "learning_rate": 8.043478260869566e-06, "loss": 0.9691, "step": 111 }, { "epoch": 0.050724637681159424, "grad_norm": 0.08430123808634124, "learning_rate": 8.115942028985508e-06, "loss": 0.9604, "step": 112 }, { "epoch": 0.051177536231884056, "grad_norm": 0.08889069907129027, "learning_rate": 8.188405797101449e-06, "loss": 0.9598, "step": 113 }, { "epoch": 0.051630434782608696, "grad_norm": 0.09710490322581038, "learning_rate": 8.260869565217392e-06, "loss": 0.9571, "step": 114 }, { "epoch": 0.052083333333333336, "grad_norm": 0.09943277984620835, "learning_rate": 8.333333333333334e-06, "loss": 0.9615, "step": 115 }, { "epoch": 0.05253623188405797, "grad_norm": 0.09041682023068774, "learning_rate": 8.405797101449275e-06, "loss": 0.9508, "step": 116 }, { "epoch": 0.05298913043478261, "grad_norm": 0.11030611127239204, "learning_rate": 8.478260869565218e-06, "loss": 0.9611, "step": 117 }, { "epoch": 0.05344202898550725, "grad_norm": 0.09854011687620831, "learning_rate": 8.55072463768116e-06, "loss": 0.9502, "step": 118 }, { "epoch": 0.05389492753623189, "grad_norm": 0.09142220557888171, "learning_rate": 8.623188405797103e-06, "loss": 0.9784, "step": 119 }, { "epoch": 0.05434782608695652, "grad_norm": 0.10993180854683386, "learning_rate": 8.695652173913044e-06, "loss": 0.9455, "step": 120 }, { "epoch": 0.05480072463768116, "grad_norm": 0.08661650366994576, "learning_rate": 8.768115942028986e-06, "loss": 0.9241, "step": 121 }, { "epoch": 0.0552536231884058, "grad_norm": 0.09356490997400345, "learning_rate": 8.840579710144929e-06, "loss": 0.9549, "step": 122 }, { "epoch": 0.05570652173913043, "grad_norm": 0.09529850476548232, "learning_rate": 8.91304347826087e-06, "loss": 0.9401, "step": 123 }, { "epoch": 0.05615942028985507, "grad_norm": 0.09608210565182848, "learning_rate": 8.985507246376812e-06, "loss": 0.9621, "step": 124 }, { "epoch": 0.05661231884057971, "grad_norm": 0.0989425400802516, "learning_rate": 9.057971014492755e-06, "loss": 0.9596, "step": 125 }, { "epoch": 0.057065217391304345, "grad_norm": 0.09983021367393163, "learning_rate": 9.130434782608697e-06, "loss": 0.9659, "step": 126 }, { "epoch": 0.057518115942028984, "grad_norm": 0.10938767347174978, "learning_rate": 9.202898550724638e-06, "loss": 0.9802, "step": 127 }, { "epoch": 0.057971014492753624, "grad_norm": 0.12393997565301197, "learning_rate": 9.275362318840581e-06, "loss": 0.9648, "step": 128 }, { "epoch": 0.058423913043478264, "grad_norm": 0.10099980061027061, "learning_rate": 9.347826086956523e-06, "loss": 0.9437, "step": 129 }, { "epoch": 0.058876811594202896, "grad_norm": 0.12458254103762074, "learning_rate": 9.420289855072464e-06, "loss": 0.9482, "step": 130 }, { "epoch": 0.059329710144927536, "grad_norm": 0.10141667856454087, "learning_rate": 9.492753623188407e-06, "loss": 0.9459, "step": 131 }, { "epoch": 0.059782608695652176, "grad_norm": 0.09497950323642038, "learning_rate": 9.565217391304349e-06, "loss": 0.9326, "step": 132 }, { "epoch": 0.06023550724637681, "grad_norm": 0.10222055966981428, "learning_rate": 9.63768115942029e-06, "loss": 0.9592, "step": 133 }, { "epoch": 0.06068840579710145, "grad_norm": 0.11247172191905425, "learning_rate": 9.710144927536233e-06, "loss": 0.9515, "step": 134 }, { "epoch": 0.06114130434782609, "grad_norm": 0.12005790525387577, "learning_rate": 9.782608695652175e-06, "loss": 0.9541, "step": 135 }, { "epoch": 0.06159420289855073, "grad_norm": 0.1029479433492832, "learning_rate": 9.855072463768118e-06, "loss": 0.9681, "step": 136 }, { "epoch": 0.06204710144927536, "grad_norm": 0.1014878320771825, "learning_rate": 9.927536231884058e-06, "loss": 0.9481, "step": 137 }, { "epoch": 0.0625, "grad_norm": 0.11131893674291352, "learning_rate": 1e-05, "loss": 0.9421, "step": 138 }, { "epoch": 0.06295289855072464, "grad_norm": 0.0961937426579682, "learning_rate": 1.0072463768115944e-05, "loss": 0.9567, "step": 139 }, { "epoch": 0.06340579710144928, "grad_norm": 0.09376746764441096, "learning_rate": 1.0144927536231885e-05, "loss": 0.9428, "step": 140 }, { "epoch": 0.06385869565217392, "grad_norm": 0.10053558438445048, "learning_rate": 1.0217391304347829e-05, "loss": 0.9424, "step": 141 }, { "epoch": 0.06431159420289854, "grad_norm": 0.0947194035836528, "learning_rate": 1.0289855072463768e-05, "loss": 0.9446, "step": 142 }, { "epoch": 0.06476449275362318, "grad_norm": 0.10108434953827919, "learning_rate": 1.036231884057971e-05, "loss": 0.9638, "step": 143 }, { "epoch": 0.06521739130434782, "grad_norm": 0.09586650201631951, "learning_rate": 1.0434782608695653e-05, "loss": 0.9634, "step": 144 }, { "epoch": 0.06567028985507246, "grad_norm": 0.1068684966895947, "learning_rate": 1.0507246376811594e-05, "loss": 0.9452, "step": 145 }, { "epoch": 0.0661231884057971, "grad_norm": 0.0969100033363503, "learning_rate": 1.0579710144927538e-05, "loss": 0.9604, "step": 146 }, { "epoch": 0.06657608695652174, "grad_norm": 0.10365210861056819, "learning_rate": 1.0652173913043479e-05, "loss": 0.9431, "step": 147 }, { "epoch": 0.06702898550724638, "grad_norm": 0.10156802148738119, "learning_rate": 1.0724637681159422e-05, "loss": 0.9637, "step": 148 }, { "epoch": 0.06748188405797101, "grad_norm": 0.10863772887107642, "learning_rate": 1.0797101449275362e-05, "loss": 0.9667, "step": 149 }, { "epoch": 0.06793478260869565, "grad_norm": 0.10782195655528112, "learning_rate": 1.0869565217391305e-05, "loss": 0.9465, "step": 150 }, { "epoch": 0.06838768115942029, "grad_norm": 0.09812922794356396, "learning_rate": 1.0942028985507247e-05, "loss": 0.9256, "step": 151 }, { "epoch": 0.06884057971014493, "grad_norm": 0.11569488602788863, "learning_rate": 1.101449275362319e-05, "loss": 0.9326, "step": 152 }, { "epoch": 0.06929347826086957, "grad_norm": 0.10446634272165661, "learning_rate": 1.1086956521739131e-05, "loss": 0.9478, "step": 153 }, { "epoch": 0.06974637681159421, "grad_norm": 0.10953831317722099, "learning_rate": 1.1159420289855074e-05, "loss": 0.9432, "step": 154 }, { "epoch": 0.07019927536231885, "grad_norm": 0.09805261881754977, "learning_rate": 1.1231884057971016e-05, "loss": 0.9507, "step": 155 }, { "epoch": 0.07065217391304347, "grad_norm": 0.13476916510262196, "learning_rate": 1.1304347826086957e-05, "loss": 0.9567, "step": 156 }, { "epoch": 0.07110507246376811, "grad_norm": 0.11682639166571869, "learning_rate": 1.1376811594202899e-05, "loss": 0.9577, "step": 157 }, { "epoch": 0.07155797101449275, "grad_norm": 0.11228304782232613, "learning_rate": 1.1449275362318842e-05, "loss": 0.9605, "step": 158 }, { "epoch": 0.07201086956521739, "grad_norm": 0.11787368956248677, "learning_rate": 1.1521739130434783e-05, "loss": 0.9466, "step": 159 }, { "epoch": 0.07246376811594203, "grad_norm": 0.10616546958822795, "learning_rate": 1.1594202898550726e-05, "loss": 0.9705, "step": 160 }, { "epoch": 0.07291666666666667, "grad_norm": 0.11965833789700073, "learning_rate": 1.1666666666666668e-05, "loss": 0.9406, "step": 161 }, { "epoch": 0.07336956521739131, "grad_norm": 0.1187623915149869, "learning_rate": 1.1739130434782611e-05, "loss": 0.9354, "step": 162 }, { "epoch": 0.07382246376811594, "grad_norm": 0.09116403045095085, "learning_rate": 1.181159420289855e-05, "loss": 0.9563, "step": 163 }, { "epoch": 0.07427536231884058, "grad_norm": 0.1303259037871339, "learning_rate": 1.1884057971014494e-05, "loss": 0.9485, "step": 164 }, { "epoch": 0.07472826086956522, "grad_norm": 0.13682502448666398, "learning_rate": 1.1956521739130435e-05, "loss": 0.9541, "step": 165 }, { "epoch": 0.07518115942028986, "grad_norm": 0.09783014473073308, "learning_rate": 1.2028985507246379e-05, "loss": 0.9133, "step": 166 }, { "epoch": 0.0756340579710145, "grad_norm": 0.13643534735209908, "learning_rate": 1.210144927536232e-05, "loss": 0.9666, "step": 167 }, { "epoch": 0.07608695652173914, "grad_norm": 0.32749928212287377, "learning_rate": 1.2173913043478263e-05, "loss": 0.9537, "step": 168 }, { "epoch": 0.07653985507246377, "grad_norm": 0.10981563555026268, "learning_rate": 1.2246376811594205e-05, "loss": 0.9347, "step": 169 }, { "epoch": 0.0769927536231884, "grad_norm": 0.12027132509647498, "learning_rate": 1.2318840579710144e-05, "loss": 0.9401, "step": 170 }, { "epoch": 0.07744565217391304, "grad_norm": 0.12186155106078453, "learning_rate": 1.2391304347826088e-05, "loss": 0.9753, "step": 171 }, { "epoch": 0.07789855072463768, "grad_norm": 0.11738724341584886, "learning_rate": 1.2463768115942029e-05, "loss": 0.9619, "step": 172 }, { "epoch": 0.07835144927536232, "grad_norm": 0.10813964133580656, "learning_rate": 1.2536231884057972e-05, "loss": 0.9508, "step": 173 }, { "epoch": 0.07880434782608696, "grad_norm": 0.1352227952633904, "learning_rate": 1.2608695652173915e-05, "loss": 0.9729, "step": 174 }, { "epoch": 0.0792572463768116, "grad_norm": 0.10954845559065535, "learning_rate": 1.2681159420289857e-05, "loss": 0.9353, "step": 175 }, { "epoch": 0.07971014492753623, "grad_norm": 0.12841702318076, "learning_rate": 1.2753623188405797e-05, "loss": 0.944, "step": 176 }, { "epoch": 0.08016304347826086, "grad_norm": 0.1088819318890285, "learning_rate": 1.282608695652174e-05, "loss": 0.9327, "step": 177 }, { "epoch": 0.0806159420289855, "grad_norm": 0.13103804393970553, "learning_rate": 1.2898550724637681e-05, "loss": 0.9583, "step": 178 }, { "epoch": 0.08106884057971014, "grad_norm": 0.12923061196542246, "learning_rate": 1.2971014492753624e-05, "loss": 0.9291, "step": 179 }, { "epoch": 0.08152173913043478, "grad_norm": 0.10964148121699822, "learning_rate": 1.3043478260869566e-05, "loss": 0.9495, "step": 180 }, { "epoch": 0.08197463768115942, "grad_norm": 0.11084421716484436, "learning_rate": 1.3115942028985509e-05, "loss": 0.9399, "step": 181 }, { "epoch": 0.08242753623188406, "grad_norm": 0.13711081613785803, "learning_rate": 1.318840579710145e-05, "loss": 0.9433, "step": 182 }, { "epoch": 0.08288043478260869, "grad_norm": 0.1334418275965735, "learning_rate": 1.3260869565217392e-05, "loss": 0.9619, "step": 183 }, { "epoch": 0.08333333333333333, "grad_norm": 0.1155503398313002, "learning_rate": 1.3333333333333333e-05, "loss": 0.9571, "step": 184 }, { "epoch": 0.08378623188405797, "grad_norm": 0.10891774400058392, "learning_rate": 1.3405797101449276e-05, "loss": 0.9572, "step": 185 }, { "epoch": 0.08423913043478261, "grad_norm": 0.10928153497898811, "learning_rate": 1.3478260869565218e-05, "loss": 0.9331, "step": 186 }, { "epoch": 0.08469202898550725, "grad_norm": 0.14074942742349866, "learning_rate": 1.3550724637681161e-05, "loss": 0.9399, "step": 187 }, { "epoch": 0.08514492753623189, "grad_norm": 0.13253115353279682, "learning_rate": 1.3623188405797103e-05, "loss": 0.9492, "step": 188 }, { "epoch": 0.08559782608695653, "grad_norm": 0.10961056084168423, "learning_rate": 1.3695652173913046e-05, "loss": 0.963, "step": 189 }, { "epoch": 0.08605072463768115, "grad_norm": 0.12313425082017569, "learning_rate": 1.3768115942028985e-05, "loss": 0.907, "step": 190 }, { "epoch": 0.08650362318840579, "grad_norm": 0.1360360005616908, "learning_rate": 1.3840579710144929e-05, "loss": 0.9606, "step": 191 }, { "epoch": 0.08695652173913043, "grad_norm": 0.1233286867208591, "learning_rate": 1.391304347826087e-05, "loss": 0.9357, "step": 192 }, { "epoch": 0.08740942028985507, "grad_norm": 0.13298503543057608, "learning_rate": 1.3985507246376813e-05, "loss": 0.9418, "step": 193 }, { "epoch": 0.08786231884057971, "grad_norm": 0.15204343408344176, "learning_rate": 1.4057971014492755e-05, "loss": 0.981, "step": 194 }, { "epoch": 0.08831521739130435, "grad_norm": 0.13740418199246163, "learning_rate": 1.4130434782608698e-05, "loss": 0.9478, "step": 195 }, { "epoch": 0.08876811594202899, "grad_norm": 0.11704360214926617, "learning_rate": 1.420289855072464e-05, "loss": 0.9394, "step": 196 }, { "epoch": 0.08922101449275362, "grad_norm": 0.14429342985702276, "learning_rate": 1.427536231884058e-05, "loss": 0.9552, "step": 197 }, { "epoch": 0.08967391304347826, "grad_norm": 0.14580623051090327, "learning_rate": 1.4347826086956522e-05, "loss": 0.9536, "step": 198 }, { "epoch": 0.0901268115942029, "grad_norm": 0.11923646046203237, "learning_rate": 1.4420289855072465e-05, "loss": 0.952, "step": 199 }, { "epoch": 0.09057971014492754, "grad_norm": 0.13703521407100558, "learning_rate": 1.4492753623188407e-05, "loss": 0.9578, "step": 200 }, { "epoch": 0.09103260869565218, "grad_norm": 0.13886754609811655, "learning_rate": 1.456521739130435e-05, "loss": 0.9356, "step": 201 }, { "epoch": 0.09148550724637682, "grad_norm": 0.14108124406427575, "learning_rate": 1.4637681159420291e-05, "loss": 0.9451, "step": 202 }, { "epoch": 0.09193840579710146, "grad_norm": 0.12769994324971434, "learning_rate": 1.4710144927536235e-05, "loss": 0.9615, "step": 203 }, { "epoch": 0.09239130434782608, "grad_norm": 0.11969930448385514, "learning_rate": 1.4782608695652174e-05, "loss": 0.9404, "step": 204 }, { "epoch": 0.09284420289855072, "grad_norm": 0.12564504652099373, "learning_rate": 1.4855072463768116e-05, "loss": 0.9472, "step": 205 }, { "epoch": 0.09329710144927536, "grad_norm": 0.11052959948181404, "learning_rate": 1.4927536231884059e-05, "loss": 0.9475, "step": 206 }, { "epoch": 0.09375, "grad_norm": 0.15057151982638298, "learning_rate": 1.5000000000000002e-05, "loss": 0.9582, "step": 207 }, { "epoch": 0.09420289855072464, "grad_norm": 0.2012390086095643, "learning_rate": 1.5072463768115944e-05, "loss": 0.9403, "step": 208 }, { "epoch": 0.09465579710144928, "grad_norm": 0.12840527345176922, "learning_rate": 1.5144927536231887e-05, "loss": 0.9474, "step": 209 }, { "epoch": 0.09510869565217392, "grad_norm": 0.1514346618248996, "learning_rate": 1.5217391304347828e-05, "loss": 0.9553, "step": 210 }, { "epoch": 0.09556159420289854, "grad_norm": 0.13134623977125215, "learning_rate": 1.5289855072463768e-05, "loss": 0.9443, "step": 211 }, { "epoch": 0.09601449275362318, "grad_norm": 0.13719753340296761, "learning_rate": 1.536231884057971e-05, "loss": 0.9489, "step": 212 }, { "epoch": 0.09646739130434782, "grad_norm": 0.12773574733790208, "learning_rate": 1.5434782608695654e-05, "loss": 0.9368, "step": 213 }, { "epoch": 0.09692028985507246, "grad_norm": 0.1299685009906771, "learning_rate": 1.5507246376811594e-05, "loss": 0.9449, "step": 214 }, { "epoch": 0.0973731884057971, "grad_norm": 0.12134771811871609, "learning_rate": 1.5579710144927537e-05, "loss": 0.9423, "step": 215 }, { "epoch": 0.09782608695652174, "grad_norm": 0.1255547339305373, "learning_rate": 1.565217391304348e-05, "loss": 0.941, "step": 216 }, { "epoch": 0.09827898550724638, "grad_norm": 0.1628973140469441, "learning_rate": 1.5724637681159423e-05, "loss": 0.9378, "step": 217 }, { "epoch": 0.09873188405797101, "grad_norm": 0.11997014003360325, "learning_rate": 1.5797101449275363e-05, "loss": 0.9607, "step": 218 }, { "epoch": 0.09918478260869565, "grad_norm": 0.12618948752827347, "learning_rate": 1.5869565217391306e-05, "loss": 0.942, "step": 219 }, { "epoch": 0.09963768115942029, "grad_norm": 0.10322218177861399, "learning_rate": 1.5942028985507246e-05, "loss": 0.9551, "step": 220 }, { "epoch": 0.10009057971014493, "grad_norm": 0.12237278974684077, "learning_rate": 1.601449275362319e-05, "loss": 0.954, "step": 221 }, { "epoch": 0.10054347826086957, "grad_norm": 0.1315650247012288, "learning_rate": 1.6086956521739132e-05, "loss": 0.9225, "step": 222 }, { "epoch": 0.10099637681159421, "grad_norm": 0.1380189482851502, "learning_rate": 1.6159420289855076e-05, "loss": 0.9565, "step": 223 }, { "epoch": 0.10144927536231885, "grad_norm": 0.12329922987482161, "learning_rate": 1.6231884057971015e-05, "loss": 0.9284, "step": 224 }, { "epoch": 0.10190217391304347, "grad_norm": 0.13710299683804547, "learning_rate": 1.630434782608696e-05, "loss": 0.9674, "step": 225 }, { "epoch": 0.10235507246376811, "grad_norm": 0.1336060551558503, "learning_rate": 1.6376811594202898e-05, "loss": 0.9356, "step": 226 }, { "epoch": 0.10280797101449275, "grad_norm": 0.15701330125321702, "learning_rate": 1.644927536231884e-05, "loss": 0.9583, "step": 227 }, { "epoch": 0.10326086956521739, "grad_norm": 0.12989026736154238, "learning_rate": 1.6521739130434785e-05, "loss": 0.9218, "step": 228 }, { "epoch": 0.10371376811594203, "grad_norm": 0.14851778758212586, "learning_rate": 1.6594202898550728e-05, "loss": 0.9388, "step": 229 }, { "epoch": 0.10416666666666667, "grad_norm": 0.17861417736210244, "learning_rate": 1.6666666666666667e-05, "loss": 0.9462, "step": 230 }, { "epoch": 0.10461956521739131, "grad_norm": 0.13841474180106383, "learning_rate": 1.673913043478261e-05, "loss": 0.9405, "step": 231 }, { "epoch": 0.10507246376811594, "grad_norm": 0.16636930905166672, "learning_rate": 1.681159420289855e-05, "loss": 0.9606, "step": 232 }, { "epoch": 0.10552536231884058, "grad_norm": 0.15640330308226594, "learning_rate": 1.6884057971014494e-05, "loss": 0.974, "step": 233 }, { "epoch": 0.10597826086956522, "grad_norm": 0.11551879727010676, "learning_rate": 1.6956521739130437e-05, "loss": 0.9295, "step": 234 }, { "epoch": 0.10643115942028986, "grad_norm": 0.17172010782821376, "learning_rate": 1.702898550724638e-05, "loss": 0.9479, "step": 235 }, { "epoch": 0.1068840579710145, "grad_norm": 0.12934455461559896, "learning_rate": 1.710144927536232e-05, "loss": 0.9423, "step": 236 }, { "epoch": 0.10733695652173914, "grad_norm": 0.12305328605192535, "learning_rate": 1.7173913043478263e-05, "loss": 0.9414, "step": 237 }, { "epoch": 0.10778985507246377, "grad_norm": 0.14703922748283618, "learning_rate": 1.7246376811594206e-05, "loss": 0.9746, "step": 238 }, { "epoch": 0.1082427536231884, "grad_norm": 0.1324966973069392, "learning_rate": 1.7318840579710146e-05, "loss": 0.962, "step": 239 }, { "epoch": 0.10869565217391304, "grad_norm": 0.12459083192401356, "learning_rate": 1.739130434782609e-05, "loss": 0.9747, "step": 240 }, { "epoch": 0.10914855072463768, "grad_norm": 0.12471917691474652, "learning_rate": 1.746376811594203e-05, "loss": 0.9512, "step": 241 }, { "epoch": 0.10960144927536232, "grad_norm": 0.1966645980435686, "learning_rate": 1.7536231884057972e-05, "loss": 0.9505, "step": 242 }, { "epoch": 0.11005434782608696, "grad_norm": 0.13713052679470125, "learning_rate": 1.7608695652173915e-05, "loss": 0.9447, "step": 243 }, { "epoch": 0.1105072463768116, "grad_norm": 0.12520103408963573, "learning_rate": 1.7681159420289858e-05, "loss": 0.9681, "step": 244 }, { "epoch": 0.11096014492753623, "grad_norm": 0.18456657291003253, "learning_rate": 1.7753623188405798e-05, "loss": 0.9387, "step": 245 }, { "epoch": 0.11141304347826086, "grad_norm": 0.14256083024115104, "learning_rate": 1.782608695652174e-05, "loss": 0.9366, "step": 246 }, { "epoch": 0.1118659420289855, "grad_norm": 0.17471580692887037, "learning_rate": 1.789855072463768e-05, "loss": 0.9408, "step": 247 }, { "epoch": 0.11231884057971014, "grad_norm": 0.13549572074480482, "learning_rate": 1.7971014492753624e-05, "loss": 0.9436, "step": 248 }, { "epoch": 0.11277173913043478, "grad_norm": 0.16201064073903787, "learning_rate": 1.8043478260869567e-05, "loss": 0.9431, "step": 249 }, { "epoch": 0.11322463768115942, "grad_norm": 0.15630947059023456, "learning_rate": 1.811594202898551e-05, "loss": 0.9628, "step": 250 }, { "epoch": 0.11367753623188406, "grad_norm": 0.14543911893451755, "learning_rate": 1.818840579710145e-05, "loss": 0.9518, "step": 251 }, { "epoch": 0.11413043478260869, "grad_norm": 0.18369139751137772, "learning_rate": 1.8260869565217393e-05, "loss": 0.945, "step": 252 }, { "epoch": 0.11458333333333333, "grad_norm": 0.14328053827043316, "learning_rate": 1.8333333333333333e-05, "loss": 0.9323, "step": 253 }, { "epoch": 0.11503623188405797, "grad_norm": 0.16492261684391443, "learning_rate": 1.8405797101449276e-05, "loss": 0.927, "step": 254 }, { "epoch": 0.11548913043478261, "grad_norm": 0.14259212621987552, "learning_rate": 1.847826086956522e-05, "loss": 0.9424, "step": 255 }, { "epoch": 0.11594202898550725, "grad_norm": 0.1501501846355763, "learning_rate": 1.8550724637681162e-05, "loss": 0.9463, "step": 256 }, { "epoch": 0.11639492753623189, "grad_norm": 0.16500532796975442, "learning_rate": 1.8623188405797102e-05, "loss": 0.9465, "step": 257 }, { "epoch": 0.11684782608695653, "grad_norm": 0.17806326248653956, "learning_rate": 1.8695652173913045e-05, "loss": 0.9339, "step": 258 }, { "epoch": 0.11730072463768115, "grad_norm": 0.15592342804692685, "learning_rate": 1.8768115942028985e-05, "loss": 0.9448, "step": 259 }, { "epoch": 0.11775362318840579, "grad_norm": 0.16185899822891098, "learning_rate": 1.8840579710144928e-05, "loss": 0.9425, "step": 260 }, { "epoch": 0.11820652173913043, "grad_norm": 0.1881976749604095, "learning_rate": 1.891304347826087e-05, "loss": 0.9631, "step": 261 }, { "epoch": 0.11865942028985507, "grad_norm": 0.12816320780379226, "learning_rate": 1.8985507246376814e-05, "loss": 0.9115, "step": 262 }, { "epoch": 0.11911231884057971, "grad_norm": 0.19070475115509514, "learning_rate": 1.9057971014492754e-05, "loss": 0.9459, "step": 263 }, { "epoch": 0.11956521739130435, "grad_norm": 0.17526081852353573, "learning_rate": 1.9130434782608697e-05, "loss": 0.9353, "step": 264 }, { "epoch": 0.12001811594202899, "grad_norm": 0.16939456146775636, "learning_rate": 1.920289855072464e-05, "loss": 0.9565, "step": 265 }, { "epoch": 0.12047101449275362, "grad_norm": 0.13256131399971324, "learning_rate": 1.927536231884058e-05, "loss": 0.943, "step": 266 }, { "epoch": 0.12092391304347826, "grad_norm": 0.16324401142605618, "learning_rate": 1.9347826086956523e-05, "loss": 0.9583, "step": 267 }, { "epoch": 0.1213768115942029, "grad_norm": 0.1486387525293964, "learning_rate": 1.9420289855072467e-05, "loss": 0.9213, "step": 268 }, { "epoch": 0.12182971014492754, "grad_norm": 0.2220926605427484, "learning_rate": 1.9492753623188406e-05, "loss": 0.9362, "step": 269 }, { "epoch": 0.12228260869565218, "grad_norm": 0.2414231461114835, "learning_rate": 1.956521739130435e-05, "loss": 0.9453, "step": 270 }, { "epoch": 0.12273550724637682, "grad_norm": 0.21876103686538814, "learning_rate": 1.9637681159420293e-05, "loss": 0.9469, "step": 271 }, { "epoch": 0.12318840579710146, "grad_norm": 0.16793391640858965, "learning_rate": 1.9710144927536236e-05, "loss": 0.9353, "step": 272 }, { "epoch": 0.12364130434782608, "grad_norm": 0.14242528225561357, "learning_rate": 1.9782608695652176e-05, "loss": 0.9323, "step": 273 }, { "epoch": 0.12409420289855072, "grad_norm": 0.22713238110927972, "learning_rate": 1.9855072463768115e-05, "loss": 0.9206, "step": 274 }, { "epoch": 0.12454710144927536, "grad_norm": 0.2739268706696986, "learning_rate": 1.992753623188406e-05, "loss": 0.9305, "step": 275 }, { "epoch": 0.125, "grad_norm": 0.2708238799523322, "learning_rate": 2e-05, "loss": 0.9215, "step": 276 }, { "epoch": 0.12545289855072464, "grad_norm": 0.16930621012544828, "learning_rate": 2.0072463768115945e-05, "loss": 0.9343, "step": 277 }, { "epoch": 0.12590579710144928, "grad_norm": 0.1630401388006593, "learning_rate": 2.0144927536231888e-05, "loss": 0.9241, "step": 278 }, { "epoch": 0.12635869565217392, "grad_norm": 0.26737936574934806, "learning_rate": 2.0217391304347828e-05, "loss": 0.9345, "step": 279 }, { "epoch": 0.12681159420289856, "grad_norm": 0.2890589763962627, "learning_rate": 2.028985507246377e-05, "loss": 0.94, "step": 280 }, { "epoch": 0.1272644927536232, "grad_norm": 0.2296574670634346, "learning_rate": 2.0362318840579714e-05, "loss": 0.9592, "step": 281 }, { "epoch": 0.12771739130434784, "grad_norm": 0.1504853775143383, "learning_rate": 2.0434782608695657e-05, "loss": 0.9236, "step": 282 }, { "epoch": 0.12817028985507245, "grad_norm": 0.1681667627412393, "learning_rate": 2.0507246376811594e-05, "loss": 0.9174, "step": 283 }, { "epoch": 0.1286231884057971, "grad_norm": 0.25267320219832423, "learning_rate": 2.0579710144927537e-05, "loss": 0.9217, "step": 284 }, { "epoch": 0.12907608695652173, "grad_norm": 0.2602241171576273, "learning_rate": 2.065217391304348e-05, "loss": 0.958, "step": 285 }, { "epoch": 0.12952898550724637, "grad_norm": 0.23784523010327532, "learning_rate": 2.072463768115942e-05, "loss": 0.9572, "step": 286 }, { "epoch": 0.129981884057971, "grad_norm": 0.2542782398655837, "learning_rate": 2.0797101449275363e-05, "loss": 0.924, "step": 287 }, { "epoch": 0.13043478260869565, "grad_norm": 0.169878338110069, "learning_rate": 2.0869565217391306e-05, "loss": 0.9286, "step": 288 }, { "epoch": 0.1308876811594203, "grad_norm": 0.2891079169847563, "learning_rate": 2.094202898550725e-05, "loss": 0.9282, "step": 289 }, { "epoch": 0.13134057971014493, "grad_norm": 0.22490821936295402, "learning_rate": 2.101449275362319e-05, "loss": 0.922, "step": 290 }, { "epoch": 0.13179347826086957, "grad_norm": 0.28254316121003986, "learning_rate": 2.1086956521739132e-05, "loss": 0.9493, "step": 291 }, { "epoch": 0.1322463768115942, "grad_norm": 0.2505549337359319, "learning_rate": 2.1159420289855075e-05, "loss": 0.9456, "step": 292 }, { "epoch": 0.13269927536231885, "grad_norm": 0.2972998595613215, "learning_rate": 2.1231884057971018e-05, "loss": 0.9291, "step": 293 }, { "epoch": 0.1331521739130435, "grad_norm": 0.21029112932740032, "learning_rate": 2.1304347826086958e-05, "loss": 0.9299, "step": 294 }, { "epoch": 0.13360507246376813, "grad_norm": 0.23696796616848453, "learning_rate": 2.13768115942029e-05, "loss": 0.9375, "step": 295 }, { "epoch": 0.13405797101449277, "grad_norm": 0.2159961484637905, "learning_rate": 2.1449275362318844e-05, "loss": 0.9291, "step": 296 }, { "epoch": 0.13451086956521738, "grad_norm": 0.20172414905578395, "learning_rate": 2.152173913043478e-05, "loss": 0.9283, "step": 297 }, { "epoch": 0.13496376811594202, "grad_norm": 0.15177359667882742, "learning_rate": 2.1594202898550724e-05, "loss": 0.9374, "step": 298 }, { "epoch": 0.13541666666666666, "grad_norm": 0.1811786917363714, "learning_rate": 2.1666666666666667e-05, "loss": 0.9544, "step": 299 }, { "epoch": 0.1358695652173913, "grad_norm": 0.17392727884005757, "learning_rate": 2.173913043478261e-05, "loss": 0.9314, "step": 300 }, { "epoch": 0.13632246376811594, "grad_norm": 0.16037469164406917, "learning_rate": 2.1811594202898553e-05, "loss": 0.9381, "step": 301 }, { "epoch": 0.13677536231884058, "grad_norm": 0.21130317422965653, "learning_rate": 2.1884057971014493e-05, "loss": 0.9417, "step": 302 }, { "epoch": 0.13722826086956522, "grad_norm": 0.2591231864018542, "learning_rate": 2.1956521739130436e-05, "loss": 0.9546, "step": 303 }, { "epoch": 0.13768115942028986, "grad_norm": 0.24643097853834386, "learning_rate": 2.202898550724638e-05, "loss": 0.9313, "step": 304 }, { "epoch": 0.1381340579710145, "grad_norm": 0.18353473544663343, "learning_rate": 2.2101449275362323e-05, "loss": 0.9243, "step": 305 }, { "epoch": 0.13858695652173914, "grad_norm": 0.21159129894845677, "learning_rate": 2.2173913043478262e-05, "loss": 0.9273, "step": 306 }, { "epoch": 0.13903985507246377, "grad_norm": 0.18024438387417543, "learning_rate": 2.2246376811594205e-05, "loss": 0.9222, "step": 307 }, { "epoch": 0.13949275362318841, "grad_norm": 0.23613214410041858, "learning_rate": 2.231884057971015e-05, "loss": 0.9484, "step": 308 }, { "epoch": 0.13994565217391305, "grad_norm": 0.19726987360081902, "learning_rate": 2.2391304347826092e-05, "loss": 0.9492, "step": 309 }, { "epoch": 0.1403985507246377, "grad_norm": 0.23430139015738868, "learning_rate": 2.246376811594203e-05, "loss": 0.9571, "step": 310 }, { "epoch": 0.1408514492753623, "grad_norm": 0.23542669350235668, "learning_rate": 2.253623188405797e-05, "loss": 0.9298, "step": 311 }, { "epoch": 0.14130434782608695, "grad_norm": 0.22295381402209047, "learning_rate": 2.2608695652173914e-05, "loss": 0.9179, "step": 312 }, { "epoch": 0.14175724637681159, "grad_norm": 0.16423585978946656, "learning_rate": 2.2681159420289854e-05, "loss": 0.932, "step": 313 }, { "epoch": 0.14221014492753623, "grad_norm": 0.19162461355638544, "learning_rate": 2.2753623188405797e-05, "loss": 0.933, "step": 314 }, { "epoch": 0.14266304347826086, "grad_norm": 0.19115460371597134, "learning_rate": 2.282608695652174e-05, "loss": 0.9235, "step": 315 }, { "epoch": 0.1431159420289855, "grad_norm": 0.22087040957171758, "learning_rate": 2.2898550724637684e-05, "loss": 0.9432, "step": 316 }, { "epoch": 0.14356884057971014, "grad_norm": 0.24108572538436587, "learning_rate": 2.2971014492753623e-05, "loss": 0.9222, "step": 317 }, { "epoch": 0.14402173913043478, "grad_norm": 0.3312526895866374, "learning_rate": 2.3043478260869567e-05, "loss": 0.9523, "step": 318 }, { "epoch": 0.14447463768115942, "grad_norm": 0.48454147945940085, "learning_rate": 2.311594202898551e-05, "loss": 0.9328, "step": 319 }, { "epoch": 0.14492753623188406, "grad_norm": 0.6856577600979537, "learning_rate": 2.3188405797101453e-05, "loss": 0.937, "step": 320 }, { "epoch": 0.1453804347826087, "grad_norm": 0.9078308309898353, "learning_rate": 2.3260869565217396e-05, "loss": 0.9316, "step": 321 }, { "epoch": 0.14583333333333334, "grad_norm": 0.9650938630474529, "learning_rate": 2.3333333333333336e-05, "loss": 0.9371, "step": 322 }, { "epoch": 0.14628623188405798, "grad_norm": 0.6793558774036353, "learning_rate": 2.340579710144928e-05, "loss": 0.9338, "step": 323 }, { "epoch": 0.14673913043478262, "grad_norm": 0.23087393915953844, "learning_rate": 2.3478260869565222e-05, "loss": 0.9445, "step": 324 }, { "epoch": 0.14719202898550723, "grad_norm": 0.3965126335472709, "learning_rate": 2.355072463768116e-05, "loss": 0.9548, "step": 325 }, { "epoch": 0.14764492753623187, "grad_norm": 0.6847245165033827, "learning_rate": 2.36231884057971e-05, "loss": 0.9339, "step": 326 }, { "epoch": 0.1480978260869565, "grad_norm": 0.72343123275752, "learning_rate": 2.3695652173913045e-05, "loss": 0.943, "step": 327 }, { "epoch": 0.14855072463768115, "grad_norm": 0.4275629563236616, "learning_rate": 2.3768115942028988e-05, "loss": 0.9304, "step": 328 }, { "epoch": 0.1490036231884058, "grad_norm": 0.1722013163029811, "learning_rate": 2.3840579710144928e-05, "loss": 0.9427, "step": 329 }, { "epoch": 0.14945652173913043, "grad_norm": 0.38660707192928884, "learning_rate": 2.391304347826087e-05, "loss": 0.9425, "step": 330 }, { "epoch": 0.14990942028985507, "grad_norm": 0.596008977310972, "learning_rate": 2.3985507246376814e-05, "loss": 0.9206, "step": 331 }, { "epoch": 0.1503623188405797, "grad_norm": 0.6536321805902559, "learning_rate": 2.4057971014492757e-05, "loss": 0.9538, "step": 332 }, { "epoch": 0.15081521739130435, "grad_norm": 0.3120616487029901, "learning_rate": 2.4130434782608697e-05, "loss": 0.9151, "step": 333 }, { "epoch": 0.151268115942029, "grad_norm": 0.2726806991581015, "learning_rate": 2.420289855072464e-05, "loss": 0.9277, "step": 334 }, { "epoch": 0.15172101449275363, "grad_norm": 0.5051506838902795, "learning_rate": 2.4275362318840583e-05, "loss": 0.9383, "step": 335 }, { "epoch": 0.15217391304347827, "grad_norm": 0.6409773741604906, "learning_rate": 2.4347826086956526e-05, "loss": 0.9387, "step": 336 }, { "epoch": 0.1526268115942029, "grad_norm": 0.5646434490146108, "learning_rate": 2.4420289855072466e-05, "loss": 0.9334, "step": 337 }, { "epoch": 0.15307971014492755, "grad_norm": 0.24537762694360465, "learning_rate": 2.449275362318841e-05, "loss": 0.93, "step": 338 }, { "epoch": 0.15353260869565216, "grad_norm": 0.27498753581353697, "learning_rate": 2.456521739130435e-05, "loss": 0.9433, "step": 339 }, { "epoch": 0.1539855072463768, "grad_norm": 0.5429110797227636, "learning_rate": 2.463768115942029e-05, "loss": 0.9354, "step": 340 }, { "epoch": 0.15443840579710144, "grad_norm": 0.5885038253315437, "learning_rate": 2.4710144927536232e-05, "loss": 0.9336, "step": 341 }, { "epoch": 0.15489130434782608, "grad_norm": 0.40852883709368504, "learning_rate": 2.4782608695652175e-05, "loss": 0.9575, "step": 342 }, { "epoch": 0.15534420289855072, "grad_norm": 0.15324672774503187, "learning_rate": 2.4855072463768118e-05, "loss": 0.9337, "step": 343 }, { "epoch": 0.15579710144927536, "grad_norm": 0.3611907730692182, "learning_rate": 2.4927536231884058e-05, "loss": 0.9415, "step": 344 }, { "epoch": 0.15625, "grad_norm": 0.45433143627935174, "learning_rate": 2.5e-05, "loss": 0.9377, "step": 345 }, { "epoch": 0.15670289855072464, "grad_norm": 0.39459930767357143, "learning_rate": 2.5072463768115944e-05, "loss": 0.9213, "step": 346 }, { "epoch": 0.15715579710144928, "grad_norm": 0.29290719723995823, "learning_rate": 2.5144927536231888e-05, "loss": 0.9342, "step": 347 }, { "epoch": 0.15760869565217392, "grad_norm": 0.16164087646143188, "learning_rate": 2.521739130434783e-05, "loss": 0.9283, "step": 348 }, { "epoch": 0.15806159420289856, "grad_norm": 0.4017109861204689, "learning_rate": 2.528985507246377e-05, "loss": 0.9419, "step": 349 }, { "epoch": 0.1585144927536232, "grad_norm": 0.40726991325880235, "learning_rate": 2.5362318840579714e-05, "loss": 0.9243, "step": 350 }, { "epoch": 0.15896739130434784, "grad_norm": 0.3110331252086983, "learning_rate": 2.5434782608695657e-05, "loss": 0.9415, "step": 351 }, { "epoch": 0.15942028985507245, "grad_norm": 0.1641008898652895, "learning_rate": 2.5507246376811593e-05, "loss": 0.9172, "step": 352 }, { "epoch": 0.1598731884057971, "grad_norm": 0.2571847933470991, "learning_rate": 2.5579710144927536e-05, "loss": 0.9309, "step": 353 }, { "epoch": 0.16032608695652173, "grad_norm": 0.34008675212420114, "learning_rate": 2.565217391304348e-05, "loss": 0.9438, "step": 354 }, { "epoch": 0.16077898550724637, "grad_norm": 0.38110595322445023, "learning_rate": 2.5724637681159423e-05, "loss": 0.9211, "step": 355 }, { "epoch": 0.161231884057971, "grad_norm": 0.28160623141332625, "learning_rate": 2.5797101449275362e-05, "loss": 0.9341, "step": 356 }, { "epoch": 0.16168478260869565, "grad_norm": 0.19064337003407902, "learning_rate": 2.5869565217391305e-05, "loss": 0.9073, "step": 357 }, { "epoch": 0.1621376811594203, "grad_norm": 0.207144770527691, "learning_rate": 2.594202898550725e-05, "loss": 0.9322, "step": 358 }, { "epoch": 0.16259057971014493, "grad_norm": 0.23480297946931467, "learning_rate": 2.6014492753623192e-05, "loss": 0.9287, "step": 359 }, { "epoch": 0.16304347826086957, "grad_norm": 0.30351838946853393, "learning_rate": 2.608695652173913e-05, "loss": 0.9553, "step": 360 }, { "epoch": 0.1634963768115942, "grad_norm": 0.31537760196375225, "learning_rate": 2.6159420289855075e-05, "loss": 0.9353, "step": 361 }, { "epoch": 0.16394927536231885, "grad_norm": 0.21901924770556577, "learning_rate": 2.6231884057971018e-05, "loss": 0.9458, "step": 362 }, { "epoch": 0.1644021739130435, "grad_norm": 0.22599262127623565, "learning_rate": 2.630434782608696e-05, "loss": 0.9252, "step": 363 }, { "epoch": 0.16485507246376813, "grad_norm": 0.16389384001091545, "learning_rate": 2.63768115942029e-05, "loss": 0.9261, "step": 364 }, { "epoch": 0.16530797101449277, "grad_norm": 0.24799491421091613, "learning_rate": 2.6449275362318844e-05, "loss": 0.9093, "step": 365 }, { "epoch": 0.16576086956521738, "grad_norm": 0.3123280486717328, "learning_rate": 2.6521739130434784e-05, "loss": 0.9327, "step": 366 }, { "epoch": 0.16621376811594202, "grad_norm": 0.40439740404859764, "learning_rate": 2.6594202898550723e-05, "loss": 0.9603, "step": 367 }, { "epoch": 0.16666666666666666, "grad_norm": 0.43412840080654125, "learning_rate": 2.6666666666666667e-05, "loss": 0.9446, "step": 368 }, { "epoch": 0.1671195652173913, "grad_norm": 0.37894072102349596, "learning_rate": 2.673913043478261e-05, "loss": 0.9366, "step": 369 }, { "epoch": 0.16757246376811594, "grad_norm": 0.2130940163038555, "learning_rate": 2.6811594202898553e-05, "loss": 0.9395, "step": 370 }, { "epoch": 0.16802536231884058, "grad_norm": 0.2567552770772861, "learning_rate": 2.6884057971014496e-05, "loss": 0.9464, "step": 371 }, { "epoch": 0.16847826086956522, "grad_norm": 0.2713028600308338, "learning_rate": 2.6956521739130436e-05, "loss": 0.9399, "step": 372 }, { "epoch": 0.16893115942028986, "grad_norm": 0.3357514820058956, "learning_rate": 2.702898550724638e-05, "loss": 0.9203, "step": 373 }, { "epoch": 0.1693840579710145, "grad_norm": 0.37164162220829045, "learning_rate": 2.7101449275362322e-05, "loss": 0.9461, "step": 374 }, { "epoch": 0.16983695652173914, "grad_norm": 0.2576853952398292, "learning_rate": 2.7173913043478265e-05, "loss": 0.936, "step": 375 }, { "epoch": 0.17028985507246377, "grad_norm": 0.34326412981810595, "learning_rate": 2.7246376811594205e-05, "loss": 0.9553, "step": 376 }, { "epoch": 0.17074275362318841, "grad_norm": 0.28822235322430106, "learning_rate": 2.7318840579710148e-05, "loss": 0.966, "step": 377 }, { "epoch": 0.17119565217391305, "grad_norm": 0.26348850458984396, "learning_rate": 2.739130434782609e-05, "loss": 0.9263, "step": 378 }, { "epoch": 0.1716485507246377, "grad_norm": 0.2970789774651534, "learning_rate": 2.7463768115942034e-05, "loss": 0.9235, "step": 379 }, { "epoch": 0.1721014492753623, "grad_norm": 0.303444755821078, "learning_rate": 2.753623188405797e-05, "loss": 0.9382, "step": 380 }, { "epoch": 0.17255434782608695, "grad_norm": 0.27192637742825665, "learning_rate": 2.7608695652173914e-05, "loss": 0.9298, "step": 381 }, { "epoch": 0.17300724637681159, "grad_norm": 0.334593614467794, "learning_rate": 2.7681159420289857e-05, "loss": 0.9599, "step": 382 }, { "epoch": 0.17346014492753623, "grad_norm": 0.3243143468243264, "learning_rate": 2.7753623188405797e-05, "loss": 0.9397, "step": 383 }, { "epoch": 0.17391304347826086, "grad_norm": 0.2906443337752772, "learning_rate": 2.782608695652174e-05, "loss": 0.9298, "step": 384 }, { "epoch": 0.1743659420289855, "grad_norm": 0.4057267493145982, "learning_rate": 2.7898550724637683e-05, "loss": 0.9421, "step": 385 }, { "epoch": 0.17481884057971014, "grad_norm": 0.3525511773392057, "learning_rate": 2.7971014492753626e-05, "loss": 0.9516, "step": 386 }, { "epoch": 0.17527173913043478, "grad_norm": 0.252685106557326, "learning_rate": 2.8043478260869566e-05, "loss": 0.9339, "step": 387 }, { "epoch": 0.17572463768115942, "grad_norm": 0.23751306241357154, "learning_rate": 2.811594202898551e-05, "loss": 0.9402, "step": 388 }, { "epoch": 0.17617753623188406, "grad_norm": 0.22514057522155564, "learning_rate": 2.8188405797101452e-05, "loss": 0.9377, "step": 389 }, { "epoch": 0.1766304347826087, "grad_norm": 0.2933348532579443, "learning_rate": 2.8260869565217396e-05, "loss": 0.9413, "step": 390 }, { "epoch": 0.17708333333333334, "grad_norm": 0.39709138188018595, "learning_rate": 2.833333333333334e-05, "loss": 0.9405, "step": 391 }, { "epoch": 0.17753623188405798, "grad_norm": 0.5359551205726306, "learning_rate": 2.840579710144928e-05, "loss": 0.9213, "step": 392 }, { "epoch": 0.17798913043478262, "grad_norm": 0.6500341802060967, "learning_rate": 2.847826086956522e-05, "loss": 0.9309, "step": 393 }, { "epoch": 0.17844202898550723, "grad_norm": 0.7929825228551707, "learning_rate": 2.855072463768116e-05, "loss": 0.9482, "step": 394 }, { "epoch": 0.17889492753623187, "grad_norm": 0.8834289798838847, "learning_rate": 2.86231884057971e-05, "loss": 0.9452, "step": 395 }, { "epoch": 0.1793478260869565, "grad_norm": 0.9358858290894887, "learning_rate": 2.8695652173913044e-05, "loss": 0.9443, "step": 396 }, { "epoch": 0.17980072463768115, "grad_norm": 0.7927994841699783, "learning_rate": 2.8768115942028988e-05, "loss": 0.9525, "step": 397 }, { "epoch": 0.1802536231884058, "grad_norm": 0.49413455704480347, "learning_rate": 2.884057971014493e-05, "loss": 0.919, "step": 398 }, { "epoch": 0.18070652173913043, "grad_norm": 0.26506346667587505, "learning_rate": 2.891304347826087e-05, "loss": 0.9279, "step": 399 }, { "epoch": 0.18115942028985507, "grad_norm": 0.29170088076454126, "learning_rate": 2.8985507246376814e-05, "loss": 0.9307, "step": 400 }, { "epoch": 0.1816123188405797, "grad_norm": 0.4823705422943487, "learning_rate": 2.9057971014492757e-05, "loss": 0.9465, "step": 401 }, { "epoch": 0.18206521739130435, "grad_norm": 0.5073446641676506, "learning_rate": 2.91304347826087e-05, "loss": 0.9295, "step": 402 }, { "epoch": 0.182518115942029, "grad_norm": 0.4223742359640241, "learning_rate": 2.920289855072464e-05, "loss": 0.9206, "step": 403 }, { "epoch": 0.18297101449275363, "grad_norm": 0.32587059382293065, "learning_rate": 2.9275362318840583e-05, "loss": 0.9331, "step": 404 }, { "epoch": 0.18342391304347827, "grad_norm": 0.3387103527798451, "learning_rate": 2.9347826086956526e-05, "loss": 0.9335, "step": 405 }, { "epoch": 0.1838768115942029, "grad_norm": 0.30445539087583334, "learning_rate": 2.942028985507247e-05, "loss": 0.9263, "step": 406 }, { "epoch": 0.18432971014492755, "grad_norm": 0.3617318800269031, "learning_rate": 2.949275362318841e-05, "loss": 0.9346, "step": 407 }, { "epoch": 0.18478260869565216, "grad_norm": 0.5668425797305667, "learning_rate": 2.956521739130435e-05, "loss": 0.9163, "step": 408 }, { "epoch": 0.1852355072463768, "grad_norm": 0.6753986893104724, "learning_rate": 2.9637681159420292e-05, "loss": 0.9451, "step": 409 }, { "epoch": 0.18568840579710144, "grad_norm": 0.6935870972756306, "learning_rate": 2.971014492753623e-05, "loss": 0.9264, "step": 410 }, { "epoch": 0.18614130434782608, "grad_norm": 0.6928137405161524, "learning_rate": 2.9782608695652175e-05, "loss": 0.9532, "step": 411 }, { "epoch": 0.18659420289855072, "grad_norm": 0.6337304789806183, "learning_rate": 2.9855072463768118e-05, "loss": 0.9105, "step": 412 }, { "epoch": 0.18704710144927536, "grad_norm": 0.4465188311574424, "learning_rate": 2.992753623188406e-05, "loss": 0.9305, "step": 413 }, { "epoch": 0.1875, "grad_norm": 0.3821778844935118, "learning_rate": 3.0000000000000004e-05, "loss": 0.9389, "step": 414 }, { "epoch": 0.18795289855072464, "grad_norm": 0.2350984459811697, "learning_rate": 3.0072463768115944e-05, "loss": 0.9219, "step": 415 }, { "epoch": 0.18840579710144928, "grad_norm": 0.36603156654804925, "learning_rate": 3.0144927536231887e-05, "loss": 0.9247, "step": 416 }, { "epoch": 0.18885869565217392, "grad_norm": 0.5482773419055748, "learning_rate": 3.021739130434783e-05, "loss": 0.9486, "step": 417 }, { "epoch": 0.18931159420289856, "grad_norm": 0.6735331495890149, "learning_rate": 3.0289855072463773e-05, "loss": 0.9277, "step": 418 }, { "epoch": 0.1897644927536232, "grad_norm": 0.667741918648474, "learning_rate": 3.0362318840579713e-05, "loss": 0.9195, "step": 419 }, { "epoch": 0.19021739130434784, "grad_norm": 0.43853960279601206, "learning_rate": 3.0434782608695656e-05, "loss": 0.9207, "step": 420 }, { "epoch": 0.19067028985507245, "grad_norm": 0.27462552995134315, "learning_rate": 3.0507246376811596e-05, "loss": 0.9295, "step": 421 }, { "epoch": 0.1911231884057971, "grad_norm": 0.326739947815201, "learning_rate": 3.0579710144927536e-05, "loss": 0.9059, "step": 422 }, { "epoch": 0.19157608695652173, "grad_norm": 0.39785678252738393, "learning_rate": 3.065217391304348e-05, "loss": 0.9269, "step": 423 }, { "epoch": 0.19202898550724637, "grad_norm": 0.4259842646513344, "learning_rate": 3.072463768115942e-05, "loss": 0.9356, "step": 424 }, { "epoch": 0.192481884057971, "grad_norm": 0.5276969905504205, "learning_rate": 3.079710144927536e-05, "loss": 0.9471, "step": 425 }, { "epoch": 0.19293478260869565, "grad_norm": 0.7199846110195925, "learning_rate": 3.086956521739131e-05, "loss": 0.9292, "step": 426 }, { "epoch": 0.1933876811594203, "grad_norm": 0.6843496181771409, "learning_rate": 3.094202898550725e-05, "loss": 0.9502, "step": 427 }, { "epoch": 0.19384057971014493, "grad_norm": 0.6574641170181403, "learning_rate": 3.101449275362319e-05, "loss": 0.9378, "step": 428 }, { "epoch": 0.19429347826086957, "grad_norm": 0.7374183549364719, "learning_rate": 3.1086956521739134e-05, "loss": 0.9327, "step": 429 }, { "epoch": 0.1947463768115942, "grad_norm": 0.8512072712225506, "learning_rate": 3.1159420289855074e-05, "loss": 0.9402, "step": 430 }, { "epoch": 0.19519927536231885, "grad_norm": 0.7293482517710056, "learning_rate": 3.123188405797102e-05, "loss": 0.9436, "step": 431 }, { "epoch": 0.1956521739130435, "grad_norm": 0.5365294191631025, "learning_rate": 3.130434782608696e-05, "loss": 0.9311, "step": 432 }, { "epoch": 0.19610507246376813, "grad_norm": 0.34772675796098756, "learning_rate": 3.13768115942029e-05, "loss": 0.9286, "step": 433 }, { "epoch": 0.19655797101449277, "grad_norm": 0.41749702502972796, "learning_rate": 3.144927536231885e-05, "loss": 0.9252, "step": 434 }, { "epoch": 0.19701086956521738, "grad_norm": 0.3119411836846022, "learning_rate": 3.152173913043478e-05, "loss": 0.9212, "step": 435 }, { "epoch": 0.19746376811594202, "grad_norm": 0.4273539473714838, "learning_rate": 3.1594202898550726e-05, "loss": 0.905, "step": 436 }, { "epoch": 0.19791666666666666, "grad_norm": 0.32682979102057497, "learning_rate": 3.1666666666666666e-05, "loss": 0.9258, "step": 437 }, { "epoch": 0.1983695652173913, "grad_norm": 0.38992718043999375, "learning_rate": 3.173913043478261e-05, "loss": 0.9228, "step": 438 }, { "epoch": 0.19882246376811594, "grad_norm": 0.49625927931442654, "learning_rate": 3.181159420289855e-05, "loss": 0.9222, "step": 439 }, { "epoch": 0.19927536231884058, "grad_norm": 0.5310694003035054, "learning_rate": 3.188405797101449e-05, "loss": 0.9438, "step": 440 }, { "epoch": 0.19972826086956522, "grad_norm": 0.4790470455533784, "learning_rate": 3.195652173913044e-05, "loss": 0.9445, "step": 441 }, { "epoch": 0.20018115942028986, "grad_norm": 0.40089670585699877, "learning_rate": 3.202898550724638e-05, "loss": 0.9239, "step": 442 }, { "epoch": 0.2006340579710145, "grad_norm": 0.31863594216781094, "learning_rate": 3.2101449275362325e-05, "loss": 0.9348, "step": 443 }, { "epoch": 0.20108695652173914, "grad_norm": 0.32548862383748517, "learning_rate": 3.2173913043478265e-05, "loss": 0.9369, "step": 444 }, { "epoch": 0.20153985507246377, "grad_norm": 0.2613583300649018, "learning_rate": 3.2246376811594205e-05, "loss": 0.9324, "step": 445 }, { "epoch": 0.20199275362318841, "grad_norm": 0.35558829120574964, "learning_rate": 3.231884057971015e-05, "loss": 0.9317, "step": 446 }, { "epoch": 0.20244565217391305, "grad_norm": 0.30512189168873205, "learning_rate": 3.239130434782609e-05, "loss": 0.9548, "step": 447 }, { "epoch": 0.2028985507246377, "grad_norm": 0.30108685720289574, "learning_rate": 3.246376811594203e-05, "loss": 0.9323, "step": 448 }, { "epoch": 0.2033514492753623, "grad_norm": 0.29621086270732033, "learning_rate": 3.253623188405797e-05, "loss": 0.935, "step": 449 }, { "epoch": 0.20380434782608695, "grad_norm": 0.4047731155761714, "learning_rate": 3.260869565217392e-05, "loss": 0.9457, "step": 450 }, { "epoch": 0.20425724637681159, "grad_norm": 0.45316732386135145, "learning_rate": 3.268115942028986e-05, "loss": 0.9211, "step": 451 }, { "epoch": 0.20471014492753623, "grad_norm": 0.5010515302542429, "learning_rate": 3.2753623188405796e-05, "loss": 0.9315, "step": 452 }, { "epoch": 0.20516304347826086, "grad_norm": 0.5752808713449907, "learning_rate": 3.282608695652174e-05, "loss": 0.9303, "step": 453 }, { "epoch": 0.2056159420289855, "grad_norm": 0.6488664262963243, "learning_rate": 3.289855072463768e-05, "loss": 0.9352, "step": 454 }, { "epoch": 0.20606884057971014, "grad_norm": 0.6953403456325224, "learning_rate": 3.297101449275362e-05, "loss": 0.9402, "step": 455 }, { "epoch": 0.20652173913043478, "grad_norm": 0.8728759022328848, "learning_rate": 3.304347826086957e-05, "loss": 0.9174, "step": 456 }, { "epoch": 0.20697463768115942, "grad_norm": 1.01774203775395, "learning_rate": 3.311594202898551e-05, "loss": 0.9459, "step": 457 }, { "epoch": 0.20742753623188406, "grad_norm": 1.0259920565205412, "learning_rate": 3.3188405797101455e-05, "loss": 0.9234, "step": 458 }, { "epoch": 0.2078804347826087, "grad_norm": 0.8354141134446039, "learning_rate": 3.3260869565217395e-05, "loss": 0.9187, "step": 459 }, { "epoch": 0.20833333333333334, "grad_norm": 0.5324619823765262, "learning_rate": 3.3333333333333335e-05, "loss": 0.9249, "step": 460 }, { "epoch": 0.20878623188405798, "grad_norm": 0.41435553547606463, "learning_rate": 3.340579710144928e-05, "loss": 0.934, "step": 461 }, { "epoch": 0.20923913043478262, "grad_norm": 0.4343734193291679, "learning_rate": 3.347826086956522e-05, "loss": 0.9198, "step": 462 }, { "epoch": 0.20969202898550723, "grad_norm": 0.567309687805124, "learning_rate": 3.355072463768116e-05, "loss": 0.9373, "step": 463 }, { "epoch": 0.21014492753623187, "grad_norm": 0.7944541708838385, "learning_rate": 3.36231884057971e-05, "loss": 0.9413, "step": 464 }, { "epoch": 0.2105978260869565, "grad_norm": 0.8588371864301152, "learning_rate": 3.369565217391305e-05, "loss": 0.911, "step": 465 }, { "epoch": 0.21105072463768115, "grad_norm": 0.6819909266040349, "learning_rate": 3.376811594202899e-05, "loss": 0.9313, "step": 466 }, { "epoch": 0.2115036231884058, "grad_norm": 0.6175508912555334, "learning_rate": 3.384057971014493e-05, "loss": 0.9284, "step": 467 }, { "epoch": 0.21195652173913043, "grad_norm": 0.4301702210946152, "learning_rate": 3.391304347826087e-05, "loss": 0.9306, "step": 468 }, { "epoch": 0.21240942028985507, "grad_norm": 0.3059977114282245, "learning_rate": 3.398550724637681e-05, "loss": 0.9156, "step": 469 }, { "epoch": 0.2128623188405797, "grad_norm": 0.22450917758111227, "learning_rate": 3.405797101449276e-05, "loss": 0.9266, "step": 470 }, { "epoch": 0.21331521739130435, "grad_norm": 0.2599734752526013, "learning_rate": 3.41304347826087e-05, "loss": 0.9063, "step": 471 }, { "epoch": 0.213768115942029, "grad_norm": 0.3022478177392291, "learning_rate": 3.420289855072464e-05, "loss": 0.9463, "step": 472 }, { "epoch": 0.21422101449275363, "grad_norm": 0.2870020169832932, "learning_rate": 3.4275362318840586e-05, "loss": 0.9451, "step": 473 }, { "epoch": 0.21467391304347827, "grad_norm": 0.26125992039161167, "learning_rate": 3.4347826086956526e-05, "loss": 0.939, "step": 474 }, { "epoch": 0.2151268115942029, "grad_norm": 0.3548419756013294, "learning_rate": 3.4420289855072465e-05, "loss": 0.9431, "step": 475 }, { "epoch": 0.21557971014492755, "grad_norm": 0.37432848124991286, "learning_rate": 3.449275362318841e-05, "loss": 0.9238, "step": 476 }, { "epoch": 0.21603260869565216, "grad_norm": 0.5049063255246709, "learning_rate": 3.456521739130435e-05, "loss": 0.9392, "step": 477 }, { "epoch": 0.2164855072463768, "grad_norm": 0.6042288043404334, "learning_rate": 3.463768115942029e-05, "loss": 0.8938, "step": 478 }, { "epoch": 0.21693840579710144, "grad_norm": 0.7032098524215953, "learning_rate": 3.471014492753623e-05, "loss": 0.9387, "step": 479 }, { "epoch": 0.21739130434782608, "grad_norm": 0.7808027399510651, "learning_rate": 3.478260869565218e-05, "loss": 0.9242, "step": 480 }, { "epoch": 0.21784420289855072, "grad_norm": 0.7088610191921552, "learning_rate": 3.485507246376812e-05, "loss": 0.919, "step": 481 }, { "epoch": 0.21829710144927536, "grad_norm": 0.5510979808453059, "learning_rate": 3.492753623188406e-05, "loss": 0.9248, "step": 482 }, { "epoch": 0.21875, "grad_norm": 0.35277616430618364, "learning_rate": 3.5000000000000004e-05, "loss": 0.9336, "step": 483 }, { "epoch": 0.21920289855072464, "grad_norm": 0.25589769134667895, "learning_rate": 3.5072463768115943e-05, "loss": 0.9341, "step": 484 }, { "epoch": 0.21965579710144928, "grad_norm": 0.35850454535169396, "learning_rate": 3.514492753623189e-05, "loss": 0.9472, "step": 485 }, { "epoch": 0.22010869565217392, "grad_norm": 0.32738333349503856, "learning_rate": 3.521739130434783e-05, "loss": 0.9325, "step": 486 }, { "epoch": 0.22056159420289856, "grad_norm": 0.32930011747939875, "learning_rate": 3.528985507246377e-05, "loss": 0.9465, "step": 487 }, { "epoch": 0.2210144927536232, "grad_norm": 0.3541471685991725, "learning_rate": 3.5362318840579716e-05, "loss": 0.9427, "step": 488 }, { "epoch": 0.22146739130434784, "grad_norm": 0.290732937299871, "learning_rate": 3.5434782608695656e-05, "loss": 0.9381, "step": 489 }, { "epoch": 0.22192028985507245, "grad_norm": 0.27585979199068633, "learning_rate": 3.5507246376811596e-05, "loss": 0.9273, "step": 490 }, { "epoch": 0.2223731884057971, "grad_norm": 0.3605005994307352, "learning_rate": 3.5579710144927535e-05, "loss": 0.9092, "step": 491 }, { "epoch": 0.22282608695652173, "grad_norm": 0.5266152190042416, "learning_rate": 3.565217391304348e-05, "loss": 0.9294, "step": 492 }, { "epoch": 0.22327898550724637, "grad_norm": 0.5546783076893773, "learning_rate": 3.572463768115942e-05, "loss": 0.9093, "step": 493 }, { "epoch": 0.223731884057971, "grad_norm": 0.6691285568141475, "learning_rate": 3.579710144927536e-05, "loss": 0.9412, "step": 494 }, { "epoch": 0.22418478260869565, "grad_norm": 0.7130413062748308, "learning_rate": 3.586956521739131e-05, "loss": 0.9206, "step": 495 }, { "epoch": 0.2246376811594203, "grad_norm": 0.7596537740181832, "learning_rate": 3.594202898550725e-05, "loss": 0.9006, "step": 496 }, { "epoch": 0.22509057971014493, "grad_norm": 0.8268225335938115, "learning_rate": 3.6014492753623194e-05, "loss": 0.9238, "step": 497 }, { "epoch": 0.22554347826086957, "grad_norm": 0.7775528829128788, "learning_rate": 3.6086956521739134e-05, "loss": 0.9335, "step": 498 }, { "epoch": 0.2259963768115942, "grad_norm": 0.7013745309702989, "learning_rate": 3.6159420289855074e-05, "loss": 0.9201, "step": 499 }, { "epoch": 0.22644927536231885, "grad_norm": 0.6561156575132201, "learning_rate": 3.623188405797102e-05, "loss": 0.9326, "step": 500 }, { "epoch": 0.2269021739130435, "grad_norm": 0.5079064787283271, "learning_rate": 3.630434782608696e-05, "loss": 0.9152, "step": 501 }, { "epoch": 0.22735507246376813, "grad_norm": 0.310490145858123, "learning_rate": 3.63768115942029e-05, "loss": 0.9374, "step": 502 }, { "epoch": 0.22780797101449277, "grad_norm": 0.5191405617576468, "learning_rate": 3.6449275362318846e-05, "loss": 0.9323, "step": 503 }, { "epoch": 0.22826086956521738, "grad_norm": 0.8043331271502718, "learning_rate": 3.6521739130434786e-05, "loss": 0.931, "step": 504 }, { "epoch": 0.22871376811594202, "grad_norm": 1.0627987527652478, "learning_rate": 3.6594202898550726e-05, "loss": 0.9199, "step": 505 }, { "epoch": 0.22916666666666666, "grad_norm": 0.96470073142813, "learning_rate": 3.6666666666666666e-05, "loss": 0.9263, "step": 506 }, { "epoch": 0.2296195652173913, "grad_norm": 0.7412627210527312, "learning_rate": 3.673913043478261e-05, "loss": 0.9283, "step": 507 }, { "epoch": 0.23007246376811594, "grad_norm": 0.557378647430522, "learning_rate": 3.681159420289855e-05, "loss": 0.9266, "step": 508 }, { "epoch": 0.23052536231884058, "grad_norm": 0.4289779863844831, "learning_rate": 3.68840579710145e-05, "loss": 0.9211, "step": 509 }, { "epoch": 0.23097826086956522, "grad_norm": 0.6289027124033914, "learning_rate": 3.695652173913044e-05, "loss": 0.9227, "step": 510 }, { "epoch": 0.23143115942028986, "grad_norm": 0.8761638376615567, "learning_rate": 3.702898550724638e-05, "loss": 0.9177, "step": 511 }, { "epoch": 0.2318840579710145, "grad_norm": 0.9224436828581118, "learning_rate": 3.7101449275362325e-05, "loss": 0.9335, "step": 512 }, { "epoch": 0.23233695652173914, "grad_norm": 0.9279241162092297, "learning_rate": 3.7173913043478264e-05, "loss": 0.9389, "step": 513 }, { "epoch": 0.23278985507246377, "grad_norm": 0.8344597679715599, "learning_rate": 3.7246376811594204e-05, "loss": 0.9389, "step": 514 }, { "epoch": 0.23324275362318841, "grad_norm": 0.6357334925481074, "learning_rate": 3.731884057971015e-05, "loss": 0.9342, "step": 515 }, { "epoch": 0.23369565217391305, "grad_norm": 0.4444048825983572, "learning_rate": 3.739130434782609e-05, "loss": 0.9345, "step": 516 }, { "epoch": 0.2341485507246377, "grad_norm": 0.2810325638553822, "learning_rate": 3.746376811594204e-05, "loss": 0.952, "step": 517 }, { "epoch": 0.2346014492753623, "grad_norm": 0.4282588073666022, "learning_rate": 3.753623188405797e-05, "loss": 0.948, "step": 518 }, { "epoch": 0.23505434782608695, "grad_norm": 0.5711973045939118, "learning_rate": 3.7608695652173917e-05, "loss": 0.919, "step": 519 }, { "epoch": 0.23550724637681159, "grad_norm": 0.6089860957511268, "learning_rate": 3.7681159420289856e-05, "loss": 0.9642, "step": 520 }, { "epoch": 0.23596014492753623, "grad_norm": 0.6218163461218927, "learning_rate": 3.7753623188405796e-05, "loss": 0.9313, "step": 521 }, { "epoch": 0.23641304347826086, "grad_norm": 0.5066506683632085, "learning_rate": 3.782608695652174e-05, "loss": 0.9433, "step": 522 }, { "epoch": 0.2368659420289855, "grad_norm": 0.45015951857002734, "learning_rate": 3.789855072463768e-05, "loss": 0.9378, "step": 523 }, { "epoch": 0.23731884057971014, "grad_norm": 0.27699906117942574, "learning_rate": 3.797101449275363e-05, "loss": 0.9044, "step": 524 }, { "epoch": 0.23777173913043478, "grad_norm": 0.4627307218712217, "learning_rate": 3.804347826086957e-05, "loss": 0.9152, "step": 525 }, { "epoch": 0.23822463768115942, "grad_norm": 0.6713961603038509, "learning_rate": 3.811594202898551e-05, "loss": 0.9083, "step": 526 }, { "epoch": 0.23867753623188406, "grad_norm": 0.6709333483814116, "learning_rate": 3.8188405797101455e-05, "loss": 0.9244, "step": 527 }, { "epoch": 0.2391304347826087, "grad_norm": 0.6282875607420679, "learning_rate": 3.8260869565217395e-05, "loss": 0.9113, "step": 528 }, { "epoch": 0.23958333333333334, "grad_norm": 0.5816267202774031, "learning_rate": 3.833333333333334e-05, "loss": 0.9459, "step": 529 }, { "epoch": 0.24003623188405798, "grad_norm": 0.5734501868483284, "learning_rate": 3.840579710144928e-05, "loss": 0.928, "step": 530 }, { "epoch": 0.24048913043478262, "grad_norm": 0.48232908964200866, "learning_rate": 3.847826086956522e-05, "loss": 0.9286, "step": 531 }, { "epoch": 0.24094202898550723, "grad_norm": 0.4012123545749521, "learning_rate": 3.855072463768116e-05, "loss": 0.9395, "step": 532 }, { "epoch": 0.24139492753623187, "grad_norm": 0.33859746641764665, "learning_rate": 3.86231884057971e-05, "loss": 0.9125, "step": 533 }, { "epoch": 0.2418478260869565, "grad_norm": 0.38520480411029145, "learning_rate": 3.869565217391305e-05, "loss": 0.9241, "step": 534 }, { "epoch": 0.24230072463768115, "grad_norm": 0.3726887825938112, "learning_rate": 3.876811594202899e-05, "loss": 0.9395, "step": 535 }, { "epoch": 0.2427536231884058, "grad_norm": 0.37914018641706104, "learning_rate": 3.884057971014493e-05, "loss": 0.9172, "step": 536 }, { "epoch": 0.24320652173913043, "grad_norm": 0.4365275093851793, "learning_rate": 3.891304347826087e-05, "loss": 0.9339, "step": 537 }, { "epoch": 0.24365942028985507, "grad_norm": 0.56567034758969, "learning_rate": 3.898550724637681e-05, "loss": 0.908, "step": 538 }, { "epoch": 0.2441123188405797, "grad_norm": 0.7104350035281679, "learning_rate": 3.905797101449276e-05, "loss": 0.9206, "step": 539 }, { "epoch": 0.24456521739130435, "grad_norm": 0.6928450565076238, "learning_rate": 3.91304347826087e-05, "loss": 0.9201, "step": 540 }, { "epoch": 0.245018115942029, "grad_norm": 0.6478439149761406, "learning_rate": 3.920289855072464e-05, "loss": 0.9242, "step": 541 }, { "epoch": 0.24547101449275363, "grad_norm": 0.6460192028134802, "learning_rate": 3.9275362318840585e-05, "loss": 0.9341, "step": 542 }, { "epoch": 0.24592391304347827, "grad_norm": 0.5519685856684485, "learning_rate": 3.9347826086956525e-05, "loss": 0.934, "step": 543 }, { "epoch": 0.2463768115942029, "grad_norm": 0.5388271805332832, "learning_rate": 3.942028985507247e-05, "loss": 0.9438, "step": 544 }, { "epoch": 0.24682971014492755, "grad_norm": 0.5872121562030309, "learning_rate": 3.949275362318841e-05, "loss": 0.9279, "step": 545 }, { "epoch": 0.24728260869565216, "grad_norm": 0.6715877428735448, "learning_rate": 3.956521739130435e-05, "loss": 0.9355, "step": 546 }, { "epoch": 0.2477355072463768, "grad_norm": 0.7323482564179499, "learning_rate": 3.963768115942029e-05, "loss": 0.9438, "step": 547 }, { "epoch": 0.24818840579710144, "grad_norm": 0.8033073952555922, "learning_rate": 3.971014492753623e-05, "loss": 0.9033, "step": 548 }, { "epoch": 0.24864130434782608, "grad_norm": 0.9278466315075773, "learning_rate": 3.978260869565218e-05, "loss": 0.9408, "step": 549 }, { "epoch": 0.24909420289855072, "grad_norm": 0.9480781796204151, "learning_rate": 3.985507246376812e-05, "loss": 0.9237, "step": 550 }, { "epoch": 0.24954710144927536, "grad_norm": 0.7744539996945323, "learning_rate": 3.9927536231884064e-05, "loss": 0.9502, "step": 551 }, { "epoch": 0.25, "grad_norm": 0.5050950479577715, "learning_rate": 4e-05, "loss": 0.9328, "step": 552 }, { "epoch": 0.2504528985507246, "grad_norm": 0.26356048301132623, "learning_rate": 4.007246376811594e-05, "loss": 0.9034, "step": 553 }, { "epoch": 0.2509057971014493, "grad_norm": 0.4339664370883352, "learning_rate": 4.014492753623189e-05, "loss": 0.9482, "step": 554 }, { "epoch": 0.2513586956521739, "grad_norm": 0.5889020042971652, "learning_rate": 4.021739130434782e-05, "loss": 0.905, "step": 555 }, { "epoch": 0.25181159420289856, "grad_norm": 0.7902178085113598, "learning_rate": 4.0289855072463776e-05, "loss": 0.9324, "step": 556 }, { "epoch": 0.25226449275362317, "grad_norm": 0.862707164113978, "learning_rate": 4.036231884057971e-05, "loss": 0.9276, "step": 557 }, { "epoch": 0.25271739130434784, "grad_norm": 0.7921422477913402, "learning_rate": 4.0434782608695655e-05, "loss": 0.9344, "step": 558 }, { "epoch": 0.25317028985507245, "grad_norm": 0.7695918365843383, "learning_rate": 4.0507246376811595e-05, "loss": 0.9196, "step": 559 }, { "epoch": 0.2536231884057971, "grad_norm": 0.7024117090118137, "learning_rate": 4.057971014492754e-05, "loss": 0.9247, "step": 560 }, { "epoch": 0.25407608695652173, "grad_norm": 0.527377624896444, "learning_rate": 4.065217391304348e-05, "loss": 0.919, "step": 561 }, { "epoch": 0.2545289855072464, "grad_norm": 0.3132820000944819, "learning_rate": 4.072463768115943e-05, "loss": 0.9366, "step": 562 }, { "epoch": 0.254981884057971, "grad_norm": 0.3514452548183584, "learning_rate": 4.079710144927537e-05, "loss": 0.9379, "step": 563 }, { "epoch": 0.2554347826086957, "grad_norm": 0.4828366225474412, "learning_rate": 4.0869565217391314e-05, "loss": 0.907, "step": 564 }, { "epoch": 0.2558876811594203, "grad_norm": 0.6674775224375358, "learning_rate": 4.094202898550725e-05, "loss": 0.9434, "step": 565 }, { "epoch": 0.2563405797101449, "grad_norm": 0.8305988777578028, "learning_rate": 4.101449275362319e-05, "loss": 0.9276, "step": 566 }, { "epoch": 0.25679347826086957, "grad_norm": 0.8868024237618182, "learning_rate": 4.1086956521739134e-05, "loss": 0.9216, "step": 567 }, { "epoch": 0.2572463768115942, "grad_norm": 0.7752164173294195, "learning_rate": 4.115942028985507e-05, "loss": 0.9227, "step": 568 }, { "epoch": 0.25769927536231885, "grad_norm": 0.5193621171310954, "learning_rate": 4.123188405797102e-05, "loss": 0.9247, "step": 569 }, { "epoch": 0.25815217391304346, "grad_norm": 0.3109010110509641, "learning_rate": 4.130434782608696e-05, "loss": 0.9429, "step": 570 }, { "epoch": 0.2586050724637681, "grad_norm": 0.32385884571586165, "learning_rate": 4.1376811594202906e-05, "loss": 0.9431, "step": 571 }, { "epoch": 0.25905797101449274, "grad_norm": 0.4767612115344969, "learning_rate": 4.144927536231884e-05, "loss": 0.9342, "step": 572 }, { "epoch": 0.2595108695652174, "grad_norm": 0.6019310741460687, "learning_rate": 4.1521739130434786e-05, "loss": 0.9296, "step": 573 }, { "epoch": 0.259963768115942, "grad_norm": 0.7398968309666555, "learning_rate": 4.1594202898550726e-05, "loss": 0.9268, "step": 574 }, { "epoch": 0.2604166666666667, "grad_norm": 0.8849851729783611, "learning_rate": 4.166666666666667e-05, "loss": 0.9296, "step": 575 }, { "epoch": 0.2608695652173913, "grad_norm": 0.9255443564881616, "learning_rate": 4.173913043478261e-05, "loss": 0.9444, "step": 576 }, { "epoch": 0.26132246376811596, "grad_norm": 0.809236455919035, "learning_rate": 4.181159420289856e-05, "loss": 0.9079, "step": 577 }, { "epoch": 0.2617753623188406, "grad_norm": 0.6415069150834323, "learning_rate": 4.18840579710145e-05, "loss": 0.9406, "step": 578 }, { "epoch": 0.26222826086956524, "grad_norm": 0.5464010652820669, "learning_rate": 4.1956521739130445e-05, "loss": 0.9127, "step": 579 }, { "epoch": 0.26268115942028986, "grad_norm": 0.35869840271606274, "learning_rate": 4.202898550724638e-05, "loss": 0.9216, "step": 580 }, { "epoch": 0.26313405797101447, "grad_norm": 0.3947661376432661, "learning_rate": 4.210144927536232e-05, "loss": 0.9547, "step": 581 }, { "epoch": 0.26358695652173914, "grad_norm": 0.7182123446035603, "learning_rate": 4.2173913043478264e-05, "loss": 0.9174, "step": 582 }, { "epoch": 0.26403985507246375, "grad_norm": 0.7672123158845898, "learning_rate": 4.2246376811594204e-05, "loss": 0.9317, "step": 583 }, { "epoch": 0.2644927536231884, "grad_norm": 0.6482598656332252, "learning_rate": 4.231884057971015e-05, "loss": 0.9292, "step": 584 }, { "epoch": 0.264945652173913, "grad_norm": 0.5998260761223577, "learning_rate": 4.239130434782609e-05, "loss": 0.9263, "step": 585 }, { "epoch": 0.2653985507246377, "grad_norm": 0.543063077289452, "learning_rate": 4.2463768115942037e-05, "loss": 0.92, "step": 586 }, { "epoch": 0.2658514492753623, "grad_norm": 0.408957399998936, "learning_rate": 4.253623188405797e-05, "loss": 0.9367, "step": 587 }, { "epoch": 0.266304347826087, "grad_norm": 0.334614193023632, "learning_rate": 4.2608695652173916e-05, "loss": 0.9491, "step": 588 }, { "epoch": 0.2667572463768116, "grad_norm": 0.4369764006601475, "learning_rate": 4.2681159420289856e-05, "loss": 0.9465, "step": 589 }, { "epoch": 0.26721014492753625, "grad_norm": 0.5923383110984323, "learning_rate": 4.27536231884058e-05, "loss": 0.9333, "step": 590 }, { "epoch": 0.26766304347826086, "grad_norm": 0.6209234152284735, "learning_rate": 4.282608695652174e-05, "loss": 0.9177, "step": 591 }, { "epoch": 0.26811594202898553, "grad_norm": 0.8289548056528077, "learning_rate": 4.289855072463769e-05, "loss": 0.9362, "step": 592 }, { "epoch": 0.26856884057971014, "grad_norm": 1.178768937956287, "learning_rate": 4.297101449275363e-05, "loss": 0.8949, "step": 593 }, { "epoch": 0.26902173913043476, "grad_norm": 0.6199668256196559, "learning_rate": 4.304347826086956e-05, "loss": 0.9282, "step": 594 }, { "epoch": 0.2694746376811594, "grad_norm": 1.164070128434212, "learning_rate": 4.311594202898551e-05, "loss": 0.917, "step": 595 }, { "epoch": 0.26992753623188404, "grad_norm": 1.3611303122075311, "learning_rate": 4.318840579710145e-05, "loss": 0.9134, "step": 596 }, { "epoch": 0.2703804347826087, "grad_norm": 0.7444473205584218, "learning_rate": 4.3260869565217394e-05, "loss": 0.9175, "step": 597 }, { "epoch": 0.2708333333333333, "grad_norm": 1.3829968378732744, "learning_rate": 4.3333333333333334e-05, "loss": 0.9386, "step": 598 }, { "epoch": 0.271286231884058, "grad_norm": 0.4161693245410058, "learning_rate": 4.340579710144928e-05, "loss": 0.9139, "step": 599 }, { "epoch": 0.2717391304347826, "grad_norm": 1.0789128725738861, "learning_rate": 4.347826086956522e-05, "loss": 0.9235, "step": 600 }, { "epoch": 0.27219202898550726, "grad_norm": 1.237051774214308, "learning_rate": 4.355072463768117e-05, "loss": 0.9301, "step": 601 }, { "epoch": 0.2726449275362319, "grad_norm": 0.7383139398586063, "learning_rate": 4.362318840579711e-05, "loss": 0.911, "step": 602 }, { "epoch": 0.27309782608695654, "grad_norm": 0.5495682233992588, "learning_rate": 4.369565217391305e-05, "loss": 0.9384, "step": 603 }, { "epoch": 0.27355072463768115, "grad_norm": 1.0005187185904043, "learning_rate": 4.3768115942028986e-05, "loss": 0.9156, "step": 604 }, { "epoch": 0.2740036231884058, "grad_norm": 1.0534482444211068, "learning_rate": 4.384057971014493e-05, "loss": 0.9365, "step": 605 }, { "epoch": 0.27445652173913043, "grad_norm": 0.918568606111427, "learning_rate": 4.391304347826087e-05, "loss": 0.9169, "step": 606 }, { "epoch": 0.2749094202898551, "grad_norm": 0.799506379466386, "learning_rate": 4.398550724637682e-05, "loss": 0.9356, "step": 607 }, { "epoch": 0.2753623188405797, "grad_norm": 0.5737417371967236, "learning_rate": 4.405797101449276e-05, "loss": 0.9048, "step": 608 }, { "epoch": 0.2758152173913043, "grad_norm": 0.4454452549511505, "learning_rate": 4.41304347826087e-05, "loss": 0.9172, "step": 609 }, { "epoch": 0.276268115942029, "grad_norm": 0.7857592911384289, "learning_rate": 4.4202898550724645e-05, "loss": 0.9301, "step": 610 }, { "epoch": 0.2767210144927536, "grad_norm": 0.8478024225318737, "learning_rate": 4.427536231884058e-05, "loss": 0.93, "step": 611 }, { "epoch": 0.27717391304347827, "grad_norm": 0.7798564821298769, "learning_rate": 4.4347826086956525e-05, "loss": 0.9314, "step": 612 }, { "epoch": 0.2776268115942029, "grad_norm": 0.7050902631214042, "learning_rate": 4.4420289855072464e-05, "loss": 0.9296, "step": 613 }, { "epoch": 0.27807971014492755, "grad_norm": 0.5201502129993721, "learning_rate": 4.449275362318841e-05, "loss": 0.9343, "step": 614 }, { "epoch": 0.27853260869565216, "grad_norm": 0.3073503013060165, "learning_rate": 4.456521739130435e-05, "loss": 0.9437, "step": 615 }, { "epoch": 0.27898550724637683, "grad_norm": 0.5118160933656088, "learning_rate": 4.46376811594203e-05, "loss": 0.9294, "step": 616 }, { "epoch": 0.27943840579710144, "grad_norm": 0.7034042778819419, "learning_rate": 4.471014492753624e-05, "loss": 0.9185, "step": 617 }, { "epoch": 0.2798913043478261, "grad_norm": 0.8437626153801598, "learning_rate": 4.4782608695652184e-05, "loss": 0.9419, "step": 618 }, { "epoch": 0.2803442028985507, "grad_norm": 0.8871925440364521, "learning_rate": 4.4855072463768117e-05, "loss": 0.9067, "step": 619 }, { "epoch": 0.2807971014492754, "grad_norm": 0.7713052211244551, "learning_rate": 4.492753623188406e-05, "loss": 0.9205, "step": 620 }, { "epoch": 0.28125, "grad_norm": 0.5383609113750187, "learning_rate": 4.5e-05, "loss": 0.9228, "step": 621 }, { "epoch": 0.2817028985507246, "grad_norm": 0.3136336892726604, "learning_rate": 4.507246376811594e-05, "loss": 0.9371, "step": 622 }, { "epoch": 0.2821557971014493, "grad_norm": 0.4422120503190147, "learning_rate": 4.514492753623189e-05, "loss": 0.9276, "step": 623 }, { "epoch": 0.2826086956521739, "grad_norm": 0.6349261584242981, "learning_rate": 4.521739130434783e-05, "loss": 0.9178, "step": 624 }, { "epoch": 0.28306159420289856, "grad_norm": 0.6665361588430452, "learning_rate": 4.5289855072463775e-05, "loss": 0.9183, "step": 625 }, { "epoch": 0.28351449275362317, "grad_norm": 0.5542043216459589, "learning_rate": 4.536231884057971e-05, "loss": 0.9249, "step": 626 }, { "epoch": 0.28396739130434784, "grad_norm": 0.3593333177987638, "learning_rate": 4.5434782608695655e-05, "loss": 0.9224, "step": 627 }, { "epoch": 0.28442028985507245, "grad_norm": 0.48949746772427066, "learning_rate": 4.5507246376811595e-05, "loss": 0.9355, "step": 628 }, { "epoch": 0.2848731884057971, "grad_norm": 0.7631397657989969, "learning_rate": 4.557971014492754e-05, "loss": 0.9328, "step": 629 }, { "epoch": 0.28532608695652173, "grad_norm": 0.9844155509354338, "learning_rate": 4.565217391304348e-05, "loss": 0.9078, "step": 630 }, { "epoch": 0.2857789855072464, "grad_norm": 1.065436208521894, "learning_rate": 4.572463768115943e-05, "loss": 0.9129, "step": 631 }, { "epoch": 0.286231884057971, "grad_norm": 0.9236316899396821, "learning_rate": 4.579710144927537e-05, "loss": 0.9259, "step": 632 }, { "epoch": 0.2866847826086957, "grad_norm": 0.6471139167357217, "learning_rate": 4.5869565217391314e-05, "loss": 0.9531, "step": 633 }, { "epoch": 0.2871376811594203, "grad_norm": 0.45072264503127996, "learning_rate": 4.594202898550725e-05, "loss": 0.9262, "step": 634 }, { "epoch": 0.2875905797101449, "grad_norm": 0.3885399403972921, "learning_rate": 4.601449275362319e-05, "loss": 0.9257, "step": 635 }, { "epoch": 0.28804347826086957, "grad_norm": 0.5717905277003796, "learning_rate": 4.608695652173913e-05, "loss": 0.9054, "step": 636 }, { "epoch": 0.2884963768115942, "grad_norm": 0.6572478787328633, "learning_rate": 4.615942028985507e-05, "loss": 0.9042, "step": 637 }, { "epoch": 0.28894927536231885, "grad_norm": 0.5952282017484364, "learning_rate": 4.623188405797102e-05, "loss": 0.9185, "step": 638 }, { "epoch": 0.28940217391304346, "grad_norm": 0.5478836955300973, "learning_rate": 4.630434782608696e-05, "loss": 0.9197, "step": 639 }, { "epoch": 0.2898550724637681, "grad_norm": 0.5381218983218109, "learning_rate": 4.6376811594202906e-05, "loss": 0.909, "step": 640 }, { "epoch": 0.29030797101449274, "grad_norm": 0.4160522510232186, "learning_rate": 4.644927536231884e-05, "loss": 0.9135, "step": 641 }, { "epoch": 0.2907608695652174, "grad_norm": 0.254354131809935, "learning_rate": 4.652173913043479e-05, "loss": 0.8984, "step": 642 }, { "epoch": 0.291213768115942, "grad_norm": 0.4109277661534246, "learning_rate": 4.6594202898550725e-05, "loss": 0.937, "step": 643 }, { "epoch": 0.2916666666666667, "grad_norm": 0.5135793660079604, "learning_rate": 4.666666666666667e-05, "loss": 0.922, "step": 644 }, { "epoch": 0.2921195652173913, "grad_norm": 0.5318188759916617, "learning_rate": 4.673913043478261e-05, "loss": 0.9302, "step": 645 }, { "epoch": 0.29257246376811596, "grad_norm": 0.7212877226468263, "learning_rate": 4.681159420289856e-05, "loss": 0.9143, "step": 646 }, { "epoch": 0.2930253623188406, "grad_norm": 0.7066803125257826, "learning_rate": 4.68840579710145e-05, "loss": 0.9152, "step": 647 }, { "epoch": 0.29347826086956524, "grad_norm": 0.6730636359989547, "learning_rate": 4.6956521739130444e-05, "loss": 0.9304, "step": 648 }, { "epoch": 0.29393115942028986, "grad_norm": 0.6796231146983196, "learning_rate": 4.7028985507246384e-05, "loss": 0.9148, "step": 649 }, { "epoch": 0.29438405797101447, "grad_norm": 0.6386510252939984, "learning_rate": 4.710144927536232e-05, "loss": 0.9198, "step": 650 }, { "epoch": 0.29483695652173914, "grad_norm": 0.619867181548157, "learning_rate": 4.7173913043478264e-05, "loss": 0.9161, "step": 651 }, { "epoch": 0.29528985507246375, "grad_norm": 0.6011908314124498, "learning_rate": 4.72463768115942e-05, "loss": 0.9118, "step": 652 }, { "epoch": 0.2957427536231884, "grad_norm": 0.6088538733495652, "learning_rate": 4.731884057971015e-05, "loss": 0.9323, "step": 653 }, { "epoch": 0.296195652173913, "grad_norm": 0.6740910898340932, "learning_rate": 4.739130434782609e-05, "loss": 0.923, "step": 654 }, { "epoch": 0.2966485507246377, "grad_norm": 0.7040680082536833, "learning_rate": 4.7463768115942036e-05, "loss": 0.9323, "step": 655 }, { "epoch": 0.2971014492753623, "grad_norm": 0.6835715943830138, "learning_rate": 4.7536231884057976e-05, "loss": 0.9328, "step": 656 }, { "epoch": 0.297554347826087, "grad_norm": 0.6567092311228194, "learning_rate": 4.760869565217392e-05, "loss": 0.9333, "step": 657 }, { "epoch": 0.2980072463768116, "grad_norm": 0.663180771577805, "learning_rate": 4.7681159420289855e-05, "loss": 0.9221, "step": 658 }, { "epoch": 0.29846014492753625, "grad_norm": 0.6253010621627507, "learning_rate": 4.77536231884058e-05, "loss": 0.9308, "step": 659 }, { "epoch": 0.29891304347826086, "grad_norm": 0.5367072553687278, "learning_rate": 4.782608695652174e-05, "loss": 0.9198, "step": 660 }, { "epoch": 0.29936594202898553, "grad_norm": 0.38823675469998314, "learning_rate": 4.789855072463769e-05, "loss": 0.8923, "step": 661 }, { "epoch": 0.29981884057971014, "grad_norm": 0.3927011363920026, "learning_rate": 4.797101449275363e-05, "loss": 0.9124, "step": 662 }, { "epoch": 0.30027173913043476, "grad_norm": 0.41412260322765265, "learning_rate": 4.804347826086957e-05, "loss": 0.9206, "step": 663 }, { "epoch": 0.3007246376811594, "grad_norm": 0.49441523557084777, "learning_rate": 4.8115942028985514e-05, "loss": 0.9385, "step": 664 }, { "epoch": 0.30117753623188404, "grad_norm": 0.6608993105754808, "learning_rate": 4.818840579710145e-05, "loss": 0.9188, "step": 665 }, { "epoch": 0.3016304347826087, "grad_norm": 0.7972309368110797, "learning_rate": 4.8260869565217394e-05, "loss": 0.918, "step": 666 }, { "epoch": 0.3020833333333333, "grad_norm": 0.9046855830887534, "learning_rate": 4.8333333333333334e-05, "loss": 0.9304, "step": 667 }, { "epoch": 0.302536231884058, "grad_norm": 0.9267138966685878, "learning_rate": 4.840579710144928e-05, "loss": 0.9361, "step": 668 }, { "epoch": 0.3029891304347826, "grad_norm": 0.9850365988936514, "learning_rate": 4.847826086956522e-05, "loss": 0.9251, "step": 669 }, { "epoch": 0.30344202898550726, "grad_norm": 0.9721026919372673, "learning_rate": 4.8550724637681166e-05, "loss": 0.9042, "step": 670 }, { "epoch": 0.3038949275362319, "grad_norm": 0.8957396751589238, "learning_rate": 4.8623188405797106e-05, "loss": 0.9343, "step": 671 }, { "epoch": 0.30434782608695654, "grad_norm": 0.6792057644654098, "learning_rate": 4.869565217391305e-05, "loss": 0.922, "step": 672 }, { "epoch": 0.30480072463768115, "grad_norm": 0.44363044390096207, "learning_rate": 4.8768115942028986e-05, "loss": 0.9215, "step": 673 }, { "epoch": 0.3052536231884058, "grad_norm": 0.33297767701699726, "learning_rate": 4.884057971014493e-05, "loss": 0.8966, "step": 674 }, { "epoch": 0.30570652173913043, "grad_norm": 0.38588329809438887, "learning_rate": 4.891304347826087e-05, "loss": 0.943, "step": 675 }, { "epoch": 0.3061594202898551, "grad_norm": 0.5614387605060912, "learning_rate": 4.898550724637682e-05, "loss": 0.9239, "step": 676 }, { "epoch": 0.3066123188405797, "grad_norm": 0.6492984253680717, "learning_rate": 4.905797101449276e-05, "loss": 0.9437, "step": 677 }, { "epoch": 0.3070652173913043, "grad_norm": 0.6177142486462757, "learning_rate": 4.91304347826087e-05, "loss": 0.9401, "step": 678 }, { "epoch": 0.307518115942029, "grad_norm": 0.6976296137791465, "learning_rate": 4.9202898550724645e-05, "loss": 0.9144, "step": 679 }, { "epoch": 0.3079710144927536, "grad_norm": 0.840447322780435, "learning_rate": 4.927536231884058e-05, "loss": 0.9413, "step": 680 }, { "epoch": 0.30842391304347827, "grad_norm": 0.807812854088942, "learning_rate": 4.9347826086956524e-05, "loss": 0.9081, "step": 681 }, { "epoch": 0.3088768115942029, "grad_norm": 0.693646109803869, "learning_rate": 4.9420289855072464e-05, "loss": 0.9304, "step": 682 }, { "epoch": 0.30932971014492755, "grad_norm": 0.5303594031996604, "learning_rate": 4.949275362318841e-05, "loss": 0.9212, "step": 683 }, { "epoch": 0.30978260869565216, "grad_norm": 0.4217477958712451, "learning_rate": 4.956521739130435e-05, "loss": 0.9335, "step": 684 }, { "epoch": 0.31023550724637683, "grad_norm": 0.4037546185876441, "learning_rate": 4.96376811594203e-05, "loss": 0.9401, "step": 685 }, { "epoch": 0.31068840579710144, "grad_norm": 0.4322946349770858, "learning_rate": 4.9710144927536237e-05, "loss": 0.9004, "step": 686 }, { "epoch": 0.3111413043478261, "grad_norm": 0.392849858373755, "learning_rate": 4.978260869565218e-05, "loss": 0.9368, "step": 687 }, { "epoch": 0.3115942028985507, "grad_norm": 0.3817263875706807, "learning_rate": 4.9855072463768116e-05, "loss": 0.916, "step": 688 }, { "epoch": 0.3120471014492754, "grad_norm": 0.39629920790536416, "learning_rate": 4.992753623188407e-05, "loss": 0.9299, "step": 689 }, { "epoch": 0.3125, "grad_norm": 0.3232403754056721, "learning_rate": 5e-05, "loss": 0.9208, "step": 690 }, { "epoch": 0.3129528985507246, "grad_norm": 0.34339607190264493, "learning_rate": 5.007246376811594e-05, "loss": 0.9136, "step": 691 }, { "epoch": 0.3134057971014493, "grad_norm": 0.293512843675919, "learning_rate": 5.014492753623189e-05, "loss": 0.9472, "step": 692 }, { "epoch": 0.3138586956521739, "grad_norm": 0.2896144661835008, "learning_rate": 5.021739130434783e-05, "loss": 0.936, "step": 693 }, { "epoch": 0.31431159420289856, "grad_norm": 0.29183537027656187, "learning_rate": 5.0289855072463775e-05, "loss": 0.9272, "step": 694 }, { "epoch": 0.31476449275362317, "grad_norm": 0.399170359868466, "learning_rate": 5.0362318840579715e-05, "loss": 0.911, "step": 695 }, { "epoch": 0.31521739130434784, "grad_norm": 0.6582377921203338, "learning_rate": 5.043478260869566e-05, "loss": 0.9155, "step": 696 }, { "epoch": 0.31567028985507245, "grad_norm": 0.9782289142878703, "learning_rate": 5.0507246376811594e-05, "loss": 0.9352, "step": 697 }, { "epoch": 0.3161231884057971, "grad_norm": 1.3559982346119157, "learning_rate": 5.057971014492754e-05, "loss": 0.943, "step": 698 }, { "epoch": 0.31657608695652173, "grad_norm": 0.6219648625662635, "learning_rate": 5.065217391304348e-05, "loss": 0.9369, "step": 699 }, { "epoch": 0.3170289855072464, "grad_norm": 0.4860427729998535, "learning_rate": 5.072463768115943e-05, "loss": 0.942, "step": 700 }, { "epoch": 0.317481884057971, "grad_norm": 0.9817600609821924, "learning_rate": 5.079710144927537e-05, "loss": 0.9293, "step": 701 }, { "epoch": 0.3179347826086957, "grad_norm": 1.4752704126891198, "learning_rate": 5.0869565217391313e-05, "loss": 0.9145, "step": 702 }, { "epoch": 0.3183876811594203, "grad_norm": 0.4873773423475212, "learning_rate": 5.094202898550725e-05, "loss": 0.916, "step": 703 }, { "epoch": 0.3188405797101449, "grad_norm": 1.111479085977121, "learning_rate": 5.1014492753623186e-05, "loss": 0.9285, "step": 704 }, { "epoch": 0.31929347826086957, "grad_norm": 1.3298531667889064, "learning_rate": 5.108695652173913e-05, "loss": 0.9077, "step": 705 }, { "epoch": 0.3197463768115942, "grad_norm": 0.5764188071112252, "learning_rate": 5.115942028985507e-05, "loss": 0.9236, "step": 706 }, { "epoch": 0.32019927536231885, "grad_norm": 1.08414994499212, "learning_rate": 5.123188405797102e-05, "loss": 0.9568, "step": 707 }, { "epoch": 0.32065217391304346, "grad_norm": 1.0553728532094824, "learning_rate": 5.130434782608696e-05, "loss": 0.9258, "step": 708 }, { "epoch": 0.3211050724637681, "grad_norm": 0.9470654660465803, "learning_rate": 5.1376811594202905e-05, "loss": 0.9417, "step": 709 }, { "epoch": 0.32155797101449274, "grad_norm": 0.9625700522355546, "learning_rate": 5.1449275362318845e-05, "loss": 0.9383, "step": 710 }, { "epoch": 0.3220108695652174, "grad_norm": 0.7336906282783408, "learning_rate": 5.152173913043479e-05, "loss": 0.9223, "step": 711 }, { "epoch": 0.322463768115942, "grad_norm": 0.5258440513771282, "learning_rate": 5.1594202898550725e-05, "loss": 0.9212, "step": 712 }, { "epoch": 0.3229166666666667, "grad_norm": 0.5488186822582767, "learning_rate": 5.166666666666667e-05, "loss": 0.9301, "step": 713 }, { "epoch": 0.3233695652173913, "grad_norm": 0.41538784763433695, "learning_rate": 5.173913043478261e-05, "loss": 0.9209, "step": 714 }, { "epoch": 0.32382246376811596, "grad_norm": 0.5302624294004918, "learning_rate": 5.181159420289856e-05, "loss": 0.9249, "step": 715 }, { "epoch": 0.3242753623188406, "grad_norm": 0.5893730610380232, "learning_rate": 5.18840579710145e-05, "loss": 0.9098, "step": 716 }, { "epoch": 0.32472826086956524, "grad_norm": 0.43385987540069215, "learning_rate": 5.1956521739130444e-05, "loss": 0.8972, "step": 717 }, { "epoch": 0.32518115942028986, "grad_norm": 0.5952773623822503, "learning_rate": 5.2028985507246384e-05, "loss": 0.9314, "step": 718 }, { "epoch": 0.32563405797101447, "grad_norm": 0.6860383357994213, "learning_rate": 5.2101449275362317e-05, "loss": 0.9182, "step": 719 }, { "epoch": 0.32608695652173914, "grad_norm": 0.6278278841094974, "learning_rate": 5.217391304347826e-05, "loss": 0.9046, "step": 720 }, { "epoch": 0.32653985507246375, "grad_norm": 0.8758570701364514, "learning_rate": 5.22463768115942e-05, "loss": 0.9169, "step": 721 }, { "epoch": 0.3269927536231884, "grad_norm": 1.1250436418603524, "learning_rate": 5.231884057971015e-05, "loss": 0.9256, "step": 722 }, { "epoch": 0.327445652173913, "grad_norm": 0.8909818806598895, "learning_rate": 5.239130434782609e-05, "loss": 0.9278, "step": 723 }, { "epoch": 0.3278985507246377, "grad_norm": 0.6629418514428254, "learning_rate": 5.2463768115942036e-05, "loss": 0.9028, "step": 724 }, { "epoch": 0.3283514492753623, "grad_norm": 0.4111637008253621, "learning_rate": 5.2536231884057975e-05, "loss": 0.9126, "step": 725 }, { "epoch": 0.328804347826087, "grad_norm": 0.536488915946722, "learning_rate": 5.260869565217392e-05, "loss": 0.934, "step": 726 }, { "epoch": 0.3292572463768116, "grad_norm": 0.5214625060288273, "learning_rate": 5.2681159420289855e-05, "loss": 0.92, "step": 727 }, { "epoch": 0.32971014492753625, "grad_norm": 0.6642900962258159, "learning_rate": 5.27536231884058e-05, "loss": 0.9386, "step": 728 }, { "epoch": 0.33016304347826086, "grad_norm": 0.8176143718621248, "learning_rate": 5.282608695652174e-05, "loss": 0.9205, "step": 729 }, { "epoch": 0.33061594202898553, "grad_norm": 0.7909340916834415, "learning_rate": 5.289855072463769e-05, "loss": 0.9146, "step": 730 }, { "epoch": 0.33106884057971014, "grad_norm": 0.8969384934815398, "learning_rate": 5.297101449275363e-05, "loss": 0.9381, "step": 731 }, { "epoch": 0.33152173913043476, "grad_norm": 0.9291910032104993, "learning_rate": 5.304347826086957e-05, "loss": 0.9277, "step": 732 }, { "epoch": 0.3319746376811594, "grad_norm": 0.8876746014951236, "learning_rate": 5.3115942028985514e-05, "loss": 0.9194, "step": 733 }, { "epoch": 0.33242753623188404, "grad_norm": 0.8704743245518505, "learning_rate": 5.318840579710145e-05, "loss": 0.9299, "step": 734 }, { "epoch": 0.3328804347826087, "grad_norm": 0.6841454963333649, "learning_rate": 5.32608695652174e-05, "loss": 0.9191, "step": 735 }, { "epoch": 0.3333333333333333, "grad_norm": 0.5331066625739037, "learning_rate": 5.333333333333333e-05, "loss": 0.9443, "step": 736 }, { "epoch": 0.333786231884058, "grad_norm": 0.4620130682947224, "learning_rate": 5.340579710144928e-05, "loss": 0.9177, "step": 737 }, { "epoch": 0.3342391304347826, "grad_norm": 0.5016750668443272, "learning_rate": 5.347826086956522e-05, "loss": 0.9417, "step": 738 }, { "epoch": 0.33469202898550726, "grad_norm": 0.6561081030471221, "learning_rate": 5.3550724637681166e-05, "loss": 0.9203, "step": 739 }, { "epoch": 0.3351449275362319, "grad_norm": 0.8573563296223922, "learning_rate": 5.3623188405797106e-05, "loss": 0.9434, "step": 740 }, { "epoch": 0.33559782608695654, "grad_norm": 1.0221289003862446, "learning_rate": 5.369565217391305e-05, "loss": 0.9054, "step": 741 }, { "epoch": 0.33605072463768115, "grad_norm": 0.9503067218575459, "learning_rate": 5.376811594202899e-05, "loss": 0.9095, "step": 742 }, { "epoch": 0.3365036231884058, "grad_norm": 0.8458331262848362, "learning_rate": 5.384057971014494e-05, "loss": 0.9101, "step": 743 }, { "epoch": 0.33695652173913043, "grad_norm": 0.627943761751433, "learning_rate": 5.391304347826087e-05, "loss": 0.9265, "step": 744 }, { "epoch": 0.3374094202898551, "grad_norm": 0.5703381823640932, "learning_rate": 5.398550724637682e-05, "loss": 0.9374, "step": 745 }, { "epoch": 0.3378623188405797, "grad_norm": 0.49846247728832865, "learning_rate": 5.405797101449276e-05, "loss": 0.9304, "step": 746 }, { "epoch": 0.3383152173913043, "grad_norm": 0.3678164048499409, "learning_rate": 5.41304347826087e-05, "loss": 0.9192, "step": 747 }, { "epoch": 0.338768115942029, "grad_norm": 0.2935222183708258, "learning_rate": 5.4202898550724644e-05, "loss": 0.9077, "step": 748 }, { "epoch": 0.3392210144927536, "grad_norm": 0.5337479824183873, "learning_rate": 5.4275362318840584e-05, "loss": 0.9084, "step": 749 }, { "epoch": 0.33967391304347827, "grad_norm": 0.6070547919604432, "learning_rate": 5.434782608695653e-05, "loss": 0.9254, "step": 750 }, { "epoch": 0.3401268115942029, "grad_norm": 0.671190031687911, "learning_rate": 5.4420289855072464e-05, "loss": 0.9331, "step": 751 }, { "epoch": 0.34057971014492755, "grad_norm": 0.609471499158221, "learning_rate": 5.449275362318841e-05, "loss": 0.915, "step": 752 }, { "epoch": 0.34103260869565216, "grad_norm": 0.7473471662156222, "learning_rate": 5.456521739130435e-05, "loss": 0.935, "step": 753 }, { "epoch": 0.34148550724637683, "grad_norm": 0.9967342198341036, "learning_rate": 5.4637681159420296e-05, "loss": 0.9345, "step": 754 }, { "epoch": 0.34193840579710144, "grad_norm": 1.2346558908982626, "learning_rate": 5.4710144927536236e-05, "loss": 0.9141, "step": 755 }, { "epoch": 0.3423913043478261, "grad_norm": 0.7180696266967821, "learning_rate": 5.478260869565218e-05, "loss": 0.9194, "step": 756 }, { "epoch": 0.3428442028985507, "grad_norm": 0.46823286837316785, "learning_rate": 5.485507246376812e-05, "loss": 0.9442, "step": 757 }, { "epoch": 0.3432971014492754, "grad_norm": 0.4828645679312603, "learning_rate": 5.492753623188407e-05, "loss": 0.9081, "step": 758 }, { "epoch": 0.34375, "grad_norm": 0.6447171503729541, "learning_rate": 5.5e-05, "loss": 0.9241, "step": 759 }, { "epoch": 0.3442028985507246, "grad_norm": 0.8036966656378958, "learning_rate": 5.507246376811594e-05, "loss": 0.9068, "step": 760 }, { "epoch": 0.3446557971014493, "grad_norm": 1.0016817623348662, "learning_rate": 5.514492753623189e-05, "loss": 0.9142, "step": 761 }, { "epoch": 0.3451086956521739, "grad_norm": 1.1230784185388565, "learning_rate": 5.521739130434783e-05, "loss": 0.9176, "step": 762 }, { "epoch": 0.34556159420289856, "grad_norm": 0.6924243585651431, "learning_rate": 5.5289855072463775e-05, "loss": 0.9232, "step": 763 }, { "epoch": 0.34601449275362317, "grad_norm": 0.3393650949837041, "learning_rate": 5.5362318840579714e-05, "loss": 0.9252, "step": 764 }, { "epoch": 0.34646739130434784, "grad_norm": 0.5554679217528196, "learning_rate": 5.543478260869566e-05, "loss": 0.9127, "step": 765 }, { "epoch": 0.34692028985507245, "grad_norm": 0.8037083483390457, "learning_rate": 5.5507246376811594e-05, "loss": 0.9438, "step": 766 }, { "epoch": 0.3473731884057971, "grad_norm": 0.906314236588192, "learning_rate": 5.557971014492754e-05, "loss": 0.9173, "step": 767 }, { "epoch": 0.34782608695652173, "grad_norm": 0.9761090843613012, "learning_rate": 5.565217391304348e-05, "loss": 0.9246, "step": 768 }, { "epoch": 0.3482789855072464, "grad_norm": 0.9518760531442182, "learning_rate": 5.572463768115943e-05, "loss": 0.9163, "step": 769 }, { "epoch": 0.348731884057971, "grad_norm": 0.9016093239943741, "learning_rate": 5.5797101449275366e-05, "loss": 0.9485, "step": 770 }, { "epoch": 0.3491847826086957, "grad_norm": 0.8242158511343614, "learning_rate": 5.586956521739131e-05, "loss": 0.9236, "step": 771 }, { "epoch": 0.3496376811594203, "grad_norm": 0.8549746208856619, "learning_rate": 5.594202898550725e-05, "loss": 0.9027, "step": 772 }, { "epoch": 0.3500905797101449, "grad_norm": 0.796574607312666, "learning_rate": 5.6014492753623186e-05, "loss": 0.91, "step": 773 }, { "epoch": 0.35054347826086957, "grad_norm": 0.7498474674984773, "learning_rate": 5.608695652173913e-05, "loss": 0.9294, "step": 774 }, { "epoch": 0.3509963768115942, "grad_norm": 0.8061957462369375, "learning_rate": 5.615942028985507e-05, "loss": 0.916, "step": 775 }, { "epoch": 0.35144927536231885, "grad_norm": 0.7614945153940957, "learning_rate": 5.623188405797102e-05, "loss": 0.9294, "step": 776 }, { "epoch": 0.35190217391304346, "grad_norm": 0.5889418961710694, "learning_rate": 5.630434782608696e-05, "loss": 0.9274, "step": 777 }, { "epoch": 0.3523550724637681, "grad_norm": 0.49760918326521697, "learning_rate": 5.6376811594202905e-05, "loss": 0.9208, "step": 778 }, { "epoch": 0.35280797101449274, "grad_norm": 0.453123374760682, "learning_rate": 5.6449275362318845e-05, "loss": 0.9258, "step": 779 }, { "epoch": 0.3532608695652174, "grad_norm": 0.3286274366997841, "learning_rate": 5.652173913043479e-05, "loss": 0.9365, "step": 780 }, { "epoch": 0.353713768115942, "grad_norm": 0.40233202255790906, "learning_rate": 5.6594202898550724e-05, "loss": 0.9148, "step": 781 }, { "epoch": 0.3541666666666667, "grad_norm": 0.4613734352486288, "learning_rate": 5.666666666666668e-05, "loss": 0.9274, "step": 782 }, { "epoch": 0.3546195652173913, "grad_norm": 0.45004042739164557, "learning_rate": 5.673913043478261e-05, "loss": 0.902, "step": 783 }, { "epoch": 0.35507246376811596, "grad_norm": 0.5288174003311418, "learning_rate": 5.681159420289856e-05, "loss": 0.9242, "step": 784 }, { "epoch": 0.3555253623188406, "grad_norm": 0.6151673008749914, "learning_rate": 5.68840579710145e-05, "loss": 0.9281, "step": 785 }, { "epoch": 0.35597826086956524, "grad_norm": 0.6285490650334814, "learning_rate": 5.695652173913044e-05, "loss": 0.9354, "step": 786 }, { "epoch": 0.35643115942028986, "grad_norm": 0.5864614523332362, "learning_rate": 5.702898550724638e-05, "loss": 0.9207, "step": 787 }, { "epoch": 0.35688405797101447, "grad_norm": 0.5834760494100859, "learning_rate": 5.710144927536232e-05, "loss": 0.9197, "step": 788 }, { "epoch": 0.35733695652173914, "grad_norm": 0.6941200496551331, "learning_rate": 5.717391304347827e-05, "loss": 0.9169, "step": 789 }, { "epoch": 0.35778985507246375, "grad_norm": 0.9443108079089927, "learning_rate": 5.72463768115942e-05, "loss": 0.9343, "step": 790 }, { "epoch": 0.3582427536231884, "grad_norm": 1.2255560222454693, "learning_rate": 5.731884057971015e-05, "loss": 0.9224, "step": 791 }, { "epoch": 0.358695652173913, "grad_norm": 0.8319994255860775, "learning_rate": 5.739130434782609e-05, "loss": 0.9404, "step": 792 }, { "epoch": 0.3591485507246377, "grad_norm": 0.7197082171430691, "learning_rate": 5.7463768115942035e-05, "loss": 0.9329, "step": 793 }, { "epoch": 0.3596014492753623, "grad_norm": 0.5139806655915944, "learning_rate": 5.7536231884057975e-05, "loss": 0.9138, "step": 794 }, { "epoch": 0.360054347826087, "grad_norm": 0.5122473390909567, "learning_rate": 5.760869565217392e-05, "loss": 0.9242, "step": 795 }, { "epoch": 0.3605072463768116, "grad_norm": 0.6344094967173998, "learning_rate": 5.768115942028986e-05, "loss": 0.9289, "step": 796 }, { "epoch": 0.36096014492753625, "grad_norm": 0.5107304815462909, "learning_rate": 5.775362318840581e-05, "loss": 0.9384, "step": 797 }, { "epoch": 0.36141304347826086, "grad_norm": 0.5388555935451755, "learning_rate": 5.782608695652174e-05, "loss": 0.9189, "step": 798 }, { "epoch": 0.36186594202898553, "grad_norm": 0.8303621273665002, "learning_rate": 5.789855072463769e-05, "loss": 0.9328, "step": 799 }, { "epoch": 0.36231884057971014, "grad_norm": 1.0767741496423753, "learning_rate": 5.797101449275363e-05, "loss": 0.9336, "step": 800 }, { "epoch": 0.36277173913043476, "grad_norm": 0.8685301544097956, "learning_rate": 5.804347826086957e-05, "loss": 0.9425, "step": 801 }, { "epoch": 0.3632246376811594, "grad_norm": 0.744160774815502, "learning_rate": 5.8115942028985513e-05, "loss": 0.9077, "step": 802 }, { "epoch": 0.36367753623188404, "grad_norm": 0.731030284552946, "learning_rate": 5.818840579710145e-05, "loss": 0.9341, "step": 803 }, { "epoch": 0.3641304347826087, "grad_norm": 0.6506205096569821, "learning_rate": 5.82608695652174e-05, "loss": 0.9166, "step": 804 }, { "epoch": 0.3645833333333333, "grad_norm": 0.6747298624337144, "learning_rate": 5.833333333333333e-05, "loss": 0.9224, "step": 805 }, { "epoch": 0.365036231884058, "grad_norm": 0.6733786593624173, "learning_rate": 5.840579710144928e-05, "loss": 0.9418, "step": 806 }, { "epoch": 0.3654891304347826, "grad_norm": 0.6641341768979752, "learning_rate": 5.847826086956522e-05, "loss": 0.9257, "step": 807 }, { "epoch": 0.36594202898550726, "grad_norm": 0.7866971476449214, "learning_rate": 5.8550724637681166e-05, "loss": 0.9262, "step": 808 }, { "epoch": 0.3663949275362319, "grad_norm": 0.9247963912615627, "learning_rate": 5.8623188405797105e-05, "loss": 0.9168, "step": 809 }, { "epoch": 0.36684782608695654, "grad_norm": 0.9986180074636626, "learning_rate": 5.869565217391305e-05, "loss": 0.9326, "step": 810 }, { "epoch": 0.36730072463768115, "grad_norm": 1.1114396294248683, "learning_rate": 5.876811594202899e-05, "loss": 0.9121, "step": 811 }, { "epoch": 0.3677536231884058, "grad_norm": 0.834806548535889, "learning_rate": 5.884057971014494e-05, "loss": 0.9377, "step": 812 }, { "epoch": 0.36820652173913043, "grad_norm": 0.6834477956033476, "learning_rate": 5.891304347826087e-05, "loss": 0.9283, "step": 813 }, { "epoch": 0.3686594202898551, "grad_norm": 0.4210228163592639, "learning_rate": 5.898550724637682e-05, "loss": 0.9213, "step": 814 }, { "epoch": 0.3691123188405797, "grad_norm": 0.3315454547206217, "learning_rate": 5.905797101449276e-05, "loss": 0.9303, "step": 815 }, { "epoch": 0.3695652173913043, "grad_norm": 0.6345592508876119, "learning_rate": 5.91304347826087e-05, "loss": 0.914, "step": 816 }, { "epoch": 0.370018115942029, "grad_norm": 0.8901234765780052, "learning_rate": 5.9202898550724644e-05, "loss": 0.9133, "step": 817 }, { "epoch": 0.3704710144927536, "grad_norm": 1.1654682285673577, "learning_rate": 5.9275362318840584e-05, "loss": 0.9176, "step": 818 }, { "epoch": 0.37092391304347827, "grad_norm": 0.7130383672507745, "learning_rate": 5.934782608695653e-05, "loss": 0.914, "step": 819 }, { "epoch": 0.3713768115942029, "grad_norm": 0.6246495881578696, "learning_rate": 5.942028985507246e-05, "loss": 0.9459, "step": 820 }, { "epoch": 0.37182971014492755, "grad_norm": 0.7109012946407147, "learning_rate": 5.949275362318841e-05, "loss": 0.9095, "step": 821 }, { "epoch": 0.37228260869565216, "grad_norm": 1.0518383859887173, "learning_rate": 5.956521739130435e-05, "loss": 0.9285, "step": 822 }, { "epoch": 0.37273550724637683, "grad_norm": 1.066137904773048, "learning_rate": 5.9637681159420296e-05, "loss": 0.9411, "step": 823 }, { "epoch": 0.37318840579710144, "grad_norm": 0.8328945266731379, "learning_rate": 5.9710144927536236e-05, "loss": 0.9219, "step": 824 }, { "epoch": 0.3736413043478261, "grad_norm": 0.7577655907519397, "learning_rate": 5.978260869565218e-05, "loss": 0.9185, "step": 825 }, { "epoch": 0.3740942028985507, "grad_norm": 0.6956703333094114, "learning_rate": 5.985507246376812e-05, "loss": 0.9261, "step": 826 }, { "epoch": 0.3745471014492754, "grad_norm": 0.6410114434000443, "learning_rate": 5.992753623188407e-05, "loss": 0.9049, "step": 827 }, { "epoch": 0.375, "grad_norm": 0.7244659630006194, "learning_rate": 6.000000000000001e-05, "loss": 0.9046, "step": 828 }, { "epoch": 0.3754528985507246, "grad_norm": 0.9430751782683358, "learning_rate": 6.007246376811594e-05, "loss": 0.9234, "step": 829 }, { "epoch": 0.3759057971014493, "grad_norm": 1.1914059460782238, "learning_rate": 6.014492753623189e-05, "loss": 0.9167, "step": 830 }, { "epoch": 0.3763586956521739, "grad_norm": 0.6655565768315193, "learning_rate": 6.021739130434783e-05, "loss": 0.8987, "step": 831 }, { "epoch": 0.37681159420289856, "grad_norm": 0.5468791046668192, "learning_rate": 6.0289855072463774e-05, "loss": 0.9232, "step": 832 }, { "epoch": 0.37726449275362317, "grad_norm": 0.7542155378109126, "learning_rate": 6.0362318840579714e-05, "loss": 0.9405, "step": 833 }, { "epoch": 0.37771739130434784, "grad_norm": 1.019163339922988, "learning_rate": 6.043478260869566e-05, "loss": 0.9123, "step": 834 }, { "epoch": 0.37817028985507245, "grad_norm": 1.1682250951755635, "learning_rate": 6.05072463768116e-05, "loss": 0.9275, "step": 835 }, { "epoch": 0.3786231884057971, "grad_norm": 0.699333936397831, "learning_rate": 6.057971014492755e-05, "loss": 0.9369, "step": 836 }, { "epoch": 0.37907608695652173, "grad_norm": 0.5776092822676793, "learning_rate": 6.065217391304348e-05, "loss": 0.9536, "step": 837 }, { "epoch": 0.3795289855072464, "grad_norm": 0.9607740302097482, "learning_rate": 6.0724637681159426e-05, "loss": 0.9539, "step": 838 }, { "epoch": 0.379981884057971, "grad_norm": 1.27286152357286, "learning_rate": 6.0797101449275366e-05, "loss": 0.9128, "step": 839 }, { "epoch": 0.3804347826086957, "grad_norm": 0.6090179731988755, "learning_rate": 6.086956521739131e-05, "loss": 0.9416, "step": 840 }, { "epoch": 0.3808876811594203, "grad_norm": 0.861481683836181, "learning_rate": 6.094202898550725e-05, "loss": 0.9249, "step": 841 }, { "epoch": 0.3813405797101449, "grad_norm": 1.2382288562068786, "learning_rate": 6.101449275362319e-05, "loss": 0.9304, "step": 842 }, { "epoch": 0.38179347826086957, "grad_norm": 0.6580985155935395, "learning_rate": 6.108695652173914e-05, "loss": 0.9404, "step": 843 }, { "epoch": 0.3822463768115942, "grad_norm": 0.8448752798388766, "learning_rate": 6.115942028985507e-05, "loss": 0.9388, "step": 844 }, { "epoch": 0.38269927536231885, "grad_norm": 0.94641346707926, "learning_rate": 6.123188405797102e-05, "loss": 0.925, "step": 845 }, { "epoch": 0.38315217391304346, "grad_norm": 0.9361219279872579, "learning_rate": 6.130434782608696e-05, "loss": 0.9114, "step": 846 }, { "epoch": 0.3836050724637681, "grad_norm": 1.0906019079106222, "learning_rate": 6.137681159420291e-05, "loss": 0.9584, "step": 847 }, { "epoch": 0.38405797101449274, "grad_norm": 0.784657213781793, "learning_rate": 6.144927536231884e-05, "loss": 0.9158, "step": 848 }, { "epoch": 0.3845108695652174, "grad_norm": 0.6841559442079151, "learning_rate": 6.152173913043479e-05, "loss": 0.9258, "step": 849 }, { "epoch": 0.384963768115942, "grad_norm": 0.8350062724173873, "learning_rate": 6.159420289855072e-05, "loss": 0.9378, "step": 850 }, { "epoch": 0.3854166666666667, "grad_norm": 0.7648788831542045, "learning_rate": 6.166666666666667e-05, "loss": 0.9355, "step": 851 }, { "epoch": 0.3858695652173913, "grad_norm": 0.5880118429080179, "learning_rate": 6.173913043478262e-05, "loss": 0.9077, "step": 852 }, { "epoch": 0.38632246376811596, "grad_norm": 0.8886372588190083, "learning_rate": 6.181159420289856e-05, "loss": 0.9506, "step": 853 }, { "epoch": 0.3867753623188406, "grad_norm": 0.8262461933861756, "learning_rate": 6.18840579710145e-05, "loss": 0.9147, "step": 854 }, { "epoch": 0.38722826086956524, "grad_norm": 0.7814707245672656, "learning_rate": 6.195652173913044e-05, "loss": 0.9169, "step": 855 }, { "epoch": 0.38768115942028986, "grad_norm": 0.9013455430452738, "learning_rate": 6.202898550724638e-05, "loss": 0.901, "step": 856 }, { "epoch": 0.38813405797101447, "grad_norm": 0.7647510313238609, "learning_rate": 6.210144927536232e-05, "loss": 0.923, "step": 857 }, { "epoch": 0.38858695652173914, "grad_norm": 0.6172343640813666, "learning_rate": 6.217391304347827e-05, "loss": 0.9249, "step": 858 }, { "epoch": 0.38903985507246375, "grad_norm": 0.6377368859855269, "learning_rate": 6.22463768115942e-05, "loss": 0.9189, "step": 859 }, { "epoch": 0.3894927536231884, "grad_norm": 0.609981603365794, "learning_rate": 6.231884057971015e-05, "loss": 0.921, "step": 860 }, { "epoch": 0.389945652173913, "grad_norm": 0.5883323520908961, "learning_rate": 6.23913043478261e-05, "loss": 0.9205, "step": 861 }, { "epoch": 0.3903985507246377, "grad_norm": 0.7080912165617972, "learning_rate": 6.246376811594204e-05, "loss": 0.9354, "step": 862 }, { "epoch": 0.3908514492753623, "grad_norm": 0.8143413606696357, "learning_rate": 6.253623188405797e-05, "loss": 0.9274, "step": 863 }, { "epoch": 0.391304347826087, "grad_norm": 0.7856798578167528, "learning_rate": 6.260869565217392e-05, "loss": 0.9254, "step": 864 }, { "epoch": 0.3917572463768116, "grad_norm": 0.9118391758509089, "learning_rate": 6.268115942028985e-05, "loss": 0.9044, "step": 865 }, { "epoch": 0.39221014492753625, "grad_norm": 1.102858014588657, "learning_rate": 6.27536231884058e-05, "loss": 0.9124, "step": 866 }, { "epoch": 0.39266304347826086, "grad_norm": 0.9635148671776067, "learning_rate": 6.282608695652175e-05, "loss": 0.9075, "step": 867 }, { "epoch": 0.39311594202898553, "grad_norm": 0.7002176631246055, "learning_rate": 6.28985507246377e-05, "loss": 0.9058, "step": 868 }, { "epoch": 0.39356884057971014, "grad_norm": 0.5924956641959584, "learning_rate": 6.297101449275363e-05, "loss": 0.9183, "step": 869 }, { "epoch": 0.39402173913043476, "grad_norm": 0.6014334431115024, "learning_rate": 6.304347826086956e-05, "loss": 0.9125, "step": 870 }, { "epoch": 0.3944746376811594, "grad_norm": 0.49032518317299606, "learning_rate": 6.31159420289855e-05, "loss": 0.9343, "step": 871 }, { "epoch": 0.39492753623188404, "grad_norm": 0.3411430048387732, "learning_rate": 6.318840579710145e-05, "loss": 0.9211, "step": 872 }, { "epoch": 0.3953804347826087, "grad_norm": 0.500137919336021, "learning_rate": 6.32608695652174e-05, "loss": 0.924, "step": 873 }, { "epoch": 0.3958333333333333, "grad_norm": 0.5150798254831087, "learning_rate": 6.333333333333333e-05, "loss": 0.9099, "step": 874 }, { "epoch": 0.396286231884058, "grad_norm": 0.545833079791146, "learning_rate": 6.340579710144928e-05, "loss": 0.9192, "step": 875 }, { "epoch": 0.3967391304347826, "grad_norm": 0.681362382458427, "learning_rate": 6.347826086956523e-05, "loss": 0.9248, "step": 876 }, { "epoch": 0.39719202898550726, "grad_norm": 0.7955000142723473, "learning_rate": 6.355072463768117e-05, "loss": 0.9203, "step": 877 }, { "epoch": 0.3976449275362319, "grad_norm": 1.1138702476642113, "learning_rate": 6.36231884057971e-05, "loss": 0.9068, "step": 878 }, { "epoch": 0.39809782608695654, "grad_norm": 0.9924065934480686, "learning_rate": 6.369565217391305e-05, "loss": 0.9319, "step": 879 }, { "epoch": 0.39855072463768115, "grad_norm": 0.8640867587680572, "learning_rate": 6.376811594202898e-05, "loss": 0.9176, "step": 880 }, { "epoch": 0.3990036231884058, "grad_norm": 0.8735719406037776, "learning_rate": 6.384057971014493e-05, "loss": 0.9252, "step": 881 }, { "epoch": 0.39945652173913043, "grad_norm": 0.9714923068229586, "learning_rate": 6.391304347826088e-05, "loss": 0.906, "step": 882 }, { "epoch": 0.3999094202898551, "grad_norm": 0.7967274058954389, "learning_rate": 6.398550724637682e-05, "loss": 0.9164, "step": 883 }, { "epoch": 0.4003623188405797, "grad_norm": 0.6493572216744268, "learning_rate": 6.405797101449276e-05, "loss": 0.9357, "step": 884 }, { "epoch": 0.4008152173913043, "grad_norm": 0.5513749872225944, "learning_rate": 6.41304347826087e-05, "loss": 0.9308, "step": 885 }, { "epoch": 0.401268115942029, "grad_norm": 0.4159465480183515, "learning_rate": 6.420289855072465e-05, "loss": 0.9518, "step": 886 }, { "epoch": 0.4017210144927536, "grad_norm": 0.4793746818722763, "learning_rate": 6.427536231884058e-05, "loss": 0.9053, "step": 887 }, { "epoch": 0.40217391304347827, "grad_norm": 0.4418515591507283, "learning_rate": 6.434782608695653e-05, "loss": 0.9097, "step": 888 }, { "epoch": 0.4026268115942029, "grad_norm": 0.27418171562997284, "learning_rate": 6.442028985507246e-05, "loss": 0.902, "step": 889 }, { "epoch": 0.40307971014492755, "grad_norm": 0.4019409683734201, "learning_rate": 6.449275362318841e-05, "loss": 0.9077, "step": 890 }, { "epoch": 0.40353260869565216, "grad_norm": 0.45309516711393577, "learning_rate": 6.456521739130436e-05, "loss": 0.9213, "step": 891 }, { "epoch": 0.40398550724637683, "grad_norm": 0.4852494963334456, "learning_rate": 6.46376811594203e-05, "loss": 0.9194, "step": 892 }, { "epoch": 0.40443840579710144, "grad_norm": 0.7373691679391835, "learning_rate": 6.471014492753624e-05, "loss": 0.9169, "step": 893 }, { "epoch": 0.4048913043478261, "grad_norm": 1.0930185518167335, "learning_rate": 6.478260869565218e-05, "loss": 0.911, "step": 894 }, { "epoch": 0.4053442028985507, "grad_norm": 1.1787416139148392, "learning_rate": 6.485507246376811e-05, "loss": 0.9142, "step": 895 }, { "epoch": 0.4057971014492754, "grad_norm": 0.7278287006838742, "learning_rate": 6.492753623188406e-05, "loss": 0.9056, "step": 896 }, { "epoch": 0.40625, "grad_norm": 0.5121019657047748, "learning_rate": 6.500000000000001e-05, "loss": 0.9519, "step": 897 }, { "epoch": 0.4067028985507246, "grad_norm": 0.5884809453613482, "learning_rate": 6.507246376811594e-05, "loss": 0.9158, "step": 898 }, { "epoch": 0.4071557971014493, "grad_norm": 0.48525790384916967, "learning_rate": 6.514492753623189e-05, "loss": 0.9141, "step": 899 }, { "epoch": 0.4076086956521739, "grad_norm": 0.763449130639772, "learning_rate": 6.521739130434783e-05, "loss": 0.9122, "step": 900 }, { "epoch": 0.40806159420289856, "grad_norm": 1.0984952002747503, "learning_rate": 6.528985507246378e-05, "loss": 0.9316, "step": 901 }, { "epoch": 0.40851449275362317, "grad_norm": 1.1068259364992579, "learning_rate": 6.536231884057971e-05, "loss": 0.9071, "step": 902 }, { "epoch": 0.40896739130434784, "grad_norm": 0.8071472361024119, "learning_rate": 6.543478260869566e-05, "loss": 0.9247, "step": 903 }, { "epoch": 0.40942028985507245, "grad_norm": 0.7006274601477807, "learning_rate": 6.550724637681159e-05, "loss": 0.9207, "step": 904 }, { "epoch": 0.4098731884057971, "grad_norm": 0.6494089038211784, "learning_rate": 6.557971014492754e-05, "loss": 0.934, "step": 905 }, { "epoch": 0.41032608695652173, "grad_norm": 0.5023112564868442, "learning_rate": 6.565217391304349e-05, "loss": 0.8964, "step": 906 }, { "epoch": 0.4107789855072464, "grad_norm": 0.5984921899989641, "learning_rate": 6.572463768115943e-05, "loss": 0.9318, "step": 907 }, { "epoch": 0.411231884057971, "grad_norm": 0.6134669433013203, "learning_rate": 6.579710144927537e-05, "loss": 0.9292, "step": 908 }, { "epoch": 0.4116847826086957, "grad_norm": 0.6532569327236463, "learning_rate": 6.586956521739131e-05, "loss": 0.939, "step": 909 }, { "epoch": 0.4121376811594203, "grad_norm": 0.6951323643602763, "learning_rate": 6.594202898550725e-05, "loss": 0.9042, "step": 910 }, { "epoch": 0.4125905797101449, "grad_norm": 0.6879900324477595, "learning_rate": 6.601449275362319e-05, "loss": 0.9295, "step": 911 }, { "epoch": 0.41304347826086957, "grad_norm": 0.7053609390157485, "learning_rate": 6.608695652173914e-05, "loss": 0.9054, "step": 912 }, { "epoch": 0.4134963768115942, "grad_norm": 0.7875030363838577, "learning_rate": 6.615942028985507e-05, "loss": 0.9255, "step": 913 }, { "epoch": 0.41394927536231885, "grad_norm": 0.955896341886391, "learning_rate": 6.623188405797102e-05, "loss": 0.9323, "step": 914 }, { "epoch": 0.41440217391304346, "grad_norm": 1.1280830772363544, "learning_rate": 6.630434782608696e-05, "loss": 0.9421, "step": 915 }, { "epoch": 0.4148550724637681, "grad_norm": 0.9357512047541908, "learning_rate": 6.637681159420291e-05, "loss": 0.9331, "step": 916 }, { "epoch": 0.41530797101449274, "grad_norm": 0.970700212916741, "learning_rate": 6.644927536231884e-05, "loss": 0.9095, "step": 917 }, { "epoch": 0.4157608695652174, "grad_norm": 1.1069678272421755, "learning_rate": 6.652173913043479e-05, "loss": 0.9292, "step": 918 }, { "epoch": 0.416213768115942, "grad_norm": 0.8823973455172384, "learning_rate": 6.659420289855072e-05, "loss": 0.9155, "step": 919 }, { "epoch": 0.4166666666666667, "grad_norm": 0.8362165607763274, "learning_rate": 6.666666666666667e-05, "loss": 0.9127, "step": 920 }, { "epoch": 0.4171195652173913, "grad_norm": 0.8991020045374146, "learning_rate": 6.673913043478262e-05, "loss": 0.9373, "step": 921 }, { "epoch": 0.41757246376811596, "grad_norm": 1.0015747166063051, "learning_rate": 6.681159420289856e-05, "loss": 0.9116, "step": 922 }, { "epoch": 0.4180253623188406, "grad_norm": 0.8662933434630848, "learning_rate": 6.68840579710145e-05, "loss": 0.9403, "step": 923 }, { "epoch": 0.41847826086956524, "grad_norm": 0.6312819248330661, "learning_rate": 6.695652173913044e-05, "loss": 0.9305, "step": 924 }, { "epoch": 0.41893115942028986, "grad_norm": 0.7382306318858888, "learning_rate": 6.702898550724639e-05, "loss": 0.9369, "step": 925 }, { "epoch": 0.41938405797101447, "grad_norm": 0.644844065994813, "learning_rate": 6.710144927536232e-05, "loss": 0.9091, "step": 926 }, { "epoch": 0.41983695652173914, "grad_norm": 0.4028249597891008, "learning_rate": 6.717391304347827e-05, "loss": 0.9356, "step": 927 }, { "epoch": 0.42028985507246375, "grad_norm": 0.3860814757407999, "learning_rate": 6.72463768115942e-05, "loss": 0.9223, "step": 928 }, { "epoch": 0.4207427536231884, "grad_norm": 0.5326490610198997, "learning_rate": 6.731884057971015e-05, "loss": 0.9406, "step": 929 }, { "epoch": 0.421195652173913, "grad_norm": 0.5877349234545799, "learning_rate": 6.73913043478261e-05, "loss": 0.9261, "step": 930 }, { "epoch": 0.4216485507246377, "grad_norm": 0.5046551062493211, "learning_rate": 6.746376811594204e-05, "loss": 0.9048, "step": 931 }, { "epoch": 0.4221014492753623, "grad_norm": 0.5613765206445445, "learning_rate": 6.753623188405797e-05, "loss": 0.9227, "step": 932 }, { "epoch": 0.422554347826087, "grad_norm": 0.7490850241560071, "learning_rate": 6.760869565217392e-05, "loss": 0.9308, "step": 933 }, { "epoch": 0.4230072463768116, "grad_norm": 0.9751465422112041, "learning_rate": 6.768115942028985e-05, "loss": 0.9471, "step": 934 }, { "epoch": 0.42346014492753625, "grad_norm": 1.3227336947121506, "learning_rate": 6.77536231884058e-05, "loss": 0.9418, "step": 935 }, { "epoch": 0.42391304347826086, "grad_norm": 0.7741999637708612, "learning_rate": 6.782608695652175e-05, "loss": 0.9072, "step": 936 }, { "epoch": 0.42436594202898553, "grad_norm": 0.7256003121082848, "learning_rate": 6.78985507246377e-05, "loss": 0.9022, "step": 937 }, { "epoch": 0.42481884057971014, "grad_norm": 0.6705980186654901, "learning_rate": 6.797101449275363e-05, "loss": 0.9205, "step": 938 }, { "epoch": 0.42527173913043476, "grad_norm": 0.5296473255158926, "learning_rate": 6.804347826086957e-05, "loss": 0.9105, "step": 939 }, { "epoch": 0.4257246376811594, "grad_norm": 0.6849442633175473, "learning_rate": 6.811594202898552e-05, "loss": 0.9055, "step": 940 }, { "epoch": 0.42617753623188404, "grad_norm": 0.9728423670349149, "learning_rate": 6.818840579710145e-05, "loss": 0.909, "step": 941 }, { "epoch": 0.4266304347826087, "grad_norm": 1.3630425731865772, "learning_rate": 6.82608695652174e-05, "loss": 0.929, "step": 942 }, { "epoch": 0.4270833333333333, "grad_norm": 0.5266273790454641, "learning_rate": 6.833333333333333e-05, "loss": 0.8922, "step": 943 }, { "epoch": 0.427536231884058, "grad_norm": 0.7957392576060599, "learning_rate": 6.840579710144928e-05, "loss": 0.9068, "step": 944 }, { "epoch": 0.4279891304347826, "grad_norm": 1.2369361232682567, "learning_rate": 6.847826086956522e-05, "loss": 0.897, "step": 945 }, { "epoch": 0.42844202898550726, "grad_norm": 0.8269266502982685, "learning_rate": 6.855072463768117e-05, "loss": 0.9167, "step": 946 }, { "epoch": 0.4288949275362319, "grad_norm": 0.9118636478084193, "learning_rate": 6.86231884057971e-05, "loss": 0.9215, "step": 947 }, { "epoch": 0.42934782608695654, "grad_norm": 0.8677390647208949, "learning_rate": 6.869565217391305e-05, "loss": 0.917, "step": 948 }, { "epoch": 0.42980072463768115, "grad_norm": 0.7096643716051181, "learning_rate": 6.876811594202898e-05, "loss": 0.9087, "step": 949 }, { "epoch": 0.4302536231884058, "grad_norm": 0.5960619698925215, "learning_rate": 6.884057971014493e-05, "loss": 0.93, "step": 950 }, { "epoch": 0.43070652173913043, "grad_norm": 0.7526764975389967, "learning_rate": 6.891304347826088e-05, "loss": 0.9132, "step": 951 }, { "epoch": 0.4311594202898551, "grad_norm": 0.7452297204756178, "learning_rate": 6.898550724637682e-05, "loss": 0.9243, "step": 952 }, { "epoch": 0.4316123188405797, "grad_norm": 0.6317068467878939, "learning_rate": 6.905797101449276e-05, "loss": 0.9198, "step": 953 }, { "epoch": 0.4320652173913043, "grad_norm": 0.6348284881024303, "learning_rate": 6.91304347826087e-05, "loss": 0.9208, "step": 954 }, { "epoch": 0.432518115942029, "grad_norm": 0.7587859276794776, "learning_rate": 6.920289855072465e-05, "loss": 0.9204, "step": 955 }, { "epoch": 0.4329710144927536, "grad_norm": 1.0427102274226085, "learning_rate": 6.927536231884058e-05, "loss": 0.9208, "step": 956 }, { "epoch": 0.43342391304347827, "grad_norm": 1.267444982371234, "learning_rate": 6.934782608695653e-05, "loss": 0.9389, "step": 957 }, { "epoch": 0.4338768115942029, "grad_norm": 0.707703874122727, "learning_rate": 6.942028985507246e-05, "loss": 0.8974, "step": 958 }, { "epoch": 0.43432971014492755, "grad_norm": 0.6931417782582984, "learning_rate": 6.949275362318841e-05, "loss": 0.9167, "step": 959 }, { "epoch": 0.43478260869565216, "grad_norm": 0.5950218221631695, "learning_rate": 6.956521739130436e-05, "loss": 0.9096, "step": 960 }, { "epoch": 0.43523550724637683, "grad_norm": 0.6653303570642178, "learning_rate": 6.96376811594203e-05, "loss": 0.9243, "step": 961 }, { "epoch": 0.43568840579710144, "grad_norm": 1.0677694284357204, "learning_rate": 6.971014492753623e-05, "loss": 0.9119, "step": 962 }, { "epoch": 0.4361413043478261, "grad_norm": 1.1213407835070501, "learning_rate": 6.978260869565218e-05, "loss": 0.9472, "step": 963 }, { "epoch": 0.4365942028985507, "grad_norm": 0.8312535902981099, "learning_rate": 6.985507246376811e-05, "loss": 0.9274, "step": 964 }, { "epoch": 0.4370471014492754, "grad_norm": 0.8276180277980286, "learning_rate": 6.992753623188407e-05, "loss": 0.9241, "step": 965 }, { "epoch": 0.4375, "grad_norm": 0.8522412930208745, "learning_rate": 7.000000000000001e-05, "loss": 0.9131, "step": 966 }, { "epoch": 0.4379528985507246, "grad_norm": 0.8641093447401348, "learning_rate": 7.007246376811594e-05, "loss": 0.9343, "step": 967 }, { "epoch": 0.4384057971014493, "grad_norm": 0.895663611436977, "learning_rate": 7.014492753623189e-05, "loss": 0.9337, "step": 968 }, { "epoch": 0.4388586956521739, "grad_norm": 1.116498361423817, "learning_rate": 7.021739130434783e-05, "loss": 0.9086, "step": 969 }, { "epoch": 0.43931159420289856, "grad_norm": 0.9841421491953964, "learning_rate": 7.028985507246378e-05, "loss": 0.9541, "step": 970 }, { "epoch": 0.43976449275362317, "grad_norm": 0.8255108567803334, "learning_rate": 7.036231884057971e-05, "loss": 0.9122, "step": 971 }, { "epoch": 0.44021739130434784, "grad_norm": 0.9237729372948424, "learning_rate": 7.043478260869566e-05, "loss": 0.9357, "step": 972 }, { "epoch": 0.44067028985507245, "grad_norm": 0.9990329147614334, "learning_rate": 7.050724637681159e-05, "loss": 0.9212, "step": 973 }, { "epoch": 0.4411231884057971, "grad_norm": 0.9464069913055916, "learning_rate": 7.057971014492754e-05, "loss": 0.9104, "step": 974 }, { "epoch": 0.44157608695652173, "grad_norm": 0.8895711428056373, "learning_rate": 7.065217391304349e-05, "loss": 0.9251, "step": 975 }, { "epoch": 0.4420289855072464, "grad_norm": 0.9627730529516607, "learning_rate": 7.072463768115943e-05, "loss": 0.9357, "step": 976 }, { "epoch": 0.442481884057971, "grad_norm": 0.9780355033974067, "learning_rate": 7.079710144927537e-05, "loss": 0.939, "step": 977 }, { "epoch": 0.4429347826086957, "grad_norm": 0.9850498072091269, "learning_rate": 7.086956521739131e-05, "loss": 0.8998, "step": 978 }, { "epoch": 0.4433876811594203, "grad_norm": 0.9350830755689827, "learning_rate": 7.094202898550726e-05, "loss": 0.9147, "step": 979 }, { "epoch": 0.4438405797101449, "grad_norm": 0.9466967094079825, "learning_rate": 7.101449275362319e-05, "loss": 0.9174, "step": 980 }, { "epoch": 0.44429347826086957, "grad_norm": 0.9226330122747225, "learning_rate": 7.108695652173914e-05, "loss": 0.9198, "step": 981 }, { "epoch": 0.4447463768115942, "grad_norm": 0.8597269656696592, "learning_rate": 7.115942028985507e-05, "loss": 0.9242, "step": 982 }, { "epoch": 0.44519927536231885, "grad_norm": 0.7095187546778, "learning_rate": 7.123188405797102e-05, "loss": 0.9045, "step": 983 }, { "epoch": 0.44565217391304346, "grad_norm": 0.5708355878055634, "learning_rate": 7.130434782608696e-05, "loss": 0.9106, "step": 984 }, { "epoch": 0.4461050724637681, "grad_norm": 0.42193951195917495, "learning_rate": 7.137681159420291e-05, "loss": 0.9227, "step": 985 }, { "epoch": 0.44655797101449274, "grad_norm": 0.282734061064534, "learning_rate": 7.144927536231884e-05, "loss": 0.9132, "step": 986 }, { "epoch": 0.4470108695652174, "grad_norm": 0.43699022976598145, "learning_rate": 7.152173913043479e-05, "loss": 0.9304, "step": 987 }, { "epoch": 0.447463768115942, "grad_norm": 0.5798038237734011, "learning_rate": 7.159420289855072e-05, "loss": 0.9295, "step": 988 }, { "epoch": 0.4479166666666667, "grad_norm": 0.7330541578236338, "learning_rate": 7.166666666666667e-05, "loss": 0.9204, "step": 989 }, { "epoch": 0.4483695652173913, "grad_norm": 0.9779602394140603, "learning_rate": 7.173913043478262e-05, "loss": 0.9159, "step": 990 }, { "epoch": 0.44882246376811596, "grad_norm": 1.4369842924907272, "learning_rate": 7.181159420289856e-05, "loss": 0.9113, "step": 991 }, { "epoch": 0.4492753623188406, "grad_norm": 0.3628922407016222, "learning_rate": 7.18840579710145e-05, "loss": 0.9261, "step": 992 }, { "epoch": 0.44972826086956524, "grad_norm": 1.3068969264707517, "learning_rate": 7.195652173913044e-05, "loss": 0.9415, "step": 993 }, { "epoch": 0.45018115942028986, "grad_norm": 1.3803604505210523, "learning_rate": 7.202898550724639e-05, "loss": 0.925, "step": 994 }, { "epoch": 0.45063405797101447, "grad_norm": 0.6201557509650881, "learning_rate": 7.210144927536232e-05, "loss": 0.9201, "step": 995 }, { "epoch": 0.45108695652173914, "grad_norm": 0.5711198554681983, "learning_rate": 7.217391304347827e-05, "loss": 0.9115, "step": 996 }, { "epoch": 0.45153985507246375, "grad_norm": 0.6551658460974549, "learning_rate": 7.22463768115942e-05, "loss": 0.9207, "step": 997 }, { "epoch": 0.4519927536231884, "grad_norm": 0.8352452874665124, "learning_rate": 7.231884057971015e-05, "loss": 0.9151, "step": 998 }, { "epoch": 0.452445652173913, "grad_norm": 1.0045907751539649, "learning_rate": 7.23913043478261e-05, "loss": 0.9326, "step": 999 }, { "epoch": 0.4528985507246377, "grad_norm": 1.2808112662283098, "learning_rate": 7.246376811594204e-05, "loss": 0.9051, "step": 1000 }, { "epoch": 0.4533514492753623, "grad_norm": 0.602715905935519, "learning_rate": 7.253623188405797e-05, "loss": 0.9155, "step": 1001 }, { "epoch": 0.453804347826087, "grad_norm": 0.6514215301352799, "learning_rate": 7.260869565217392e-05, "loss": 0.9134, "step": 1002 }, { "epoch": 0.4542572463768116, "grad_norm": 1.0049523353710352, "learning_rate": 7.268115942028985e-05, "loss": 0.92, "step": 1003 }, { "epoch": 0.45471014492753625, "grad_norm": 1.156427002027996, "learning_rate": 7.27536231884058e-05, "loss": 0.9354, "step": 1004 }, { "epoch": 0.45516304347826086, "grad_norm": 0.6871565973414852, "learning_rate": 7.282608695652175e-05, "loss": 0.9249, "step": 1005 }, { "epoch": 0.45561594202898553, "grad_norm": 0.5931325231159821, "learning_rate": 7.289855072463769e-05, "loss": 0.9279, "step": 1006 }, { "epoch": 0.45606884057971014, "grad_norm": 0.796721877454686, "learning_rate": 7.297101449275363e-05, "loss": 0.921, "step": 1007 }, { "epoch": 0.45652173913043476, "grad_norm": 0.7990678070518293, "learning_rate": 7.304347826086957e-05, "loss": 0.9302, "step": 1008 }, { "epoch": 0.4569746376811594, "grad_norm": 0.6232050012037776, "learning_rate": 7.311594202898552e-05, "loss": 0.9365, "step": 1009 }, { "epoch": 0.45742753623188404, "grad_norm": 0.5752376882599053, "learning_rate": 7.318840579710145e-05, "loss": 0.9266, "step": 1010 }, { "epoch": 0.4578804347826087, "grad_norm": 0.7466508545942488, "learning_rate": 7.32608695652174e-05, "loss": 0.9154, "step": 1011 }, { "epoch": 0.4583333333333333, "grad_norm": 0.7093725813792349, "learning_rate": 7.333333333333333e-05, "loss": 0.9415, "step": 1012 }, { "epoch": 0.458786231884058, "grad_norm": 1.0658023051496002, "learning_rate": 7.340579710144928e-05, "loss": 0.9165, "step": 1013 }, { "epoch": 0.4592391304347826, "grad_norm": 1.3274724394113324, "learning_rate": 7.347826086956522e-05, "loss": 0.8935, "step": 1014 }, { "epoch": 0.45969202898550726, "grad_norm": 0.6408949480520323, "learning_rate": 7.355072463768117e-05, "loss": 0.9305, "step": 1015 }, { "epoch": 0.4601449275362319, "grad_norm": 0.9340721077242315, "learning_rate": 7.36231884057971e-05, "loss": 0.9139, "step": 1016 }, { "epoch": 0.46059782608695654, "grad_norm": 1.1075460807040816, "learning_rate": 7.369565217391305e-05, "loss": 0.9165, "step": 1017 }, { "epoch": 0.46105072463768115, "grad_norm": 1.1353020923043404, "learning_rate": 7.3768115942029e-05, "loss": 0.9373, "step": 1018 }, { "epoch": 0.4615036231884058, "grad_norm": 0.8840206667569946, "learning_rate": 7.384057971014494e-05, "loss": 0.9287, "step": 1019 }, { "epoch": 0.46195652173913043, "grad_norm": 0.7747927924171875, "learning_rate": 7.391304347826088e-05, "loss": 0.8978, "step": 1020 }, { "epoch": 0.4624094202898551, "grad_norm": 1.0371898939876119, "learning_rate": 7.398550724637682e-05, "loss": 0.9515, "step": 1021 }, { "epoch": 0.4628623188405797, "grad_norm": 0.9035160729055464, "learning_rate": 7.405797101449276e-05, "loss": 0.9134, "step": 1022 }, { "epoch": 0.4633152173913043, "grad_norm": 0.5225632092464614, "learning_rate": 7.41304347826087e-05, "loss": 0.9082, "step": 1023 }, { "epoch": 0.463768115942029, "grad_norm": 0.6188162813343251, "learning_rate": 7.420289855072465e-05, "loss": 0.9163, "step": 1024 }, { "epoch": 0.4642210144927536, "grad_norm": 0.5494051844137956, "learning_rate": 7.427536231884058e-05, "loss": 0.9227, "step": 1025 }, { "epoch": 0.46467391304347827, "grad_norm": 0.6120030596506838, "learning_rate": 7.434782608695653e-05, "loss": 0.9147, "step": 1026 }, { "epoch": 0.4651268115942029, "grad_norm": 0.9460490398236653, "learning_rate": 7.442028985507246e-05, "loss": 0.9225, "step": 1027 }, { "epoch": 0.46557971014492755, "grad_norm": 1.239589246143075, "learning_rate": 7.449275362318841e-05, "loss": 0.9177, "step": 1028 }, { "epoch": 0.46603260869565216, "grad_norm": 0.7530293114312808, "learning_rate": 7.456521739130435e-05, "loss": 0.9396, "step": 1029 }, { "epoch": 0.46648550724637683, "grad_norm": 0.8438135786623439, "learning_rate": 7.46376811594203e-05, "loss": 0.8973, "step": 1030 }, { "epoch": 0.46693840579710144, "grad_norm": 0.6432742637110426, "learning_rate": 7.471014492753623e-05, "loss": 0.9353, "step": 1031 }, { "epoch": 0.4673913043478261, "grad_norm": 0.5140673099608868, "learning_rate": 7.478260869565218e-05, "loss": 0.9269, "step": 1032 }, { "epoch": 0.4678442028985507, "grad_norm": 0.8281549827835201, "learning_rate": 7.485507246376813e-05, "loss": 0.9375, "step": 1033 }, { "epoch": 0.4682971014492754, "grad_norm": 0.6847411593166745, "learning_rate": 7.492753623188407e-05, "loss": 0.9152, "step": 1034 }, { "epoch": 0.46875, "grad_norm": 0.5928447591378778, "learning_rate": 7.500000000000001e-05, "loss": 0.933, "step": 1035 }, { "epoch": 0.4692028985507246, "grad_norm": 0.8610032309895078, "learning_rate": 7.507246376811594e-05, "loss": 0.9002, "step": 1036 }, { "epoch": 0.4696557971014493, "grad_norm": 1.1556637780597883, "learning_rate": 7.514492753623189e-05, "loss": 0.9174, "step": 1037 }, { "epoch": 0.4701086956521739, "grad_norm": 0.9818567630590692, "learning_rate": 7.521739130434783e-05, "loss": 0.912, "step": 1038 }, { "epoch": 0.47056159420289856, "grad_norm": 1.0220370750021777, "learning_rate": 7.528985507246378e-05, "loss": 0.9203, "step": 1039 }, { "epoch": 0.47101449275362317, "grad_norm": 1.0763035093053577, "learning_rate": 7.536231884057971e-05, "loss": 0.921, "step": 1040 }, { "epoch": 0.47146739130434784, "grad_norm": 1.0176644451019972, "learning_rate": 7.543478260869566e-05, "loss": 0.9323, "step": 1041 }, { "epoch": 0.47192028985507245, "grad_norm": 0.8442428116633446, "learning_rate": 7.550724637681159e-05, "loss": 0.9349, "step": 1042 }, { "epoch": 0.4723731884057971, "grad_norm": 0.8520627959346452, "learning_rate": 7.557971014492754e-05, "loss": 0.9352, "step": 1043 }, { "epoch": 0.47282608695652173, "grad_norm": 0.8394919593280677, "learning_rate": 7.565217391304349e-05, "loss": 0.9184, "step": 1044 }, { "epoch": 0.4732789855072464, "grad_norm": 0.9941490050021993, "learning_rate": 7.572463768115943e-05, "loss": 0.924, "step": 1045 }, { "epoch": 0.473731884057971, "grad_norm": 1.2158101309089118, "learning_rate": 7.579710144927536e-05, "loss": 0.9542, "step": 1046 }, { "epoch": 0.4741847826086957, "grad_norm": 0.7300259024958792, "learning_rate": 7.586956521739131e-05, "loss": 0.9209, "step": 1047 }, { "epoch": 0.4746376811594203, "grad_norm": 0.5264520313666778, "learning_rate": 7.594202898550726e-05, "loss": 0.9194, "step": 1048 }, { "epoch": 0.4750905797101449, "grad_norm": 0.5362393585456904, "learning_rate": 7.601449275362319e-05, "loss": 0.9303, "step": 1049 }, { "epoch": 0.47554347826086957, "grad_norm": 0.5774478381251648, "learning_rate": 7.608695652173914e-05, "loss": 0.9187, "step": 1050 }, { "epoch": 0.4759963768115942, "grad_norm": 0.8339273612617631, "learning_rate": 7.615942028985507e-05, "loss": 0.9372, "step": 1051 }, { "epoch": 0.47644927536231885, "grad_norm": 1.3839215684323485, "learning_rate": 7.623188405797102e-05, "loss": 0.921, "step": 1052 }, { "epoch": 0.47690217391304346, "grad_norm": 0.7126420873616887, "learning_rate": 7.630434782608696e-05, "loss": 0.8902, "step": 1053 }, { "epoch": 0.4773550724637681, "grad_norm": 0.5393302688197543, "learning_rate": 7.637681159420291e-05, "loss": 0.9277, "step": 1054 }, { "epoch": 0.47780797101449274, "grad_norm": 0.976675338917996, "learning_rate": 7.644927536231884e-05, "loss": 0.9238, "step": 1055 }, { "epoch": 0.4782608695652174, "grad_norm": 1.6251340416277418, "learning_rate": 7.652173913043479e-05, "loss": 0.9015, "step": 1056 }, { "epoch": 0.478713768115942, "grad_norm": 0.5514377956009725, "learning_rate": 7.659420289855072e-05, "loss": 0.9301, "step": 1057 }, { "epoch": 0.4791666666666667, "grad_norm": 1.8755349392818184, "learning_rate": 7.666666666666668e-05, "loss": 0.9278, "step": 1058 }, { "epoch": 0.4796195652173913, "grad_norm": 0.6403102478391672, "learning_rate": 7.673913043478262e-05, "loss": 0.9146, "step": 1059 }, { "epoch": 0.48007246376811596, "grad_norm": 1.7573447824741033, "learning_rate": 7.681159420289856e-05, "loss": 0.9324, "step": 1060 }, { "epoch": 0.4805253623188406, "grad_norm": 0.8972397730811255, "learning_rate": 7.68840579710145e-05, "loss": 0.9143, "step": 1061 }, { "epoch": 0.48097826086956524, "grad_norm": 1.8795123676602592, "learning_rate": 7.695652173913044e-05, "loss": 0.9433, "step": 1062 }, { "epoch": 0.48143115942028986, "grad_norm": 1.4181546996248235, "learning_rate": 7.702898550724639e-05, "loss": 0.9185, "step": 1063 }, { "epoch": 0.48188405797101447, "grad_norm": 1.4174125269335978, "learning_rate": 7.710144927536232e-05, "loss": 0.9332, "step": 1064 }, { "epoch": 0.48233695652173914, "grad_norm": 1.3510956953372835, "learning_rate": 7.717391304347827e-05, "loss": 0.928, "step": 1065 }, { "epoch": 0.48278985507246375, "grad_norm": 1.0608211927479057, "learning_rate": 7.72463768115942e-05, "loss": 0.9352, "step": 1066 }, { "epoch": 0.4832427536231884, "grad_norm": 0.9720236863405595, "learning_rate": 7.731884057971015e-05, "loss": 0.9331, "step": 1067 }, { "epoch": 0.483695652173913, "grad_norm": 0.9586036306289841, "learning_rate": 7.73913043478261e-05, "loss": 0.9352, "step": 1068 }, { "epoch": 0.4841485507246377, "grad_norm": 1.069265184198433, "learning_rate": 7.746376811594204e-05, "loss": 0.9087, "step": 1069 }, { "epoch": 0.4846014492753623, "grad_norm": 0.8556398021702576, "learning_rate": 7.753623188405797e-05, "loss": 0.9278, "step": 1070 }, { "epoch": 0.485054347826087, "grad_norm": 0.7267047103348371, "learning_rate": 7.760869565217392e-05, "loss": 0.9388, "step": 1071 }, { "epoch": 0.4855072463768116, "grad_norm": 0.8384409120823576, "learning_rate": 7.768115942028987e-05, "loss": 0.8959, "step": 1072 }, { "epoch": 0.48596014492753625, "grad_norm": 0.6243259205267621, "learning_rate": 7.775362318840581e-05, "loss": 0.9342, "step": 1073 }, { "epoch": 0.48641304347826086, "grad_norm": 0.8119439286536587, "learning_rate": 7.782608695652175e-05, "loss": 0.9129, "step": 1074 }, { "epoch": 0.48686594202898553, "grad_norm": 0.9187127025759521, "learning_rate": 7.789855072463769e-05, "loss": 0.9042, "step": 1075 }, { "epoch": 0.48731884057971014, "grad_norm": 1.0339631723128295, "learning_rate": 7.797101449275363e-05, "loss": 0.9157, "step": 1076 }, { "epoch": 0.48777173913043476, "grad_norm": 0.5425953549430041, "learning_rate": 7.804347826086957e-05, "loss": 0.921, "step": 1077 }, { "epoch": 0.4882246376811594, "grad_norm": 0.8187822196133742, "learning_rate": 7.811594202898552e-05, "loss": 0.9203, "step": 1078 }, { "epoch": 0.48867753623188404, "grad_norm": 0.497727525380424, "learning_rate": 7.818840579710145e-05, "loss": 0.9204, "step": 1079 }, { "epoch": 0.4891304347826087, "grad_norm": 0.6740530688697867, "learning_rate": 7.82608695652174e-05, "loss": 0.8873, "step": 1080 }, { "epoch": 0.4895833333333333, "grad_norm": 0.7554790789905333, "learning_rate": 7.833333333333333e-05, "loss": 0.9124, "step": 1081 }, { "epoch": 0.490036231884058, "grad_norm": 0.9311975131748791, "learning_rate": 7.840579710144928e-05, "loss": 0.9158, "step": 1082 }, { "epoch": 0.4904891304347826, "grad_norm": 1.03574070701002, "learning_rate": 7.847826086956522e-05, "loss": 0.9289, "step": 1083 }, { "epoch": 0.49094202898550726, "grad_norm": 1.0362070949547961, "learning_rate": 7.855072463768117e-05, "loss": 0.9285, "step": 1084 }, { "epoch": 0.4913949275362319, "grad_norm": 0.9722531137211095, "learning_rate": 7.86231884057971e-05, "loss": 0.9353, "step": 1085 }, { "epoch": 0.49184782608695654, "grad_norm": 0.7955462770719468, "learning_rate": 7.869565217391305e-05, "loss": 0.9203, "step": 1086 }, { "epoch": 0.49230072463768115, "grad_norm": 0.7537975826076027, "learning_rate": 7.8768115942029e-05, "loss": 0.904, "step": 1087 }, { "epoch": 0.4927536231884058, "grad_norm": 0.760925711498545, "learning_rate": 7.884057971014494e-05, "loss": 0.9154, "step": 1088 }, { "epoch": 0.49320652173913043, "grad_norm": 0.5998327801095166, "learning_rate": 7.891304347826088e-05, "loss": 0.9182, "step": 1089 }, { "epoch": 0.4936594202898551, "grad_norm": 0.44809229797508154, "learning_rate": 7.898550724637682e-05, "loss": 0.9207, "step": 1090 }, { "epoch": 0.4941123188405797, "grad_norm": 0.38185847119152483, "learning_rate": 7.905797101449276e-05, "loss": 0.9111, "step": 1091 }, { "epoch": 0.4945652173913043, "grad_norm": 0.3916632880404342, "learning_rate": 7.91304347826087e-05, "loss": 0.9245, "step": 1092 }, { "epoch": 0.495018115942029, "grad_norm": 0.4220133548227791, "learning_rate": 7.920289855072465e-05, "loss": 0.9153, "step": 1093 }, { "epoch": 0.4954710144927536, "grad_norm": 0.393344885162379, "learning_rate": 7.927536231884058e-05, "loss": 0.9013, "step": 1094 }, { "epoch": 0.49592391304347827, "grad_norm": 0.541661168388871, "learning_rate": 7.934782608695653e-05, "loss": 0.9044, "step": 1095 }, { "epoch": 0.4963768115942029, "grad_norm": 0.7707616502446689, "learning_rate": 7.942028985507246e-05, "loss": 0.9472, "step": 1096 }, { "epoch": 0.49682971014492755, "grad_norm": 1.0169036475172542, "learning_rate": 7.949275362318841e-05, "loss": 0.8959, "step": 1097 }, { "epoch": 0.49728260869565216, "grad_norm": 1.412902097358644, "learning_rate": 7.956521739130435e-05, "loss": 0.9246, "step": 1098 }, { "epoch": 0.49773550724637683, "grad_norm": 0.5775378163927202, "learning_rate": 7.96376811594203e-05, "loss": 0.9195, "step": 1099 }, { "epoch": 0.49818840579710144, "grad_norm": 0.6600606607012302, "learning_rate": 7.971014492753623e-05, "loss": 0.9119, "step": 1100 }, { "epoch": 0.4986413043478261, "grad_norm": 1.206778460560871, "learning_rate": 7.978260869565218e-05, "loss": 0.9313, "step": 1101 }, { "epoch": 0.4990942028985507, "grad_norm": 1.0105544258092762, "learning_rate": 7.985507246376813e-05, "loss": 0.925, "step": 1102 }, { "epoch": 0.4995471014492754, "grad_norm": 1.2014250165681521, "learning_rate": 7.992753623188407e-05, "loss": 0.8979, "step": 1103 }, { "epoch": 0.5, "grad_norm": 0.8261057150824941, "learning_rate": 8e-05, "loss": 0.9203, "step": 1104 }, { "epoch": 0.5004528985507246, "grad_norm": 0.649103803340008, "learning_rate": 7.999999800056831e-05, "loss": 0.9343, "step": 1105 }, { "epoch": 0.5009057971014492, "grad_norm": 0.6307259736944943, "learning_rate": 7.999999200227343e-05, "loss": 0.9025, "step": 1106 }, { "epoch": 0.501358695652174, "grad_norm": 0.7514715018741468, "learning_rate": 7.999998200511597e-05, "loss": 0.9055, "step": 1107 }, { "epoch": 0.5018115942028986, "grad_norm": 0.9079170838034792, "learning_rate": 7.999996800909693e-05, "loss": 0.937, "step": 1108 }, { "epoch": 0.5022644927536232, "grad_norm": 1.061390579540053, "learning_rate": 7.999995001421769e-05, "loss": 0.9285, "step": 1109 }, { "epoch": 0.5027173913043478, "grad_norm": 1.0970857972262267, "learning_rate": 7.999992802048006e-05, "loss": 0.9009, "step": 1110 }, { "epoch": 0.5031702898550725, "grad_norm": 0.9793846830526332, "learning_rate": 7.999990202788624e-05, "loss": 0.9231, "step": 1111 }, { "epoch": 0.5036231884057971, "grad_norm": 1.0423957516126894, "learning_rate": 7.999987203643883e-05, "loss": 0.931, "step": 1112 }, { "epoch": 0.5040760869565217, "grad_norm": 0.9800668164495734, "learning_rate": 7.999983804614083e-05, "loss": 0.9257, "step": 1113 }, { "epoch": 0.5045289855072463, "grad_norm": 1.048044669898651, "learning_rate": 7.999980005699563e-05, "loss": 0.9236, "step": 1114 }, { "epoch": 0.5049818840579711, "grad_norm": 0.9635514781487307, "learning_rate": 7.999975806900705e-05, "loss": 0.9158, "step": 1115 }, { "epoch": 0.5054347826086957, "grad_norm": 0.8933962429627315, "learning_rate": 7.999971208217926e-05, "loss": 0.9073, "step": 1116 }, { "epoch": 0.5058876811594203, "grad_norm": 0.8978374322077182, "learning_rate": 7.999966209651687e-05, "loss": 0.8873, "step": 1117 }, { "epoch": 0.5063405797101449, "grad_norm": 0.8216770465302271, "learning_rate": 7.999960811202487e-05, "loss": 0.9031, "step": 1118 }, { "epoch": 0.5067934782608695, "grad_norm": 0.6958218874982438, "learning_rate": 7.999955012870866e-05, "loss": 0.9424, "step": 1119 }, { "epoch": 0.5072463768115942, "grad_norm": 0.6779977092759469, "learning_rate": 7.999948814657406e-05, "loss": 0.9355, "step": 1120 }, { "epoch": 0.5076992753623188, "grad_norm": 0.7709376885194463, "learning_rate": 7.999942216562722e-05, "loss": 0.9457, "step": 1121 }, { "epoch": 0.5081521739130435, "grad_norm": 1.0383919812123266, "learning_rate": 7.999935218587477e-05, "loss": 0.9311, "step": 1122 }, { "epoch": 0.5086050724637681, "grad_norm": 1.2233554981081656, "learning_rate": 7.99992782073237e-05, "loss": 0.9194, "step": 1123 }, { "epoch": 0.5090579710144928, "grad_norm": 0.6226153558042128, "learning_rate": 7.99992002299814e-05, "loss": 0.9039, "step": 1124 }, { "epoch": 0.5095108695652174, "grad_norm": 0.7739058179384181, "learning_rate": 7.999911825385567e-05, "loss": 0.9009, "step": 1125 }, { "epoch": 0.509963768115942, "grad_norm": 1.0169196441027557, "learning_rate": 7.999903227895471e-05, "loss": 0.9164, "step": 1126 }, { "epoch": 0.5104166666666666, "grad_norm": 1.095929578051187, "learning_rate": 7.99989423052871e-05, "loss": 0.9193, "step": 1127 }, { "epoch": 0.5108695652173914, "grad_norm": 0.9614000898991459, "learning_rate": 7.999884833286186e-05, "loss": 0.9074, "step": 1128 }, { "epoch": 0.511322463768116, "grad_norm": 1.1065821194858305, "learning_rate": 7.999875036168835e-05, "loss": 0.9002, "step": 1129 }, { "epoch": 0.5117753623188406, "grad_norm": 1.0990284577107718, "learning_rate": 7.999864839177639e-05, "loss": 0.9098, "step": 1130 }, { "epoch": 0.5122282608695652, "grad_norm": 0.8725855842656955, "learning_rate": 7.999854242313617e-05, "loss": 0.931, "step": 1131 }, { "epoch": 0.5126811594202898, "grad_norm": 0.7273060838858465, "learning_rate": 7.999843245577828e-05, "loss": 0.9319, "step": 1132 }, { "epoch": 0.5131340579710145, "grad_norm": 0.827118314374097, "learning_rate": 7.99983184897137e-05, "loss": 0.9145, "step": 1133 }, { "epoch": 0.5135869565217391, "grad_norm": 0.8168569287896776, "learning_rate": 7.999820052495386e-05, "loss": 0.9002, "step": 1134 }, { "epoch": 0.5140398550724637, "grad_norm": 0.7129071106164161, "learning_rate": 7.999807856151052e-05, "loss": 0.9162, "step": 1135 }, { "epoch": 0.5144927536231884, "grad_norm": 0.8333073906763239, "learning_rate": 7.999795259939589e-05, "loss": 0.8918, "step": 1136 }, { "epoch": 0.5149456521739131, "grad_norm": 1.030956896681823, "learning_rate": 7.999782263862255e-05, "loss": 0.8933, "step": 1137 }, { "epoch": 0.5153985507246377, "grad_norm": 1.150980602413913, "learning_rate": 7.99976886792035e-05, "loss": 0.9185, "step": 1138 }, { "epoch": 0.5158514492753623, "grad_norm": 0.8638927025805793, "learning_rate": 7.999755072115213e-05, "loss": 0.9148, "step": 1139 }, { "epoch": 0.5163043478260869, "grad_norm": 0.777021950505541, "learning_rate": 7.999740876448222e-05, "loss": 0.8981, "step": 1140 }, { "epoch": 0.5167572463768116, "grad_norm": 0.4869512105469568, "learning_rate": 7.999726280920799e-05, "loss": 0.9345, "step": 1141 }, { "epoch": 0.5172101449275363, "grad_norm": 0.6668894084281048, "learning_rate": 7.999711285534402e-05, "loss": 0.9276, "step": 1142 }, { "epoch": 0.5176630434782609, "grad_norm": 1.3537222431693863, "learning_rate": 7.999695890290529e-05, "loss": 0.9209, "step": 1143 }, { "epoch": 0.5181159420289855, "grad_norm": 0.6309016145317473, "learning_rate": 7.99968009519072e-05, "loss": 0.9013, "step": 1144 }, { "epoch": 0.5185688405797102, "grad_norm": 0.8515357448731641, "learning_rate": 7.999663900236555e-05, "loss": 0.8959, "step": 1145 }, { "epoch": 0.5190217391304348, "grad_norm": 1.359726930750473, "learning_rate": 7.99964730542965e-05, "loss": 0.9086, "step": 1146 }, { "epoch": 0.5194746376811594, "grad_norm": 0.5020459154536636, "learning_rate": 7.999630310771667e-05, "loss": 0.9034, "step": 1147 }, { "epoch": 0.519927536231884, "grad_norm": 1.2287091255560254, "learning_rate": 7.999612916264302e-05, "loss": 0.9072, "step": 1148 }, { "epoch": 0.5203804347826086, "grad_norm": 0.9329260743520377, "learning_rate": 7.999595121909298e-05, "loss": 0.92, "step": 1149 }, { "epoch": 0.5208333333333334, "grad_norm": 0.654878141708257, "learning_rate": 7.99957692770843e-05, "loss": 0.9114, "step": 1150 }, { "epoch": 0.521286231884058, "grad_norm": 0.5498552585583566, "learning_rate": 7.99955833366352e-05, "loss": 0.9085, "step": 1151 }, { "epoch": 0.5217391304347826, "grad_norm": 0.7226707452720723, "learning_rate": 7.999539339776424e-05, "loss": 0.9371, "step": 1152 }, { "epoch": 0.5221920289855072, "grad_norm": 0.931101263453234, "learning_rate": 7.999519946049045e-05, "loss": 0.9204, "step": 1153 }, { "epoch": 0.5226449275362319, "grad_norm": 1.47962886237914, "learning_rate": 7.999500152483318e-05, "loss": 0.9286, "step": 1154 }, { "epoch": 0.5230978260869565, "grad_norm": 0.6031196539412971, "learning_rate": 7.999479959081223e-05, "loss": 0.9049, "step": 1155 }, { "epoch": 0.5235507246376812, "grad_norm": 0.8812207756583474, "learning_rate": 7.999459365844779e-05, "loss": 0.9123, "step": 1156 }, { "epoch": 0.5240036231884058, "grad_norm": 1.6352290988876588, "learning_rate": 7.999438372776043e-05, "loss": 0.9116, "step": 1157 }, { "epoch": 0.5244565217391305, "grad_norm": 0.6123807656168353, "learning_rate": 7.999416979877117e-05, "loss": 0.8971, "step": 1158 }, { "epoch": 0.5249094202898551, "grad_norm": 1.485525619823534, "learning_rate": 7.999395187150137e-05, "loss": 0.9269, "step": 1159 }, { "epoch": 0.5253623188405797, "grad_norm": 0.5061812430431317, "learning_rate": 7.999372994597282e-05, "loss": 0.9092, "step": 1160 }, { "epoch": 0.5258152173913043, "grad_norm": 1.5019647097526978, "learning_rate": 7.999350402220772e-05, "loss": 0.9102, "step": 1161 }, { "epoch": 0.5262681159420289, "grad_norm": 0.6096166927492497, "learning_rate": 7.999327410022866e-05, "loss": 0.9098, "step": 1162 }, { "epoch": 0.5267210144927537, "grad_norm": 1.4602591598400674, "learning_rate": 7.99930401800586e-05, "loss": 0.9278, "step": 1163 }, { "epoch": 0.5271739130434783, "grad_norm": 0.9290478356824898, "learning_rate": 7.999280226172094e-05, "loss": 0.9142, "step": 1164 }, { "epoch": 0.5276268115942029, "grad_norm": 1.4779517379801153, "learning_rate": 7.999256034523946e-05, "loss": 0.9437, "step": 1165 }, { "epoch": 0.5280797101449275, "grad_norm": 1.1669151955341075, "learning_rate": 7.999231443063835e-05, "loss": 0.9074, "step": 1166 }, { "epoch": 0.5285326086956522, "grad_norm": 1.2468130322516235, "learning_rate": 7.999206451794219e-05, "loss": 0.9164, "step": 1167 }, { "epoch": 0.5289855072463768, "grad_norm": 1.1589314619165167, "learning_rate": 7.999181060717597e-05, "loss": 0.9418, "step": 1168 }, { "epoch": 0.5294384057971014, "grad_norm": 0.7779198410669995, "learning_rate": 7.999155269836508e-05, "loss": 0.9077, "step": 1169 }, { "epoch": 0.529891304347826, "grad_norm": 1.1876987668388819, "learning_rate": 7.999129079153529e-05, "loss": 0.9348, "step": 1170 }, { "epoch": 0.5303442028985508, "grad_norm": 0.546982556711736, "learning_rate": 7.99910248867128e-05, "loss": 0.917, "step": 1171 }, { "epoch": 0.5307971014492754, "grad_norm": 0.9633790230075614, "learning_rate": 7.999075498392416e-05, "loss": 0.9097, "step": 1172 }, { "epoch": 0.53125, "grad_norm": 0.7509901810836326, "learning_rate": 7.999048108319636e-05, "loss": 0.9306, "step": 1173 }, { "epoch": 0.5317028985507246, "grad_norm": 0.7419730265796716, "learning_rate": 7.999020318455682e-05, "loss": 0.9109, "step": 1174 }, { "epoch": 0.5321557971014492, "grad_norm": 0.7700742643528627, "learning_rate": 7.99899212880333e-05, "loss": 0.9088, "step": 1175 }, { "epoch": 0.532608695652174, "grad_norm": 0.5969271635108835, "learning_rate": 7.998963539365397e-05, "loss": 0.9263, "step": 1176 }, { "epoch": 0.5330615942028986, "grad_norm": 0.6718512875521578, "learning_rate": 7.998934550144743e-05, "loss": 0.9229, "step": 1177 }, { "epoch": 0.5335144927536232, "grad_norm": 0.49567477250771713, "learning_rate": 7.998905161144264e-05, "loss": 0.8952, "step": 1178 }, { "epoch": 0.5339673913043478, "grad_norm": 0.5669749926623829, "learning_rate": 7.9988753723669e-05, "loss": 0.9259, "step": 1179 }, { "epoch": 0.5344202898550725, "grad_norm": 0.5537627669187503, "learning_rate": 7.998845183815628e-05, "loss": 0.8857, "step": 1180 }, { "epoch": 0.5348731884057971, "grad_norm": 0.6552995137316542, "learning_rate": 7.998814595493466e-05, "loss": 0.9224, "step": 1181 }, { "epoch": 0.5353260869565217, "grad_norm": 0.5230990940856795, "learning_rate": 7.998783607403473e-05, "loss": 0.9286, "step": 1182 }, { "epoch": 0.5357789855072463, "grad_norm": 0.6220821123360049, "learning_rate": 7.998752219548746e-05, "loss": 0.9196, "step": 1183 }, { "epoch": 0.5362318840579711, "grad_norm": 0.6139325512288845, "learning_rate": 7.998720431932423e-05, "loss": 0.9182, "step": 1184 }, { "epoch": 0.5366847826086957, "grad_norm": 0.759022070677022, "learning_rate": 7.99868824455768e-05, "loss": 0.9227, "step": 1185 }, { "epoch": 0.5371376811594203, "grad_norm": 0.9946256896838849, "learning_rate": 7.998655657427739e-05, "loss": 0.939, "step": 1186 }, { "epoch": 0.5375905797101449, "grad_norm": 1.2692106821721967, "learning_rate": 7.998622670545855e-05, "loss": 0.8916, "step": 1187 }, { "epoch": 0.5380434782608695, "grad_norm": 0.6246917775555398, "learning_rate": 7.998589283915326e-05, "loss": 0.9057, "step": 1188 }, { "epoch": 0.5384963768115942, "grad_norm": 0.45836817142236874, "learning_rate": 7.99855549753949e-05, "loss": 0.9115, "step": 1189 }, { "epoch": 0.5389492753623188, "grad_norm": 0.5361391593749351, "learning_rate": 7.998521311421725e-05, "loss": 0.9075, "step": 1190 }, { "epoch": 0.5394021739130435, "grad_norm": 0.7387837853041448, "learning_rate": 7.998486725565448e-05, "loss": 0.9301, "step": 1191 }, { "epoch": 0.5398550724637681, "grad_norm": 1.1227050461806913, "learning_rate": 7.998451739974117e-05, "loss": 0.9132, "step": 1192 }, { "epoch": 0.5403079710144928, "grad_norm": 0.9270341679336637, "learning_rate": 7.99841635465123e-05, "loss": 0.9111, "step": 1193 }, { "epoch": 0.5407608695652174, "grad_norm": 0.6777392182596637, "learning_rate": 7.998380569600324e-05, "loss": 0.9199, "step": 1194 }, { "epoch": 0.541213768115942, "grad_norm": 0.4705972092013831, "learning_rate": 7.998344384824975e-05, "loss": 0.9068, "step": 1195 }, { "epoch": 0.5416666666666666, "grad_norm": 0.3638200544840724, "learning_rate": 7.998307800328803e-05, "loss": 0.8996, "step": 1196 }, { "epoch": 0.5421195652173914, "grad_norm": 0.2937598566277626, "learning_rate": 7.998270816115464e-05, "loss": 0.8995, "step": 1197 }, { "epoch": 0.542572463768116, "grad_norm": 0.45656740798269746, "learning_rate": 7.998233432188656e-05, "loss": 0.9114, "step": 1198 }, { "epoch": 0.5430253623188406, "grad_norm": 0.6469600434442493, "learning_rate": 7.998195648552116e-05, "loss": 0.9159, "step": 1199 }, { "epoch": 0.5434782608695652, "grad_norm": 0.8440948704968727, "learning_rate": 7.998157465209622e-05, "loss": 0.9152, "step": 1200 }, { "epoch": 0.5439311594202898, "grad_norm": 0.9909645176010141, "learning_rate": 7.998118882164989e-05, "loss": 0.9174, "step": 1201 }, { "epoch": 0.5443840579710145, "grad_norm": 1.191133137267815, "learning_rate": 7.998079899422076e-05, "loss": 0.9185, "step": 1202 }, { "epoch": 0.5448369565217391, "grad_norm": 0.7516538084189905, "learning_rate": 7.998040516984782e-05, "loss": 0.9231, "step": 1203 }, { "epoch": 0.5452898550724637, "grad_norm": 0.5208465983609138, "learning_rate": 7.99800073485704e-05, "loss": 0.9162, "step": 1204 }, { "epoch": 0.5457427536231884, "grad_norm": 0.37007123934144964, "learning_rate": 7.99796055304283e-05, "loss": 0.9159, "step": 1205 }, { "epoch": 0.5461956521739131, "grad_norm": 0.31535205656008974, "learning_rate": 7.997919971546169e-05, "loss": 0.9197, "step": 1206 }, { "epoch": 0.5466485507246377, "grad_norm": 0.5209378746391773, "learning_rate": 7.997878990371111e-05, "loss": 0.9154, "step": 1207 }, { "epoch": 0.5471014492753623, "grad_norm": 0.6863675575399619, "learning_rate": 7.997837609521756e-05, "loss": 0.8966, "step": 1208 }, { "epoch": 0.5475543478260869, "grad_norm": 0.8354913782498246, "learning_rate": 7.99779582900224e-05, "loss": 0.9129, "step": 1209 }, { "epoch": 0.5480072463768116, "grad_norm": 0.9641897369036977, "learning_rate": 7.99775364881674e-05, "loss": 0.9016, "step": 1210 }, { "epoch": 0.5484601449275363, "grad_norm": 1.1313065029780422, "learning_rate": 7.997711068969474e-05, "loss": 0.9268, "step": 1211 }, { "epoch": 0.5489130434782609, "grad_norm": 0.7872941891965322, "learning_rate": 7.997668089464696e-05, "loss": 0.9134, "step": 1212 }, { "epoch": 0.5493659420289855, "grad_norm": 0.5563721545326498, "learning_rate": 7.997624710306706e-05, "loss": 0.9303, "step": 1213 }, { "epoch": 0.5498188405797102, "grad_norm": 0.47520528159628495, "learning_rate": 7.997580931499837e-05, "loss": 0.9113, "step": 1214 }, { "epoch": 0.5502717391304348, "grad_norm": 0.4516932833046107, "learning_rate": 7.997536753048469e-05, "loss": 0.8954, "step": 1215 }, { "epoch": 0.5507246376811594, "grad_norm": 0.7290078706087741, "learning_rate": 7.997492174957016e-05, "loss": 0.9401, "step": 1216 }, { "epoch": 0.551177536231884, "grad_norm": 1.0627236099666186, "learning_rate": 7.997447197229937e-05, "loss": 0.9126, "step": 1217 }, { "epoch": 0.5516304347826086, "grad_norm": 1.0928134462280203, "learning_rate": 7.997401819871726e-05, "loss": 0.9283, "step": 1218 }, { "epoch": 0.5520833333333334, "grad_norm": 0.8890492975901262, "learning_rate": 7.997356042886921e-05, "loss": 0.9104, "step": 1219 }, { "epoch": 0.552536231884058, "grad_norm": 0.7884429640986806, "learning_rate": 7.997309866280098e-05, "loss": 0.9034, "step": 1220 }, { "epoch": 0.5529891304347826, "grad_norm": 0.6537300000826751, "learning_rate": 7.997263290055873e-05, "loss": 0.893, "step": 1221 }, { "epoch": 0.5534420289855072, "grad_norm": 0.5114807584066792, "learning_rate": 7.997216314218903e-05, "loss": 0.9157, "step": 1222 }, { "epoch": 0.5538949275362319, "grad_norm": 0.40545023758410215, "learning_rate": 7.997168938773884e-05, "loss": 0.9146, "step": 1223 }, { "epoch": 0.5543478260869565, "grad_norm": 0.3959916873939735, "learning_rate": 7.997121163725553e-05, "loss": 0.8755, "step": 1224 }, { "epoch": 0.5548007246376812, "grad_norm": 0.5355151575345123, "learning_rate": 7.997072989078684e-05, "loss": 0.8941, "step": 1225 }, { "epoch": 0.5552536231884058, "grad_norm": 0.6187705893535176, "learning_rate": 7.997024414838095e-05, "loss": 0.899, "step": 1226 }, { "epoch": 0.5557065217391305, "grad_norm": 0.6611551931861095, "learning_rate": 7.996975441008642e-05, "loss": 0.9201, "step": 1227 }, { "epoch": 0.5561594202898551, "grad_norm": 0.8209057604990417, "learning_rate": 7.99692606759522e-05, "loss": 0.91, "step": 1228 }, { "epoch": 0.5566123188405797, "grad_norm": 1.0320943268675111, "learning_rate": 7.996876294602765e-05, "loss": 0.9234, "step": 1229 }, { "epoch": 0.5570652173913043, "grad_norm": 1.080575627477833, "learning_rate": 7.996826122036252e-05, "loss": 0.9158, "step": 1230 }, { "epoch": 0.5575181159420289, "grad_norm": 0.8676481064107101, "learning_rate": 7.9967755499007e-05, "loss": 0.9353, "step": 1231 }, { "epoch": 0.5579710144927537, "grad_norm": 0.6854485137241147, "learning_rate": 7.996724578201162e-05, "loss": 0.9164, "step": 1232 }, { "epoch": 0.5584239130434783, "grad_norm": 0.5529058847505076, "learning_rate": 7.996673206942735e-05, "loss": 0.9261, "step": 1233 }, { "epoch": 0.5588768115942029, "grad_norm": 0.5569211207357228, "learning_rate": 7.996621436130553e-05, "loss": 0.9095, "step": 1234 }, { "epoch": 0.5593297101449275, "grad_norm": 0.5832796052869625, "learning_rate": 7.996569265769795e-05, "loss": 0.9111, "step": 1235 }, { "epoch": 0.5597826086956522, "grad_norm": 0.5166404207714961, "learning_rate": 7.996516695865674e-05, "loss": 0.9298, "step": 1236 }, { "epoch": 0.5602355072463768, "grad_norm": 0.5004621840510608, "learning_rate": 7.996463726423445e-05, "loss": 0.9089, "step": 1237 }, { "epoch": 0.5606884057971014, "grad_norm": 0.5951321317315531, "learning_rate": 7.996410357448404e-05, "loss": 0.9123, "step": 1238 }, { "epoch": 0.561141304347826, "grad_norm": 0.6685371786952595, "learning_rate": 7.996356588945887e-05, "loss": 0.9299, "step": 1239 }, { "epoch": 0.5615942028985508, "grad_norm": 0.9591387099557539, "learning_rate": 7.99630242092127e-05, "loss": 0.9174, "step": 1240 }, { "epoch": 0.5620471014492754, "grad_norm": 1.2873878184636585, "learning_rate": 7.996247853379965e-05, "loss": 0.9119, "step": 1241 }, { "epoch": 0.5625, "grad_norm": 0.6025964844512687, "learning_rate": 7.996192886327432e-05, "loss": 0.9008, "step": 1242 }, { "epoch": 0.5629528985507246, "grad_norm": 0.40458244309767005, "learning_rate": 7.996137519769163e-05, "loss": 0.9247, "step": 1243 }, { "epoch": 0.5634057971014492, "grad_norm": 0.7232085910910507, "learning_rate": 7.996081753710693e-05, "loss": 0.9057, "step": 1244 }, { "epoch": 0.563858695652174, "grad_norm": 1.0825686668996692, "learning_rate": 7.996025588157598e-05, "loss": 0.9177, "step": 1245 }, { "epoch": 0.5643115942028986, "grad_norm": 1.1551446195803339, "learning_rate": 7.995969023115493e-05, "loss": 0.9247, "step": 1246 }, { "epoch": 0.5647644927536232, "grad_norm": 0.7392660239626504, "learning_rate": 7.995912058590031e-05, "loss": 0.9058, "step": 1247 }, { "epoch": 0.5652173913043478, "grad_norm": 0.5997589110510154, "learning_rate": 7.995854694586911e-05, "loss": 0.9107, "step": 1248 }, { "epoch": 0.5656702898550725, "grad_norm": 0.7274442258873419, "learning_rate": 7.995796931111864e-05, "loss": 0.9238, "step": 1249 }, { "epoch": 0.5661231884057971, "grad_norm": 0.6628124124229372, "learning_rate": 7.995738768170666e-05, "loss": 0.9337, "step": 1250 }, { "epoch": 0.5665760869565217, "grad_norm": 0.5267775732753363, "learning_rate": 7.995680205769132e-05, "loss": 0.9141, "step": 1251 }, { "epoch": 0.5670289855072463, "grad_norm": 0.48778513596800216, "learning_rate": 7.995621243913115e-05, "loss": 0.9131, "step": 1252 }, { "epoch": 0.5674818840579711, "grad_norm": 0.5823671330965113, "learning_rate": 7.995561882608512e-05, "loss": 0.9223, "step": 1253 }, { "epoch": 0.5679347826086957, "grad_norm": 0.6067504065991304, "learning_rate": 7.995502121861255e-05, "loss": 0.9177, "step": 1254 }, { "epoch": 0.5683876811594203, "grad_norm": 0.5904622007916123, "learning_rate": 7.995441961677319e-05, "loss": 0.8962, "step": 1255 }, { "epoch": 0.5688405797101449, "grad_norm": 0.60995506762879, "learning_rate": 7.99538140206272e-05, "loss": 0.9079, "step": 1256 }, { "epoch": 0.5692934782608695, "grad_norm": 0.6895153649131889, "learning_rate": 7.995320443023511e-05, "loss": 0.9122, "step": 1257 }, { "epoch": 0.5697463768115942, "grad_norm": 0.9298940090419585, "learning_rate": 7.995259084565786e-05, "loss": 0.9173, "step": 1258 }, { "epoch": 0.5701992753623188, "grad_norm": 1.1158890409897142, "learning_rate": 7.995197326695678e-05, "loss": 0.9116, "step": 1259 }, { "epoch": 0.5706521739130435, "grad_norm": 1.0049816536186065, "learning_rate": 7.995135169419365e-05, "loss": 0.9101, "step": 1260 }, { "epoch": 0.5711050724637681, "grad_norm": 0.8911073773028056, "learning_rate": 7.995072612743055e-05, "loss": 0.9108, "step": 1261 }, { "epoch": 0.5715579710144928, "grad_norm": 0.8385308773410439, "learning_rate": 7.995009656673008e-05, "loss": 0.9042, "step": 1262 }, { "epoch": 0.5720108695652174, "grad_norm": 0.8308811269291819, "learning_rate": 7.994946301215514e-05, "loss": 0.9186, "step": 1263 }, { "epoch": 0.572463768115942, "grad_norm": 0.9264097312105701, "learning_rate": 7.994882546376908e-05, "loss": 0.9168, "step": 1264 }, { "epoch": 0.5729166666666666, "grad_norm": 0.9560861242809028, "learning_rate": 7.994818392163563e-05, "loss": 0.9115, "step": 1265 }, { "epoch": 0.5733695652173914, "grad_norm": 0.9221332983740232, "learning_rate": 7.994753838581892e-05, "loss": 0.9295, "step": 1266 }, { "epoch": 0.573822463768116, "grad_norm": 0.8560859198899263, "learning_rate": 7.994688885638351e-05, "loss": 0.9224, "step": 1267 }, { "epoch": 0.5742753623188406, "grad_norm": 0.699365122308952, "learning_rate": 7.99462353333943e-05, "loss": 0.8902, "step": 1268 }, { "epoch": 0.5747282608695652, "grad_norm": 0.6894623218103528, "learning_rate": 7.994557781691668e-05, "loss": 0.9057, "step": 1269 }, { "epoch": 0.5751811594202898, "grad_norm": 0.6531943206583773, "learning_rate": 7.994491630701633e-05, "loss": 0.9057, "step": 1270 }, { "epoch": 0.5756340579710145, "grad_norm": 0.6068295622986929, "learning_rate": 7.994425080375939e-05, "loss": 0.9451, "step": 1271 }, { "epoch": 0.5760869565217391, "grad_norm": 0.5454451269996418, "learning_rate": 7.994358130721238e-05, "loss": 0.9319, "step": 1272 }, { "epoch": 0.5765398550724637, "grad_norm": 0.6078832773491369, "learning_rate": 7.994290781744229e-05, "loss": 0.9104, "step": 1273 }, { "epoch": 0.5769927536231884, "grad_norm": 0.7327736786275292, "learning_rate": 7.99422303345164e-05, "loss": 0.9082, "step": 1274 }, { "epoch": 0.5774456521739131, "grad_norm": 0.7895668233304675, "learning_rate": 7.994154885850244e-05, "loss": 0.9199, "step": 1275 }, { "epoch": 0.5778985507246377, "grad_norm": 0.9049155812463641, "learning_rate": 7.994086338946855e-05, "loss": 0.9258, "step": 1276 }, { "epoch": 0.5783514492753623, "grad_norm": 1.06545736080362, "learning_rate": 7.994017392748325e-05, "loss": 0.9222, "step": 1277 }, { "epoch": 0.5788043478260869, "grad_norm": 0.9390750363622388, "learning_rate": 7.993948047261549e-05, "loss": 0.9144, "step": 1278 }, { "epoch": 0.5792572463768116, "grad_norm": 0.758349691416432, "learning_rate": 7.993878302493456e-05, "loss": 0.9125, "step": 1279 }, { "epoch": 0.5797101449275363, "grad_norm": 0.5398821179594426, "learning_rate": 7.993808158451022e-05, "loss": 0.916, "step": 1280 }, { "epoch": 0.5801630434782609, "grad_norm": 0.41720168201058117, "learning_rate": 7.993737615141255e-05, "loss": 0.8878, "step": 1281 }, { "epoch": 0.5806159420289855, "grad_norm": 0.2586866898137774, "learning_rate": 7.993666672571213e-05, "loss": 0.9106, "step": 1282 }, { "epoch": 0.5810688405797102, "grad_norm": 0.4009014853942388, "learning_rate": 7.993595330747983e-05, "loss": 0.9159, "step": 1283 }, { "epoch": 0.5815217391304348, "grad_norm": 0.5000919604839852, "learning_rate": 7.993523589678701e-05, "loss": 0.905, "step": 1284 }, { "epoch": 0.5819746376811594, "grad_norm": 0.6923152177670936, "learning_rate": 7.993451449370537e-05, "loss": 0.9166, "step": 1285 }, { "epoch": 0.582427536231884, "grad_norm": 0.9329920862272895, "learning_rate": 7.993378909830702e-05, "loss": 0.9087, "step": 1286 }, { "epoch": 0.5828804347826086, "grad_norm": 1.222726044127117, "learning_rate": 7.993305971066451e-05, "loss": 0.9124, "step": 1287 }, { "epoch": 0.5833333333333334, "grad_norm": 0.6928071048192764, "learning_rate": 7.993232633085074e-05, "loss": 0.9447, "step": 1288 }, { "epoch": 0.583786231884058, "grad_norm": 0.47672560416125964, "learning_rate": 7.993158895893902e-05, "loss": 0.8946, "step": 1289 }, { "epoch": 0.5842391304347826, "grad_norm": 0.38096475959854526, "learning_rate": 7.993084759500308e-05, "loss": 0.9356, "step": 1290 }, { "epoch": 0.5846920289855072, "grad_norm": 0.40466968751196397, "learning_rate": 7.993010223911702e-05, "loss": 0.9379, "step": 1291 }, { "epoch": 0.5851449275362319, "grad_norm": 0.5539037705976694, "learning_rate": 7.992935289135538e-05, "loss": 0.9236, "step": 1292 }, { "epoch": 0.5855978260869565, "grad_norm": 0.6754393295985838, "learning_rate": 7.992859955179306e-05, "loss": 0.9216, "step": 1293 }, { "epoch": 0.5860507246376812, "grad_norm": 0.903400791195944, "learning_rate": 7.992784222050536e-05, "loss": 0.9179, "step": 1294 }, { "epoch": 0.5865036231884058, "grad_norm": 1.182559691252768, "learning_rate": 7.9927080897568e-05, "loss": 0.889, "step": 1295 }, { "epoch": 0.5869565217391305, "grad_norm": 0.7866458724521241, "learning_rate": 7.992631558305709e-05, "loss": 0.929, "step": 1296 }, { "epoch": 0.5874094202898551, "grad_norm": 0.47657958052781835, "learning_rate": 7.992554627704915e-05, "loss": 0.914, "step": 1297 }, { "epoch": 0.5878623188405797, "grad_norm": 0.36357353000416254, "learning_rate": 7.992477297962108e-05, "loss": 0.904, "step": 1298 }, { "epoch": 0.5883152173913043, "grad_norm": 0.5198349722987449, "learning_rate": 7.99239956908502e-05, "loss": 0.9086, "step": 1299 }, { "epoch": 0.5887681159420289, "grad_norm": 0.6336834531330658, "learning_rate": 7.992321441081419e-05, "loss": 0.8975, "step": 1300 }, { "epoch": 0.5892210144927537, "grad_norm": 0.7814189433282106, "learning_rate": 7.992242913959118e-05, "loss": 0.9387, "step": 1301 }, { "epoch": 0.5896739130434783, "grad_norm": 0.9038972552284389, "learning_rate": 7.992163987725967e-05, "loss": 0.8907, "step": 1302 }, { "epoch": 0.5901268115942029, "grad_norm": 1.0908316534461102, "learning_rate": 7.992084662389854e-05, "loss": 0.9377, "step": 1303 }, { "epoch": 0.5905797101449275, "grad_norm": 1.0034235793612158, "learning_rate": 7.992004937958714e-05, "loss": 0.9266, "step": 1304 }, { "epoch": 0.5910326086956522, "grad_norm": 1.045823154062548, "learning_rate": 7.991924814440513e-05, "loss": 0.9116, "step": 1305 }, { "epoch": 0.5914855072463768, "grad_norm": 0.8663658429961425, "learning_rate": 7.991844291843263e-05, "loss": 0.9141, "step": 1306 }, { "epoch": 0.5919384057971014, "grad_norm": 0.8201797240809109, "learning_rate": 7.991763370175013e-05, "loss": 0.9192, "step": 1307 }, { "epoch": 0.592391304347826, "grad_norm": 0.9219653533144914, "learning_rate": 7.991682049443853e-05, "loss": 0.9083, "step": 1308 }, { "epoch": 0.5928442028985508, "grad_norm": 1.0032233290776178, "learning_rate": 7.991600329657914e-05, "loss": 0.9063, "step": 1309 }, { "epoch": 0.5932971014492754, "grad_norm": 1.0652848914719582, "learning_rate": 7.991518210825364e-05, "loss": 0.9363, "step": 1310 }, { "epoch": 0.59375, "grad_norm": 0.9572541868191728, "learning_rate": 7.991435692954414e-05, "loss": 0.9276, "step": 1311 }, { "epoch": 0.5942028985507246, "grad_norm": 0.8943666294081023, "learning_rate": 7.991352776053314e-05, "loss": 0.9269, "step": 1312 }, { "epoch": 0.5946557971014492, "grad_norm": 0.9110177988753985, "learning_rate": 7.991269460130351e-05, "loss": 0.9107, "step": 1313 }, { "epoch": 0.595108695652174, "grad_norm": 0.9714216684992028, "learning_rate": 7.991185745193854e-05, "loss": 0.936, "step": 1314 }, { "epoch": 0.5955615942028986, "grad_norm": 1.16500562213777, "learning_rate": 7.991101631252195e-05, "loss": 0.908, "step": 1315 }, { "epoch": 0.5960144927536232, "grad_norm": 0.6936880173953078, "learning_rate": 7.991017118313781e-05, "loss": 0.9219, "step": 1316 }, { "epoch": 0.5964673913043478, "grad_norm": 0.6346669238512774, "learning_rate": 7.990932206387062e-05, "loss": 0.9401, "step": 1317 }, { "epoch": 0.5969202898550725, "grad_norm": 0.602028120311264, "learning_rate": 7.990846895480525e-05, "loss": 0.9273, "step": 1318 }, { "epoch": 0.5973731884057971, "grad_norm": 0.6069551809432329, "learning_rate": 7.9907611856027e-05, "loss": 0.9002, "step": 1319 }, { "epoch": 0.5978260869565217, "grad_norm": 0.8287867524292072, "learning_rate": 7.990675076762158e-05, "loss": 0.896, "step": 1320 }, { "epoch": 0.5982789855072463, "grad_norm": 0.9969276852497234, "learning_rate": 7.990588568967502e-05, "loss": 0.9124, "step": 1321 }, { "epoch": 0.5987318840579711, "grad_norm": 1.1912765079497947, "learning_rate": 7.990501662227384e-05, "loss": 0.9183, "step": 1322 }, { "epoch": 0.5991847826086957, "grad_norm": 0.7668232663806914, "learning_rate": 7.990414356550492e-05, "loss": 0.9311, "step": 1323 }, { "epoch": 0.5996376811594203, "grad_norm": 0.6899065311113128, "learning_rate": 7.990326651945552e-05, "loss": 0.9101, "step": 1324 }, { "epoch": 0.6000905797101449, "grad_norm": 0.6786034180970838, "learning_rate": 7.990238548421334e-05, "loss": 0.9143, "step": 1325 }, { "epoch": 0.6005434782608695, "grad_norm": 0.5531330753472103, "learning_rate": 7.990150045986645e-05, "loss": 0.8837, "step": 1326 }, { "epoch": 0.6009963768115942, "grad_norm": 0.6280217278038568, "learning_rate": 7.990061144650334e-05, "loss": 0.9206, "step": 1327 }, { "epoch": 0.6014492753623188, "grad_norm": 0.8122273419729146, "learning_rate": 7.989971844421286e-05, "loss": 0.898, "step": 1328 }, { "epoch": 0.6019021739130435, "grad_norm": 0.9445115191952115, "learning_rate": 7.98988214530843e-05, "loss": 0.9153, "step": 1329 }, { "epoch": 0.6023550724637681, "grad_norm": 1.176511587121097, "learning_rate": 7.989792047320736e-05, "loss": 0.9149, "step": 1330 }, { "epoch": 0.6028079710144928, "grad_norm": 0.9411059384840856, "learning_rate": 7.989701550467207e-05, "loss": 0.9148, "step": 1331 }, { "epoch": 0.6032608695652174, "grad_norm": 0.7582741428883195, "learning_rate": 7.989610654756891e-05, "loss": 0.9138, "step": 1332 }, { "epoch": 0.603713768115942, "grad_norm": 0.5121763957577531, "learning_rate": 7.989519360198878e-05, "loss": 0.9118, "step": 1333 }, { "epoch": 0.6041666666666666, "grad_norm": 0.4580657989098245, "learning_rate": 7.98942766680229e-05, "loss": 0.9025, "step": 1334 }, { "epoch": 0.6046195652173914, "grad_norm": 0.6740328237204297, "learning_rate": 7.989335574576297e-05, "loss": 0.8907, "step": 1335 }, { "epoch": 0.605072463768116, "grad_norm": 0.7555917584370975, "learning_rate": 7.989243083530106e-05, "loss": 0.908, "step": 1336 }, { "epoch": 0.6055253623188406, "grad_norm": 0.7887490980345763, "learning_rate": 7.989150193672962e-05, "loss": 0.9138, "step": 1337 }, { "epoch": 0.6059782608695652, "grad_norm": 0.8626098246612909, "learning_rate": 7.989056905014152e-05, "loss": 0.9173, "step": 1338 }, { "epoch": 0.6064311594202898, "grad_norm": 0.8742434057076901, "learning_rate": 7.988963217563001e-05, "loss": 0.8997, "step": 1339 }, { "epoch": 0.6068840579710145, "grad_norm": 0.9176963757752605, "learning_rate": 7.988869131328876e-05, "loss": 0.92, "step": 1340 }, { "epoch": 0.6073369565217391, "grad_norm": 0.8877982140017808, "learning_rate": 7.988774646321186e-05, "loss": 0.9146, "step": 1341 }, { "epoch": 0.6077898550724637, "grad_norm": 0.920757239419765, "learning_rate": 7.988679762549369e-05, "loss": 0.9143, "step": 1342 }, { "epoch": 0.6082427536231884, "grad_norm": 0.8983222615123794, "learning_rate": 7.988584480022919e-05, "loss": 0.8924, "step": 1343 }, { "epoch": 0.6086956521739131, "grad_norm": 0.7838386396234941, "learning_rate": 7.988488798751358e-05, "loss": 0.8982, "step": 1344 }, { "epoch": 0.6091485507246377, "grad_norm": 0.7110275772555308, "learning_rate": 7.98839271874425e-05, "loss": 0.9206, "step": 1345 }, { "epoch": 0.6096014492753623, "grad_norm": 0.6338587361902931, "learning_rate": 7.988296240011203e-05, "loss": 0.898, "step": 1346 }, { "epoch": 0.6100543478260869, "grad_norm": 0.4203301835949531, "learning_rate": 7.98819936256186e-05, "loss": 0.9213, "step": 1347 }, { "epoch": 0.6105072463768116, "grad_norm": 0.43147543142536815, "learning_rate": 7.988102086405907e-05, "loss": 0.9075, "step": 1348 }, { "epoch": 0.6109601449275363, "grad_norm": 0.5447867427905031, "learning_rate": 7.98800441155307e-05, "loss": 0.9181, "step": 1349 }, { "epoch": 0.6114130434782609, "grad_norm": 0.5266233624287802, "learning_rate": 7.98790633801311e-05, "loss": 0.8946, "step": 1350 }, { "epoch": 0.6118659420289855, "grad_norm": 0.505948500775946, "learning_rate": 7.987807865795836e-05, "loss": 0.9418, "step": 1351 }, { "epoch": 0.6123188405797102, "grad_norm": 0.6100724249680882, "learning_rate": 7.987708994911091e-05, "loss": 0.912, "step": 1352 }, { "epoch": 0.6127717391304348, "grad_norm": 0.7929166506552539, "learning_rate": 7.987609725368757e-05, "loss": 0.9391, "step": 1353 }, { "epoch": 0.6132246376811594, "grad_norm": 0.8366041006590712, "learning_rate": 7.987510057178761e-05, "loss": 0.8945, "step": 1354 }, { "epoch": 0.613677536231884, "grad_norm": 0.8459029183155867, "learning_rate": 7.987409990351065e-05, "loss": 0.9184, "step": 1355 }, { "epoch": 0.6141304347826086, "grad_norm": 1.0581966346190141, "learning_rate": 7.987309524895674e-05, "loss": 0.9084, "step": 1356 }, { "epoch": 0.6145833333333334, "grad_norm": 1.0694595387596082, "learning_rate": 7.987208660822631e-05, "loss": 0.8949, "step": 1357 }, { "epoch": 0.615036231884058, "grad_norm": 0.7781965059148739, "learning_rate": 7.98710739814202e-05, "loss": 0.8973, "step": 1358 }, { "epoch": 0.6154891304347826, "grad_norm": 0.5000826623552204, "learning_rate": 7.987005736863964e-05, "loss": 0.9112, "step": 1359 }, { "epoch": 0.6159420289855072, "grad_norm": 0.3691059785709693, "learning_rate": 7.986903676998628e-05, "loss": 0.9178, "step": 1360 }, { "epoch": 0.6163949275362319, "grad_norm": 0.5040714426265225, "learning_rate": 7.986801218556212e-05, "loss": 0.9033, "step": 1361 }, { "epoch": 0.6168478260869565, "grad_norm": 0.6600762258005699, "learning_rate": 7.986698361546962e-05, "loss": 0.9126, "step": 1362 }, { "epoch": 0.6173007246376812, "grad_norm": 0.8049629499035965, "learning_rate": 7.986595105981159e-05, "loss": 0.8825, "step": 1363 }, { "epoch": 0.6177536231884058, "grad_norm": 0.8120607270620845, "learning_rate": 7.986491451869125e-05, "loss": 0.9212, "step": 1364 }, { "epoch": 0.6182065217391305, "grad_norm": 0.6382512226200158, "learning_rate": 7.986387399221224e-05, "loss": 0.9103, "step": 1365 }, { "epoch": 0.6186594202898551, "grad_norm": 0.6030895604973564, "learning_rate": 7.986282948047857e-05, "loss": 0.8998, "step": 1366 }, { "epoch": 0.6191123188405797, "grad_norm": 0.5687389385685332, "learning_rate": 7.986178098359468e-05, "loss": 0.9214, "step": 1367 }, { "epoch": 0.6195652173913043, "grad_norm": 0.7643472914708405, "learning_rate": 7.986072850166538e-05, "loss": 0.9202, "step": 1368 }, { "epoch": 0.6200181159420289, "grad_norm": 0.9717056880957099, "learning_rate": 7.985967203479588e-05, "loss": 0.9006, "step": 1369 }, { "epoch": 0.6204710144927537, "grad_norm": 1.1607742733301176, "learning_rate": 7.98586115830918e-05, "loss": 0.8985, "step": 1370 }, { "epoch": 0.6209239130434783, "grad_norm": 0.7961747772390056, "learning_rate": 7.985754714665916e-05, "loss": 0.8969, "step": 1371 }, { "epoch": 0.6213768115942029, "grad_norm": 0.6060774750006008, "learning_rate": 7.985647872560437e-05, "loss": 0.9058, "step": 1372 }, { "epoch": 0.6218297101449275, "grad_norm": 0.576979521575575, "learning_rate": 7.985540632003425e-05, "loss": 0.9226, "step": 1373 }, { "epoch": 0.6222826086956522, "grad_norm": 0.42712322612368586, "learning_rate": 7.985432993005601e-05, "loss": 0.9122, "step": 1374 }, { "epoch": 0.6227355072463768, "grad_norm": 0.4352112187631184, "learning_rate": 7.985324955577724e-05, "loss": 0.9153, "step": 1375 }, { "epoch": 0.6231884057971014, "grad_norm": 0.6682948221798753, "learning_rate": 7.985216519730597e-05, "loss": 0.9407, "step": 1376 }, { "epoch": 0.623641304347826, "grad_norm": 0.8402020026691697, "learning_rate": 7.985107685475058e-05, "loss": 0.9206, "step": 1377 }, { "epoch": 0.6240942028985508, "grad_norm": 0.9645539350689619, "learning_rate": 7.98499845282199e-05, "loss": 0.907, "step": 1378 }, { "epoch": 0.6245471014492754, "grad_norm": 1.1949987657239962, "learning_rate": 7.984888821782312e-05, "loss": 0.9181, "step": 1379 }, { "epoch": 0.625, "grad_norm": 0.7431344217566518, "learning_rate": 7.984778792366983e-05, "loss": 0.9003, "step": 1380 }, { "epoch": 0.6254528985507246, "grad_norm": 0.57432705064337, "learning_rate": 7.984668364587004e-05, "loss": 0.8975, "step": 1381 }, { "epoch": 0.6259057971014492, "grad_norm": 0.5367003418883942, "learning_rate": 7.984557538453415e-05, "loss": 0.9312, "step": 1382 }, { "epoch": 0.626358695652174, "grad_norm": 0.4835619161461902, "learning_rate": 7.984446313977293e-05, "loss": 0.893, "step": 1383 }, { "epoch": 0.6268115942028986, "grad_norm": 0.37426162243566935, "learning_rate": 7.984334691169761e-05, "loss": 0.914, "step": 1384 }, { "epoch": 0.6272644927536232, "grad_norm": 0.4394535015102046, "learning_rate": 7.984222670041975e-05, "loss": 0.9109, "step": 1385 }, { "epoch": 0.6277173913043478, "grad_norm": 0.5747817584045118, "learning_rate": 7.984110250605136e-05, "loss": 0.9121, "step": 1386 }, { "epoch": 0.6281702898550725, "grad_norm": 0.5412624163573944, "learning_rate": 7.983997432870481e-05, "loss": 0.9159, "step": 1387 }, { "epoch": 0.6286231884057971, "grad_norm": 0.5691356139394489, "learning_rate": 7.98388421684929e-05, "loss": 0.9018, "step": 1388 }, { "epoch": 0.6290760869565217, "grad_norm": 0.6148253291061031, "learning_rate": 7.983770602552881e-05, "loss": 0.9113, "step": 1389 }, { "epoch": 0.6295289855072463, "grad_norm": 0.6299250474378052, "learning_rate": 7.983656589992612e-05, "loss": 0.9091, "step": 1390 }, { "epoch": 0.6299818840579711, "grad_norm": 0.44742945044612076, "learning_rate": 7.98354217917988e-05, "loss": 0.9127, "step": 1391 }, { "epoch": 0.6304347826086957, "grad_norm": 0.36612415937606035, "learning_rate": 7.983427370126126e-05, "loss": 0.8927, "step": 1392 }, { "epoch": 0.6308876811594203, "grad_norm": 0.5149784636487857, "learning_rate": 7.983312162842824e-05, "loss": 0.8903, "step": 1393 }, { "epoch": 0.6313405797101449, "grad_norm": 0.5820802244684028, "learning_rate": 7.983196557341493e-05, "loss": 0.9157, "step": 1394 }, { "epoch": 0.6317934782608695, "grad_norm": 0.6798980218452901, "learning_rate": 7.983080553633691e-05, "loss": 0.8895, "step": 1395 }, { "epoch": 0.6322463768115942, "grad_norm": 0.9764452741747511, "learning_rate": 7.982964151731014e-05, "loss": 0.9331, "step": 1396 }, { "epoch": 0.6326992753623188, "grad_norm": 1.3014049572461677, "learning_rate": 7.9828473516451e-05, "loss": 0.898, "step": 1397 }, { "epoch": 0.6331521739130435, "grad_norm": 0.6543786038107252, "learning_rate": 7.982730153387623e-05, "loss": 0.9154, "step": 1398 }, { "epoch": 0.6336050724637681, "grad_norm": 0.4165433945143613, "learning_rate": 7.982612556970302e-05, "loss": 0.917, "step": 1399 }, { "epoch": 0.6340579710144928, "grad_norm": 0.4392809989799357, "learning_rate": 7.982494562404893e-05, "loss": 0.9051, "step": 1400 }, { "epoch": 0.6345108695652174, "grad_norm": 0.6006218530588678, "learning_rate": 7.982376169703192e-05, "loss": 0.9036, "step": 1401 }, { "epoch": 0.634963768115942, "grad_norm": 0.9543120194760116, "learning_rate": 7.982257378877034e-05, "loss": 0.9251, "step": 1402 }, { "epoch": 0.6354166666666666, "grad_norm": 1.2935330269071224, "learning_rate": 7.982138189938296e-05, "loss": 0.9231, "step": 1403 }, { "epoch": 0.6358695652173914, "grad_norm": 0.6364858254678657, "learning_rate": 7.982018602898893e-05, "loss": 0.8961, "step": 1404 }, { "epoch": 0.636322463768116, "grad_norm": 0.5777213074720529, "learning_rate": 7.98189861777078e-05, "loss": 0.907, "step": 1405 }, { "epoch": 0.6367753623188406, "grad_norm": 0.6350609366332146, "learning_rate": 7.981778234565953e-05, "loss": 0.937, "step": 1406 }, { "epoch": 0.6372282608695652, "grad_norm": 0.7466532360645445, "learning_rate": 7.981657453296445e-05, "loss": 0.9048, "step": 1407 }, { "epoch": 0.6376811594202898, "grad_norm": 1.0685984447875374, "learning_rate": 7.981536273974332e-05, "loss": 0.9118, "step": 1408 }, { "epoch": 0.6381340579710145, "grad_norm": 1.1669100957524723, "learning_rate": 7.98141469661173e-05, "loss": 0.9204, "step": 1409 }, { "epoch": 0.6385869565217391, "grad_norm": 0.7191329772878565, "learning_rate": 7.98129272122079e-05, "loss": 0.9208, "step": 1410 }, { "epoch": 0.6390398550724637, "grad_norm": 0.5130232648087233, "learning_rate": 7.981170347813708e-05, "loss": 0.9212, "step": 1411 }, { "epoch": 0.6394927536231884, "grad_norm": 0.40522072642861273, "learning_rate": 7.981047576402718e-05, "loss": 0.9032, "step": 1412 }, { "epoch": 0.6399456521739131, "grad_norm": 0.45541124346471523, "learning_rate": 7.980924407000094e-05, "loss": 0.9423, "step": 1413 }, { "epoch": 0.6403985507246377, "grad_norm": 0.577439983103026, "learning_rate": 7.980800839618149e-05, "loss": 0.914, "step": 1414 }, { "epoch": 0.6408514492753623, "grad_norm": 0.743132827108993, "learning_rate": 7.980676874269235e-05, "loss": 0.8973, "step": 1415 }, { "epoch": 0.6413043478260869, "grad_norm": 0.9622823145764873, "learning_rate": 7.980552510965744e-05, "loss": 0.9102, "step": 1416 }, { "epoch": 0.6417572463768116, "grad_norm": 1.2464780640375734, "learning_rate": 7.980427749720113e-05, "loss": 0.8985, "step": 1417 }, { "epoch": 0.6422101449275363, "grad_norm": 0.6868364238005741, "learning_rate": 7.980302590544812e-05, "loss": 0.9053, "step": 1418 }, { "epoch": 0.6426630434782609, "grad_norm": 0.5175005058808695, "learning_rate": 7.980177033452354e-05, "loss": 0.9159, "step": 1419 }, { "epoch": 0.6431159420289855, "grad_norm": 0.6806324985678436, "learning_rate": 7.980051078455291e-05, "loss": 0.9122, "step": 1420 }, { "epoch": 0.6435688405797102, "grad_norm": 0.6823596866946602, "learning_rate": 7.979924725566214e-05, "loss": 0.9244, "step": 1421 }, { "epoch": 0.6440217391304348, "grad_norm": 0.9032828378713708, "learning_rate": 7.979797974797755e-05, "loss": 0.9175, "step": 1422 }, { "epoch": 0.6444746376811594, "grad_norm": 1.3620205875399645, "learning_rate": 7.979670826162587e-05, "loss": 0.9188, "step": 1423 }, { "epoch": 0.644927536231884, "grad_norm": 0.6352772331660473, "learning_rate": 7.979543279673419e-05, "loss": 0.9275, "step": 1424 }, { "epoch": 0.6453804347826086, "grad_norm": 0.5358670742306314, "learning_rate": 7.979415335343005e-05, "loss": 0.9062, "step": 1425 }, { "epoch": 0.6458333333333334, "grad_norm": 1.0556126907986607, "learning_rate": 7.979286993184134e-05, "loss": 0.9089, "step": 1426 }, { "epoch": 0.646286231884058, "grad_norm": 1.3428307004870488, "learning_rate": 7.979158253209634e-05, "loss": 0.9299, "step": 1427 }, { "epoch": 0.6467391304347826, "grad_norm": 0.389020771308368, "learning_rate": 7.97902911543238e-05, "loss": 0.9015, "step": 1428 }, { "epoch": 0.6471920289855072, "grad_norm": 0.7128004376719899, "learning_rate": 7.97889957986528e-05, "loss": 0.9215, "step": 1429 }, { "epoch": 0.6476449275362319, "grad_norm": 1.2374348493193408, "learning_rate": 7.978769646521283e-05, "loss": 0.9136, "step": 1430 }, { "epoch": 0.6480978260869565, "grad_norm": 0.7851861542137576, "learning_rate": 7.97863931541338e-05, "loss": 0.9267, "step": 1431 }, { "epoch": 0.6485507246376812, "grad_norm": 0.5187825715150548, "learning_rate": 7.978508586554599e-05, "loss": 0.9124, "step": 1432 }, { "epoch": 0.6490036231884058, "grad_norm": 0.5069677790780107, "learning_rate": 7.978377459958012e-05, "loss": 0.8983, "step": 1433 }, { "epoch": 0.6494565217391305, "grad_norm": 0.809853494853227, "learning_rate": 7.978245935636725e-05, "loss": 0.9088, "step": 1434 }, { "epoch": 0.6499094202898551, "grad_norm": 0.9918820974703227, "learning_rate": 7.978114013603888e-05, "loss": 0.9119, "step": 1435 }, { "epoch": 0.6503623188405797, "grad_norm": 0.922659159984025, "learning_rate": 7.977981693872688e-05, "loss": 0.914, "step": 1436 }, { "epoch": 0.6508152173913043, "grad_norm": 0.8598280242491507, "learning_rate": 7.977848976456356e-05, "loss": 0.8972, "step": 1437 }, { "epoch": 0.6512681159420289, "grad_norm": 0.7169874848979553, "learning_rate": 7.977715861368157e-05, "loss": 0.9053, "step": 1438 }, { "epoch": 0.6517210144927537, "grad_norm": 0.5435851510279257, "learning_rate": 7.977582348621402e-05, "loss": 0.9011, "step": 1439 }, { "epoch": 0.6521739130434783, "grad_norm": 0.6067196372885288, "learning_rate": 7.977448438229435e-05, "loss": 0.9477, "step": 1440 }, { "epoch": 0.6526268115942029, "grad_norm": 0.4365173006314243, "learning_rate": 7.977314130205646e-05, "loss": 0.9157, "step": 1441 }, { "epoch": 0.6530797101449275, "grad_norm": 0.39561411868329277, "learning_rate": 7.97717942456346e-05, "loss": 0.9092, "step": 1442 }, { "epoch": 0.6535326086956522, "grad_norm": 0.3638751953698047, "learning_rate": 7.977044321316344e-05, "loss": 0.9191, "step": 1443 }, { "epoch": 0.6539855072463768, "grad_norm": 0.36670939510385075, "learning_rate": 7.976908820477805e-05, "loss": 0.9162, "step": 1444 }, { "epoch": 0.6544384057971014, "grad_norm": 0.37937498935004343, "learning_rate": 7.97677292206139e-05, "loss": 0.9149, "step": 1445 }, { "epoch": 0.654891304347826, "grad_norm": 0.39664295846603526, "learning_rate": 7.976636626080685e-05, "loss": 0.8949, "step": 1446 }, { "epoch": 0.6553442028985508, "grad_norm": 0.39315333836054195, "learning_rate": 7.976499932549314e-05, "loss": 0.9448, "step": 1447 }, { "epoch": 0.6557971014492754, "grad_norm": 0.47445489664362883, "learning_rate": 7.976362841480944e-05, "loss": 0.9185, "step": 1448 }, { "epoch": 0.65625, "grad_norm": 0.45271777340608443, "learning_rate": 7.976225352889278e-05, "loss": 0.9245, "step": 1449 }, { "epoch": 0.6567028985507246, "grad_norm": 0.6187545084907004, "learning_rate": 7.976087466788064e-05, "loss": 0.9347, "step": 1450 }, { "epoch": 0.6571557971014492, "grad_norm": 0.8013878599334633, "learning_rate": 7.975949183191086e-05, "loss": 0.9132, "step": 1451 }, { "epoch": 0.657608695652174, "grad_norm": 0.9501920952488339, "learning_rate": 7.975810502112167e-05, "loss": 0.9251, "step": 1452 }, { "epoch": 0.6580615942028986, "grad_norm": 1.1936809720843113, "learning_rate": 7.975671423565171e-05, "loss": 0.9163, "step": 1453 }, { "epoch": 0.6585144927536232, "grad_norm": 0.9324170695761356, "learning_rate": 7.975531947564005e-05, "loss": 0.9201, "step": 1454 }, { "epoch": 0.6589673913043478, "grad_norm": 0.8101763900802027, "learning_rate": 7.975392074122608e-05, "loss": 0.8969, "step": 1455 }, { "epoch": 0.6594202898550725, "grad_norm": 0.6344799133342441, "learning_rate": 7.975251803254967e-05, "loss": 0.9097, "step": 1456 }, { "epoch": 0.6598731884057971, "grad_norm": 0.5192800310087873, "learning_rate": 7.975111134975104e-05, "loss": 0.9183, "step": 1457 }, { "epoch": 0.6603260869565217, "grad_norm": 0.48766661535710676, "learning_rate": 7.974970069297082e-05, "loss": 0.9076, "step": 1458 }, { "epoch": 0.6607789855072463, "grad_norm": 0.5174058956649525, "learning_rate": 7.974828606235002e-05, "loss": 0.9235, "step": 1459 }, { "epoch": 0.6612318840579711, "grad_norm": 0.5068153065746605, "learning_rate": 7.974686745803008e-05, "loss": 0.9072, "step": 1460 }, { "epoch": 0.6616847826086957, "grad_norm": 0.5992293441741817, "learning_rate": 7.974544488015282e-05, "loss": 0.898, "step": 1461 }, { "epoch": 0.6621376811594203, "grad_norm": 0.7891257816164943, "learning_rate": 7.974401832886046e-05, "loss": 0.935, "step": 1462 }, { "epoch": 0.6625905797101449, "grad_norm": 0.9897537247766093, "learning_rate": 7.974258780429559e-05, "loss": 0.9162, "step": 1463 }, { "epoch": 0.6630434782608695, "grad_norm": 0.9769683147614128, "learning_rate": 7.974115330660125e-05, "loss": 0.9264, "step": 1464 }, { "epoch": 0.6634963768115942, "grad_norm": 0.9444117849929153, "learning_rate": 7.973971483592085e-05, "loss": 0.9141, "step": 1465 }, { "epoch": 0.6639492753623188, "grad_norm": 1.1084785250278741, "learning_rate": 7.973827239239816e-05, "loss": 0.9841, "step": 1466 }, { "epoch": 0.6644021739130435, "grad_norm": 0.7550349265908893, "learning_rate": 7.973682597617743e-05, "loss": 0.9117, "step": 1467 }, { "epoch": 0.6648550724637681, "grad_norm": 0.5651984321653999, "learning_rate": 7.973537558740323e-05, "loss": 0.9057, "step": 1468 }, { "epoch": 0.6653079710144928, "grad_norm": 0.5012547967876538, "learning_rate": 7.973392122622056e-05, "loss": 0.8816, "step": 1469 }, { "epoch": 0.6657608695652174, "grad_norm": 0.44955717567026976, "learning_rate": 7.973246289277483e-05, "loss": 0.8903, "step": 1470 }, { "epoch": 0.666213768115942, "grad_norm": 0.5071182014946579, "learning_rate": 7.973100058721182e-05, "loss": 0.914, "step": 1471 }, { "epoch": 0.6666666666666666, "grad_norm": 0.6363944977761146, "learning_rate": 7.972953430967773e-05, "loss": 0.9116, "step": 1472 }, { "epoch": 0.6671195652173914, "grad_norm": 0.766298778815955, "learning_rate": 7.972806406031914e-05, "loss": 0.9017, "step": 1473 }, { "epoch": 0.667572463768116, "grad_norm": 0.9714936781537574, "learning_rate": 7.972658983928301e-05, "loss": 0.9342, "step": 1474 }, { "epoch": 0.6680253623188406, "grad_norm": 1.0631656258682383, "learning_rate": 7.972511164671675e-05, "loss": 0.9155, "step": 1475 }, { "epoch": 0.6684782608695652, "grad_norm": 0.875609464387028, "learning_rate": 7.972362948276814e-05, "loss": 0.9003, "step": 1476 }, { "epoch": 0.6689311594202898, "grad_norm": 0.7593379235955361, "learning_rate": 7.972214334758534e-05, "loss": 0.9107, "step": 1477 }, { "epoch": 0.6693840579710145, "grad_norm": 0.7732756109279533, "learning_rate": 7.972065324131692e-05, "loss": 0.9217, "step": 1478 }, { "epoch": 0.6698369565217391, "grad_norm": 0.842026544221249, "learning_rate": 7.971915916411185e-05, "loss": 0.9104, "step": 1479 }, { "epoch": 0.6702898550724637, "grad_norm": 0.8025470260141838, "learning_rate": 7.97176611161195e-05, "loss": 0.9015, "step": 1480 }, { "epoch": 0.6707427536231884, "grad_norm": 0.7524496066255576, "learning_rate": 7.971615909748963e-05, "loss": 0.9184, "step": 1481 }, { "epoch": 0.6711956521739131, "grad_norm": 0.6821090514038275, "learning_rate": 7.97146531083724e-05, "loss": 0.9318, "step": 1482 }, { "epoch": 0.6716485507246377, "grad_norm": 0.8088676381380535, "learning_rate": 7.971314314891838e-05, "loss": 0.924, "step": 1483 }, { "epoch": 0.6721014492753623, "grad_norm": 0.8913222346968458, "learning_rate": 7.971162921927848e-05, "loss": 0.9224, "step": 1484 }, { "epoch": 0.6725543478260869, "grad_norm": 0.9227251452617751, "learning_rate": 7.97101113196041e-05, "loss": 0.9144, "step": 1485 }, { "epoch": 0.6730072463768116, "grad_norm": 0.9591148762984631, "learning_rate": 7.970858945004695e-05, "loss": 0.9187, "step": 1486 }, { "epoch": 0.6734601449275363, "grad_norm": 1.0652847624597095, "learning_rate": 7.97070636107592e-05, "loss": 0.9141, "step": 1487 }, { "epoch": 0.6739130434782609, "grad_norm": 0.8874185926812314, "learning_rate": 7.970553380189338e-05, "loss": 0.9064, "step": 1488 }, { "epoch": 0.6743659420289855, "grad_norm": 0.6723299106140963, "learning_rate": 7.970400002360243e-05, "loss": 0.9286, "step": 1489 }, { "epoch": 0.6748188405797102, "grad_norm": 0.42600085341088145, "learning_rate": 7.970246227603966e-05, "loss": 0.9194, "step": 1490 }, { "epoch": 0.6752717391304348, "grad_norm": 0.39410773142900773, "learning_rate": 7.970092055935886e-05, "loss": 0.9251, "step": 1491 }, { "epoch": 0.6757246376811594, "grad_norm": 0.48077648910134946, "learning_rate": 7.969937487371409e-05, "loss": 0.8778, "step": 1492 }, { "epoch": 0.676177536231884, "grad_norm": 0.5184340162297307, "learning_rate": 7.969782521925992e-05, "loss": 0.9153, "step": 1493 }, { "epoch": 0.6766304347826086, "grad_norm": 0.8245794504025894, "learning_rate": 7.969627159615124e-05, "loss": 0.9034, "step": 1494 }, { "epoch": 0.6770833333333334, "grad_norm": 1.1652818244711876, "learning_rate": 7.96947140045434e-05, "loss": 0.9209, "step": 1495 }, { "epoch": 0.677536231884058, "grad_norm": 0.711381186971893, "learning_rate": 7.969315244459209e-05, "loss": 0.8967, "step": 1496 }, { "epoch": 0.6779891304347826, "grad_norm": 0.4093811310002057, "learning_rate": 7.969158691645343e-05, "loss": 0.9182, "step": 1497 }, { "epoch": 0.6784420289855072, "grad_norm": 0.46928801938494524, "learning_rate": 7.969001742028394e-05, "loss": 0.9363, "step": 1498 }, { "epoch": 0.6788949275362319, "grad_norm": 0.5164625694845608, "learning_rate": 7.96884439562405e-05, "loss": 0.8812, "step": 1499 }, { "epoch": 0.6793478260869565, "grad_norm": 0.6930611591948698, "learning_rate": 7.968686652448043e-05, "loss": 0.9051, "step": 1500 }, { "epoch": 0.6798007246376812, "grad_norm": 0.8879333173780413, "learning_rate": 7.968528512516143e-05, "loss": 0.9175, "step": 1501 }, { "epoch": 0.6802536231884058, "grad_norm": 1.0227506427318924, "learning_rate": 7.96836997584416e-05, "loss": 0.9069, "step": 1502 }, { "epoch": 0.6807065217391305, "grad_norm": 1.0524285455350586, "learning_rate": 7.968211042447939e-05, "loss": 0.8978, "step": 1503 }, { "epoch": 0.6811594202898551, "grad_norm": 0.9283468868231672, "learning_rate": 7.968051712343374e-05, "loss": 0.9131, "step": 1504 }, { "epoch": 0.6816123188405797, "grad_norm": 0.7584757138622678, "learning_rate": 7.967891985546391e-05, "loss": 0.8817, "step": 1505 }, { "epoch": 0.6820652173913043, "grad_norm": 0.5667673401648299, "learning_rate": 7.967731862072959e-05, "loss": 0.9019, "step": 1506 }, { "epoch": 0.6825181159420289, "grad_norm": 0.46818677771243444, "learning_rate": 7.967571341939085e-05, "loss": 0.928, "step": 1507 }, { "epoch": 0.6829710144927537, "grad_norm": 0.3947956626336851, "learning_rate": 7.967410425160817e-05, "loss": 0.9284, "step": 1508 }, { "epoch": 0.6834239130434783, "grad_norm": 0.27452716037035263, "learning_rate": 7.96724911175424e-05, "loss": 0.9123, "step": 1509 }, { "epoch": 0.6838768115942029, "grad_norm": 0.33565020596717976, "learning_rate": 7.967087401735486e-05, "loss": 0.9225, "step": 1510 }, { "epoch": 0.6843297101449275, "grad_norm": 0.3516643476715455, "learning_rate": 7.966925295120715e-05, "loss": 0.9033, "step": 1511 }, { "epoch": 0.6847826086956522, "grad_norm": 0.3802834485129694, "learning_rate": 7.966762791926138e-05, "loss": 0.911, "step": 1512 }, { "epoch": 0.6852355072463768, "grad_norm": 0.3926723912389705, "learning_rate": 7.966599892168e-05, "loss": 0.9277, "step": 1513 }, { "epoch": 0.6856884057971014, "grad_norm": 0.3754200079820878, "learning_rate": 7.966436595862583e-05, "loss": 0.9058, "step": 1514 }, { "epoch": 0.686141304347826, "grad_norm": 0.4641207296587077, "learning_rate": 7.966272903026214e-05, "loss": 0.9157, "step": 1515 }, { "epoch": 0.6865942028985508, "grad_norm": 0.6459968335401927, "learning_rate": 7.966108813675258e-05, "loss": 0.9103, "step": 1516 }, { "epoch": 0.6870471014492754, "grad_norm": 0.729263521525964, "learning_rate": 7.965944327826119e-05, "loss": 0.9048, "step": 1517 }, { "epoch": 0.6875, "grad_norm": 0.8238611324023053, "learning_rate": 7.965779445495243e-05, "loss": 0.9205, "step": 1518 }, { "epoch": 0.6879528985507246, "grad_norm": 1.1394084435822398, "learning_rate": 7.965614166699108e-05, "loss": 0.9257, "step": 1519 }, { "epoch": 0.6884057971014492, "grad_norm": 1.01644407351119, "learning_rate": 7.965448491454243e-05, "loss": 0.917, "step": 1520 }, { "epoch": 0.688858695652174, "grad_norm": 0.8632431169704454, "learning_rate": 7.96528241977721e-05, "loss": 0.9158, "step": 1521 }, { "epoch": 0.6893115942028986, "grad_norm": 0.7096870110476933, "learning_rate": 7.965115951684608e-05, "loss": 0.8939, "step": 1522 }, { "epoch": 0.6897644927536232, "grad_norm": 0.678336587308397, "learning_rate": 7.964949087193081e-05, "loss": 0.9212, "step": 1523 }, { "epoch": 0.6902173913043478, "grad_norm": 0.7614187186534526, "learning_rate": 7.96478182631931e-05, "loss": 0.9083, "step": 1524 }, { "epoch": 0.6906702898550725, "grad_norm": 0.8899490009271966, "learning_rate": 7.964614169080018e-05, "loss": 0.8824, "step": 1525 }, { "epoch": 0.6911231884057971, "grad_norm": 1.0042938093475793, "learning_rate": 7.964446115491966e-05, "loss": 0.9094, "step": 1526 }, { "epoch": 0.6915760869565217, "grad_norm": 1.0777176666866677, "learning_rate": 7.964277665571952e-05, "loss": 0.9099, "step": 1527 }, { "epoch": 0.6920289855072463, "grad_norm": 0.8988136206306355, "learning_rate": 7.96410881933682e-05, "loss": 0.9119, "step": 1528 }, { "epoch": 0.6924818840579711, "grad_norm": 0.8313798318399104, "learning_rate": 7.963939576803445e-05, "loss": 0.914, "step": 1529 }, { "epoch": 0.6929347826086957, "grad_norm": 0.6607904749883841, "learning_rate": 7.96376993798875e-05, "loss": 0.8907, "step": 1530 }, { "epoch": 0.6933876811594203, "grad_norm": 0.46874523861507256, "learning_rate": 7.963599902909694e-05, "loss": 0.9031, "step": 1531 }, { "epoch": 0.6938405797101449, "grad_norm": 0.45040529554771064, "learning_rate": 7.963429471583274e-05, "loss": 0.9091, "step": 1532 }, { "epoch": 0.6942934782608695, "grad_norm": 0.4948659683174892, "learning_rate": 7.96325864402653e-05, "loss": 0.903, "step": 1533 }, { "epoch": 0.6947463768115942, "grad_norm": 0.466193700398689, "learning_rate": 7.963087420256538e-05, "loss": 0.9082, "step": 1534 }, { "epoch": 0.6951992753623188, "grad_norm": 0.5592294829420039, "learning_rate": 7.962915800290417e-05, "loss": 0.9118, "step": 1535 }, { "epoch": 0.6956521739130435, "grad_norm": 0.646665213509921, "learning_rate": 7.962743784145323e-05, "loss": 0.8861, "step": 1536 }, { "epoch": 0.6961050724637681, "grad_norm": 0.7294603682284448, "learning_rate": 7.962571371838455e-05, "loss": 0.9152, "step": 1537 }, { "epoch": 0.6965579710144928, "grad_norm": 0.8812470747392689, "learning_rate": 7.962398563387047e-05, "loss": 0.905, "step": 1538 }, { "epoch": 0.6970108695652174, "grad_norm": 0.9566284300813973, "learning_rate": 7.962225358808375e-05, "loss": 0.9146, "step": 1539 }, { "epoch": 0.697463768115942, "grad_norm": 0.9553939796790669, "learning_rate": 7.962051758119756e-05, "loss": 0.9027, "step": 1540 }, { "epoch": 0.6979166666666666, "grad_norm": 1.0070761356757754, "learning_rate": 7.961877761338545e-05, "loss": 0.8996, "step": 1541 }, { "epoch": 0.6983695652173914, "grad_norm": 1.0512747803037867, "learning_rate": 7.961703368482134e-05, "loss": 0.9345, "step": 1542 }, { "epoch": 0.698822463768116, "grad_norm": 0.8220671999468921, "learning_rate": 7.96152857956796e-05, "loss": 0.9163, "step": 1543 }, { "epoch": 0.6992753623188406, "grad_norm": 0.6128259474604768, "learning_rate": 7.961353394613497e-05, "loss": 0.9137, "step": 1544 }, { "epoch": 0.6997282608695652, "grad_norm": 0.5445148256659295, "learning_rate": 7.961177813636258e-05, "loss": 0.9026, "step": 1545 }, { "epoch": 0.7001811594202898, "grad_norm": 0.3769804878158883, "learning_rate": 7.961001836653795e-05, "loss": 0.909, "step": 1546 }, { "epoch": 0.7006340579710145, "grad_norm": 0.2989937040652264, "learning_rate": 7.960825463683702e-05, "loss": 0.9077, "step": 1547 }, { "epoch": 0.7010869565217391, "grad_norm": 0.2758404999876314, "learning_rate": 7.960648694743612e-05, "loss": 0.9019, "step": 1548 }, { "epoch": 0.7015398550724637, "grad_norm": 0.3733078017208397, "learning_rate": 7.960471529851194e-05, "loss": 0.913, "step": 1549 }, { "epoch": 0.7019927536231884, "grad_norm": 0.47408870610612897, "learning_rate": 7.960293969024164e-05, "loss": 0.9331, "step": 1550 }, { "epoch": 0.7024456521739131, "grad_norm": 0.545087372533199, "learning_rate": 7.960116012280267e-05, "loss": 0.8969, "step": 1551 }, { "epoch": 0.7028985507246377, "grad_norm": 0.6550921333279605, "learning_rate": 7.959937659637298e-05, "loss": 0.9121, "step": 1552 }, { "epoch": 0.7033514492753623, "grad_norm": 0.780750100836875, "learning_rate": 7.959758911113087e-05, "loss": 0.9059, "step": 1553 }, { "epoch": 0.7038043478260869, "grad_norm": 1.0074280730456804, "learning_rate": 7.959579766725503e-05, "loss": 0.9005, "step": 1554 }, { "epoch": 0.7042572463768116, "grad_norm": 1.238777060896414, "learning_rate": 7.959400226492455e-05, "loss": 0.9251, "step": 1555 }, { "epoch": 0.7047101449275363, "grad_norm": 0.6213677202846072, "learning_rate": 7.959220290431891e-05, "loss": 0.9085, "step": 1556 }, { "epoch": 0.7051630434782609, "grad_norm": 0.34121336642238503, "learning_rate": 7.959039958561801e-05, "loss": 0.9047, "step": 1557 }, { "epoch": 0.7056159420289855, "grad_norm": 0.6481763583912108, "learning_rate": 7.958859230900214e-05, "loss": 0.9193, "step": 1558 }, { "epoch": 0.7060688405797102, "grad_norm": 0.9949441351862774, "learning_rate": 7.958678107465194e-05, "loss": 0.8963, "step": 1559 }, { "epoch": 0.7065217391304348, "grad_norm": 1.2725457423841513, "learning_rate": 7.958496588274853e-05, "loss": 0.922, "step": 1560 }, { "epoch": 0.7069746376811594, "grad_norm": 0.49103132327011584, "learning_rate": 7.958314673347334e-05, "loss": 0.9088, "step": 1561 }, { "epoch": 0.707427536231884, "grad_norm": 0.7116675356028408, "learning_rate": 7.958132362700826e-05, "loss": 0.9275, "step": 1562 }, { "epoch": 0.7078804347826086, "grad_norm": 1.1744286451931185, "learning_rate": 7.957949656353551e-05, "loss": 0.9393, "step": 1563 }, { "epoch": 0.7083333333333334, "grad_norm": 0.938221045754262, "learning_rate": 7.957766554323778e-05, "loss": 0.9116, "step": 1564 }, { "epoch": 0.708786231884058, "grad_norm": 0.8068109245316364, "learning_rate": 7.957583056629812e-05, "loss": 0.9078, "step": 1565 }, { "epoch": 0.7092391304347826, "grad_norm": 0.7928669265959305, "learning_rate": 7.957399163289996e-05, "loss": 0.9026, "step": 1566 }, { "epoch": 0.7096920289855072, "grad_norm": 0.6826130943828623, "learning_rate": 7.957214874322714e-05, "loss": 0.9049, "step": 1567 }, { "epoch": 0.7101449275362319, "grad_norm": 0.5551390831505416, "learning_rate": 7.957030189746391e-05, "loss": 0.8962, "step": 1568 }, { "epoch": 0.7105978260869565, "grad_norm": 0.45684729018259373, "learning_rate": 7.956845109579488e-05, "loss": 0.93, "step": 1569 }, { "epoch": 0.7110507246376812, "grad_norm": 0.4116226170619623, "learning_rate": 7.956659633840512e-05, "loss": 0.8843, "step": 1570 }, { "epoch": 0.7115036231884058, "grad_norm": 0.46756074670627523, "learning_rate": 7.956473762548001e-05, "loss": 0.9012, "step": 1571 }, { "epoch": 0.7119565217391305, "grad_norm": 0.36754299063604673, "learning_rate": 7.956287495720538e-05, "loss": 0.9046, "step": 1572 }, { "epoch": 0.7124094202898551, "grad_norm": 0.3722578373283463, "learning_rate": 7.956100833376745e-05, "loss": 0.8976, "step": 1573 }, { "epoch": 0.7128623188405797, "grad_norm": 0.4525098501581898, "learning_rate": 7.955913775535282e-05, "loss": 0.9105, "step": 1574 }, { "epoch": 0.7133152173913043, "grad_norm": 0.5665674697221582, "learning_rate": 7.955726322214852e-05, "loss": 0.9235, "step": 1575 }, { "epoch": 0.7137681159420289, "grad_norm": 0.7747358362174089, "learning_rate": 7.955538473434191e-05, "loss": 0.9179, "step": 1576 }, { "epoch": 0.7142210144927537, "grad_norm": 1.011192904419723, "learning_rate": 7.955350229212082e-05, "loss": 0.9118, "step": 1577 }, { "epoch": 0.7146739130434783, "grad_norm": 0.9571779029254225, "learning_rate": 7.955161589567343e-05, "loss": 0.913, "step": 1578 }, { "epoch": 0.7151268115942029, "grad_norm": 1.0846982598185524, "learning_rate": 7.954972554518832e-05, "loss": 0.9159, "step": 1579 }, { "epoch": 0.7155797101449275, "grad_norm": 0.9487266404901215, "learning_rate": 7.954783124085447e-05, "loss": 0.9083, "step": 1580 }, { "epoch": 0.7160326086956522, "grad_norm": 0.8785488198843768, "learning_rate": 7.954593298286128e-05, "loss": 0.9002, "step": 1581 }, { "epoch": 0.7164855072463768, "grad_norm": 0.7323769966610493, "learning_rate": 7.954403077139849e-05, "loss": 0.9024, "step": 1582 }, { "epoch": 0.7169384057971014, "grad_norm": 0.5805490031549809, "learning_rate": 7.954212460665628e-05, "loss": 0.9003, "step": 1583 }, { "epoch": 0.717391304347826, "grad_norm": 0.4989972818673647, "learning_rate": 7.954021448882522e-05, "loss": 0.9248, "step": 1584 }, { "epoch": 0.7178442028985508, "grad_norm": 0.42102955270713294, "learning_rate": 7.953830041809625e-05, "loss": 0.8984, "step": 1585 }, { "epoch": 0.7182971014492754, "grad_norm": 0.41173892913444277, "learning_rate": 7.953638239466074e-05, "loss": 0.898, "step": 1586 }, { "epoch": 0.71875, "grad_norm": 0.5608221826206937, "learning_rate": 7.953446041871044e-05, "loss": 0.9115, "step": 1587 }, { "epoch": 0.7192028985507246, "grad_norm": 0.8143310241706798, "learning_rate": 7.953253449043747e-05, "loss": 0.8992, "step": 1588 }, { "epoch": 0.7196557971014492, "grad_norm": 1.0568871970220413, "learning_rate": 7.95306046100344e-05, "loss": 0.9048, "step": 1589 }, { "epoch": 0.720108695652174, "grad_norm": 1.048485673656128, "learning_rate": 7.952867077769412e-05, "loss": 0.9424, "step": 1590 }, { "epoch": 0.7205615942028986, "grad_norm": 0.8541497724003552, "learning_rate": 7.952673299360999e-05, "loss": 0.9205, "step": 1591 }, { "epoch": 0.7210144927536232, "grad_norm": 0.7271022398036845, "learning_rate": 7.952479125797575e-05, "loss": 0.899, "step": 1592 }, { "epoch": 0.7214673913043478, "grad_norm": 0.5887080607234358, "learning_rate": 7.952284557098547e-05, "loss": 0.9233, "step": 1593 }, { "epoch": 0.7219202898550725, "grad_norm": 0.5102618608079893, "learning_rate": 7.952089593283371e-05, "loss": 0.9073, "step": 1594 }, { "epoch": 0.7223731884057971, "grad_norm": 0.3693915835088139, "learning_rate": 7.951894234371534e-05, "loss": 0.9146, "step": 1595 }, { "epoch": 0.7228260869565217, "grad_norm": 0.4858067969648147, "learning_rate": 7.951698480382569e-05, "loss": 0.9059, "step": 1596 }, { "epoch": 0.7232789855072463, "grad_norm": 0.7637458297968653, "learning_rate": 7.951502331336044e-05, "loss": 0.9092, "step": 1597 }, { "epoch": 0.7237318840579711, "grad_norm": 0.8661341418416378, "learning_rate": 7.95130578725157e-05, "loss": 0.9245, "step": 1598 }, { "epoch": 0.7241847826086957, "grad_norm": 0.946917004358759, "learning_rate": 7.951108848148794e-05, "loss": 0.9204, "step": 1599 }, { "epoch": 0.7246376811594203, "grad_norm": 1.0409768254705527, "learning_rate": 7.950911514047407e-05, "loss": 0.9043, "step": 1600 }, { "epoch": 0.7250905797101449, "grad_norm": 0.8640417443639444, "learning_rate": 7.950713784967134e-05, "loss": 0.9143, "step": 1601 }, { "epoch": 0.7255434782608695, "grad_norm": 0.6102348267639451, "learning_rate": 7.950515660927745e-05, "loss": 0.9168, "step": 1602 }, { "epoch": 0.7259963768115942, "grad_norm": 0.42833105837335644, "learning_rate": 7.950317141949043e-05, "loss": 0.9192, "step": 1603 }, { "epoch": 0.7264492753623188, "grad_norm": 0.5088645596914736, "learning_rate": 7.950118228050879e-05, "loss": 0.941, "step": 1604 }, { "epoch": 0.7269021739130435, "grad_norm": 0.5750486607299815, "learning_rate": 7.949918919253134e-05, "loss": 0.8997, "step": 1605 }, { "epoch": 0.7273550724637681, "grad_norm": 0.7052717098825094, "learning_rate": 7.949719215575737e-05, "loss": 0.924, "step": 1606 }, { "epoch": 0.7278079710144928, "grad_norm": 0.6988369880160127, "learning_rate": 7.94951911703865e-05, "loss": 0.9289, "step": 1607 }, { "epoch": 0.7282608695652174, "grad_norm": 0.6916911147078508, "learning_rate": 7.94931862366188e-05, "loss": 0.9181, "step": 1608 }, { "epoch": 0.728713768115942, "grad_norm": 0.6777347957249225, "learning_rate": 7.949117735465467e-05, "loss": 0.8956, "step": 1609 }, { "epoch": 0.7291666666666666, "grad_norm": 0.6853048651905429, "learning_rate": 7.948916452469497e-05, "loss": 0.8989, "step": 1610 }, { "epoch": 0.7296195652173914, "grad_norm": 0.62042635990192, "learning_rate": 7.948714774694092e-05, "loss": 0.8931, "step": 1611 }, { "epoch": 0.730072463768116, "grad_norm": 0.5822174701551568, "learning_rate": 7.948512702159413e-05, "loss": 0.9375, "step": 1612 }, { "epoch": 0.7305253623188406, "grad_norm": 0.4789889744145628, "learning_rate": 7.948310234885661e-05, "loss": 0.9126, "step": 1613 }, { "epoch": 0.7309782608695652, "grad_norm": 0.46721068720303716, "learning_rate": 7.94810737289308e-05, "loss": 0.9271, "step": 1614 }, { "epoch": 0.7314311594202898, "grad_norm": 0.3875258666716423, "learning_rate": 7.94790411620195e-05, "loss": 0.9129, "step": 1615 }, { "epoch": 0.7318840579710145, "grad_norm": 0.45895317086001663, "learning_rate": 7.947700464832587e-05, "loss": 0.8892, "step": 1616 }, { "epoch": 0.7323369565217391, "grad_norm": 0.5085101431327985, "learning_rate": 7.947496418805353e-05, "loss": 0.9081, "step": 1617 }, { "epoch": 0.7327898550724637, "grad_norm": 0.5472440716188616, "learning_rate": 7.947291978140648e-05, "loss": 0.8981, "step": 1618 }, { "epoch": 0.7332427536231884, "grad_norm": 0.5096471816948581, "learning_rate": 7.947087142858909e-05, "loss": 0.9028, "step": 1619 }, { "epoch": 0.7336956521739131, "grad_norm": 0.5508633397926751, "learning_rate": 7.946881912980612e-05, "loss": 0.9111, "step": 1620 }, { "epoch": 0.7341485507246377, "grad_norm": 0.6286627415956578, "learning_rate": 7.946676288526276e-05, "loss": 0.9329, "step": 1621 }, { "epoch": 0.7346014492753623, "grad_norm": 0.6662093921374047, "learning_rate": 7.94647026951646e-05, "loss": 0.9135, "step": 1622 }, { "epoch": 0.7350543478260869, "grad_norm": 0.6628181564091136, "learning_rate": 7.946263855971755e-05, "loss": 0.9154, "step": 1623 }, { "epoch": 0.7355072463768116, "grad_norm": 0.7054345561448908, "learning_rate": 7.946057047912798e-05, "loss": 0.9142, "step": 1624 }, { "epoch": 0.7359601449275363, "grad_norm": 0.6479105466578922, "learning_rate": 7.945849845360266e-05, "loss": 0.9013, "step": 1625 }, { "epoch": 0.7364130434782609, "grad_norm": 0.6767934774869268, "learning_rate": 7.945642248334873e-05, "loss": 0.9163, "step": 1626 }, { "epoch": 0.7368659420289855, "grad_norm": 0.7908998572949842, "learning_rate": 7.945434256857372e-05, "loss": 0.9371, "step": 1627 }, { "epoch": 0.7373188405797102, "grad_norm": 0.9434814484250643, "learning_rate": 7.945225870948555e-05, "loss": 0.9107, "step": 1628 }, { "epoch": 0.7377717391304348, "grad_norm": 1.0181741566463174, "learning_rate": 7.945017090629257e-05, "loss": 0.9005, "step": 1629 }, { "epoch": 0.7382246376811594, "grad_norm": 0.9576067453210506, "learning_rate": 7.944807915920349e-05, "loss": 0.924, "step": 1630 }, { "epoch": 0.738677536231884, "grad_norm": 0.9312414986691083, "learning_rate": 7.944598346842743e-05, "loss": 0.9238, "step": 1631 }, { "epoch": 0.7391304347826086, "grad_norm": 0.9094081759074338, "learning_rate": 7.944388383417388e-05, "loss": 0.9045, "step": 1632 }, { "epoch": 0.7395833333333334, "grad_norm": 0.8045049179216661, "learning_rate": 7.944178025665277e-05, "loss": 0.911, "step": 1633 }, { "epoch": 0.740036231884058, "grad_norm": 0.6702505806055573, "learning_rate": 7.943967273607439e-05, "loss": 0.8932, "step": 1634 }, { "epoch": 0.7404891304347826, "grad_norm": 0.484362840438258, "learning_rate": 7.943756127264942e-05, "loss": 0.9049, "step": 1635 }, { "epoch": 0.7409420289855072, "grad_norm": 0.36714189046569123, "learning_rate": 7.943544586658896e-05, "loss": 0.9125, "step": 1636 }, { "epoch": 0.7413949275362319, "grad_norm": 0.34614855580156034, "learning_rate": 7.943332651810448e-05, "loss": 0.8893, "step": 1637 }, { "epoch": 0.7418478260869565, "grad_norm": 0.3867866318411081, "learning_rate": 7.943120322740786e-05, "loss": 0.908, "step": 1638 }, { "epoch": 0.7423007246376812, "grad_norm": 0.57715311533328, "learning_rate": 7.942907599471138e-05, "loss": 0.907, "step": 1639 }, { "epoch": 0.7427536231884058, "grad_norm": 0.7988915937477076, "learning_rate": 7.942694482022768e-05, "loss": 0.9189, "step": 1640 }, { "epoch": 0.7432065217391305, "grad_norm": 0.9291050171471364, "learning_rate": 7.942480970416985e-05, "loss": 0.9194, "step": 1641 }, { "epoch": 0.7436594202898551, "grad_norm": 0.8846625116272352, "learning_rate": 7.94226706467513e-05, "loss": 0.9402, "step": 1642 }, { "epoch": 0.7441123188405797, "grad_norm": 0.9256335782126947, "learning_rate": 7.94205276481859e-05, "loss": 0.9229, "step": 1643 }, { "epoch": 0.7445652173913043, "grad_norm": 1.0130782858038867, "learning_rate": 7.941838070868787e-05, "loss": 0.9108, "step": 1644 }, { "epoch": 0.7450181159420289, "grad_norm": 1.1289449284818045, "learning_rate": 7.941622982847189e-05, "loss": 0.9374, "step": 1645 }, { "epoch": 0.7454710144927537, "grad_norm": 0.6774084024571004, "learning_rate": 7.941407500775292e-05, "loss": 0.9124, "step": 1646 }, { "epoch": 0.7459239130434783, "grad_norm": 0.31275301497705005, "learning_rate": 7.941191624674643e-05, "loss": 0.8801, "step": 1647 }, { "epoch": 0.7463768115942029, "grad_norm": 0.6889851934293386, "learning_rate": 7.940975354566823e-05, "loss": 0.923, "step": 1648 }, { "epoch": 0.7468297101449275, "grad_norm": 1.0430660176240623, "learning_rate": 7.940758690473451e-05, "loss": 0.9205, "step": 1649 }, { "epoch": 0.7472826086956522, "grad_norm": 1.027097021737211, "learning_rate": 7.940541632416187e-05, "loss": 0.9025, "step": 1650 }, { "epoch": 0.7477355072463768, "grad_norm": 0.8678632931388098, "learning_rate": 7.940324180416732e-05, "loss": 0.9198, "step": 1651 }, { "epoch": 0.7481884057971014, "grad_norm": 0.7444315551754055, "learning_rate": 7.940106334496825e-05, "loss": 0.9055, "step": 1652 }, { "epoch": 0.748641304347826, "grad_norm": 0.5977971572693496, "learning_rate": 7.939888094678244e-05, "loss": 0.9237, "step": 1653 }, { "epoch": 0.7490942028985508, "grad_norm": 0.5383875169128501, "learning_rate": 7.939669460982807e-05, "loss": 0.9211, "step": 1654 }, { "epoch": 0.7495471014492754, "grad_norm": 0.42212930778901203, "learning_rate": 7.939450433432371e-05, "loss": 0.9069, "step": 1655 }, { "epoch": 0.75, "grad_norm": 0.38679647489582186, "learning_rate": 7.939231012048833e-05, "loss": 0.8933, "step": 1656 }, { "epoch": 0.7504528985507246, "grad_norm": 0.4951817055077575, "learning_rate": 7.939011196854129e-05, "loss": 0.9088, "step": 1657 }, { "epoch": 0.7509057971014492, "grad_norm": 0.8687550741736288, "learning_rate": 7.938790987870232e-05, "loss": 0.8957, "step": 1658 }, { "epoch": 0.751358695652174, "grad_norm": 1.2513618695543318, "learning_rate": 7.938570385119159e-05, "loss": 0.9372, "step": 1659 }, { "epoch": 0.7518115942028986, "grad_norm": 0.7375824206416562, "learning_rate": 7.938349388622964e-05, "loss": 0.904, "step": 1660 }, { "epoch": 0.7522644927536232, "grad_norm": 0.7364618009408264, "learning_rate": 7.938127998403738e-05, "loss": 0.8837, "step": 1661 }, { "epoch": 0.7527173913043478, "grad_norm": 0.8006549153296776, "learning_rate": 7.937906214483618e-05, "loss": 0.9201, "step": 1662 }, { "epoch": 0.7531702898550725, "grad_norm": 0.7330740713894062, "learning_rate": 7.937684036884772e-05, "loss": 0.9126, "step": 1663 }, { "epoch": 0.7536231884057971, "grad_norm": 0.7951770228523185, "learning_rate": 7.937461465629414e-05, "loss": 0.9173, "step": 1664 }, { "epoch": 0.7540760869565217, "grad_norm": 1.2220425174123282, "learning_rate": 7.937238500739794e-05, "loss": 0.8906, "step": 1665 }, { "epoch": 0.7545289855072463, "grad_norm": 0.8574768471111481, "learning_rate": 7.9370151422382e-05, "loss": 0.9005, "step": 1666 }, { "epoch": 0.7549818840579711, "grad_norm": 0.6242481804533728, "learning_rate": 7.936791390146965e-05, "loss": 0.927, "step": 1667 }, { "epoch": 0.7554347826086957, "grad_norm": 0.43544943634187494, "learning_rate": 7.936567244488458e-05, "loss": 0.9028, "step": 1668 }, { "epoch": 0.7558876811594203, "grad_norm": 0.41494760856037194, "learning_rate": 7.936342705285084e-05, "loss": 0.9348, "step": 1669 }, { "epoch": 0.7563405797101449, "grad_norm": 0.4997672671402251, "learning_rate": 7.936117772559291e-05, "loss": 0.9009, "step": 1670 }, { "epoch": 0.7567934782608695, "grad_norm": 0.4700668411826192, "learning_rate": 7.935892446333568e-05, "loss": 0.9194, "step": 1671 }, { "epoch": 0.7572463768115942, "grad_norm": 0.5086403978230015, "learning_rate": 7.935666726630439e-05, "loss": 0.912, "step": 1672 }, { "epoch": 0.7576992753623188, "grad_norm": 0.6347166376076082, "learning_rate": 7.935440613472473e-05, "loss": 0.9084, "step": 1673 }, { "epoch": 0.7581521739130435, "grad_norm": 0.7479021852017442, "learning_rate": 7.935214106882272e-05, "loss": 0.9088, "step": 1674 }, { "epoch": 0.7586050724637681, "grad_norm": 0.83700500045262, "learning_rate": 7.934987206882479e-05, "loss": 0.9171, "step": 1675 }, { "epoch": 0.7590579710144928, "grad_norm": 0.9768437020107367, "learning_rate": 7.93475991349578e-05, "loss": 0.9261, "step": 1676 }, { "epoch": 0.7595108695652174, "grad_norm": 1.1810224617575313, "learning_rate": 7.9345322267449e-05, "loss": 0.9152, "step": 1677 }, { "epoch": 0.759963768115942, "grad_norm": 0.8511117704469221, "learning_rate": 7.934304146652596e-05, "loss": 0.9233, "step": 1678 }, { "epoch": 0.7604166666666666, "grad_norm": 0.6470957003358129, "learning_rate": 7.934075673241672e-05, "loss": 0.9068, "step": 1679 }, { "epoch": 0.7608695652173914, "grad_norm": 0.3992790109861137, "learning_rate": 7.93384680653497e-05, "loss": 0.8995, "step": 1680 }, { "epoch": 0.761322463768116, "grad_norm": 0.43818835502758186, "learning_rate": 7.933617546555367e-05, "loss": 0.9217, "step": 1681 }, { "epoch": 0.7617753623188406, "grad_norm": 0.61110582131498, "learning_rate": 7.933387893325786e-05, "loss": 0.9029, "step": 1682 }, { "epoch": 0.7622282608695652, "grad_norm": 0.6683548212792629, "learning_rate": 7.933157846869184e-05, "loss": 0.9128, "step": 1683 }, { "epoch": 0.7626811594202898, "grad_norm": 0.7286915068318099, "learning_rate": 7.932927407208561e-05, "loss": 0.891, "step": 1684 }, { "epoch": 0.7631340579710145, "grad_norm": 0.8789706037212018, "learning_rate": 7.932696574366952e-05, "loss": 0.9079, "step": 1685 }, { "epoch": 0.7635869565217391, "grad_norm": 1.1926023901630343, "learning_rate": 7.932465348367432e-05, "loss": 0.8889, "step": 1686 }, { "epoch": 0.7640398550724637, "grad_norm": 0.9932141973383182, "learning_rate": 7.932233729233123e-05, "loss": 0.9093, "step": 1687 }, { "epoch": 0.7644927536231884, "grad_norm": 0.9925038471074212, "learning_rate": 7.932001716987175e-05, "loss": 0.9275, "step": 1688 }, { "epoch": 0.7649456521739131, "grad_norm": 1.0513645585552587, "learning_rate": 7.931769311652785e-05, "loss": 0.8867, "step": 1689 }, { "epoch": 0.7653985507246377, "grad_norm": 0.771301215190776, "learning_rate": 7.931536513253186e-05, "loss": 0.8874, "step": 1690 }, { "epoch": 0.7658514492753623, "grad_norm": 0.6254962126561019, "learning_rate": 7.931303321811654e-05, "loss": 0.9132, "step": 1691 }, { "epoch": 0.7663043478260869, "grad_norm": 0.6811937707941594, "learning_rate": 7.931069737351497e-05, "loss": 0.931, "step": 1692 }, { "epoch": 0.7667572463768116, "grad_norm": 0.9075356959863129, "learning_rate": 7.930835759896069e-05, "loss": 0.9097, "step": 1693 }, { "epoch": 0.7672101449275363, "grad_norm": 0.9986590784862996, "learning_rate": 7.930601389468761e-05, "loss": 0.9107, "step": 1694 }, { "epoch": 0.7676630434782609, "grad_norm": 0.9913517494490376, "learning_rate": 7.930366626093005e-05, "loss": 0.8778, "step": 1695 }, { "epoch": 0.7681159420289855, "grad_norm": 0.9719981742322674, "learning_rate": 7.930131469792269e-05, "loss": 0.9136, "step": 1696 }, { "epoch": 0.7685688405797102, "grad_norm": 0.8307224628252935, "learning_rate": 7.92989592059006e-05, "loss": 0.9001, "step": 1697 }, { "epoch": 0.7690217391304348, "grad_norm": 0.9141393453191868, "learning_rate": 7.929659978509931e-05, "loss": 0.9116, "step": 1698 }, { "epoch": 0.7694746376811594, "grad_norm": 1.189051580862336, "learning_rate": 7.929423643575466e-05, "loss": 0.8986, "step": 1699 }, { "epoch": 0.769927536231884, "grad_norm": 0.6954226104908785, "learning_rate": 7.929186915810292e-05, "loss": 0.8951, "step": 1700 }, { "epoch": 0.7703804347826086, "grad_norm": 0.38516448122275, "learning_rate": 7.928949795238076e-05, "loss": 0.9019, "step": 1701 }, { "epoch": 0.7708333333333334, "grad_norm": 0.5751852856772914, "learning_rate": 7.928712281882523e-05, "loss": 0.8909, "step": 1702 }, { "epoch": 0.771286231884058, "grad_norm": 0.896071236667014, "learning_rate": 7.928474375767378e-05, "loss": 0.9305, "step": 1703 }, { "epoch": 0.7717391304347826, "grad_norm": 1.1452576635206257, "learning_rate": 7.928236076916423e-05, "loss": 0.9179, "step": 1704 }, { "epoch": 0.7721920289855072, "grad_norm": 0.7008751965958819, "learning_rate": 7.927997385353485e-05, "loss": 0.889, "step": 1705 }, { "epoch": 0.7726449275362319, "grad_norm": 0.4939426786301924, "learning_rate": 7.927758301102423e-05, "loss": 0.9204, "step": 1706 }, { "epoch": 0.7730978260869565, "grad_norm": 0.4986018205499359, "learning_rate": 7.927518824187139e-05, "loss": 0.8932, "step": 1707 }, { "epoch": 0.7735507246376812, "grad_norm": 0.6061207445772376, "learning_rate": 7.927278954631576e-05, "loss": 0.9085, "step": 1708 }, { "epoch": 0.7740036231884058, "grad_norm": 0.6537540737275058, "learning_rate": 7.927038692459712e-05, "loss": 0.9257, "step": 1709 }, { "epoch": 0.7744565217391305, "grad_norm": 0.7729543571397196, "learning_rate": 7.926798037695564e-05, "loss": 0.918, "step": 1710 }, { "epoch": 0.7749094202898551, "grad_norm": 0.9531458236039817, "learning_rate": 7.926556990363196e-05, "loss": 0.9054, "step": 1711 }, { "epoch": 0.7753623188405797, "grad_norm": 0.924839785261663, "learning_rate": 7.926315550486704e-05, "loss": 0.9066, "step": 1712 }, { "epoch": 0.7758152173913043, "grad_norm": 0.865447122208724, "learning_rate": 7.926073718090226e-05, "loss": 0.8979, "step": 1713 }, { "epoch": 0.7762681159420289, "grad_norm": 0.8760609243768027, "learning_rate": 7.925831493197933e-05, "loss": 0.8896, "step": 1714 }, { "epoch": 0.7767210144927537, "grad_norm": 0.8304222526506398, "learning_rate": 7.925588875834047e-05, "loss": 0.9139, "step": 1715 }, { "epoch": 0.7771739130434783, "grad_norm": 0.7949617190102332, "learning_rate": 7.92534586602282e-05, "loss": 0.9156, "step": 1716 }, { "epoch": 0.7776268115942029, "grad_norm": 0.8690276160818412, "learning_rate": 7.925102463788547e-05, "loss": 0.913, "step": 1717 }, { "epoch": 0.7780797101449275, "grad_norm": 0.9282937643175938, "learning_rate": 7.92485866915556e-05, "loss": 0.9105, "step": 1718 }, { "epoch": 0.7785326086956522, "grad_norm": 0.7980479663320144, "learning_rate": 7.924614482148233e-05, "loss": 0.8925, "step": 1719 }, { "epoch": 0.7789855072463768, "grad_norm": 0.6753921647060465, "learning_rate": 7.924369902790978e-05, "loss": 0.9304, "step": 1720 }, { "epoch": 0.7794384057971014, "grad_norm": 0.5651067361823358, "learning_rate": 7.924124931108245e-05, "loss": 0.8949, "step": 1721 }, { "epoch": 0.779891304347826, "grad_norm": 0.4467073549953192, "learning_rate": 7.923879567124523e-05, "loss": 0.9173, "step": 1722 }, { "epoch": 0.7803442028985508, "grad_norm": 0.4097555952402969, "learning_rate": 7.923633810864342e-05, "loss": 0.89, "step": 1723 }, { "epoch": 0.7807971014492754, "grad_norm": 0.44383072075798197, "learning_rate": 7.923387662352273e-05, "loss": 0.895, "step": 1724 }, { "epoch": 0.78125, "grad_norm": 0.46538637773598585, "learning_rate": 7.923141121612922e-05, "loss": 0.9218, "step": 1725 }, { "epoch": 0.7817028985507246, "grad_norm": 0.5927686973258744, "learning_rate": 7.922894188670937e-05, "loss": 0.8888, "step": 1726 }, { "epoch": 0.7821557971014492, "grad_norm": 0.7288169572542698, "learning_rate": 7.922646863551003e-05, "loss": 0.9167, "step": 1727 }, { "epoch": 0.782608695652174, "grad_norm": 0.828241641478681, "learning_rate": 7.922399146277846e-05, "loss": 0.8945, "step": 1728 }, { "epoch": 0.7830615942028986, "grad_norm": 0.9540543898367853, "learning_rate": 7.92215103687623e-05, "loss": 0.919, "step": 1729 }, { "epoch": 0.7835144927536232, "grad_norm": 1.1713385455371776, "learning_rate": 7.92190253537096e-05, "loss": 0.9089, "step": 1730 }, { "epoch": 0.7839673913043478, "grad_norm": 0.7690162579806606, "learning_rate": 7.921653641786879e-05, "loss": 0.9164, "step": 1731 }, { "epoch": 0.7844202898550725, "grad_norm": 0.4542334304678735, "learning_rate": 7.92140435614887e-05, "loss": 0.9093, "step": 1732 }, { "epoch": 0.7848731884057971, "grad_norm": 0.3571847819321863, "learning_rate": 7.921154678481852e-05, "loss": 0.8781, "step": 1733 }, { "epoch": 0.7853260869565217, "grad_norm": 0.6256971481241971, "learning_rate": 7.920904608810788e-05, "loss": 0.9001, "step": 1734 }, { "epoch": 0.7857789855072463, "grad_norm": 0.7925324857413506, "learning_rate": 7.920654147160678e-05, "loss": 0.917, "step": 1735 }, { "epoch": 0.7862318840579711, "grad_norm": 0.8820164799207086, "learning_rate": 7.920403293556558e-05, "loss": 0.9001, "step": 1736 }, { "epoch": 0.7866847826086957, "grad_norm": 0.9255824478095918, "learning_rate": 7.920152048023509e-05, "loss": 0.9226, "step": 1737 }, { "epoch": 0.7871376811594203, "grad_norm": 0.9344236277237731, "learning_rate": 7.919900410586648e-05, "loss": 0.9113, "step": 1738 }, { "epoch": 0.7875905797101449, "grad_norm": 0.8722560543509953, "learning_rate": 7.919648381271132e-05, "loss": 0.9067, "step": 1739 }, { "epoch": 0.7880434782608695, "grad_norm": 0.6740005384512736, "learning_rate": 7.919395960102155e-05, "loss": 0.908, "step": 1740 }, { "epoch": 0.7884963768115942, "grad_norm": 0.5147739876346827, "learning_rate": 7.919143147104954e-05, "loss": 0.9157, "step": 1741 }, { "epoch": 0.7889492753623188, "grad_norm": 0.4380010710160886, "learning_rate": 7.918889942304802e-05, "loss": 0.8969, "step": 1742 }, { "epoch": 0.7894021739130435, "grad_norm": 0.4257730984348674, "learning_rate": 7.918636345727012e-05, "loss": 0.9116, "step": 1743 }, { "epoch": 0.7898550724637681, "grad_norm": 0.4025645689162029, "learning_rate": 7.918382357396937e-05, "loss": 0.9133, "step": 1744 }, { "epoch": 0.7903079710144928, "grad_norm": 0.3607536223324417, "learning_rate": 7.918127977339969e-05, "loss": 0.8949, "step": 1745 }, { "epoch": 0.7907608695652174, "grad_norm": 0.46195270940299876, "learning_rate": 7.917873205581539e-05, "loss": 0.9099, "step": 1746 }, { "epoch": 0.791213768115942, "grad_norm": 0.5096600625799307, "learning_rate": 7.917618042147114e-05, "loss": 0.8821, "step": 1747 }, { "epoch": 0.7916666666666666, "grad_norm": 0.4834449141561264, "learning_rate": 7.917362487062207e-05, "loss": 0.9209, "step": 1748 }, { "epoch": 0.7921195652173914, "grad_norm": 0.5332837623810226, "learning_rate": 7.917106540352364e-05, "loss": 0.9081, "step": 1749 }, { "epoch": 0.792572463768116, "grad_norm": 0.6621229799788382, "learning_rate": 7.916850202043172e-05, "loss": 0.8986, "step": 1750 }, { "epoch": 0.7930253623188406, "grad_norm": 0.8167847187326669, "learning_rate": 7.91659347216026e-05, "loss": 0.9061, "step": 1751 }, { "epoch": 0.7934782608695652, "grad_norm": 0.9937772703996423, "learning_rate": 7.916336350729293e-05, "loss": 0.9319, "step": 1752 }, { "epoch": 0.7939311594202898, "grad_norm": 1.0615591347355875, "learning_rate": 7.916078837775973e-05, "loss": 0.9133, "step": 1753 }, { "epoch": 0.7943840579710145, "grad_norm": 0.9004187141138174, "learning_rate": 7.915820933326046e-05, "loss": 0.9296, "step": 1754 }, { "epoch": 0.7948369565217391, "grad_norm": 0.7896824875758779, "learning_rate": 7.915562637405296e-05, "loss": 0.94, "step": 1755 }, { "epoch": 0.7952898550724637, "grad_norm": 0.7929267705956629, "learning_rate": 7.915303950039545e-05, "loss": 0.9174, "step": 1756 }, { "epoch": 0.7957427536231884, "grad_norm": 0.8008007186091725, "learning_rate": 7.915044871254653e-05, "loss": 0.9082, "step": 1757 }, { "epoch": 0.7961956521739131, "grad_norm": 0.8338913046369028, "learning_rate": 7.914785401076522e-05, "loss": 0.9082, "step": 1758 }, { "epoch": 0.7966485507246377, "grad_norm": 0.7114145283222018, "learning_rate": 7.91452553953109e-05, "loss": 0.8769, "step": 1759 }, { "epoch": 0.7971014492753623, "grad_norm": 0.5400623684345788, "learning_rate": 7.914265286644337e-05, "loss": 0.8914, "step": 1760 }, { "epoch": 0.7975543478260869, "grad_norm": 0.5064579479257022, "learning_rate": 7.914004642442281e-05, "loss": 0.9281, "step": 1761 }, { "epoch": 0.7980072463768116, "grad_norm": 0.5046736383965797, "learning_rate": 7.913743606950979e-05, "loss": 0.914, "step": 1762 }, { "epoch": 0.7984601449275363, "grad_norm": 0.6557155212169683, "learning_rate": 7.913482180196526e-05, "loss": 0.9139, "step": 1763 }, { "epoch": 0.7989130434782609, "grad_norm": 0.7372389978731413, "learning_rate": 7.913220362205059e-05, "loss": 0.9164, "step": 1764 }, { "epoch": 0.7993659420289855, "grad_norm": 0.7487084776327843, "learning_rate": 7.91295815300275e-05, "loss": 0.9023, "step": 1765 }, { "epoch": 0.7998188405797102, "grad_norm": 0.8945252908507952, "learning_rate": 7.912695552615814e-05, "loss": 0.9122, "step": 1766 }, { "epoch": 0.8002717391304348, "grad_norm": 1.0862863015881996, "learning_rate": 7.912432561070503e-05, "loss": 0.9035, "step": 1767 }, { "epoch": 0.8007246376811594, "grad_norm": 0.9806694980032382, "learning_rate": 7.91216917839311e-05, "loss": 0.8887, "step": 1768 }, { "epoch": 0.801177536231884, "grad_norm": 0.9229501503719609, "learning_rate": 7.911905404609965e-05, "loss": 0.9301, "step": 1769 }, { "epoch": 0.8016304347826086, "grad_norm": 0.7430233227345847, "learning_rate": 7.911641239747438e-05, "loss": 0.9174, "step": 1770 }, { "epoch": 0.8020833333333334, "grad_norm": 0.4878701442964826, "learning_rate": 7.911376683831937e-05, "loss": 0.9127, "step": 1771 }, { "epoch": 0.802536231884058, "grad_norm": 0.34516235151008345, "learning_rate": 7.91111173688991e-05, "loss": 0.9074, "step": 1772 }, { "epoch": 0.8029891304347826, "grad_norm": 0.5245064221208808, "learning_rate": 7.910846398947847e-05, "loss": 0.928, "step": 1773 }, { "epoch": 0.8034420289855072, "grad_norm": 0.37937010324033665, "learning_rate": 7.910580670032272e-05, "loss": 0.9175, "step": 1774 }, { "epoch": 0.8038949275362319, "grad_norm": 0.43258191148768577, "learning_rate": 7.910314550169749e-05, "loss": 0.9098, "step": 1775 }, { "epoch": 0.8043478260869565, "grad_norm": 0.49759870926220284, "learning_rate": 7.910048039386885e-05, "loss": 0.8962, "step": 1776 }, { "epoch": 0.8048007246376812, "grad_norm": 0.5298239601996784, "learning_rate": 7.909781137710322e-05, "loss": 0.9161, "step": 1777 }, { "epoch": 0.8052536231884058, "grad_norm": 0.6048490066278864, "learning_rate": 7.909513845166744e-05, "loss": 0.9104, "step": 1778 }, { "epoch": 0.8057065217391305, "grad_norm": 0.6323820745631986, "learning_rate": 7.90924616178287e-05, "loss": 0.9015, "step": 1779 }, { "epoch": 0.8061594202898551, "grad_norm": 0.6986523389696562, "learning_rate": 7.908978087585465e-05, "loss": 0.93, "step": 1780 }, { "epoch": 0.8066123188405797, "grad_norm": 0.9601340703988878, "learning_rate": 7.908709622601324e-05, "loss": 0.9087, "step": 1781 }, { "epoch": 0.8070652173913043, "grad_norm": 1.2033968962856771, "learning_rate": 7.908440766857288e-05, "loss": 0.9143, "step": 1782 }, { "epoch": 0.8075181159420289, "grad_norm": 0.7592901109273297, "learning_rate": 7.908171520380236e-05, "loss": 0.8964, "step": 1783 }, { "epoch": 0.8079710144927537, "grad_norm": 0.5487927641921743, "learning_rate": 7.907901883197085e-05, "loss": 0.912, "step": 1784 }, { "epoch": 0.8084239130434783, "grad_norm": 0.5826385653451449, "learning_rate": 7.907631855334788e-05, "loss": 0.9391, "step": 1785 }, { "epoch": 0.8088768115942029, "grad_norm": 0.6141003856228455, "learning_rate": 7.907361436820345e-05, "loss": 0.8969, "step": 1786 }, { "epoch": 0.8093297101449275, "grad_norm": 0.7515048704596282, "learning_rate": 7.907090627680786e-05, "loss": 0.8917, "step": 1787 }, { "epoch": 0.8097826086956522, "grad_norm": 0.8599613148263581, "learning_rate": 7.906819427943185e-05, "loss": 0.8954, "step": 1788 }, { "epoch": 0.8102355072463768, "grad_norm": 0.674411847349825, "learning_rate": 7.906547837634655e-05, "loss": 0.9151, "step": 1789 }, { "epoch": 0.8106884057971014, "grad_norm": 0.5157940318341842, "learning_rate": 7.906275856782347e-05, "loss": 0.883, "step": 1790 }, { "epoch": 0.811141304347826, "grad_norm": 0.5444985864441825, "learning_rate": 7.906003485413454e-05, "loss": 0.8892, "step": 1791 }, { "epoch": 0.8115942028985508, "grad_norm": 0.6043868316507307, "learning_rate": 7.905730723555201e-05, "loss": 0.915, "step": 1792 }, { "epoch": 0.8120471014492754, "grad_norm": 0.6654206264935715, "learning_rate": 7.905457571234858e-05, "loss": 0.9035, "step": 1793 }, { "epoch": 0.8125, "grad_norm": 0.6404397476010313, "learning_rate": 7.905184028479734e-05, "loss": 0.9317, "step": 1794 }, { "epoch": 0.8129528985507246, "grad_norm": 0.598340826934135, "learning_rate": 7.904910095317175e-05, "loss": 0.889, "step": 1795 }, { "epoch": 0.8134057971014492, "grad_norm": 0.6498944922685825, "learning_rate": 7.904635771774564e-05, "loss": 0.9303, "step": 1796 }, { "epoch": 0.813858695652174, "grad_norm": 0.7405034970596779, "learning_rate": 7.904361057879329e-05, "loss": 0.9103, "step": 1797 }, { "epoch": 0.8143115942028986, "grad_norm": 0.7413349428139097, "learning_rate": 7.90408595365893e-05, "loss": 0.9092, "step": 1798 }, { "epoch": 0.8147644927536232, "grad_norm": 0.6191266130430921, "learning_rate": 7.903810459140873e-05, "loss": 0.9222, "step": 1799 }, { "epoch": 0.8152173913043478, "grad_norm": 0.6227367124007476, "learning_rate": 7.903534574352698e-05, "loss": 0.9094, "step": 1800 }, { "epoch": 0.8156702898550725, "grad_norm": 0.683321571070582, "learning_rate": 7.903258299321987e-05, "loss": 0.9233, "step": 1801 }, { "epoch": 0.8161231884057971, "grad_norm": 0.8007322391364707, "learning_rate": 7.902981634076357e-05, "loss": 0.9159, "step": 1802 }, { "epoch": 0.8165760869565217, "grad_norm": 0.8669755108894504, "learning_rate": 7.90270457864347e-05, "loss": 0.9201, "step": 1803 }, { "epoch": 0.8170289855072463, "grad_norm": 0.9683819230036679, "learning_rate": 7.90242713305102e-05, "loss": 0.907, "step": 1804 }, { "epoch": 0.8174818840579711, "grad_norm": 1.041646320937391, "learning_rate": 7.902149297326746e-05, "loss": 0.9118, "step": 1805 }, { "epoch": 0.8179347826086957, "grad_norm": 1.000729747301983, "learning_rate": 7.901871071498425e-05, "loss": 0.9045, "step": 1806 }, { "epoch": 0.8183876811594203, "grad_norm": 0.9359734643545571, "learning_rate": 7.901592455593869e-05, "loss": 0.9015, "step": 1807 }, { "epoch": 0.8188405797101449, "grad_norm": 0.8448977657522692, "learning_rate": 7.901313449640931e-05, "loss": 0.9035, "step": 1808 }, { "epoch": 0.8192934782608695, "grad_norm": 0.6483623730110806, "learning_rate": 7.901034053667508e-05, "loss": 0.9028, "step": 1809 }, { "epoch": 0.8197463768115942, "grad_norm": 0.5219044474282143, "learning_rate": 7.900754267701526e-05, "loss": 0.9087, "step": 1810 }, { "epoch": 0.8201992753623188, "grad_norm": 0.514078994626908, "learning_rate": 7.900474091770961e-05, "loss": 0.9293, "step": 1811 }, { "epoch": 0.8206521739130435, "grad_norm": 0.9044445836280122, "learning_rate": 7.90019352590382e-05, "loss": 0.9358, "step": 1812 }, { "epoch": 0.8211050724637681, "grad_norm": 0.48449433113980866, "learning_rate": 7.89991257012815e-05, "loss": 0.8937, "step": 1813 }, { "epoch": 0.8215579710144928, "grad_norm": 0.5517479037979581, "learning_rate": 7.899631224472042e-05, "loss": 0.9193, "step": 1814 }, { "epoch": 0.8220108695652174, "grad_norm": 0.6387368691407966, "learning_rate": 7.899349488963621e-05, "loss": 0.8976, "step": 1815 }, { "epoch": 0.822463768115942, "grad_norm": 0.7761157549062435, "learning_rate": 7.899067363631051e-05, "loss": 0.9036, "step": 1816 }, { "epoch": 0.8229166666666666, "grad_norm": 0.8361816492941122, "learning_rate": 7.89878484850254e-05, "loss": 0.9077, "step": 1817 }, { "epoch": 0.8233695652173914, "grad_norm": 0.7431821887951378, "learning_rate": 7.898501943606329e-05, "loss": 0.8857, "step": 1818 }, { "epoch": 0.823822463768116, "grad_norm": 0.656014632984969, "learning_rate": 7.8982186489707e-05, "loss": 0.8919, "step": 1819 }, { "epoch": 0.8242753623188406, "grad_norm": 0.7324891824849389, "learning_rate": 7.897934964623977e-05, "loss": 0.9172, "step": 1820 }, { "epoch": 0.8247282608695652, "grad_norm": 0.8076222363210426, "learning_rate": 7.897650890594517e-05, "loss": 0.8882, "step": 1821 }, { "epoch": 0.8251811594202898, "grad_norm": 0.7422165091720847, "learning_rate": 7.897366426910723e-05, "loss": 0.8976, "step": 1822 }, { "epoch": 0.8256340579710145, "grad_norm": 0.6409264196356641, "learning_rate": 7.89708157360103e-05, "loss": 0.8909, "step": 1823 }, { "epoch": 0.8260869565217391, "grad_norm": 0.6236019878308774, "learning_rate": 7.896796330693917e-05, "loss": 0.8823, "step": 1824 }, { "epoch": 0.8265398550724637, "grad_norm": 0.6773662730014154, "learning_rate": 7.896510698217899e-05, "loss": 0.9094, "step": 1825 }, { "epoch": 0.8269927536231884, "grad_norm": 0.811895255511281, "learning_rate": 7.896224676201533e-05, "loss": 0.9068, "step": 1826 }, { "epoch": 0.8274456521739131, "grad_norm": 1.062125874947121, "learning_rate": 7.895938264673412e-05, "loss": 0.9085, "step": 1827 }, { "epoch": 0.8278985507246377, "grad_norm": 0.9515614833263637, "learning_rate": 7.89565146366217e-05, "loss": 0.9174, "step": 1828 }, { "epoch": 0.8283514492753623, "grad_norm": 0.7811504165848246, "learning_rate": 7.895364273196475e-05, "loss": 0.8922, "step": 1829 }, { "epoch": 0.8288043478260869, "grad_norm": 0.6485300338688957, "learning_rate": 7.895076693305044e-05, "loss": 0.9185, "step": 1830 }, { "epoch": 0.8292572463768116, "grad_norm": 0.5984588190165295, "learning_rate": 7.894788724016621e-05, "loss": 0.8846, "step": 1831 }, { "epoch": 0.8297101449275363, "grad_norm": 0.6087905822685538, "learning_rate": 7.89450036536e-05, "loss": 0.9198, "step": 1832 }, { "epoch": 0.8301630434782609, "grad_norm": 0.5815586450118566, "learning_rate": 7.894211617364005e-05, "loss": 0.9115, "step": 1833 }, { "epoch": 0.8306159420289855, "grad_norm": 0.5601785201274032, "learning_rate": 7.893922480057503e-05, "loss": 0.9136, "step": 1834 }, { "epoch": 0.8310688405797102, "grad_norm": 0.6634099618863303, "learning_rate": 7.893632953469401e-05, "loss": 0.9038, "step": 1835 }, { "epoch": 0.8315217391304348, "grad_norm": 0.8051311842029186, "learning_rate": 7.893343037628643e-05, "loss": 0.9074, "step": 1836 }, { "epoch": 0.8319746376811594, "grad_norm": 0.8467330126234619, "learning_rate": 7.893052732564211e-05, "loss": 0.9072, "step": 1837 }, { "epoch": 0.832427536231884, "grad_norm": 0.8099928658049604, "learning_rate": 7.89276203830513e-05, "loss": 0.9052, "step": 1838 }, { "epoch": 0.8328804347826086, "grad_norm": 0.8497831866966743, "learning_rate": 7.892470954880458e-05, "loss": 0.9218, "step": 1839 }, { "epoch": 0.8333333333333334, "grad_norm": 0.9148636235501277, "learning_rate": 7.892179482319297e-05, "loss": 0.9031, "step": 1840 }, { "epoch": 0.833786231884058, "grad_norm": 0.9081192134952909, "learning_rate": 7.891887620650784e-05, "loss": 0.9085, "step": 1841 }, { "epoch": 0.8342391304347826, "grad_norm": 0.8182175533798222, "learning_rate": 7.8915953699041e-05, "loss": 0.9113, "step": 1842 }, { "epoch": 0.8346920289855072, "grad_norm": 0.8234216621782842, "learning_rate": 7.891302730108459e-05, "loss": 0.9445, "step": 1843 }, { "epoch": 0.8351449275362319, "grad_norm": 0.6305149779556898, "learning_rate": 7.891009701293119e-05, "loss": 0.92, "step": 1844 }, { "epoch": 0.8355978260869565, "grad_norm": 0.5264874479145689, "learning_rate": 7.890716283487373e-05, "loss": 0.9096, "step": 1845 }, { "epoch": 0.8360507246376812, "grad_norm": 0.4388620549794513, "learning_rate": 7.890422476720554e-05, "loss": 0.9352, "step": 1846 }, { "epoch": 0.8365036231884058, "grad_norm": 0.36959626076397056, "learning_rate": 7.890128281022036e-05, "loss": 0.9045, "step": 1847 }, { "epoch": 0.8369565217391305, "grad_norm": 0.4007966291845229, "learning_rate": 7.889833696421229e-05, "loss": 0.915, "step": 1848 }, { "epoch": 0.8374094202898551, "grad_norm": 0.5897820750559711, "learning_rate": 7.889538722947583e-05, "loss": 0.8995, "step": 1849 }, { "epoch": 0.8378623188405797, "grad_norm": 0.7781080995427261, "learning_rate": 7.889243360630587e-05, "loss": 0.9114, "step": 1850 }, { "epoch": 0.8383152173913043, "grad_norm": 0.8394590422382638, "learning_rate": 7.88894760949977e-05, "loss": 0.9057, "step": 1851 }, { "epoch": 0.8387681159420289, "grad_norm": 0.852053853923639, "learning_rate": 7.888651469584698e-05, "loss": 0.9086, "step": 1852 }, { "epoch": 0.8392210144927537, "grad_norm": 0.8464826361945473, "learning_rate": 7.888354940914976e-05, "loss": 0.9024, "step": 1853 }, { "epoch": 0.8396739130434783, "grad_norm": 0.8484406523493223, "learning_rate": 7.888058023520249e-05, "loss": 0.9129, "step": 1854 }, { "epoch": 0.8401268115942029, "grad_norm": 0.7842743330719198, "learning_rate": 7.8877607174302e-05, "loss": 0.9187, "step": 1855 }, { "epoch": 0.8405797101449275, "grad_norm": 0.6902767230029142, "learning_rate": 7.887463022674552e-05, "loss": 0.8979, "step": 1856 }, { "epoch": 0.8410326086956522, "grad_norm": 0.4979044290716324, "learning_rate": 7.887164939283066e-05, "loss": 0.9086, "step": 1857 }, { "epoch": 0.8414855072463768, "grad_norm": 0.3545263532161202, "learning_rate": 7.88686646728554e-05, "loss": 0.8916, "step": 1858 }, { "epoch": 0.8419384057971014, "grad_norm": 0.3126504624451456, "learning_rate": 7.886567606711815e-05, "loss": 0.8962, "step": 1859 }, { "epoch": 0.842391304347826, "grad_norm": 0.35853878284115276, "learning_rate": 7.886268357591766e-05, "loss": 0.891, "step": 1860 }, { "epoch": 0.8428442028985508, "grad_norm": 0.3848043336413898, "learning_rate": 7.885968719955312e-05, "loss": 0.9083, "step": 1861 }, { "epoch": 0.8432971014492754, "grad_norm": 0.5133161051832805, "learning_rate": 7.885668693832409e-05, "loss": 0.9017, "step": 1862 }, { "epoch": 0.84375, "grad_norm": 0.6857948328072658, "learning_rate": 7.885368279253045e-05, "loss": 0.8923, "step": 1863 }, { "epoch": 0.8442028985507246, "grad_norm": 0.9090231345948332, "learning_rate": 7.885067476247261e-05, "loss": 0.9033, "step": 1864 }, { "epoch": 0.8446557971014492, "grad_norm": 1.1724976783072338, "learning_rate": 7.884766284845124e-05, "loss": 0.9174, "step": 1865 }, { "epoch": 0.845108695652174, "grad_norm": 0.8633722072295119, "learning_rate": 7.884464705076745e-05, "loss": 0.9099, "step": 1866 }, { "epoch": 0.8455615942028986, "grad_norm": 0.8189207396559198, "learning_rate": 7.884162736972274e-05, "loss": 0.9037, "step": 1867 }, { "epoch": 0.8460144927536232, "grad_norm": 0.8451828282323013, "learning_rate": 7.883860380561901e-05, "loss": 0.9154, "step": 1868 }, { "epoch": 0.8464673913043478, "grad_norm": 0.9995111197728166, "learning_rate": 7.883557635875849e-05, "loss": 0.9149, "step": 1869 }, { "epoch": 0.8469202898550725, "grad_norm": 0.9761477772604755, "learning_rate": 7.883254502944387e-05, "loss": 0.8911, "step": 1870 }, { "epoch": 0.8473731884057971, "grad_norm": 0.8389245551402934, "learning_rate": 7.882950981797819e-05, "loss": 0.9319, "step": 1871 }, { "epoch": 0.8478260869565217, "grad_norm": 0.9421975579950924, "learning_rate": 7.882647072466488e-05, "loss": 0.8988, "step": 1872 }, { "epoch": 0.8482789855072463, "grad_norm": 1.0747590637343702, "learning_rate": 7.882342774980776e-05, "loss": 0.8933, "step": 1873 }, { "epoch": 0.8487318840579711, "grad_norm": 0.8237215437738029, "learning_rate": 7.882038089371104e-05, "loss": 0.9323, "step": 1874 }, { "epoch": 0.8491847826086957, "grad_norm": 0.6592744749296967, "learning_rate": 7.881733015667935e-05, "loss": 0.8856, "step": 1875 }, { "epoch": 0.8496376811594203, "grad_norm": 0.406631474371022, "learning_rate": 7.881427553901764e-05, "loss": 0.885, "step": 1876 }, { "epoch": 0.8500905797101449, "grad_norm": 0.5233115996234472, "learning_rate": 7.88112170410313e-05, "loss": 0.9153, "step": 1877 }, { "epoch": 0.8505434782608695, "grad_norm": 0.6609030095442695, "learning_rate": 7.880815466302609e-05, "loss": 0.9142, "step": 1878 }, { "epoch": 0.8509963768115942, "grad_norm": 0.8108529088750236, "learning_rate": 7.880508840530815e-05, "loss": 0.9004, "step": 1879 }, { "epoch": 0.8514492753623188, "grad_norm": 1.0308334159625678, "learning_rate": 7.880201826818405e-05, "loss": 0.9233, "step": 1880 }, { "epoch": 0.8519021739130435, "grad_norm": 1.0160159952807244, "learning_rate": 7.879894425196069e-05, "loss": 0.8981, "step": 1881 }, { "epoch": 0.8523550724637681, "grad_norm": 0.8174313926125739, "learning_rate": 7.879586635694539e-05, "loss": 0.9418, "step": 1882 }, { "epoch": 0.8528079710144928, "grad_norm": 0.6009944453098232, "learning_rate": 7.879278458344585e-05, "loss": 0.8949, "step": 1883 }, { "epoch": 0.8532608695652174, "grad_norm": 0.4942925252382599, "learning_rate": 7.878969893177016e-05, "loss": 0.8951, "step": 1884 }, { "epoch": 0.853713768115942, "grad_norm": 0.47929437742357445, "learning_rate": 7.87866094022268e-05, "loss": 0.8849, "step": 1885 }, { "epoch": 0.8541666666666666, "grad_norm": 0.37487440791562127, "learning_rate": 7.878351599512465e-05, "loss": 0.8995, "step": 1886 }, { "epoch": 0.8546195652173914, "grad_norm": 0.28576249011535165, "learning_rate": 7.878041871077294e-05, "loss": 0.902, "step": 1887 }, { "epoch": 0.855072463768116, "grad_norm": 0.37928664634806775, "learning_rate": 7.87773175494813e-05, "loss": 0.8881, "step": 1888 }, { "epoch": 0.8555253623188406, "grad_norm": 0.48328844651370895, "learning_rate": 7.87742125115598e-05, "loss": 0.9054, "step": 1889 }, { "epoch": 0.8559782608695652, "grad_norm": 0.5471969417986683, "learning_rate": 7.877110359731883e-05, "loss": 0.9048, "step": 1890 }, { "epoch": 0.8564311594202898, "grad_norm": 0.7103404369140709, "learning_rate": 7.876799080706919e-05, "loss": 0.903, "step": 1891 }, { "epoch": 0.8568840579710145, "grad_norm": 0.8275263195825695, "learning_rate": 7.876487414112208e-05, "loss": 0.9197, "step": 1892 }, { "epoch": 0.8573369565217391, "grad_norm": 1.0210424180631223, "learning_rate": 7.876175359978907e-05, "loss": 0.8924, "step": 1893 }, { "epoch": 0.8577898550724637, "grad_norm": 1.230458989477745, "learning_rate": 7.875862918338213e-05, "loss": 0.8868, "step": 1894 }, { "epoch": 0.8582427536231884, "grad_norm": 0.6867081486633551, "learning_rate": 7.875550089221363e-05, "loss": 0.9079, "step": 1895 }, { "epoch": 0.8586956521739131, "grad_norm": 0.4812634385881775, "learning_rate": 7.875236872659626e-05, "loss": 0.9064, "step": 1896 }, { "epoch": 0.8591485507246377, "grad_norm": 0.8135791685806301, "learning_rate": 7.87492326868432e-05, "loss": 0.8904, "step": 1897 }, { "epoch": 0.8596014492753623, "grad_norm": 1.3655103639994395, "learning_rate": 7.874609277326793e-05, "loss": 0.8916, "step": 1898 }, { "epoch": 0.8600543478260869, "grad_norm": 0.5105873193351639, "learning_rate": 7.874294898618438e-05, "loss": 0.8932, "step": 1899 }, { "epoch": 0.8605072463768116, "grad_norm": 0.7020670901565392, "learning_rate": 7.873980132590683e-05, "loss": 0.9067, "step": 1900 }, { "epoch": 0.8609601449275363, "grad_norm": 1.3233164190726145, "learning_rate": 7.873664979274993e-05, "loss": 0.9148, "step": 1901 }, { "epoch": 0.8614130434782609, "grad_norm": 0.6239680139788621, "learning_rate": 7.87334943870288e-05, "loss": 0.8991, "step": 1902 }, { "epoch": 0.8618659420289855, "grad_norm": 0.5751768750410281, "learning_rate": 7.873033510905883e-05, "loss": 0.9093, "step": 1903 }, { "epoch": 0.8623188405797102, "grad_norm": 0.7860947801334105, "learning_rate": 7.872717195915589e-05, "loss": 0.9085, "step": 1904 }, { "epoch": 0.8627717391304348, "grad_norm": 0.917811559828852, "learning_rate": 7.872400493763621e-05, "loss": 0.908, "step": 1905 }, { "epoch": 0.8632246376811594, "grad_norm": 1.0316576803907056, "learning_rate": 7.87208340448164e-05, "loss": 0.8973, "step": 1906 }, { "epoch": 0.863677536231884, "grad_norm": 1.014636954421375, "learning_rate": 7.871765928101343e-05, "loss": 0.9099, "step": 1907 }, { "epoch": 0.8641304347826086, "grad_norm": 0.8368404666474126, "learning_rate": 7.871448064654473e-05, "loss": 0.9015, "step": 1908 }, { "epoch": 0.8645833333333334, "grad_norm": 0.6297895544842885, "learning_rate": 7.871129814172805e-05, "loss": 0.9246, "step": 1909 }, { "epoch": 0.865036231884058, "grad_norm": 0.39870744250016515, "learning_rate": 7.870811176688155e-05, "loss": 0.9136, "step": 1910 }, { "epoch": 0.8654891304347826, "grad_norm": 0.4502815684667535, "learning_rate": 7.870492152232375e-05, "loss": 0.8965, "step": 1911 }, { "epoch": 0.8659420289855072, "grad_norm": 0.5320536715760216, "learning_rate": 7.870172740837364e-05, "loss": 0.9124, "step": 1912 }, { "epoch": 0.8663949275362319, "grad_norm": 0.5969261063519685, "learning_rate": 7.869852942535051e-05, "loss": 0.902, "step": 1913 }, { "epoch": 0.8668478260869565, "grad_norm": 0.597408330446369, "learning_rate": 7.869532757357408e-05, "loss": 0.9023, "step": 1914 }, { "epoch": 0.8673007246376812, "grad_norm": 0.6796617200865893, "learning_rate": 7.86921218533644e-05, "loss": 0.9026, "step": 1915 }, { "epoch": 0.8677536231884058, "grad_norm": 0.8528415947723567, "learning_rate": 7.868891226504202e-05, "loss": 0.9196, "step": 1916 }, { "epoch": 0.8682065217391305, "grad_norm": 1.0665715720976576, "learning_rate": 7.868569880892775e-05, "loss": 0.9167, "step": 1917 }, { "epoch": 0.8686594202898551, "grad_norm": 0.9552920979731115, "learning_rate": 7.868248148534288e-05, "loss": 0.9201, "step": 1918 }, { "epoch": 0.8691123188405797, "grad_norm": 0.9487936648390362, "learning_rate": 7.867926029460904e-05, "loss": 0.9205, "step": 1919 }, { "epoch": 0.8695652173913043, "grad_norm": 1.030837333066833, "learning_rate": 7.867603523704825e-05, "loss": 0.897, "step": 1920 }, { "epoch": 0.8700181159420289, "grad_norm": 1.0465896686854292, "learning_rate": 7.867280631298293e-05, "loss": 0.9071, "step": 1921 }, { "epoch": 0.8704710144927537, "grad_norm": 0.7762323925524323, "learning_rate": 7.866957352273588e-05, "loss": 0.911, "step": 1922 }, { "epoch": 0.8709239130434783, "grad_norm": 0.5366467420214596, "learning_rate": 7.866633686663029e-05, "loss": 0.875, "step": 1923 }, { "epoch": 0.8713768115942029, "grad_norm": 0.3938401225276566, "learning_rate": 7.866309634498973e-05, "loss": 0.9024, "step": 1924 }, { "epoch": 0.8718297101449275, "grad_norm": 0.5207092745481234, "learning_rate": 7.865985195813816e-05, "loss": 0.9098, "step": 1925 }, { "epoch": 0.8722826086956522, "grad_norm": 0.6694391520542791, "learning_rate": 7.865660370639993e-05, "loss": 0.8996, "step": 1926 }, { "epoch": 0.8727355072463768, "grad_norm": 0.7778173324697393, "learning_rate": 7.865335159009977e-05, "loss": 0.9054, "step": 1927 }, { "epoch": 0.8731884057971014, "grad_norm": 0.8306325320036113, "learning_rate": 7.86500956095628e-05, "loss": 0.9214, "step": 1928 }, { "epoch": 0.873641304347826, "grad_norm": 0.8201088724191654, "learning_rate": 7.864683576511454e-05, "loss": 0.8942, "step": 1929 }, { "epoch": 0.8740942028985508, "grad_norm": 0.8564499197792991, "learning_rate": 7.864357205708085e-05, "loss": 0.9143, "step": 1930 }, { "epoch": 0.8745471014492754, "grad_norm": 0.81200652561834, "learning_rate": 7.864030448578803e-05, "loss": 0.897, "step": 1931 }, { "epoch": 0.875, "grad_norm": 0.7066074721060459, "learning_rate": 7.863703305156273e-05, "loss": 0.8899, "step": 1932 }, { "epoch": 0.8754528985507246, "grad_norm": 0.540041731704535, "learning_rate": 7.863375775473203e-05, "loss": 0.9123, "step": 1933 }, { "epoch": 0.8759057971014492, "grad_norm": 0.5702928189624145, "learning_rate": 7.863047859562333e-05, "loss": 0.9009, "step": 1934 }, { "epoch": 0.876358695652174, "grad_norm": 0.5307192768877085, "learning_rate": 7.862719557456448e-05, "loss": 0.9186, "step": 1935 }, { "epoch": 0.8768115942028986, "grad_norm": 0.46776355057463215, "learning_rate": 7.862390869188367e-05, "loss": 0.907, "step": 1936 }, { "epoch": 0.8772644927536232, "grad_norm": 0.5580425954788996, "learning_rate": 7.86206179479095e-05, "loss": 0.9189, "step": 1937 }, { "epoch": 0.8777173913043478, "grad_norm": 0.7311272665459889, "learning_rate": 7.861732334297098e-05, "loss": 0.896, "step": 1938 }, { "epoch": 0.8781702898550725, "grad_norm": 0.9257198919056172, "learning_rate": 7.861402487739743e-05, "loss": 0.9062, "step": 1939 }, { "epoch": 0.8786231884057971, "grad_norm": 1.1503098595466674, "learning_rate": 7.861072255151863e-05, "loss": 0.9108, "step": 1940 }, { "epoch": 0.8790760869565217, "grad_norm": 0.8673679192343076, "learning_rate": 7.86074163656647e-05, "loss": 0.8935, "step": 1941 }, { "epoch": 0.8795289855072463, "grad_norm": 0.661234592031665, "learning_rate": 7.86041063201662e-05, "loss": 0.9126, "step": 1942 }, { "epoch": 0.8799818840579711, "grad_norm": 0.551560821614797, "learning_rate": 7.8600792415354e-05, "loss": 0.9178, "step": 1943 }, { "epoch": 0.8804347826086957, "grad_norm": 0.5147819713280612, "learning_rate": 7.859747465155941e-05, "loss": 0.9089, "step": 1944 }, { "epoch": 0.8808876811594203, "grad_norm": 0.5981990954216311, "learning_rate": 7.859415302911412e-05, "loss": 0.9217, "step": 1945 }, { "epoch": 0.8813405797101449, "grad_norm": 0.5878548336518682, "learning_rate": 7.85908275483502e-05, "loss": 0.9048, "step": 1946 }, { "epoch": 0.8817934782608695, "grad_norm": 0.5792852444537904, "learning_rate": 7.858749820960009e-05, "loss": 0.9059, "step": 1947 }, { "epoch": 0.8822463768115942, "grad_norm": 0.4879435009814253, "learning_rate": 7.858416501319666e-05, "loss": 0.904, "step": 1948 }, { "epoch": 0.8826992753623188, "grad_norm": 0.46579001896554706, "learning_rate": 7.858082795947309e-05, "loss": 0.9352, "step": 1949 }, { "epoch": 0.8831521739130435, "grad_norm": 0.647445793975872, "learning_rate": 7.8577487048763e-05, "loss": 0.9144, "step": 1950 }, { "epoch": 0.8836050724637681, "grad_norm": 0.6644648737248048, "learning_rate": 7.857414228140043e-05, "loss": 0.9042, "step": 1951 }, { "epoch": 0.8840579710144928, "grad_norm": 0.47539023633576194, "learning_rate": 7.857079365771972e-05, "loss": 0.9026, "step": 1952 }, { "epoch": 0.8845108695652174, "grad_norm": 0.45312260791412157, "learning_rate": 7.856744117805564e-05, "loss": 0.9098, "step": 1953 }, { "epoch": 0.884963768115942, "grad_norm": 0.5131522793060148, "learning_rate": 7.856408484274335e-05, "loss": 0.8962, "step": 1954 }, { "epoch": 0.8854166666666666, "grad_norm": 0.4895599700716663, "learning_rate": 7.856072465211839e-05, "loss": 0.905, "step": 1955 }, { "epoch": 0.8858695652173914, "grad_norm": 0.6214564500949696, "learning_rate": 7.855736060651668e-05, "loss": 0.9115, "step": 1956 }, { "epoch": 0.886322463768116, "grad_norm": 0.8003536142846329, "learning_rate": 7.855399270627454e-05, "loss": 0.9136, "step": 1957 }, { "epoch": 0.8867753623188406, "grad_norm": 0.8690704754615094, "learning_rate": 7.855062095172865e-05, "loss": 0.9031, "step": 1958 }, { "epoch": 0.8872282608695652, "grad_norm": 0.8932568838727191, "learning_rate": 7.854724534321609e-05, "loss": 0.9079, "step": 1959 }, { "epoch": 0.8876811594202898, "grad_norm": 1.022257839294402, "learning_rate": 7.854386588107433e-05, "loss": 0.9144, "step": 1960 }, { "epoch": 0.8881340579710145, "grad_norm": 1.1192480417365696, "learning_rate": 7.854048256564122e-05, "loss": 0.8968, "step": 1961 }, { "epoch": 0.8885869565217391, "grad_norm": 0.70984631991774, "learning_rate": 7.853709539725501e-05, "loss": 0.9306, "step": 1962 }, { "epoch": 0.8890398550724637, "grad_norm": 0.5424298611011982, "learning_rate": 7.853370437625429e-05, "loss": 0.9037, "step": 1963 }, { "epoch": 0.8894927536231884, "grad_norm": 0.5442525842682411, "learning_rate": 7.853030950297809e-05, "loss": 0.9168, "step": 1964 }, { "epoch": 0.8899456521739131, "grad_norm": 0.4902255584905581, "learning_rate": 7.852691077776579e-05, "loss": 0.8915, "step": 1965 }, { "epoch": 0.8903985507246377, "grad_norm": 0.4972037497064008, "learning_rate": 7.852350820095715e-05, "loss": 0.9083, "step": 1966 }, { "epoch": 0.8908514492753623, "grad_norm": 0.6370535890045225, "learning_rate": 7.852010177289239e-05, "loss": 0.9164, "step": 1967 }, { "epoch": 0.8913043478260869, "grad_norm": 0.8031221133936277, "learning_rate": 7.851669149391198e-05, "loss": 0.9082, "step": 1968 }, { "epoch": 0.8917572463768116, "grad_norm": 0.9029044768981582, "learning_rate": 7.851327736435689e-05, "loss": 0.9086, "step": 1969 }, { "epoch": 0.8922101449275363, "grad_norm": 0.9782521034141625, "learning_rate": 7.850985938456845e-05, "loss": 0.9184, "step": 1970 }, { "epoch": 0.8926630434782609, "grad_norm": 1.0042404730373446, "learning_rate": 7.850643755488833e-05, "loss": 0.9095, "step": 1971 }, { "epoch": 0.8931159420289855, "grad_norm": 0.9860264505503497, "learning_rate": 7.850301187565864e-05, "loss": 0.8928, "step": 1972 }, { "epoch": 0.8935688405797102, "grad_norm": 0.9411125915232506, "learning_rate": 7.849958234722183e-05, "loss": 0.8888, "step": 1973 }, { "epoch": 0.8940217391304348, "grad_norm": 0.947691899608784, "learning_rate": 7.849614896992077e-05, "loss": 0.9014, "step": 1974 }, { "epoch": 0.8944746376811594, "grad_norm": 0.8789558157646963, "learning_rate": 7.849271174409868e-05, "loss": 0.9211, "step": 1975 }, { "epoch": 0.894927536231884, "grad_norm": 0.7798085941103069, "learning_rate": 7.848927067009922e-05, "loss": 0.9218, "step": 1976 }, { "epoch": 0.8953804347826086, "grad_norm": 0.6478275161551259, "learning_rate": 7.848582574826637e-05, "loss": 0.9131, "step": 1977 }, { "epoch": 0.8958333333333334, "grad_norm": 0.6086086051530566, "learning_rate": 7.848237697894453e-05, "loss": 0.9073, "step": 1978 }, { "epoch": 0.896286231884058, "grad_norm": 0.5446717503115697, "learning_rate": 7.847892436247849e-05, "loss": 0.8971, "step": 1979 }, { "epoch": 0.8967391304347826, "grad_norm": 0.5141564237456893, "learning_rate": 7.84754678992134e-05, "loss": 0.9059, "step": 1980 }, { "epoch": 0.8971920289855072, "grad_norm": 0.4932172610759034, "learning_rate": 7.847200758949482e-05, "loss": 0.9363, "step": 1981 }, { "epoch": 0.8976449275362319, "grad_norm": 0.6392174974062094, "learning_rate": 7.846854343366868e-05, "loss": 0.8837, "step": 1982 }, { "epoch": 0.8980978260869565, "grad_norm": 0.7547808642483703, "learning_rate": 7.846507543208129e-05, "loss": 0.9035, "step": 1983 }, { "epoch": 0.8985507246376812, "grad_norm": 0.8703139723492811, "learning_rate": 7.846160358507936e-05, "loss": 0.9079, "step": 1984 }, { "epoch": 0.8990036231884058, "grad_norm": 0.9691969178748308, "learning_rate": 7.845812789300997e-05, "loss": 0.9199, "step": 1985 }, { "epoch": 0.8994565217391305, "grad_norm": 0.9376010896807209, "learning_rate": 7.845464835622059e-05, "loss": 0.9106, "step": 1986 }, { "epoch": 0.8999094202898551, "grad_norm": 0.8519480462292551, "learning_rate": 7.845116497505908e-05, "loss": 0.9124, "step": 1987 }, { "epoch": 0.9003623188405797, "grad_norm": 0.7505805471826577, "learning_rate": 7.844767774987367e-05, "loss": 0.9039, "step": 1988 }, { "epoch": 0.9008152173913043, "grad_norm": 0.6655168814823808, "learning_rate": 7.8444186681013e-05, "loss": 0.8932, "step": 1989 }, { "epoch": 0.9012681159420289, "grad_norm": 0.4644203471225832, "learning_rate": 7.844069176882606e-05, "loss": 0.8981, "step": 1990 }, { "epoch": 0.9017210144927537, "grad_norm": 0.3070596267893877, "learning_rate": 7.843719301366226e-05, "loss": 0.9099, "step": 1991 }, { "epoch": 0.9021739130434783, "grad_norm": 0.41563286403725, "learning_rate": 7.843369041587135e-05, "loss": 0.9012, "step": 1992 }, { "epoch": 0.9026268115942029, "grad_norm": 0.5974756244807449, "learning_rate": 7.843018397580351e-05, "loss": 0.9186, "step": 1993 }, { "epoch": 0.9030797101449275, "grad_norm": 0.7546018626591101, "learning_rate": 7.842667369380929e-05, "loss": 0.8855, "step": 1994 }, { "epoch": 0.9035326086956522, "grad_norm": 0.8802049631969494, "learning_rate": 7.842315957023961e-05, "loss": 0.9233, "step": 1995 }, { "epoch": 0.9039855072463768, "grad_norm": 1.12895719578658, "learning_rate": 7.841964160544577e-05, "loss": 0.9099, "step": 1996 }, { "epoch": 0.9044384057971014, "grad_norm": 0.899573490939267, "learning_rate": 7.841611979977947e-05, "loss": 0.9157, "step": 1997 }, { "epoch": 0.904891304347826, "grad_norm": 0.6639366521657567, "learning_rate": 7.841259415359282e-05, "loss": 0.888, "step": 1998 }, { "epoch": 0.9053442028985508, "grad_norm": 0.4695653266137768, "learning_rate": 7.840906466723823e-05, "loss": 0.9126, "step": 1999 }, { "epoch": 0.9057971014492754, "grad_norm": 0.46293101263653996, "learning_rate": 7.840553134106862e-05, "loss": 0.866, "step": 2000 }, { "epoch": 0.90625, "grad_norm": 0.32478379248684097, "learning_rate": 7.840199417543716e-05, "loss": 0.8883, "step": 2001 }, { "epoch": 0.9067028985507246, "grad_norm": 0.44775153698027, "learning_rate": 7.839845317069749e-05, "loss": 0.9211, "step": 2002 }, { "epoch": 0.9071557971014492, "grad_norm": 0.5330072722600072, "learning_rate": 7.839490832720361e-05, "loss": 0.8995, "step": 2003 }, { "epoch": 0.907608695652174, "grad_norm": 0.4752336232206368, "learning_rate": 7.83913596453099e-05, "loss": 0.8912, "step": 2004 }, { "epoch": 0.9080615942028986, "grad_norm": 0.4367963076283476, "learning_rate": 7.838780712537114e-05, "loss": 0.9162, "step": 2005 }, { "epoch": 0.9085144927536232, "grad_norm": 0.5208135044667824, "learning_rate": 7.838425076774247e-05, "loss": 0.889, "step": 2006 }, { "epoch": 0.9089673913043478, "grad_norm": 0.7256156042362127, "learning_rate": 7.838069057277943e-05, "loss": 0.9115, "step": 2007 }, { "epoch": 0.9094202898550725, "grad_norm": 0.8541828884386087, "learning_rate": 7.83771265408379e-05, "loss": 0.8954, "step": 2008 }, { "epoch": 0.9098731884057971, "grad_norm": 0.9512809266234022, "learning_rate": 7.837355867227426e-05, "loss": 0.8896, "step": 2009 }, { "epoch": 0.9103260869565217, "grad_norm": 0.9274328862096458, "learning_rate": 7.836998696744513e-05, "loss": 0.8873, "step": 2010 }, { "epoch": 0.9107789855072463, "grad_norm": 0.8340033021095087, "learning_rate": 7.836641142670759e-05, "loss": 0.919, "step": 2011 }, { "epoch": 0.9112318840579711, "grad_norm": 0.7572327720703492, "learning_rate": 7.836283205041913e-05, "loss": 0.9018, "step": 2012 }, { "epoch": 0.9116847826086957, "grad_norm": 0.6870707082684233, "learning_rate": 7.835924883893753e-05, "loss": 0.911, "step": 2013 }, { "epoch": 0.9121376811594203, "grad_norm": 0.586941936232824, "learning_rate": 7.835566179262106e-05, "loss": 0.9035, "step": 2014 }, { "epoch": 0.9125905797101449, "grad_norm": 0.4721242399524216, "learning_rate": 7.83520709118283e-05, "loss": 0.8955, "step": 2015 }, { "epoch": 0.9130434782608695, "grad_norm": 0.5081426710692672, "learning_rate": 7.834847619691823e-05, "loss": 0.91, "step": 2016 }, { "epoch": 0.9134963768115942, "grad_norm": 0.5310609633537797, "learning_rate": 7.834487764825022e-05, "loss": 0.8882, "step": 2017 }, { "epoch": 0.9139492753623188, "grad_norm": 0.5436106551866123, "learning_rate": 7.834127526618405e-05, "loss": 0.9251, "step": 2018 }, { "epoch": 0.9144021739130435, "grad_norm": 0.4697675651505968, "learning_rate": 7.833766905107982e-05, "loss": 0.9082, "step": 2019 }, { "epoch": 0.9148550724637681, "grad_norm": 0.3788739247288177, "learning_rate": 7.833405900329806e-05, "loss": 0.9086, "step": 2020 }, { "epoch": 0.9153079710144928, "grad_norm": 0.3283252922878422, "learning_rate": 7.833044512319968e-05, "loss": 0.888, "step": 2021 }, { "epoch": 0.9157608695652174, "grad_norm": 0.3923279164298861, "learning_rate": 7.832682741114597e-05, "loss": 0.9036, "step": 2022 }, { "epoch": 0.916213768115942, "grad_norm": 0.561197468130468, "learning_rate": 7.832320586749857e-05, "loss": 0.9054, "step": 2023 }, { "epoch": 0.9166666666666666, "grad_norm": 0.6118414815365569, "learning_rate": 7.831958049261956e-05, "loss": 0.9063, "step": 2024 }, { "epoch": 0.9171195652173914, "grad_norm": 0.731654336927263, "learning_rate": 7.831595128687136e-05, "loss": 0.8872, "step": 2025 }, { "epoch": 0.917572463768116, "grad_norm": 0.8183352336301352, "learning_rate": 7.831231825061681e-05, "loss": 0.9031, "step": 2026 }, { "epoch": 0.9180253623188406, "grad_norm": 0.9373294337136497, "learning_rate": 7.830868138421908e-05, "loss": 0.9105, "step": 2027 }, { "epoch": 0.9184782608695652, "grad_norm": 0.9815152778255495, "learning_rate": 7.830504068804178e-05, "loss": 0.8934, "step": 2028 }, { "epoch": 0.9189311594202898, "grad_norm": 0.9906200714830354, "learning_rate": 7.830139616244886e-05, "loss": 0.905, "step": 2029 }, { "epoch": 0.9193840579710145, "grad_norm": 0.9616043246101117, "learning_rate": 7.829774780780465e-05, "loss": 0.8905, "step": 2030 }, { "epoch": 0.9198369565217391, "grad_norm": 0.8134606138664129, "learning_rate": 7.829409562447392e-05, "loss": 0.9048, "step": 2031 }, { "epoch": 0.9202898550724637, "grad_norm": 0.5304197242918494, "learning_rate": 7.829043961282177e-05, "loss": 0.8993, "step": 2032 }, { "epoch": 0.9207427536231884, "grad_norm": 0.4000783058414411, "learning_rate": 7.82867797732137e-05, "loss": 0.9051, "step": 2033 }, { "epoch": 0.9211956521739131, "grad_norm": 0.3615031600115553, "learning_rate": 7.828311610601557e-05, "loss": 0.8668, "step": 2034 }, { "epoch": 0.9216485507246377, "grad_norm": 0.36045796390135054, "learning_rate": 7.827944861159367e-05, "loss": 0.8992, "step": 2035 }, { "epoch": 0.9221014492753623, "grad_norm": 0.3524867545268556, "learning_rate": 7.827577729031464e-05, "loss": 0.8992, "step": 2036 }, { "epoch": 0.9225543478260869, "grad_norm": 0.383375229486044, "learning_rate": 7.827210214254548e-05, "loss": 0.9111, "step": 2037 }, { "epoch": 0.9230072463768116, "grad_norm": 0.4759557451487254, "learning_rate": 7.826842316865363e-05, "loss": 0.8976, "step": 2038 }, { "epoch": 0.9234601449275363, "grad_norm": 0.4665264761910608, "learning_rate": 7.826474036900687e-05, "loss": 0.8911, "step": 2039 }, { "epoch": 0.9239130434782609, "grad_norm": 0.5606387463888932, "learning_rate": 7.82610537439734e-05, "loss": 0.8994, "step": 2040 }, { "epoch": 0.9243659420289855, "grad_norm": 0.6170324466600895, "learning_rate": 7.825736329392174e-05, "loss": 0.9008, "step": 2041 }, { "epoch": 0.9248188405797102, "grad_norm": 0.6820654782156065, "learning_rate": 7.825366901922085e-05, "loss": 0.8902, "step": 2042 }, { "epoch": 0.9252717391304348, "grad_norm": 0.7793007635501233, "learning_rate": 7.824997092024004e-05, "loss": 0.8874, "step": 2043 }, { "epoch": 0.9257246376811594, "grad_norm": 0.9820121966546101, "learning_rate": 7.824626899734905e-05, "loss": 0.9064, "step": 2044 }, { "epoch": 0.926177536231884, "grad_norm": 1.1806164619043114, "learning_rate": 7.824256325091793e-05, "loss": 0.8993, "step": 2045 }, { "epoch": 0.9266304347826086, "grad_norm": 0.7281012256320594, "learning_rate": 7.823885368131716e-05, "loss": 0.905, "step": 2046 }, { "epoch": 0.9270833333333334, "grad_norm": 0.4014947998915133, "learning_rate": 7.823514028891758e-05, "loss": 0.9049, "step": 2047 }, { "epoch": 0.927536231884058, "grad_norm": 0.3179232543389663, "learning_rate": 7.823142307409045e-05, "loss": 0.9024, "step": 2048 }, { "epoch": 0.9279891304347826, "grad_norm": 0.5202431501145403, "learning_rate": 7.822770203720737e-05, "loss": 0.897, "step": 2049 }, { "epoch": 0.9284420289855072, "grad_norm": 0.8332436772983846, "learning_rate": 7.822397717864035e-05, "loss": 0.9059, "step": 2050 }, { "epoch": 0.9288949275362319, "grad_norm": 1.0927690412896667, "learning_rate": 7.822024849876175e-05, "loss": 0.9174, "step": 2051 }, { "epoch": 0.9293478260869565, "grad_norm": 1.07109804111847, "learning_rate": 7.821651599794435e-05, "loss": 0.911, "step": 2052 }, { "epoch": 0.9298007246376812, "grad_norm": 0.8929643172241255, "learning_rate": 7.821277967656127e-05, "loss": 0.9106, "step": 2053 }, { "epoch": 0.9302536231884058, "grad_norm": 0.6171469656499018, "learning_rate": 7.820903953498607e-05, "loss": 0.904, "step": 2054 }, { "epoch": 0.9307065217391305, "grad_norm": 0.4748291590812278, "learning_rate": 7.820529557359263e-05, "loss": 0.9107, "step": 2055 }, { "epoch": 0.9311594202898551, "grad_norm": 0.44118027941978016, "learning_rate": 7.820154779275525e-05, "loss": 0.8899, "step": 2056 }, { "epoch": 0.9316123188405797, "grad_norm": 0.44810658389604696, "learning_rate": 7.819779619284861e-05, "loss": 0.8896, "step": 2057 }, { "epoch": 0.9320652173913043, "grad_norm": 0.3831410538010979, "learning_rate": 7.819404077424775e-05, "loss": 0.9283, "step": 2058 }, { "epoch": 0.9325181159420289, "grad_norm": 0.48993027097165026, "learning_rate": 7.819028153732811e-05, "loss": 0.8962, "step": 2059 }, { "epoch": 0.9329710144927537, "grad_norm": 0.8848688742451837, "learning_rate": 7.81865184824655e-05, "loss": 0.8909, "step": 2060 }, { "epoch": 0.9334239130434783, "grad_norm": 1.1553861755836603, "learning_rate": 7.818275161003615e-05, "loss": 0.8979, "step": 2061 }, { "epoch": 0.9338768115942029, "grad_norm": 0.8911625319257958, "learning_rate": 7.817898092041659e-05, "loss": 0.8896, "step": 2062 }, { "epoch": 0.9343297101449275, "grad_norm": 0.8214414710020764, "learning_rate": 7.817520641398381e-05, "loss": 0.9088, "step": 2063 }, { "epoch": 0.9347826086956522, "grad_norm": 0.8871454450744317, "learning_rate": 7.817142809111516e-05, "loss": 0.9017, "step": 2064 }, { "epoch": 0.9352355072463768, "grad_norm": 0.9538038339007225, "learning_rate": 7.816764595218836e-05, "loss": 0.8919, "step": 2065 }, { "epoch": 0.9356884057971014, "grad_norm": 1.0648634594515622, "learning_rate": 7.816385999758151e-05, "loss": 0.9284, "step": 2066 }, { "epoch": 0.936141304347826, "grad_norm": 0.8240009610812988, "learning_rate": 7.816007022767308e-05, "loss": 0.9254, "step": 2067 }, { "epoch": 0.9365942028985508, "grad_norm": 0.5324671828256748, "learning_rate": 7.815627664284198e-05, "loss": 0.8935, "step": 2068 }, { "epoch": 0.9370471014492754, "grad_norm": 0.418626346570576, "learning_rate": 7.815247924346743e-05, "loss": 0.9136, "step": 2069 }, { "epoch": 0.9375, "grad_norm": 0.4980082773860058, "learning_rate": 7.814867802992907e-05, "loss": 0.9115, "step": 2070 }, { "epoch": 0.9379528985507246, "grad_norm": 0.5170988071508107, "learning_rate": 7.814487300260693e-05, "loss": 0.8946, "step": 2071 }, { "epoch": 0.9384057971014492, "grad_norm": 0.48337442767159794, "learning_rate": 7.81410641618814e-05, "loss": 0.904, "step": 2072 }, { "epoch": 0.938858695652174, "grad_norm": 0.44140956781116475, "learning_rate": 7.813725150813321e-05, "loss": 0.9066, "step": 2073 }, { "epoch": 0.9393115942028986, "grad_norm": 0.45042126976969726, "learning_rate": 7.813343504174357e-05, "loss": 0.9027, "step": 2074 }, { "epoch": 0.9397644927536232, "grad_norm": 0.4906826183582538, "learning_rate": 7.812961476309401e-05, "loss": 0.8909, "step": 2075 }, { "epoch": 0.9402173913043478, "grad_norm": 0.48601004744104087, "learning_rate": 7.812579067256644e-05, "loss": 0.9041, "step": 2076 }, { "epoch": 0.9406702898550725, "grad_norm": 0.5146601337916319, "learning_rate": 7.812196277054316e-05, "loss": 0.9211, "step": 2077 }, { "epoch": 0.9411231884057971, "grad_norm": 0.506890416414368, "learning_rate": 7.811813105740684e-05, "loss": 0.8957, "step": 2078 }, { "epoch": 0.9415760869565217, "grad_norm": 0.56049186627914, "learning_rate": 7.811429553354057e-05, "loss": 0.9127, "step": 2079 }, { "epoch": 0.9420289855072463, "grad_norm": 0.6761451517801773, "learning_rate": 7.811045619932775e-05, "loss": 0.8997, "step": 2080 }, { "epoch": 0.9424818840579711, "grad_norm": 0.7952132245841467, "learning_rate": 7.810661305515227e-05, "loss": 0.9071, "step": 2081 }, { "epoch": 0.9429347826086957, "grad_norm": 0.86375212332064, "learning_rate": 7.810276610139827e-05, "loss": 0.9108, "step": 2082 }, { "epoch": 0.9433876811594203, "grad_norm": 0.9180636522106995, "learning_rate": 7.809891533845039e-05, "loss": 0.911, "step": 2083 }, { "epoch": 0.9438405797101449, "grad_norm": 0.9569212605570845, "learning_rate": 7.809506076669355e-05, "loss": 0.9141, "step": 2084 }, { "epoch": 0.9442934782608695, "grad_norm": 0.947432464556113, "learning_rate": 7.809120238651313e-05, "loss": 0.918, "step": 2085 }, { "epoch": 0.9447463768115942, "grad_norm": 0.8642813749674056, "learning_rate": 7.808734019829484e-05, "loss": 0.9027, "step": 2086 }, { "epoch": 0.9451992753623188, "grad_norm": 0.7446483482302994, "learning_rate": 7.80834742024248e-05, "loss": 0.8942, "step": 2087 }, { "epoch": 0.9456521739130435, "grad_norm": 0.5647499858900267, "learning_rate": 7.807960439928948e-05, "loss": 0.9011, "step": 2088 }, { "epoch": 0.9461050724637681, "grad_norm": 0.5213853578470702, "learning_rate": 7.807573078927577e-05, "loss": 0.9128, "step": 2089 }, { "epoch": 0.9465579710144928, "grad_norm": 0.5070685206716545, "learning_rate": 7.807185337277093e-05, "loss": 0.9322, "step": 2090 }, { "epoch": 0.9470108695652174, "grad_norm": 0.3849040891455112, "learning_rate": 7.806797215016257e-05, "loss": 0.8853, "step": 2091 }, { "epoch": 0.947463768115942, "grad_norm": 0.42149262409552907, "learning_rate": 7.806408712183873e-05, "loss": 0.9014, "step": 2092 }, { "epoch": 0.9479166666666666, "grad_norm": 0.5535931315736232, "learning_rate": 7.806019828818776e-05, "loss": 0.8848, "step": 2093 }, { "epoch": 0.9483695652173914, "grad_norm": 0.6003309216080389, "learning_rate": 7.805630564959846e-05, "loss": 0.9022, "step": 2094 }, { "epoch": 0.948822463768116, "grad_norm": 0.6803319054167197, "learning_rate": 7.805240920645998e-05, "loss": 0.9111, "step": 2095 }, { "epoch": 0.9492753623188406, "grad_norm": 0.7633683996554416, "learning_rate": 7.804850895916185e-05, "loss": 0.9037, "step": 2096 }, { "epoch": 0.9497282608695652, "grad_norm": 0.864672771856854, "learning_rate": 7.804460490809399e-05, "loss": 0.91, "step": 2097 }, { "epoch": 0.9501811594202898, "grad_norm": 0.9570567499008463, "learning_rate": 7.804069705364669e-05, "loss": 0.912, "step": 2098 }, { "epoch": 0.9506340579710145, "grad_norm": 1.0264281326766207, "learning_rate": 7.803678539621063e-05, "loss": 0.912, "step": 2099 }, { "epoch": 0.9510869565217391, "grad_norm": 0.9746231233362802, "learning_rate": 7.803286993617685e-05, "loss": 0.9171, "step": 2100 }, { "epoch": 0.9515398550724637, "grad_norm": 0.8600223197787742, "learning_rate": 7.80289506739368e-05, "loss": 0.9079, "step": 2101 }, { "epoch": 0.9519927536231884, "grad_norm": 0.7565976335136299, "learning_rate": 7.802502760988229e-05, "loss": 0.9195, "step": 2102 }, { "epoch": 0.9524456521739131, "grad_norm": 0.5035451577750346, "learning_rate": 7.802110074440551e-05, "loss": 0.9004, "step": 2103 }, { "epoch": 0.9528985507246377, "grad_norm": 0.3292881288454833, "learning_rate": 7.801717007789904e-05, "loss": 0.9089, "step": 2104 }, { "epoch": 0.9533514492753623, "grad_norm": 0.439399412516054, "learning_rate": 7.801323561075583e-05, "loss": 0.8979, "step": 2105 }, { "epoch": 0.9538043478260869, "grad_norm": 0.5217248047852303, "learning_rate": 7.800929734336922e-05, "loss": 0.8892, "step": 2106 }, { "epoch": 0.9542572463768116, "grad_norm": 0.6936991717272553, "learning_rate": 7.800535527613293e-05, "loss": 0.9081, "step": 2107 }, { "epoch": 0.9547101449275363, "grad_norm": 0.821421717942182, "learning_rate": 7.800140940944104e-05, "loss": 0.8972, "step": 2108 }, { "epoch": 0.9551630434782609, "grad_norm": 0.9168417087177829, "learning_rate": 7.799745974368804e-05, "loss": 0.8992, "step": 2109 }, { "epoch": 0.9556159420289855, "grad_norm": 0.966918975942187, "learning_rate": 7.799350627926878e-05, "loss": 0.8939, "step": 2110 }, { "epoch": 0.9560688405797102, "grad_norm": 0.9234545864040854, "learning_rate": 7.798954901657848e-05, "loss": 0.9008, "step": 2111 }, { "epoch": 0.9565217391304348, "grad_norm": 0.7921639884037811, "learning_rate": 7.798558795601278e-05, "loss": 0.8791, "step": 2112 }, { "epoch": 0.9569746376811594, "grad_norm": 0.6541462665232836, "learning_rate": 7.798162309796764e-05, "loss": 0.9095, "step": 2113 }, { "epoch": 0.957427536231884, "grad_norm": 0.5378402028372534, "learning_rate": 7.797765444283947e-05, "loss": 0.8977, "step": 2114 }, { "epoch": 0.9578804347826086, "grad_norm": 0.4366033363189359, "learning_rate": 7.797368199102501e-05, "loss": 0.8927, "step": 2115 }, { "epoch": 0.9583333333333334, "grad_norm": 0.31489232894439734, "learning_rate": 7.796970574292136e-05, "loss": 0.8892, "step": 2116 }, { "epoch": 0.958786231884058, "grad_norm": 0.5165835439244956, "learning_rate": 7.796572569892607e-05, "loss": 0.9152, "step": 2117 }, { "epoch": 0.9592391304347826, "grad_norm": 0.690235382711277, "learning_rate": 7.796174185943702e-05, "loss": 0.901, "step": 2118 }, { "epoch": 0.9596920289855072, "grad_norm": 0.7070778772017684, "learning_rate": 7.79577542248525e-05, "loss": 0.9105, "step": 2119 }, { "epoch": 0.9601449275362319, "grad_norm": 0.7055628367096572, "learning_rate": 7.795376279557111e-05, "loss": 0.9298, "step": 2120 }, { "epoch": 0.9605978260869565, "grad_norm": 0.6688276629110497, "learning_rate": 7.794976757199192e-05, "loss": 0.8914, "step": 2121 }, { "epoch": 0.9610507246376812, "grad_norm": 0.6764252875822199, "learning_rate": 7.794576855451434e-05, "loss": 0.9008, "step": 2122 }, { "epoch": 0.9615036231884058, "grad_norm": 0.6194538410194635, "learning_rate": 7.794176574353815e-05, "loss": 0.8751, "step": 2123 }, { "epoch": 0.9619565217391305, "grad_norm": 0.6386524302781527, "learning_rate": 7.79377591394635e-05, "loss": 0.9197, "step": 2124 }, { "epoch": 0.9624094202898551, "grad_norm": 0.6618316580986712, "learning_rate": 7.793374874269096e-05, "loss": 0.9065, "step": 2125 }, { "epoch": 0.9628623188405797, "grad_norm": 0.6421248705922953, "learning_rate": 7.792973455362145e-05, "loss": 0.9082, "step": 2126 }, { "epoch": 0.9633152173913043, "grad_norm": 0.685954397933943, "learning_rate": 7.792571657265626e-05, "loss": 0.8734, "step": 2127 }, { "epoch": 0.9637681159420289, "grad_norm": 0.7658423969772364, "learning_rate": 7.79216948001971e-05, "loss": 0.9055, "step": 2128 }, { "epoch": 0.9642210144927537, "grad_norm": 0.8104145063174439, "learning_rate": 7.7917669236646e-05, "loss": 0.9061, "step": 2129 }, { "epoch": 0.9646739130434783, "grad_norm": 0.9051286774173892, "learning_rate": 7.791363988240544e-05, "loss": 0.9101, "step": 2130 }, { "epoch": 0.9651268115942029, "grad_norm": 0.9519217133993346, "learning_rate": 7.79096067378782e-05, "loss": 0.9161, "step": 2131 }, { "epoch": 0.9655797101449275, "grad_norm": 0.8657871245126172, "learning_rate": 7.790556980346752e-05, "loss": 0.9062, "step": 2132 }, { "epoch": 0.9660326086956522, "grad_norm": 0.6949798402683933, "learning_rate": 7.790152907957697e-05, "loss": 0.9124, "step": 2133 }, { "epoch": 0.9664855072463768, "grad_norm": 0.5245308142347858, "learning_rate": 7.789748456661047e-05, "loss": 0.8857, "step": 2134 }, { "epoch": 0.9669384057971014, "grad_norm": 0.42507459472427, "learning_rate": 7.78934362649724e-05, "loss": 0.905, "step": 2135 }, { "epoch": 0.967391304347826, "grad_norm": 0.3261288319082701, "learning_rate": 7.788938417506746e-05, "loss": 0.8986, "step": 2136 }, { "epoch": 0.9678442028985508, "grad_norm": 0.26281355244076027, "learning_rate": 7.788532829730074e-05, "loss": 0.9016, "step": 2137 }, { "epoch": 0.9682971014492754, "grad_norm": 0.3646004941742128, "learning_rate": 7.78812686320777e-05, "loss": 0.9019, "step": 2138 }, { "epoch": 0.96875, "grad_norm": 0.45457188089955364, "learning_rate": 7.787720517980424e-05, "loss": 0.9201, "step": 2139 }, { "epoch": 0.9692028985507246, "grad_norm": 0.4803144810619339, "learning_rate": 7.787313794088653e-05, "loss": 0.8769, "step": 2140 }, { "epoch": 0.9696557971014492, "grad_norm": 0.53050586702709, "learning_rate": 7.786906691573121e-05, "loss": 0.8978, "step": 2141 }, { "epoch": 0.970108695652174, "grad_norm": 0.565227595726983, "learning_rate": 7.786499210474527e-05, "loss": 0.8969, "step": 2142 }, { "epoch": 0.9705615942028986, "grad_norm": 0.7731118813600508, "learning_rate": 7.786091350833606e-05, "loss": 0.9136, "step": 2143 }, { "epoch": 0.9710144927536232, "grad_norm": 0.9883645349628133, "learning_rate": 7.785683112691135e-05, "loss": 0.929, "step": 2144 }, { "epoch": 0.9714673913043478, "grad_norm": 1.1691703742078166, "learning_rate": 7.785274496087921e-05, "loss": 0.8906, "step": 2145 }, { "epoch": 0.9719202898550725, "grad_norm": 0.7087465286994377, "learning_rate": 7.784865501064818e-05, "loss": 0.9017, "step": 2146 }, { "epoch": 0.9723731884057971, "grad_norm": 0.5423648146800102, "learning_rate": 7.784456127662715e-05, "loss": 0.9184, "step": 2147 }, { "epoch": 0.9728260869565217, "grad_norm": 0.4785319048061186, "learning_rate": 7.784046375922536e-05, "loss": 0.9043, "step": 2148 }, { "epoch": 0.9732789855072463, "grad_norm": 0.43715832806800964, "learning_rate": 7.783636245885243e-05, "loss": 0.9355, "step": 2149 }, { "epoch": 0.9737318840579711, "grad_norm": 0.37577394310850387, "learning_rate": 7.783225737591838e-05, "loss": 0.9063, "step": 2150 }, { "epoch": 0.9741847826086957, "grad_norm": 0.4490928679760909, "learning_rate": 7.782814851083364e-05, "loss": 0.913, "step": 2151 }, { "epoch": 0.9746376811594203, "grad_norm": 0.7564463183138898, "learning_rate": 7.782403586400894e-05, "loss": 0.9093, "step": 2152 }, { "epoch": 0.9750905797101449, "grad_norm": 0.9475319336449197, "learning_rate": 7.781991943585542e-05, "loss": 0.8788, "step": 2153 }, { "epoch": 0.9755434782608695, "grad_norm": 1.100231445849726, "learning_rate": 7.781579922678463e-05, "loss": 0.9206, "step": 2154 }, { "epoch": 0.9759963768115942, "grad_norm": 1.0358821685661606, "learning_rate": 7.781167523720848e-05, "loss": 0.9174, "step": 2155 }, { "epoch": 0.9764492753623188, "grad_norm": 1.0743999701929288, "learning_rate": 7.780754746753922e-05, "loss": 0.9208, "step": 2156 }, { "epoch": 0.9769021739130435, "grad_norm": 0.8471042481715843, "learning_rate": 7.780341591818954e-05, "loss": 0.9079, "step": 2157 }, { "epoch": 0.9773550724637681, "grad_norm": 0.5912861273328422, "learning_rate": 7.779928058957247e-05, "loss": 0.8886, "step": 2158 }, { "epoch": 0.9778079710144928, "grad_norm": 0.4604738966160315, "learning_rate": 7.779514148210141e-05, "loss": 0.9227, "step": 2159 }, { "epoch": 0.9782608695652174, "grad_norm": 0.39647062512402376, "learning_rate": 7.779099859619017e-05, "loss": 0.9119, "step": 2160 }, { "epoch": 0.978713768115942, "grad_norm": 0.3772069957742732, "learning_rate": 7.778685193225292e-05, "loss": 0.9023, "step": 2161 }, { "epoch": 0.9791666666666666, "grad_norm": 0.3895187064515661, "learning_rate": 7.77827014907042e-05, "loss": 0.9221, "step": 2162 }, { "epoch": 0.9796195652173914, "grad_norm": 0.5041706377030248, "learning_rate": 7.777854727195892e-05, "loss": 0.9025, "step": 2163 }, { "epoch": 0.980072463768116, "grad_norm": 0.6279340014759656, "learning_rate": 7.777438927643244e-05, "loss": 0.905, "step": 2164 }, { "epoch": 0.9805253623188406, "grad_norm": 0.7176142563132402, "learning_rate": 7.777022750454039e-05, "loss": 0.9189, "step": 2165 }, { "epoch": 0.9809782608695652, "grad_norm": 0.7234576027816606, "learning_rate": 7.776606195669883e-05, "loss": 0.8918, "step": 2166 }, { "epoch": 0.9814311594202898, "grad_norm": 0.6978224794387285, "learning_rate": 7.776189263332423e-05, "loss": 0.9049, "step": 2167 }, { "epoch": 0.9818840579710145, "grad_norm": 0.6722613658101512, "learning_rate": 7.775771953483338e-05, "loss": 0.8907, "step": 2168 }, { "epoch": 0.9823369565217391, "grad_norm": 0.6235163581010537, "learning_rate": 7.775354266164348e-05, "loss": 0.9153, "step": 2169 }, { "epoch": 0.9827898550724637, "grad_norm": 0.5724111346209878, "learning_rate": 7.774936201417209e-05, "loss": 0.9183, "step": 2170 }, { "epoch": 0.9832427536231884, "grad_norm": 0.5970262432489175, "learning_rate": 7.774517759283715e-05, "loss": 0.8938, "step": 2171 }, { "epoch": 0.9836956521739131, "grad_norm": 1.0842745451889988, "learning_rate": 7.774098939805701e-05, "loss": 0.9077, "step": 2172 }, { "epoch": 0.9841485507246377, "grad_norm": 0.3716481786897261, "learning_rate": 7.773679743025033e-05, "loss": 0.895, "step": 2173 }, { "epoch": 0.9846014492753623, "grad_norm": 0.28679015245770895, "learning_rate": 7.773260168983624e-05, "loss": 0.8883, "step": 2174 }, { "epoch": 0.9850543478260869, "grad_norm": 0.3810407315881486, "learning_rate": 7.772840217723415e-05, "loss": 0.9052, "step": 2175 }, { "epoch": 0.9855072463768116, "grad_norm": 0.55983666608048, "learning_rate": 7.772419889286392e-05, "loss": 0.9228, "step": 2176 }, { "epoch": 0.9859601449275363, "grad_norm": 0.6422229800847092, "learning_rate": 7.771999183714575e-05, "loss": 0.9043, "step": 2177 }, { "epoch": 0.9864130434782609, "grad_norm": 0.7980559414634971, "learning_rate": 7.771578101050021e-05, "loss": 0.9076, "step": 2178 }, { "epoch": 0.9868659420289855, "grad_norm": 0.9693893447131114, "learning_rate": 7.771156641334829e-05, "loss": 0.9223, "step": 2179 }, { "epoch": 0.9873188405797102, "grad_norm": 1.099277340373941, "learning_rate": 7.770734804611129e-05, "loss": 0.9047, "step": 2180 }, { "epoch": 0.9877717391304348, "grad_norm": 0.8048030747911339, "learning_rate": 7.770312590921098e-05, "loss": 0.9018, "step": 2181 }, { "epoch": 0.9882246376811594, "grad_norm": 0.6898213073160384, "learning_rate": 7.769890000306943e-05, "loss": 0.9429, "step": 2182 }, { "epoch": 0.988677536231884, "grad_norm": 0.6240963972531942, "learning_rate": 7.769467032810909e-05, "loss": 0.9116, "step": 2183 }, { "epoch": 0.9891304347826086, "grad_norm": 0.5366677281067928, "learning_rate": 7.769043688475283e-05, "loss": 0.9143, "step": 2184 }, { "epoch": 0.9895833333333334, "grad_norm": 0.49289248837309807, "learning_rate": 7.768619967342386e-05, "loss": 0.9135, "step": 2185 }, { "epoch": 0.990036231884058, "grad_norm": 0.566862313705527, "learning_rate": 7.76819586945458e-05, "loss": 0.9065, "step": 2186 }, { "epoch": 0.9904891304347826, "grad_norm": 0.7080885143303824, "learning_rate": 7.767771394854263e-05, "loss": 0.8841, "step": 2187 }, { "epoch": 0.9909420289855072, "grad_norm": 0.8000614088219198, "learning_rate": 7.767346543583867e-05, "loss": 0.9151, "step": 2188 }, { "epoch": 0.9913949275362319, "grad_norm": 0.8206268004396964, "learning_rate": 7.766921315685868e-05, "loss": 0.8987, "step": 2189 }, { "epoch": 0.9918478260869565, "grad_norm": 0.8094995599213618, "learning_rate": 7.766495711202777e-05, "loss": 0.8925, "step": 2190 }, { "epoch": 0.9923007246376812, "grad_norm": 0.7034563142103311, "learning_rate": 7.766069730177139e-05, "loss": 0.8836, "step": 2191 }, { "epoch": 0.9927536231884058, "grad_norm": 0.5975613529499801, "learning_rate": 7.765643372651543e-05, "loss": 0.912, "step": 2192 }, { "epoch": 0.9932065217391305, "grad_norm": 0.5190484024276882, "learning_rate": 7.765216638668612e-05, "loss": 0.9112, "step": 2193 }, { "epoch": 0.9936594202898551, "grad_norm": 0.47570350944318457, "learning_rate": 7.764789528271008e-05, "loss": 0.8974, "step": 2194 }, { "epoch": 0.9941123188405797, "grad_norm": 0.5155699099254601, "learning_rate": 7.764362041501429e-05, "loss": 0.9042, "step": 2195 }, { "epoch": 0.9945652173913043, "grad_norm": 0.5963732051752392, "learning_rate": 7.763934178402611e-05, "loss": 0.9135, "step": 2196 }, { "epoch": 0.9950181159420289, "grad_norm": 0.6465246137562862, "learning_rate": 7.763505939017328e-05, "loss": 0.9129, "step": 2197 }, { "epoch": 0.9954710144927537, "grad_norm": 0.6888686708130356, "learning_rate": 7.763077323388393e-05, "loss": 0.9167, "step": 2198 }, { "epoch": 0.9959239130434783, "grad_norm": 0.662117782638028, "learning_rate": 7.762648331558656e-05, "loss": 0.919, "step": 2199 }, { "epoch": 0.9963768115942029, "grad_norm": 0.7157087244356455, "learning_rate": 7.762218963571002e-05, "loss": 0.8987, "step": 2200 }, { "epoch": 0.9968297101449275, "grad_norm": 0.7754358905080657, "learning_rate": 7.761789219468356e-05, "loss": 0.8999, "step": 2201 }, { "epoch": 0.9972826086956522, "grad_norm": 0.7996482287585124, "learning_rate": 7.76135909929368e-05, "loss": 0.903, "step": 2202 }, { "epoch": 0.9977355072463768, "grad_norm": 0.7219261850686745, "learning_rate": 7.760928603089975e-05, "loss": 0.8945, "step": 2203 }, { "epoch": 0.9981884057971014, "grad_norm": 0.694420666884844, "learning_rate": 7.760497730900279e-05, "loss": 0.9008, "step": 2204 }, { "epoch": 0.998641304347826, "grad_norm": 0.7567300876962233, "learning_rate": 7.760066482767665e-05, "loss": 0.8793, "step": 2205 }, { "epoch": 0.9990942028985508, "grad_norm": 0.8068419890012675, "learning_rate": 7.759634858735247e-05, "loss": 0.8917, "step": 2206 }, { "epoch": 0.9995471014492754, "grad_norm": 0.6898863152689747, "learning_rate": 7.759202858846174e-05, "loss": 0.883, "step": 2207 }, { "epoch": 1.0, "grad_norm": 0.549658388681855, "learning_rate": 7.758770483143634e-05, "loss": 0.8957, "step": 2208 }, { "epoch": 1.0004528985507246, "grad_norm": 0.43886989549464106, "learning_rate": 7.758337731670854e-05, "loss": 0.8738, "step": 2209 }, { "epoch": 1.0009057971014492, "grad_norm": 0.3360845721435473, "learning_rate": 7.757904604471093e-05, "loss": 0.8943, "step": 2210 }, { "epoch": 1.0013586956521738, "grad_norm": 0.3045049380887944, "learning_rate": 7.757471101587655e-05, "loss": 0.8726, "step": 2211 }, { "epoch": 1.0018115942028984, "grad_norm": 0.3698333848357665, "learning_rate": 7.757037223063877e-05, "loss": 0.8933, "step": 2212 }, { "epoch": 1.0022644927536233, "grad_norm": 0.45517954167436303, "learning_rate": 7.756602968943133e-05, "loss": 0.8916, "step": 2213 }, { "epoch": 1.002717391304348, "grad_norm": 0.43726776894070823, "learning_rate": 7.75616833926884e-05, "loss": 0.8856, "step": 2214 }, { "epoch": 1.0031702898550725, "grad_norm": 0.3937319279812047, "learning_rate": 7.755733334084443e-05, "loss": 0.8881, "step": 2215 }, { "epoch": 1.0036231884057971, "grad_norm": 0.34582409550254134, "learning_rate": 7.755297953433435e-05, "loss": 0.9168, "step": 2216 }, { "epoch": 1.0040760869565217, "grad_norm": 0.4088553696700257, "learning_rate": 7.75486219735934e-05, "loss": 0.9052, "step": 2217 }, { "epoch": 1.0045289855072463, "grad_norm": 0.48376613886420516, "learning_rate": 7.754426065905719e-05, "loss": 0.8793, "step": 2218 }, { "epoch": 1.004981884057971, "grad_norm": 0.4460505138138079, "learning_rate": 7.753989559116177e-05, "loss": 0.8777, "step": 2219 }, { "epoch": 1.0054347826086956, "grad_norm": 0.4453481072452792, "learning_rate": 7.753552677034349e-05, "loss": 0.8716, "step": 2220 }, { "epoch": 1.0058876811594204, "grad_norm": 0.5626789304616607, "learning_rate": 7.75311541970391e-05, "loss": 0.8938, "step": 2221 }, { "epoch": 1.006340579710145, "grad_norm": 0.6498810110643773, "learning_rate": 7.752677787168578e-05, "loss": 0.8811, "step": 2222 }, { "epoch": 1.0067934782608696, "grad_norm": 0.7438173259092831, "learning_rate": 7.752239779472101e-05, "loss": 0.8789, "step": 2223 }, { "epoch": 1.0072463768115942, "grad_norm": 0.8219540744357341, "learning_rate": 7.751801396658266e-05, "loss": 0.8837, "step": 2224 }, { "epoch": 1.0076992753623188, "grad_norm": 0.8484001857759251, "learning_rate": 7.7513626387709e-05, "loss": 0.8839, "step": 2225 }, { "epoch": 1.0081521739130435, "grad_norm": 0.8870191365240497, "learning_rate": 7.750923505853868e-05, "loss": 0.8871, "step": 2226 }, { "epoch": 1.008605072463768, "grad_norm": 0.9458686538807827, "learning_rate": 7.750483997951068e-05, "loss": 0.9112, "step": 2227 }, { "epoch": 1.0090579710144927, "grad_norm": 0.7631827032062082, "learning_rate": 7.75004411510644e-05, "loss": 0.8844, "step": 2228 }, { "epoch": 1.0095108695652173, "grad_norm": 0.6179646583626611, "learning_rate": 7.749603857363958e-05, "loss": 0.8973, "step": 2229 }, { "epoch": 1.0099637681159421, "grad_norm": 0.5402889227835093, "learning_rate": 7.74916322476764e-05, "loss": 0.9003, "step": 2230 }, { "epoch": 1.0104166666666667, "grad_norm": 0.5064900706230071, "learning_rate": 7.74872221736153e-05, "loss": 0.8805, "step": 2231 }, { "epoch": 1.0108695652173914, "grad_norm": 0.5111910513131989, "learning_rate": 7.748280835189721e-05, "loss": 0.8997, "step": 2232 }, { "epoch": 1.011322463768116, "grad_norm": 0.6088464389434033, "learning_rate": 7.747839078296337e-05, "loss": 0.889, "step": 2233 }, { "epoch": 1.0117753623188406, "grad_norm": 0.6776977360821357, "learning_rate": 7.747396946725541e-05, "loss": 0.8944, "step": 2234 }, { "epoch": 1.0122282608695652, "grad_norm": 0.6964335760454986, "learning_rate": 7.746954440521535e-05, "loss": 0.885, "step": 2235 }, { "epoch": 1.0126811594202898, "grad_norm": 0.6817408598653796, "learning_rate": 7.746511559728556e-05, "loss": 0.8766, "step": 2236 }, { "epoch": 1.0131340579710144, "grad_norm": 0.7853865552741826, "learning_rate": 7.746068304390878e-05, "loss": 0.8953, "step": 2237 }, { "epoch": 1.013586956521739, "grad_norm": 0.791606207152818, "learning_rate": 7.745624674552817e-05, "loss": 0.8848, "step": 2238 }, { "epoch": 1.0140398550724639, "grad_norm": 0.7119547106001769, "learning_rate": 7.745180670258722e-05, "loss": 0.8911, "step": 2239 }, { "epoch": 1.0144927536231885, "grad_norm": 0.5737237527107696, "learning_rate": 7.744736291552979e-05, "loss": 0.8872, "step": 2240 }, { "epoch": 1.014945652173913, "grad_norm": 0.4840966884011419, "learning_rate": 7.744291538480015e-05, "loss": 0.916, "step": 2241 }, { "epoch": 1.0153985507246377, "grad_norm": 0.4484076671382358, "learning_rate": 7.743846411084294e-05, "loss": 0.8845, "step": 2242 }, { "epoch": 1.0158514492753623, "grad_norm": 0.3976916338220263, "learning_rate": 7.743400909410315e-05, "loss": 0.8928, "step": 2243 }, { "epoch": 1.016304347826087, "grad_norm": 0.3488269326219715, "learning_rate": 7.742955033502613e-05, "loss": 0.8863, "step": 2244 }, { "epoch": 1.0167572463768115, "grad_norm": 0.3225014442452831, "learning_rate": 7.742508783405767e-05, "loss": 0.8931, "step": 2245 }, { "epoch": 1.0172101449275361, "grad_norm": 0.30649033424935745, "learning_rate": 7.742062159164387e-05, "loss": 0.8714, "step": 2246 }, { "epoch": 1.017663043478261, "grad_norm": 0.30831710250822664, "learning_rate": 7.741615160823122e-05, "loss": 0.9077, "step": 2247 }, { "epoch": 1.0181159420289856, "grad_norm": 0.39436638053477163, "learning_rate": 7.74116778842666e-05, "loss": 0.8877, "step": 2248 }, { "epoch": 1.0185688405797102, "grad_norm": 0.5145522323583774, "learning_rate": 7.740720042019727e-05, "loss": 0.8949, "step": 2249 }, { "epoch": 1.0190217391304348, "grad_norm": 0.5938765889073946, "learning_rate": 7.740271921647085e-05, "loss": 0.8732, "step": 2250 }, { "epoch": 1.0194746376811594, "grad_norm": 0.7656891526463047, "learning_rate": 7.73982342735353e-05, "loss": 0.8897, "step": 2251 }, { "epoch": 1.019927536231884, "grad_norm": 0.9062964152214268, "learning_rate": 7.739374559183901e-05, "loss": 0.9103, "step": 2252 }, { "epoch": 1.0203804347826086, "grad_norm": 1.006109546886707, "learning_rate": 7.738925317183072e-05, "loss": 0.8945, "step": 2253 }, { "epoch": 1.0208333333333333, "grad_norm": 1.169601113358342, "learning_rate": 7.738475701395955e-05, "loss": 0.9038, "step": 2254 }, { "epoch": 1.0212862318840579, "grad_norm": 0.7387712871498835, "learning_rate": 7.738025711867496e-05, "loss": 0.8921, "step": 2255 }, { "epoch": 1.0217391304347827, "grad_norm": 0.49183412279855626, "learning_rate": 7.737575348642685e-05, "loss": 0.8747, "step": 2256 }, { "epoch": 1.0221920289855073, "grad_norm": 0.4762284764995665, "learning_rate": 7.737124611766543e-05, "loss": 0.8832, "step": 2257 }, { "epoch": 1.022644927536232, "grad_norm": 0.6862850783278787, "learning_rate": 7.736673501284131e-05, "loss": 0.8925, "step": 2258 }, { "epoch": 1.0230978260869565, "grad_norm": 0.8368412756468655, "learning_rate": 7.736222017240549e-05, "loss": 0.8989, "step": 2259 }, { "epoch": 1.0235507246376812, "grad_norm": 0.8531601036040132, "learning_rate": 7.735770159680931e-05, "loss": 0.8764, "step": 2260 }, { "epoch": 1.0240036231884058, "grad_norm": 0.8381085810536463, "learning_rate": 7.735317928650451e-05, "loss": 0.9081, "step": 2261 }, { "epoch": 1.0244565217391304, "grad_norm": 0.705397433048298, "learning_rate": 7.734865324194317e-05, "loss": 0.899, "step": 2262 }, { "epoch": 1.024909420289855, "grad_norm": 0.6258127596200551, "learning_rate": 7.734412346357779e-05, "loss": 0.8893, "step": 2263 }, { "epoch": 1.0253623188405796, "grad_norm": 0.6668398607034223, "learning_rate": 7.73395899518612e-05, "loss": 0.9082, "step": 2264 }, { "epoch": 1.0258152173913044, "grad_norm": 0.6581318932951122, "learning_rate": 7.733505270724665e-05, "loss": 0.8955, "step": 2265 }, { "epoch": 1.026268115942029, "grad_norm": 0.7298623529783602, "learning_rate": 7.733051173018772e-05, "loss": 0.8967, "step": 2266 }, { "epoch": 1.0267210144927537, "grad_norm": 0.7496454076259262, "learning_rate": 7.732596702113837e-05, "loss": 0.8861, "step": 2267 }, { "epoch": 1.0271739130434783, "grad_norm": 0.6645184725435862, "learning_rate": 7.732141858055294e-05, "loss": 0.9117, "step": 2268 }, { "epoch": 1.0276268115942029, "grad_norm": 0.6513887921326186, "learning_rate": 7.731686640888617e-05, "loss": 0.9048, "step": 2269 }, { "epoch": 1.0280797101449275, "grad_norm": 0.6409427625608902, "learning_rate": 7.731231050659312e-05, "loss": 0.8904, "step": 2270 }, { "epoch": 1.028532608695652, "grad_norm": 0.5982598559893084, "learning_rate": 7.730775087412926e-05, "loss": 0.8686, "step": 2271 }, { "epoch": 1.0289855072463767, "grad_norm": 0.5912196935860569, "learning_rate": 7.730318751195043e-05, "loss": 0.891, "step": 2272 }, { "epoch": 1.0294384057971016, "grad_norm": 0.5919185832271039, "learning_rate": 7.729862042051283e-05, "loss": 0.8788, "step": 2273 }, { "epoch": 1.0298913043478262, "grad_norm": 0.4681397793943598, "learning_rate": 7.729404960027305e-05, "loss": 0.8926, "step": 2274 }, { "epoch": 1.0303442028985508, "grad_norm": 0.40087047263275843, "learning_rate": 7.728947505168803e-05, "loss": 0.9053, "step": 2275 }, { "epoch": 1.0307971014492754, "grad_norm": 0.4712578583051231, "learning_rate": 7.72848967752151e-05, "loss": 0.8604, "step": 2276 }, { "epoch": 1.03125, "grad_norm": 0.4511999595826908, "learning_rate": 7.728031477131195e-05, "loss": 0.8978, "step": 2277 }, { "epoch": 1.0317028985507246, "grad_norm": 0.3538972457100566, "learning_rate": 7.727572904043666e-05, "loss": 0.8874, "step": 2278 }, { "epoch": 1.0321557971014492, "grad_norm": 0.34742094462882184, "learning_rate": 7.727113958304767e-05, "loss": 0.8942, "step": 2279 }, { "epoch": 1.0326086956521738, "grad_norm": 0.4396439055927605, "learning_rate": 7.726654639960379e-05, "loss": 0.9038, "step": 2280 }, { "epoch": 1.0330615942028984, "grad_norm": 0.55803033442929, "learning_rate": 7.726194949056421e-05, "loss": 0.9012, "step": 2281 }, { "epoch": 1.0335144927536233, "grad_norm": 0.6000446325695301, "learning_rate": 7.725734885638852e-05, "loss": 0.8806, "step": 2282 }, { "epoch": 1.033967391304348, "grad_norm": 0.6434746097408198, "learning_rate": 7.725274449753659e-05, "loss": 0.8935, "step": 2283 }, { "epoch": 1.0344202898550725, "grad_norm": 0.776210233314016, "learning_rate": 7.724813641446878e-05, "loss": 0.8957, "step": 2284 }, { "epoch": 1.0348731884057971, "grad_norm": 0.947829056182448, "learning_rate": 7.724352460764574e-05, "loss": 0.9046, "step": 2285 }, { "epoch": 1.0353260869565217, "grad_norm": 1.0896371888755225, "learning_rate": 7.723890907752854e-05, "loss": 0.8781, "step": 2286 }, { "epoch": 1.0357789855072463, "grad_norm": 0.8365813330663285, "learning_rate": 7.723428982457857e-05, "loss": 0.9151, "step": 2287 }, { "epoch": 1.036231884057971, "grad_norm": 0.6364557349403086, "learning_rate": 7.722966684925766e-05, "loss": 0.8863, "step": 2288 }, { "epoch": 1.0366847826086956, "grad_norm": 0.49807436869314653, "learning_rate": 7.722504015202795e-05, "loss": 0.8764, "step": 2289 }, { "epoch": 1.0371376811594204, "grad_norm": 0.35019669141193716, "learning_rate": 7.722040973335199e-05, "loss": 0.904, "step": 2290 }, { "epoch": 1.037590579710145, "grad_norm": 0.3387299139872143, "learning_rate": 7.72157755936927e-05, "loss": 0.9305, "step": 2291 }, { "epoch": 1.0380434782608696, "grad_norm": 0.4683175890050408, "learning_rate": 7.721113773351333e-05, "loss": 0.9048, "step": 2292 }, { "epoch": 1.0384963768115942, "grad_norm": 0.6071533476480511, "learning_rate": 7.720649615327757e-05, "loss": 0.8923, "step": 2293 }, { "epoch": 1.0389492753623188, "grad_norm": 0.7020979153015001, "learning_rate": 7.720185085344942e-05, "loss": 0.8924, "step": 2294 }, { "epoch": 1.0394021739130435, "grad_norm": 0.7631023669003375, "learning_rate": 7.71972018344933e-05, "loss": 0.8786, "step": 2295 }, { "epoch": 1.039855072463768, "grad_norm": 0.8038958883215559, "learning_rate": 7.719254909687396e-05, "loss": 0.894, "step": 2296 }, { "epoch": 1.0403079710144927, "grad_norm": 0.8714585317528404, "learning_rate": 7.718789264105655e-05, "loss": 0.8999, "step": 2297 }, { "epoch": 1.0407608695652173, "grad_norm": 0.8590608273605552, "learning_rate": 7.71832324675066e-05, "loss": 0.9048, "step": 2298 }, { "epoch": 1.0412137681159421, "grad_norm": 0.8612332583773858, "learning_rate": 7.717856857668996e-05, "loss": 0.9031, "step": 2299 }, { "epoch": 1.0416666666666667, "grad_norm": 0.853626791611736, "learning_rate": 7.71739009690729e-05, "loss": 0.8943, "step": 2300 }, { "epoch": 1.0421195652173914, "grad_norm": 0.8417954182349082, "learning_rate": 7.716922964512206e-05, "loss": 0.8697, "step": 2301 }, { "epoch": 1.042572463768116, "grad_norm": 0.6840482426695462, "learning_rate": 7.716455460530444e-05, "loss": 0.8885, "step": 2302 }, { "epoch": 1.0430253623188406, "grad_norm": 0.4668712651317937, "learning_rate": 7.71598758500874e-05, "loss": 0.8886, "step": 2303 }, { "epoch": 1.0434782608695652, "grad_norm": 0.34754626532067734, "learning_rate": 7.715519337993869e-05, "loss": 0.9016, "step": 2304 }, { "epoch": 1.0439311594202898, "grad_norm": 0.44578012688481156, "learning_rate": 7.715050719532642e-05, "loss": 0.876, "step": 2305 }, { "epoch": 1.0443840579710144, "grad_norm": 0.7901106463728861, "learning_rate": 7.714581729671907e-05, "loss": 0.9062, "step": 2306 }, { "epoch": 1.044836956521739, "grad_norm": 1.1723706565649061, "learning_rate": 7.71411236845855e-05, "loss": 0.8896, "step": 2307 }, { "epoch": 1.0452898550724639, "grad_norm": 0.8811378708481702, "learning_rate": 7.713642635939493e-05, "loss": 0.8885, "step": 2308 }, { "epoch": 1.0457427536231885, "grad_norm": 0.7742414130417807, "learning_rate": 7.7131725321617e-05, "loss": 0.8881, "step": 2309 }, { "epoch": 1.046195652173913, "grad_norm": 0.8134220025435568, "learning_rate": 7.712702057172164e-05, "loss": 0.8888, "step": 2310 }, { "epoch": 1.0466485507246377, "grad_norm": 0.9053003943076131, "learning_rate": 7.712231211017917e-05, "loss": 0.8712, "step": 2311 }, { "epoch": 1.0471014492753623, "grad_norm": 1.0233878308239581, "learning_rate": 7.711759993746037e-05, "loss": 0.8732, "step": 2312 }, { "epoch": 1.047554347826087, "grad_norm": 0.8974558655356941, "learning_rate": 7.711288405403628e-05, "loss": 0.8895, "step": 2313 }, { "epoch": 1.0480072463768115, "grad_norm": 0.7478456518036504, "learning_rate": 7.710816446037835e-05, "loss": 0.8887, "step": 2314 }, { "epoch": 1.0484601449275361, "grad_norm": 0.49212177090304293, "learning_rate": 7.710344115695841e-05, "loss": 0.904, "step": 2315 }, { "epoch": 1.048913043478261, "grad_norm": 0.4721015088065946, "learning_rate": 7.709871414424869e-05, "loss": 0.8823, "step": 2316 }, { "epoch": 1.0493659420289856, "grad_norm": 0.6050850855338306, "learning_rate": 7.709398342272169e-05, "loss": 0.9011, "step": 2317 }, { "epoch": 1.0498188405797102, "grad_norm": 0.5992886053401605, "learning_rate": 7.708924899285042e-05, "loss": 0.867, "step": 2318 }, { "epoch": 1.0502717391304348, "grad_norm": 0.5025050494345249, "learning_rate": 7.708451085510813e-05, "loss": 0.9, "step": 2319 }, { "epoch": 1.0507246376811594, "grad_norm": 0.5139298057576684, "learning_rate": 7.707976900996854e-05, "loss": 0.8775, "step": 2320 }, { "epoch": 1.051177536231884, "grad_norm": 0.6277534315489014, "learning_rate": 7.707502345790568e-05, "loss": 0.8915, "step": 2321 }, { "epoch": 1.0516304347826086, "grad_norm": 0.7581826662453066, "learning_rate": 7.707027419939398e-05, "loss": 0.8883, "step": 2322 }, { "epoch": 1.0520833333333333, "grad_norm": 0.8487734135483486, "learning_rate": 7.706552123490822e-05, "loss": 0.8975, "step": 2323 }, { "epoch": 1.0525362318840579, "grad_norm": 0.9499165349459711, "learning_rate": 7.706076456492356e-05, "loss": 0.8855, "step": 2324 }, { "epoch": 1.0529891304347827, "grad_norm": 1.0671331298974944, "learning_rate": 7.705600418991556e-05, "loss": 0.9015, "step": 2325 }, { "epoch": 1.0534420289855073, "grad_norm": 0.9369441607436143, "learning_rate": 7.705124011036008e-05, "loss": 0.9073, "step": 2326 }, { "epoch": 1.053894927536232, "grad_norm": 0.7181483310073715, "learning_rate": 7.704647232673342e-05, "loss": 0.8976, "step": 2327 }, { "epoch": 1.0543478260869565, "grad_norm": 0.6198967951946955, "learning_rate": 7.704170083951221e-05, "loss": 0.897, "step": 2328 }, { "epoch": 1.0548007246376812, "grad_norm": 0.6627181779456826, "learning_rate": 7.703692564917348e-05, "loss": 0.8912, "step": 2329 }, { "epoch": 1.0552536231884058, "grad_norm": 0.8041933423713262, "learning_rate": 7.70321467561946e-05, "loss": 0.8756, "step": 2330 }, { "epoch": 1.0557065217391304, "grad_norm": 0.7769910345690151, "learning_rate": 7.702736416105333e-05, "loss": 0.8994, "step": 2331 }, { "epoch": 1.056159420289855, "grad_norm": 0.7761616743370342, "learning_rate": 7.702257786422778e-05, "loss": 0.869, "step": 2332 }, { "epoch": 1.0566123188405796, "grad_norm": 0.7217985514944651, "learning_rate": 7.701778786619646e-05, "loss": 0.9002, "step": 2333 }, { "epoch": 1.0570652173913044, "grad_norm": 0.7437889836706068, "learning_rate": 7.701299416743823e-05, "loss": 0.9109, "step": 2334 }, { "epoch": 1.057518115942029, "grad_norm": 0.8639808282939666, "learning_rate": 7.700819676843232e-05, "loss": 0.9022, "step": 2335 }, { "epoch": 1.0579710144927537, "grad_norm": 1.0023991624824309, "learning_rate": 7.700339566965833e-05, "loss": 0.9017, "step": 2336 }, { "epoch": 1.0584239130434783, "grad_norm": 1.0532812670402085, "learning_rate": 7.699859087159624e-05, "loss": 0.9039, "step": 2337 }, { "epoch": 1.0588768115942029, "grad_norm": 1.0120163198145944, "learning_rate": 7.699378237472639e-05, "loss": 0.9103, "step": 2338 }, { "epoch": 1.0593297101449275, "grad_norm": 1.0384429472475596, "learning_rate": 7.698897017952949e-05, "loss": 0.9019, "step": 2339 }, { "epoch": 1.059782608695652, "grad_norm": 0.9685424550155335, "learning_rate": 7.698415428648663e-05, "loss": 0.8803, "step": 2340 }, { "epoch": 1.0602355072463767, "grad_norm": 0.9349121319796454, "learning_rate": 7.697933469607927e-05, "loss": 0.8971, "step": 2341 }, { "epoch": 1.0606884057971016, "grad_norm": 0.9181582494634662, "learning_rate": 7.69745114087892e-05, "loss": 0.8933, "step": 2342 }, { "epoch": 1.0611413043478262, "grad_norm": 0.9943825797441826, "learning_rate": 7.696968442509866e-05, "loss": 0.8578, "step": 2343 }, { "epoch": 1.0615942028985508, "grad_norm": 0.9422279123768553, "learning_rate": 7.696485374549017e-05, "loss": 0.884, "step": 2344 }, { "epoch": 1.0620471014492754, "grad_norm": 0.7210999622024662, "learning_rate": 7.696001937044668e-05, "loss": 0.8839, "step": 2345 }, { "epoch": 1.0625, "grad_norm": 0.6120120445098435, "learning_rate": 7.695518130045147e-05, "loss": 0.9006, "step": 2346 }, { "epoch": 1.0629528985507246, "grad_norm": 0.4994536463525578, "learning_rate": 7.695033953598825e-05, "loss": 0.8835, "step": 2347 }, { "epoch": 1.0634057971014492, "grad_norm": 0.421179475844814, "learning_rate": 7.694549407754102e-05, "loss": 0.9071, "step": 2348 }, { "epoch": 1.0638586956521738, "grad_norm": 0.39596893317202897, "learning_rate": 7.69406449255942e-05, "loss": 0.8951, "step": 2349 }, { "epoch": 1.0643115942028984, "grad_norm": 0.5383723609312483, "learning_rate": 7.693579208063256e-05, "loss": 0.8757, "step": 2350 }, { "epoch": 1.0647644927536233, "grad_norm": 0.6547454215298161, "learning_rate": 7.693093554314128e-05, "loss": 0.888, "step": 2351 }, { "epoch": 1.065217391304348, "grad_norm": 0.6788267465616081, "learning_rate": 7.692607531360583e-05, "loss": 0.8885, "step": 2352 }, { "epoch": 1.0656702898550725, "grad_norm": 0.7061817965818302, "learning_rate": 7.692121139251213e-05, "loss": 0.9106, "step": 2353 }, { "epoch": 1.0661231884057971, "grad_norm": 0.5871857300510459, "learning_rate": 7.691634378034641e-05, "loss": 0.8766, "step": 2354 }, { "epoch": 1.0665760869565217, "grad_norm": 0.592354688289403, "learning_rate": 7.691147247759532e-05, "loss": 0.8762, "step": 2355 }, { "epoch": 1.0670289855072463, "grad_norm": 0.6966792538582084, "learning_rate": 7.690659748474583e-05, "loss": 0.8863, "step": 2356 }, { "epoch": 1.067481884057971, "grad_norm": 0.7242946475870478, "learning_rate": 7.690171880228529e-05, "loss": 0.8781, "step": 2357 }, { "epoch": 1.0679347826086956, "grad_norm": 0.7282649749869557, "learning_rate": 7.689683643070145e-05, "loss": 0.893, "step": 2358 }, { "epoch": 1.0683876811594204, "grad_norm": 0.8648464459017934, "learning_rate": 7.689195037048242e-05, "loss": 0.8874, "step": 2359 }, { "epoch": 1.068840579710145, "grad_norm": 1.0723191698164622, "learning_rate": 7.688706062211664e-05, "loss": 0.9062, "step": 2360 }, { "epoch": 1.0692934782608696, "grad_norm": 0.9552162672935571, "learning_rate": 7.688216718609296e-05, "loss": 0.8719, "step": 2361 }, { "epoch": 1.0697463768115942, "grad_norm": 0.7875976490205137, "learning_rate": 7.687727006290058e-05, "loss": 0.9078, "step": 2362 }, { "epoch": 1.0701992753623188, "grad_norm": 0.5860748045925367, "learning_rate": 7.687236925302908e-05, "loss": 0.8886, "step": 2363 }, { "epoch": 1.0706521739130435, "grad_norm": 0.38936258020211034, "learning_rate": 7.686746475696838e-05, "loss": 0.9028, "step": 2364 }, { "epoch": 1.071105072463768, "grad_norm": 0.26741444252661334, "learning_rate": 7.686255657520882e-05, "loss": 0.8853, "step": 2365 }, { "epoch": 1.0715579710144927, "grad_norm": 0.3643747743568967, "learning_rate": 7.685764470824106e-05, "loss": 0.8873, "step": 2366 }, { "epoch": 1.0720108695652173, "grad_norm": 0.4947989569414989, "learning_rate": 7.685272915655617e-05, "loss": 0.8947, "step": 2367 }, { "epoch": 1.0724637681159421, "grad_norm": 0.6475873985671657, "learning_rate": 7.684780992064553e-05, "loss": 0.902, "step": 2368 }, { "epoch": 1.0729166666666667, "grad_norm": 0.8560114260339472, "learning_rate": 7.684288700100095e-05, "loss": 0.9278, "step": 2369 }, { "epoch": 1.0733695652173914, "grad_norm": 1.0743926619298811, "learning_rate": 7.683796039811457e-05, "loss": 0.8977, "step": 2370 }, { "epoch": 1.073822463768116, "grad_norm": 0.9956404010211579, "learning_rate": 7.68330301124789e-05, "loss": 0.8917, "step": 2371 }, { "epoch": 1.0742753623188406, "grad_norm": 0.922928970542657, "learning_rate": 7.682809614458686e-05, "loss": 0.9036, "step": 2372 }, { "epoch": 1.0747282608695652, "grad_norm": 0.768681402106881, "learning_rate": 7.682315849493169e-05, "loss": 0.8904, "step": 2373 }, { "epoch": 1.0751811594202898, "grad_norm": 0.568242298158612, "learning_rate": 7.681821716400701e-05, "loss": 0.903, "step": 2374 }, { "epoch": 1.0756340579710144, "grad_norm": 0.4205695494826044, "learning_rate": 7.681327215230682e-05, "loss": 0.9036, "step": 2375 }, { "epoch": 1.0760869565217392, "grad_norm": 0.2961027388704721, "learning_rate": 7.680832346032547e-05, "loss": 0.8944, "step": 2376 }, { "epoch": 1.0765398550724639, "grad_norm": 0.38932915441745614, "learning_rate": 7.680337108855771e-05, "loss": 0.8725, "step": 2377 }, { "epoch": 1.0769927536231885, "grad_norm": 0.5839504465767931, "learning_rate": 7.679841503749861e-05, "loss": 0.9162, "step": 2378 }, { "epoch": 1.077445652173913, "grad_norm": 0.7554495309502608, "learning_rate": 7.679345530764365e-05, "loss": 0.8771, "step": 2379 }, { "epoch": 1.0778985507246377, "grad_norm": 1.0541051593613386, "learning_rate": 7.678849189948865e-05, "loss": 0.8757, "step": 2380 }, { "epoch": 1.0783514492753623, "grad_norm": 1.043147294560753, "learning_rate": 7.678352481352983e-05, "loss": 0.912, "step": 2381 }, { "epoch": 1.078804347826087, "grad_norm": 0.7629449914314885, "learning_rate": 7.677855405026375e-05, "loss": 0.8898, "step": 2382 }, { "epoch": 1.0792572463768115, "grad_norm": 0.5374492357341856, "learning_rate": 7.677357961018732e-05, "loss": 0.8897, "step": 2383 }, { "epoch": 1.0797101449275361, "grad_norm": 0.40074130012234865, "learning_rate": 7.676860149379789e-05, "loss": 0.8908, "step": 2384 }, { "epoch": 1.0801630434782608, "grad_norm": 0.3585418179343435, "learning_rate": 7.676361970159308e-05, "loss": 0.891, "step": 2385 }, { "epoch": 1.0806159420289856, "grad_norm": 0.4481139868081754, "learning_rate": 7.675863423407097e-05, "loss": 0.8842, "step": 2386 }, { "epoch": 1.0810688405797102, "grad_norm": 0.6931305773540279, "learning_rate": 7.675364509172993e-05, "loss": 0.884, "step": 2387 }, { "epoch": 1.0815217391304348, "grad_norm": 0.8097016820233438, "learning_rate": 7.674865227506876e-05, "loss": 0.8987, "step": 2388 }, { "epoch": 1.0819746376811594, "grad_norm": 0.791003283221729, "learning_rate": 7.67436557845866e-05, "loss": 0.9092, "step": 2389 }, { "epoch": 1.082427536231884, "grad_norm": 0.8052282249726614, "learning_rate": 7.673865562078292e-05, "loss": 0.8936, "step": 2390 }, { "epoch": 1.0828804347826086, "grad_norm": 0.7919367473221481, "learning_rate": 7.673365178415764e-05, "loss": 0.9014, "step": 2391 }, { "epoch": 1.0833333333333333, "grad_norm": 0.7131411372313886, "learning_rate": 7.672864427521097e-05, "loss": 0.9003, "step": 2392 }, { "epoch": 1.0837862318840579, "grad_norm": 0.7689134032583537, "learning_rate": 7.672363309444353e-05, "loss": 0.91, "step": 2393 }, { "epoch": 1.0842391304347827, "grad_norm": 0.6634085818466351, "learning_rate": 7.671861824235631e-05, "loss": 0.8614, "step": 2394 }, { "epoch": 1.0846920289855073, "grad_norm": 0.4846201184993437, "learning_rate": 7.671359971945062e-05, "loss": 0.8969, "step": 2395 }, { "epoch": 1.085144927536232, "grad_norm": 0.4238980615377807, "learning_rate": 7.67085775262282e-05, "loss": 0.8851, "step": 2396 }, { "epoch": 1.0855978260869565, "grad_norm": 0.4535780593542466, "learning_rate": 7.670355166319111e-05, "loss": 0.9149, "step": 2397 }, { "epoch": 1.0860507246376812, "grad_norm": 0.46556506037020534, "learning_rate": 7.669852213084181e-05, "loss": 0.8862, "step": 2398 }, { "epoch": 1.0865036231884058, "grad_norm": 0.4686475532362961, "learning_rate": 7.669348892968308e-05, "loss": 0.888, "step": 2399 }, { "epoch": 1.0869565217391304, "grad_norm": 0.3556494961274013, "learning_rate": 7.668845206021812e-05, "loss": 0.8959, "step": 2400 }, { "epoch": 1.087409420289855, "grad_norm": 0.4963533997038759, "learning_rate": 7.668341152295048e-05, "loss": 0.8913, "step": 2401 }, { "epoch": 1.0878623188405796, "grad_norm": 0.7134871883026699, "learning_rate": 7.667836731838406e-05, "loss": 0.8895, "step": 2402 }, { "epoch": 1.0883152173913044, "grad_norm": 0.7188975180350747, "learning_rate": 7.667331944702313e-05, "loss": 0.8805, "step": 2403 }, { "epoch": 1.088768115942029, "grad_norm": 0.7095449240439043, "learning_rate": 7.666826790937235e-05, "loss": 0.877, "step": 2404 }, { "epoch": 1.0892210144927537, "grad_norm": 0.815103280117197, "learning_rate": 7.666321270593671e-05, "loss": 0.8786, "step": 2405 }, { "epoch": 1.0896739130434783, "grad_norm": 0.9909141163573981, "learning_rate": 7.66581538372216e-05, "loss": 0.8789, "step": 2406 }, { "epoch": 1.0901268115942029, "grad_norm": 1.1127693707874458, "learning_rate": 7.665309130373276e-05, "loss": 0.8718, "step": 2407 }, { "epoch": 1.0905797101449275, "grad_norm": 0.829088430705116, "learning_rate": 7.664802510597631e-05, "loss": 0.8821, "step": 2408 }, { "epoch": 1.091032608695652, "grad_norm": 0.6411530426872737, "learning_rate": 7.664295524445872e-05, "loss": 0.9082, "step": 2409 }, { "epoch": 1.0914855072463767, "grad_norm": 0.47094028288123474, "learning_rate": 7.663788171968681e-05, "loss": 0.9009, "step": 2410 }, { "epoch": 1.0919384057971016, "grad_norm": 0.3435251050292715, "learning_rate": 7.663280453216782e-05, "loss": 0.8999, "step": 2411 }, { "epoch": 1.0923913043478262, "grad_norm": 0.39330776616061525, "learning_rate": 7.66277236824093e-05, "loss": 0.8835, "step": 2412 }, { "epoch": 1.0928442028985508, "grad_norm": 0.43179182148560263, "learning_rate": 7.662263917091922e-05, "loss": 0.8992, "step": 2413 }, { "epoch": 1.0932971014492754, "grad_norm": 0.43175580573281647, "learning_rate": 7.661755099820586e-05, "loss": 0.8664, "step": 2414 }, { "epoch": 1.09375, "grad_norm": 0.5139289508591898, "learning_rate": 7.66124591647779e-05, "loss": 0.8839, "step": 2415 }, { "epoch": 1.0942028985507246, "grad_norm": 0.6594085114825056, "learning_rate": 7.660736367114437e-05, "loss": 0.8779, "step": 2416 }, { "epoch": 1.0946557971014492, "grad_norm": 0.8966904992128094, "learning_rate": 7.66022645178147e-05, "loss": 0.8964, "step": 2417 }, { "epoch": 1.0951086956521738, "grad_norm": 1.1000117253612818, "learning_rate": 7.659716170529863e-05, "loss": 0.8974, "step": 2418 }, { "epoch": 1.0955615942028984, "grad_norm": 0.8530929651203062, "learning_rate": 7.659205523410633e-05, "loss": 0.8809, "step": 2419 }, { "epoch": 1.0960144927536233, "grad_norm": 0.7638397236198307, "learning_rate": 7.658694510474826e-05, "loss": 0.8793, "step": 2420 }, { "epoch": 1.096467391304348, "grad_norm": 0.6833491509835794, "learning_rate": 7.658183131773533e-05, "loss": 0.8921, "step": 2421 }, { "epoch": 1.0969202898550725, "grad_norm": 0.5880505400721283, "learning_rate": 7.657671387357875e-05, "loss": 0.8886, "step": 2422 }, { "epoch": 1.0973731884057971, "grad_norm": 0.5308278274602037, "learning_rate": 7.65715927727901e-05, "loss": 0.8934, "step": 2423 }, { "epoch": 1.0978260869565217, "grad_norm": 0.4913407548644938, "learning_rate": 7.65664680158814e-05, "loss": 0.8876, "step": 2424 }, { "epoch": 1.0982789855072463, "grad_norm": 0.4854403106289595, "learning_rate": 7.656133960336495e-05, "loss": 0.911, "step": 2425 }, { "epoch": 1.098731884057971, "grad_norm": 0.5003456229515114, "learning_rate": 7.655620753575342e-05, "loss": 0.8955, "step": 2426 }, { "epoch": 1.0991847826086956, "grad_norm": 0.49173911088426236, "learning_rate": 7.655107181355992e-05, "loss": 0.8974, "step": 2427 }, { "epoch": 1.0996376811594204, "grad_norm": 0.40092512774343536, "learning_rate": 7.654593243729782e-05, "loss": 0.8948, "step": 2428 }, { "epoch": 1.100090579710145, "grad_norm": 0.3622193087005715, "learning_rate": 7.654078940748096e-05, "loss": 0.9003, "step": 2429 }, { "epoch": 1.1005434782608696, "grad_norm": 0.43740083152283427, "learning_rate": 7.653564272462348e-05, "loss": 0.8762, "step": 2430 }, { "epoch": 1.1009963768115942, "grad_norm": 0.4983914571775104, "learning_rate": 7.65304923892399e-05, "loss": 0.8993, "step": 2431 }, { "epoch": 1.1014492753623188, "grad_norm": 0.5271686999848458, "learning_rate": 7.65253384018451e-05, "loss": 0.8702, "step": 2432 }, { "epoch": 1.1019021739130435, "grad_norm": 0.3363431654736905, "learning_rate": 7.652018076295435e-05, "loss": 0.9074, "step": 2433 }, { "epoch": 1.102355072463768, "grad_norm": 0.3522934570226333, "learning_rate": 7.651501947308326e-05, "loss": 0.9012, "step": 2434 }, { "epoch": 1.1028079710144927, "grad_norm": 0.42478741250549135, "learning_rate": 7.65098545327478e-05, "loss": 0.8687, "step": 2435 }, { "epoch": 1.1032608695652173, "grad_norm": 0.5228621693578419, "learning_rate": 7.650468594246433e-05, "loss": 0.9003, "step": 2436 }, { "epoch": 1.1037137681159421, "grad_norm": 0.5990758966559646, "learning_rate": 7.649951370274956e-05, "loss": 0.8948, "step": 2437 }, { "epoch": 1.1041666666666667, "grad_norm": 0.7224712855593457, "learning_rate": 7.649433781412058e-05, "loss": 0.8859, "step": 2438 }, { "epoch": 1.1046195652173914, "grad_norm": 0.806574836674065, "learning_rate": 7.64891582770948e-05, "loss": 0.9059, "step": 2439 }, { "epoch": 1.105072463768116, "grad_norm": 0.8268796710183348, "learning_rate": 7.648397509219005e-05, "loss": 0.9018, "step": 2440 }, { "epoch": 1.1055253623188406, "grad_norm": 0.9068581667705562, "learning_rate": 7.64787882599245e-05, "loss": 0.8839, "step": 2441 }, { "epoch": 1.1059782608695652, "grad_norm": 1.0187399116213656, "learning_rate": 7.647359778081669e-05, "loss": 0.932, "step": 2442 }, { "epoch": 1.1064311594202898, "grad_norm": 0.9916800143476062, "learning_rate": 7.64684036553855e-05, "loss": 0.9094, "step": 2443 }, { "epoch": 1.1068840579710144, "grad_norm": 1.0264409884969559, "learning_rate": 7.646320588415021e-05, "loss": 0.9196, "step": 2444 }, { "epoch": 1.1073369565217392, "grad_norm": 0.8420499639338559, "learning_rate": 7.645800446763044e-05, "loss": 0.9126, "step": 2445 }, { "epoch": 1.1077898550724639, "grad_norm": 0.6165666991899653, "learning_rate": 7.64527994063462e-05, "loss": 0.8894, "step": 2446 }, { "epoch": 1.1082427536231885, "grad_norm": 0.3713271046297714, "learning_rate": 7.644759070081783e-05, "loss": 0.8953, "step": 2447 }, { "epoch": 1.108695652173913, "grad_norm": 0.3480407454769848, "learning_rate": 7.644237835156608e-05, "loss": 0.8966, "step": 2448 }, { "epoch": 1.1091485507246377, "grad_norm": 0.44835764092143443, "learning_rate": 7.6437162359112e-05, "loss": 0.9149, "step": 2449 }, { "epoch": 1.1096014492753623, "grad_norm": 0.5315739055497892, "learning_rate": 7.643194272397708e-05, "loss": 0.8765, "step": 2450 }, { "epoch": 1.110054347826087, "grad_norm": 0.6055693016015007, "learning_rate": 7.642671944668309e-05, "loss": 0.8961, "step": 2451 }, { "epoch": 1.1105072463768115, "grad_norm": 0.6069850114499381, "learning_rate": 7.642149252775224e-05, "loss": 0.8885, "step": 2452 }, { "epoch": 1.1109601449275361, "grad_norm": 0.5470571127165366, "learning_rate": 7.641626196770707e-05, "loss": 0.8868, "step": 2453 }, { "epoch": 1.1114130434782608, "grad_norm": 0.5862229649639707, "learning_rate": 7.64110277670705e-05, "loss": 0.8763, "step": 2454 }, { "epoch": 1.1118659420289856, "grad_norm": 0.701297328044464, "learning_rate": 7.640578992636576e-05, "loss": 0.8724, "step": 2455 }, { "epoch": 1.1123188405797102, "grad_norm": 0.6368454521435064, "learning_rate": 7.640054844611654e-05, "loss": 0.9026, "step": 2456 }, { "epoch": 1.1127717391304348, "grad_norm": 0.4651412805155573, "learning_rate": 7.639530332684678e-05, "loss": 0.88, "step": 2457 }, { "epoch": 1.1132246376811594, "grad_norm": 0.3960343802361894, "learning_rate": 7.63900545690809e-05, "loss": 0.9162, "step": 2458 }, { "epoch": 1.113677536231884, "grad_norm": 0.479259343029708, "learning_rate": 7.638480217334357e-05, "loss": 0.8841, "step": 2459 }, { "epoch": 1.1141304347826086, "grad_norm": 0.4436192731227122, "learning_rate": 7.637954614015993e-05, "loss": 0.8868, "step": 2460 }, { "epoch": 1.1145833333333333, "grad_norm": 0.44808785278528196, "learning_rate": 7.637428647005541e-05, "loss": 0.9007, "step": 2461 }, { "epoch": 1.1150362318840579, "grad_norm": 0.5455265940648613, "learning_rate": 7.636902316355583e-05, "loss": 0.9029, "step": 2462 }, { "epoch": 1.1154891304347827, "grad_norm": 0.6315364647037296, "learning_rate": 7.636375622118738e-05, "loss": 0.8905, "step": 2463 }, { "epoch": 1.1159420289855073, "grad_norm": 0.5517727705458376, "learning_rate": 7.63584856434766e-05, "loss": 0.8866, "step": 2464 }, { "epoch": 1.116394927536232, "grad_norm": 0.5103286514082994, "learning_rate": 7.635321143095038e-05, "loss": 0.8924, "step": 2465 }, { "epoch": 1.1168478260869565, "grad_norm": 0.5861804398200701, "learning_rate": 7.634793358413602e-05, "loss": 0.9111, "step": 2466 }, { "epoch": 1.1173007246376812, "grad_norm": 0.7541229011496231, "learning_rate": 7.634265210356113e-05, "loss": 0.8821, "step": 2467 }, { "epoch": 1.1177536231884058, "grad_norm": 0.877824973013904, "learning_rate": 7.633736698975373e-05, "loss": 0.8861, "step": 2468 }, { "epoch": 1.1182065217391304, "grad_norm": 1.0130086374651162, "learning_rate": 7.633207824324215e-05, "loss": 0.8971, "step": 2469 }, { "epoch": 1.118659420289855, "grad_norm": 1.0132566265535254, "learning_rate": 7.632678586455517e-05, "loss": 0.8615, "step": 2470 }, { "epoch": 1.1191123188405796, "grad_norm": 0.8390353450664416, "learning_rate": 7.63214898542218e-05, "loss": 0.8803, "step": 2471 }, { "epoch": 1.1195652173913044, "grad_norm": 0.7206403224185345, "learning_rate": 7.631619021277156e-05, "loss": 0.903, "step": 2472 }, { "epoch": 1.120018115942029, "grad_norm": 0.7134023662486058, "learning_rate": 7.631088694073423e-05, "loss": 0.8928, "step": 2473 }, { "epoch": 1.1204710144927537, "grad_norm": 0.6635625149658081, "learning_rate": 7.630558003864e-05, "loss": 0.866, "step": 2474 }, { "epoch": 1.1209239130434783, "grad_norm": 0.5248389357530167, "learning_rate": 7.63002695070194e-05, "loss": 0.904, "step": 2475 }, { "epoch": 1.1213768115942029, "grad_norm": 0.3512972059884818, "learning_rate": 7.629495534640333e-05, "loss": 0.904, "step": 2476 }, { "epoch": 1.1218297101449275, "grad_norm": 0.3723300058651289, "learning_rate": 7.628963755732306e-05, "loss": 0.8921, "step": 2477 }, { "epoch": 1.122282608695652, "grad_norm": 0.4140672775028071, "learning_rate": 7.628431614031022e-05, "loss": 0.894, "step": 2478 }, { "epoch": 1.1227355072463767, "grad_norm": 0.5383378476198766, "learning_rate": 7.627899109589679e-05, "loss": 0.9136, "step": 2479 }, { "epoch": 1.1231884057971016, "grad_norm": 0.5460695388206266, "learning_rate": 7.627366242461515e-05, "loss": 0.8863, "step": 2480 }, { "epoch": 1.1236413043478262, "grad_norm": 0.5601570368155379, "learning_rate": 7.626833012699799e-05, "loss": 0.8927, "step": 2481 }, { "epoch": 1.1240942028985508, "grad_norm": 0.7398690798677015, "learning_rate": 7.626299420357839e-05, "loss": 0.9065, "step": 2482 }, { "epoch": 1.1245471014492754, "grad_norm": 0.9687697771271125, "learning_rate": 7.625765465488979e-05, "loss": 0.8988, "step": 2483 }, { "epoch": 1.125, "grad_norm": 1.1271490238854074, "learning_rate": 7.625231148146601e-05, "loss": 0.8894, "step": 2484 }, { "epoch": 1.1254528985507246, "grad_norm": 0.7449205772056653, "learning_rate": 7.624696468384119e-05, "loss": 0.8681, "step": 2485 }, { "epoch": 1.1259057971014492, "grad_norm": 0.42699358913301705, "learning_rate": 7.624161426254989e-05, "loss": 0.9092, "step": 2486 }, { "epoch": 1.1263586956521738, "grad_norm": 0.3347114438925639, "learning_rate": 7.623626021812697e-05, "loss": 0.8919, "step": 2487 }, { "epoch": 1.1268115942028984, "grad_norm": 0.5166594721222098, "learning_rate": 7.62309025511077e-05, "loss": 0.9073, "step": 2488 }, { "epoch": 1.1272644927536233, "grad_norm": 0.6672242559978696, "learning_rate": 7.62255412620277e-05, "loss": 0.9015, "step": 2489 }, { "epoch": 1.127717391304348, "grad_norm": 0.8029673060098185, "learning_rate": 7.622017635142291e-05, "loss": 0.9081, "step": 2490 }, { "epoch": 1.1281702898550725, "grad_norm": 0.9457946522547075, "learning_rate": 7.621480781982972e-05, "loss": 0.8784, "step": 2491 }, { "epoch": 1.1286231884057971, "grad_norm": 1.0259568013750917, "learning_rate": 7.620943566778478e-05, "loss": 0.8865, "step": 2492 }, { "epoch": 1.1290760869565217, "grad_norm": 0.8795037282714062, "learning_rate": 7.62040598958252e-05, "loss": 0.8967, "step": 2493 }, { "epoch": 1.1295289855072463, "grad_norm": 0.7474390885964508, "learning_rate": 7.619868050448838e-05, "loss": 0.892, "step": 2494 }, { "epoch": 1.129981884057971, "grad_norm": 0.5844799591798387, "learning_rate": 7.619329749431208e-05, "loss": 0.8707, "step": 2495 }, { "epoch": 1.1304347826086956, "grad_norm": 0.3907782089697615, "learning_rate": 7.618791086583451e-05, "loss": 0.8888, "step": 2496 }, { "epoch": 1.1308876811594204, "grad_norm": 0.31157572452074617, "learning_rate": 7.618252061959413e-05, "loss": 0.9147, "step": 2497 }, { "epoch": 1.131340579710145, "grad_norm": 0.4283864669336336, "learning_rate": 7.617712675612983e-05, "loss": 0.8645, "step": 2498 }, { "epoch": 1.1317934782608696, "grad_norm": 0.514064568765625, "learning_rate": 7.617172927598084e-05, "loss": 0.9123, "step": 2499 }, { "epoch": 1.1322463768115942, "grad_norm": 0.536086262363734, "learning_rate": 7.616632817968676e-05, "loss": 0.8946, "step": 2500 }, { "epoch": 1.1326992753623188, "grad_norm": 0.6040217165815842, "learning_rate": 7.616092346778754e-05, "loss": 0.9119, "step": 2501 }, { "epoch": 1.1331521739130435, "grad_norm": 0.7356416374913362, "learning_rate": 7.61555151408235e-05, "loss": 0.8697, "step": 2502 }, { "epoch": 1.133605072463768, "grad_norm": 0.8083644757022987, "learning_rate": 7.61501031993353e-05, "loss": 0.88, "step": 2503 }, { "epoch": 1.1340579710144927, "grad_norm": 0.7515933483769626, "learning_rate": 7.614468764386402e-05, "loss": 0.8848, "step": 2504 }, { "epoch": 1.1345108695652173, "grad_norm": 0.6779816386256707, "learning_rate": 7.613926847495104e-05, "loss": 0.8972, "step": 2505 }, { "epoch": 1.134963768115942, "grad_norm": 0.6298070275748077, "learning_rate": 7.613384569313812e-05, "loss": 0.911, "step": 2506 }, { "epoch": 1.1354166666666667, "grad_norm": 0.6304290742538686, "learning_rate": 7.612841929896737e-05, "loss": 0.8811, "step": 2507 }, { "epoch": 1.1358695652173914, "grad_norm": 0.6159385489408726, "learning_rate": 7.612298929298132e-05, "loss": 0.8829, "step": 2508 }, { "epoch": 1.136322463768116, "grad_norm": 0.5674222323798198, "learning_rate": 7.611755567572276e-05, "loss": 0.9196, "step": 2509 }, { "epoch": 1.1367753623188406, "grad_norm": 0.5423885393595258, "learning_rate": 7.611211844773496e-05, "loss": 0.8694, "step": 2510 }, { "epoch": 1.1372282608695652, "grad_norm": 0.5677010568322601, "learning_rate": 7.610667760956143e-05, "loss": 0.9062, "step": 2511 }, { "epoch": 1.1376811594202898, "grad_norm": 0.5161877124751486, "learning_rate": 7.610123316174612e-05, "loss": 0.8804, "step": 2512 }, { "epoch": 1.1381340579710144, "grad_norm": 0.3499597185579186, "learning_rate": 7.609578510483334e-05, "loss": 0.905, "step": 2513 }, { "epoch": 1.1385869565217392, "grad_norm": 0.3276521058837357, "learning_rate": 7.609033343936771e-05, "loss": 0.8926, "step": 2514 }, { "epoch": 1.1390398550724639, "grad_norm": 0.34751456907092326, "learning_rate": 7.608487816589426e-05, "loss": 0.8934, "step": 2515 }, { "epoch": 1.1394927536231885, "grad_norm": 0.40924787789925493, "learning_rate": 7.607941928495837e-05, "loss": 0.8972, "step": 2516 }, { "epoch": 1.139945652173913, "grad_norm": 0.5096523502451058, "learning_rate": 7.607395679710575e-05, "loss": 0.8805, "step": 2517 }, { "epoch": 1.1403985507246377, "grad_norm": 0.593856270169815, "learning_rate": 7.60684907028825e-05, "loss": 0.8975, "step": 2518 }, { "epoch": 1.1408514492753623, "grad_norm": 0.6412348425362039, "learning_rate": 7.60630210028351e-05, "loss": 0.8749, "step": 2519 }, { "epoch": 1.141304347826087, "grad_norm": 0.6946045532834807, "learning_rate": 7.605754769751033e-05, "loss": 0.8976, "step": 2520 }, { "epoch": 1.1417572463768115, "grad_norm": 0.7332116278328763, "learning_rate": 7.605207078745537e-05, "loss": 0.8826, "step": 2521 }, { "epoch": 1.1422101449275361, "grad_norm": 0.7846046399061328, "learning_rate": 7.604659027321778e-05, "loss": 0.9027, "step": 2522 }, { "epoch": 1.1426630434782608, "grad_norm": 0.8261962954359371, "learning_rate": 7.604110615534542e-05, "loss": 0.8918, "step": 2523 }, { "epoch": 1.1431159420289856, "grad_norm": 0.8331758614341671, "learning_rate": 7.603561843438657e-05, "loss": 0.8813, "step": 2524 }, { "epoch": 1.1435688405797102, "grad_norm": 0.7304727981369862, "learning_rate": 7.603012711088985e-05, "loss": 0.8808, "step": 2525 }, { "epoch": 1.1440217391304348, "grad_norm": 0.6307892357052814, "learning_rate": 7.602463218540423e-05, "loss": 0.9009, "step": 2526 }, { "epoch": 1.1444746376811594, "grad_norm": 0.4951801617061492, "learning_rate": 7.601913365847905e-05, "loss": 0.8828, "step": 2527 }, { "epoch": 1.144927536231884, "grad_norm": 0.4032650042442616, "learning_rate": 7.601363153066399e-05, "loss": 0.8692, "step": 2528 }, { "epoch": 1.1453804347826086, "grad_norm": 0.3803438155967065, "learning_rate": 7.600812580250912e-05, "loss": 0.9065, "step": 2529 }, { "epoch": 1.1458333333333333, "grad_norm": 0.3762632677365636, "learning_rate": 7.600261647456485e-05, "loss": 0.8833, "step": 2530 }, { "epoch": 1.146286231884058, "grad_norm": 0.2912141256063736, "learning_rate": 7.599710354738197e-05, "loss": 0.9006, "step": 2531 }, { "epoch": 1.1467391304347827, "grad_norm": 0.3795081508725028, "learning_rate": 7.59915870215116e-05, "loss": 0.8985, "step": 2532 }, { "epoch": 1.1471920289855073, "grad_norm": 0.5216748996735571, "learning_rate": 7.598606689750524e-05, "loss": 0.8803, "step": 2533 }, { "epoch": 1.147644927536232, "grad_norm": 0.6613528646277568, "learning_rate": 7.598054317591474e-05, "loss": 0.8964, "step": 2534 }, { "epoch": 1.1480978260869565, "grad_norm": 0.7738691430265123, "learning_rate": 7.597501585729234e-05, "loss": 0.9043, "step": 2535 }, { "epoch": 1.1485507246376812, "grad_norm": 0.8150179202826427, "learning_rate": 7.596948494219059e-05, "loss": 0.8902, "step": 2536 }, { "epoch": 1.1490036231884058, "grad_norm": 0.8185332383830002, "learning_rate": 7.596395043116243e-05, "loss": 0.9105, "step": 2537 }, { "epoch": 1.1494565217391304, "grad_norm": 0.794767828030164, "learning_rate": 7.595841232476116e-05, "loss": 0.8952, "step": 2538 }, { "epoch": 1.149909420289855, "grad_norm": 0.7646188050367311, "learning_rate": 7.595287062354042e-05, "loss": 0.9124, "step": 2539 }, { "epoch": 1.1503623188405796, "grad_norm": 0.7134259316343581, "learning_rate": 7.594732532805425e-05, "loss": 0.8651, "step": 2540 }, { "epoch": 1.1508152173913044, "grad_norm": 0.5886552036348137, "learning_rate": 7.594177643885697e-05, "loss": 0.8815, "step": 2541 }, { "epoch": 1.151268115942029, "grad_norm": 0.4080658371683358, "learning_rate": 7.593622395650338e-05, "loss": 0.8967, "step": 2542 }, { "epoch": 1.1517210144927537, "grad_norm": 0.28785229197260476, "learning_rate": 7.593066788154851e-05, "loss": 0.8898, "step": 2543 }, { "epoch": 1.1521739130434783, "grad_norm": 0.39111898618242547, "learning_rate": 7.592510821454785e-05, "loss": 0.8853, "step": 2544 }, { "epoch": 1.1526268115942029, "grad_norm": 0.5507484303174613, "learning_rate": 7.59195449560572e-05, "loss": 0.8925, "step": 2545 }, { "epoch": 1.1530797101449275, "grad_norm": 0.6432519180674104, "learning_rate": 7.59139781066327e-05, "loss": 0.8791, "step": 2546 }, { "epoch": 1.153532608695652, "grad_norm": 0.6929365286877063, "learning_rate": 7.590840766683091e-05, "loss": 0.9048, "step": 2547 }, { "epoch": 1.1539855072463767, "grad_norm": 0.7351042327421395, "learning_rate": 7.590283363720871e-05, "loss": 0.8946, "step": 2548 }, { "epoch": 1.1544384057971016, "grad_norm": 0.6905286720260807, "learning_rate": 7.589725601832332e-05, "loss": 0.9045, "step": 2549 }, { "epoch": 1.1548913043478262, "grad_norm": 0.6141892039817093, "learning_rate": 7.589167481073238e-05, "loss": 0.915, "step": 2550 }, { "epoch": 1.1553442028985508, "grad_norm": 0.5753542886067238, "learning_rate": 7.588609001499383e-05, "loss": 0.8899, "step": 2551 }, { "epoch": 1.1557971014492754, "grad_norm": 0.5412829492915106, "learning_rate": 7.588050163166599e-05, "loss": 0.8759, "step": 2552 }, { "epoch": 1.15625, "grad_norm": 0.4990921687588306, "learning_rate": 7.587490966130754e-05, "loss": 0.9116, "step": 2553 }, { "epoch": 1.1567028985507246, "grad_norm": 0.4491741672588447, "learning_rate": 7.586931410447752e-05, "loss": 0.8828, "step": 2554 }, { "epoch": 1.1571557971014492, "grad_norm": 0.4033629968757986, "learning_rate": 7.586371496173535e-05, "loss": 0.8865, "step": 2555 }, { "epoch": 1.1576086956521738, "grad_norm": 0.3014799662515984, "learning_rate": 7.585811223364075e-05, "loss": 0.8894, "step": 2556 }, { "epoch": 1.1580615942028984, "grad_norm": 0.3761633355334381, "learning_rate": 7.585250592075385e-05, "loss": 0.8783, "step": 2557 }, { "epoch": 1.1585144927536233, "grad_norm": 0.5231581449730617, "learning_rate": 7.584689602363512e-05, "loss": 0.8906, "step": 2558 }, { "epoch": 1.158967391304348, "grad_norm": 0.5539644883803063, "learning_rate": 7.584128254284538e-05, "loss": 0.8825, "step": 2559 }, { "epoch": 1.1594202898550725, "grad_norm": 0.6619960095776685, "learning_rate": 7.583566547894584e-05, "loss": 0.9013, "step": 2560 }, { "epoch": 1.1598731884057971, "grad_norm": 0.9644467260581376, "learning_rate": 7.583004483249804e-05, "loss": 0.8787, "step": 2561 }, { "epoch": 1.1603260869565217, "grad_norm": 1.1354165280751018, "learning_rate": 7.582442060406386e-05, "loss": 0.9137, "step": 2562 }, { "epoch": 1.1607789855072463, "grad_norm": 0.7286525756003915, "learning_rate": 7.581879279420561e-05, "loss": 0.8855, "step": 2563 }, { "epoch": 1.161231884057971, "grad_norm": 0.4491674605578995, "learning_rate": 7.581316140348587e-05, "loss": 0.8892, "step": 2564 }, { "epoch": 1.1616847826086956, "grad_norm": 0.3525651534351678, "learning_rate": 7.580752643246763e-05, "loss": 0.9012, "step": 2565 }, { "epoch": 1.1621376811594204, "grad_norm": 0.42573346611593854, "learning_rate": 7.580188788171425e-05, "loss": 0.9049, "step": 2566 }, { "epoch": 1.162590579710145, "grad_norm": 0.544938625514511, "learning_rate": 7.579624575178939e-05, "loss": 0.9007, "step": 2567 }, { "epoch": 1.1630434782608696, "grad_norm": 0.7996431006958452, "learning_rate": 7.579060004325712e-05, "loss": 0.9242, "step": 2568 }, { "epoch": 1.1634963768115942, "grad_norm": 0.9939643582733471, "learning_rate": 7.578495075668186e-05, "loss": 0.9054, "step": 2569 }, { "epoch": 1.1639492753623188, "grad_norm": 1.1191041244156241, "learning_rate": 7.577929789262835e-05, "loss": 0.885, "step": 2570 }, { "epoch": 1.1644021739130435, "grad_norm": 0.7017552156704125, "learning_rate": 7.577364145166175e-05, "loss": 0.9041, "step": 2571 }, { "epoch": 1.164855072463768, "grad_norm": 0.41704527225332483, "learning_rate": 7.57679814343475e-05, "loss": 0.8833, "step": 2572 }, { "epoch": 1.1653079710144927, "grad_norm": 0.5558378323119352, "learning_rate": 7.576231784125147e-05, "loss": 0.897, "step": 2573 }, { "epoch": 1.1657608695652173, "grad_norm": 0.6109373655934287, "learning_rate": 7.575665067293988e-05, "loss": 0.9014, "step": 2574 }, { "epoch": 1.166213768115942, "grad_norm": 0.7964956290536187, "learning_rate": 7.575097992997924e-05, "loss": 0.9066, "step": 2575 }, { "epoch": 1.1666666666666667, "grad_norm": 0.9667752832695374, "learning_rate": 7.57453056129365e-05, "loss": 0.9129, "step": 2576 }, { "epoch": 1.1671195652173914, "grad_norm": 1.0743212017422483, "learning_rate": 7.57396277223789e-05, "loss": 0.8772, "step": 2577 }, { "epoch": 1.167572463768116, "grad_norm": 0.869325226491849, "learning_rate": 7.573394625887409e-05, "loss": 0.8944, "step": 2578 }, { "epoch": 1.1680253623188406, "grad_norm": 0.6232094183486608, "learning_rate": 7.572826122299005e-05, "loss": 0.8992, "step": 2579 }, { "epoch": 1.1684782608695652, "grad_norm": 0.44798119995419156, "learning_rate": 7.572257261529511e-05, "loss": 0.884, "step": 2580 }, { "epoch": 1.1689311594202898, "grad_norm": 0.5009056704357616, "learning_rate": 7.571688043635799e-05, "loss": 0.9051, "step": 2581 }, { "epoch": 1.1693840579710144, "grad_norm": 0.7496399613972604, "learning_rate": 7.571118468674773e-05, "loss": 0.8962, "step": 2582 }, { "epoch": 1.1698369565217392, "grad_norm": 0.8654321780176453, "learning_rate": 7.570548536703375e-05, "loss": 0.9025, "step": 2583 }, { "epoch": 1.1702898550724639, "grad_norm": 0.8270944349161389, "learning_rate": 7.569978247778581e-05, "loss": 0.9003, "step": 2584 }, { "epoch": 1.1707427536231885, "grad_norm": 0.7368481464756134, "learning_rate": 7.569407601957404e-05, "loss": 0.8994, "step": 2585 }, { "epoch": 1.171195652173913, "grad_norm": 0.6939068174682275, "learning_rate": 7.568836599296893e-05, "loss": 0.881, "step": 2586 }, { "epoch": 1.1716485507246377, "grad_norm": 0.6146683452383318, "learning_rate": 7.568265239854133e-05, "loss": 0.9156, "step": 2587 }, { "epoch": 1.1721014492753623, "grad_norm": 0.43791646446546934, "learning_rate": 7.567693523686241e-05, "loss": 0.8947, "step": 2588 }, { "epoch": 1.172554347826087, "grad_norm": 0.35816654157827044, "learning_rate": 7.567121450850376e-05, "loss": 0.9012, "step": 2589 }, { "epoch": 1.1730072463768115, "grad_norm": 0.37895325962370285, "learning_rate": 7.566549021403725e-05, "loss": 0.8999, "step": 2590 }, { "epoch": 1.1734601449275361, "grad_norm": 0.4480481635091382, "learning_rate": 7.565976235403518e-05, "loss": 0.8864, "step": 2591 }, { "epoch": 1.1739130434782608, "grad_norm": 0.5087855432632371, "learning_rate": 7.565403092907015e-05, "loss": 0.8957, "step": 2592 }, { "epoch": 1.1743659420289856, "grad_norm": 0.6073887079027738, "learning_rate": 7.564829593971516e-05, "loss": 0.8833, "step": 2593 }, { "epoch": 1.1748188405797102, "grad_norm": 0.5669971580486368, "learning_rate": 7.564255738654354e-05, "loss": 0.9057, "step": 2594 }, { "epoch": 1.1752717391304348, "grad_norm": 0.4167176926248499, "learning_rate": 7.563681527012897e-05, "loss": 0.9126, "step": 2595 }, { "epoch": 1.1757246376811594, "grad_norm": 0.5059231724173837, "learning_rate": 7.56310695910455e-05, "loss": 0.8899, "step": 2596 }, { "epoch": 1.176177536231884, "grad_norm": 0.6607522356938671, "learning_rate": 7.562532034986755e-05, "loss": 0.9166, "step": 2597 }, { "epoch": 1.1766304347826086, "grad_norm": 0.5777714950173894, "learning_rate": 7.561956754716986e-05, "loss": 0.8935, "step": 2598 }, { "epoch": 1.1770833333333333, "grad_norm": 0.3649917727478265, "learning_rate": 7.561381118352757e-05, "loss": 0.8791, "step": 2599 }, { "epoch": 1.177536231884058, "grad_norm": 0.4079466070080207, "learning_rate": 7.560805125951612e-05, "loss": 0.9032, "step": 2600 }, { "epoch": 1.1779891304347827, "grad_norm": 0.4809094978576903, "learning_rate": 7.560228777571138e-05, "loss": 0.888, "step": 2601 }, { "epoch": 1.1784420289855073, "grad_norm": 0.4952244957973529, "learning_rate": 7.559652073268951e-05, "loss": 0.9017, "step": 2602 }, { "epoch": 1.178894927536232, "grad_norm": 0.481591373983726, "learning_rate": 7.559075013102705e-05, "loss": 0.8901, "step": 2603 }, { "epoch": 1.1793478260869565, "grad_norm": 0.5488176657677853, "learning_rate": 7.558497597130091e-05, "loss": 0.8792, "step": 2604 }, { "epoch": 1.1798007246376812, "grad_norm": 0.6421986052067599, "learning_rate": 7.557919825408833e-05, "loss": 0.8904, "step": 2605 }, { "epoch": 1.1802536231884058, "grad_norm": 0.8872182897349734, "learning_rate": 7.557341697996692e-05, "loss": 0.8703, "step": 2606 }, { "epoch": 1.1807065217391304, "grad_norm": 1.1375025760255746, "learning_rate": 7.556763214951464e-05, "loss": 0.9029, "step": 2607 }, { "epoch": 1.181159420289855, "grad_norm": 0.8819691909449644, "learning_rate": 7.556184376330981e-05, "loss": 0.9233, "step": 2608 }, { "epoch": 1.1816123188405796, "grad_norm": 0.6509793603957518, "learning_rate": 7.555605182193111e-05, "loss": 0.8778, "step": 2609 }, { "epoch": 1.1820652173913044, "grad_norm": 0.44931568710779957, "learning_rate": 7.555025632595758e-05, "loss": 0.9049, "step": 2610 }, { "epoch": 1.182518115942029, "grad_norm": 0.46643343582298347, "learning_rate": 7.554445727596858e-05, "loss": 0.899, "step": 2611 }, { "epoch": 1.1829710144927537, "grad_norm": 0.4144449235703896, "learning_rate": 7.553865467254386e-05, "loss": 0.8691, "step": 2612 }, { "epoch": 1.1834239130434783, "grad_norm": 0.34981050084965903, "learning_rate": 7.55328485162635e-05, "loss": 0.8812, "step": 2613 }, { "epoch": 1.1838768115942029, "grad_norm": 0.4795249953643573, "learning_rate": 7.552703880770798e-05, "loss": 0.9009, "step": 2614 }, { "epoch": 1.1843297101449275, "grad_norm": 0.5758610364035313, "learning_rate": 7.552122554745811e-05, "loss": 0.9006, "step": 2615 }, { "epoch": 1.184782608695652, "grad_norm": 0.6797149516640045, "learning_rate": 7.551540873609502e-05, "loss": 0.8703, "step": 2616 }, { "epoch": 1.1852355072463767, "grad_norm": 0.7608968521369416, "learning_rate": 7.550958837420023e-05, "loss": 0.8872, "step": 2617 }, { "epoch": 1.1856884057971016, "grad_norm": 0.7414354837106685, "learning_rate": 7.550376446235563e-05, "loss": 0.8951, "step": 2618 }, { "epoch": 1.1861413043478262, "grad_norm": 0.6861630955221227, "learning_rate": 7.549793700114344e-05, "loss": 0.8967, "step": 2619 }, { "epoch": 1.1865942028985508, "grad_norm": 0.6593075812703139, "learning_rate": 7.549210599114624e-05, "loss": 0.9014, "step": 2620 }, { "epoch": 1.1870471014492754, "grad_norm": 0.5454799914741062, "learning_rate": 7.548627143294695e-05, "loss": 0.8893, "step": 2621 }, { "epoch": 1.1875, "grad_norm": 0.4419273951238063, "learning_rate": 7.548043332712887e-05, "loss": 0.9092, "step": 2622 }, { "epoch": 1.1879528985507246, "grad_norm": 0.3874318460672211, "learning_rate": 7.547459167427566e-05, "loss": 0.911, "step": 2623 }, { "epoch": 1.1884057971014492, "grad_norm": 0.4756135956511833, "learning_rate": 7.546874647497129e-05, "loss": 0.9189, "step": 2624 }, { "epoch": 1.1888586956521738, "grad_norm": 0.32118041999679525, "learning_rate": 7.546289772980014e-05, "loss": 0.8943, "step": 2625 }, { "epoch": 1.1893115942028984, "grad_norm": 0.3095012147448859, "learning_rate": 7.545704543934691e-05, "loss": 0.8802, "step": 2626 }, { "epoch": 1.1897644927536233, "grad_norm": 0.3689027152037398, "learning_rate": 7.545118960419666e-05, "loss": 0.8789, "step": 2627 }, { "epoch": 1.190217391304348, "grad_norm": 0.45965845997059396, "learning_rate": 7.54453302249348e-05, "loss": 0.8911, "step": 2628 }, { "epoch": 1.1906702898550725, "grad_norm": 0.5848408917493472, "learning_rate": 7.543946730214713e-05, "loss": 0.885, "step": 2629 }, { "epoch": 1.1911231884057971, "grad_norm": 0.6211930455963546, "learning_rate": 7.543360083641974e-05, "loss": 0.89, "step": 2630 }, { "epoch": 1.1915760869565217, "grad_norm": 0.6386157953856603, "learning_rate": 7.542773082833912e-05, "loss": 0.8946, "step": 2631 }, { "epoch": 1.1920289855072463, "grad_norm": 0.8158136145636982, "learning_rate": 7.542185727849211e-05, "loss": 0.8879, "step": 2632 }, { "epoch": 1.192481884057971, "grad_norm": 1.0191702869570016, "learning_rate": 7.541598018746591e-05, "loss": 0.8764, "step": 2633 }, { "epoch": 1.1929347826086956, "grad_norm": 1.082295985872742, "learning_rate": 7.541009955584803e-05, "loss": 0.8765, "step": 2634 }, { "epoch": 1.1933876811594204, "grad_norm": 0.8615809521550089, "learning_rate": 7.54042153842264e-05, "loss": 0.8935, "step": 2635 }, { "epoch": 1.193840579710145, "grad_norm": 0.6574729689849301, "learning_rate": 7.539832767318925e-05, "loss": 0.8705, "step": 2636 }, { "epoch": 1.1942934782608696, "grad_norm": 0.5360237839790978, "learning_rate": 7.53924364233252e-05, "loss": 0.8935, "step": 2637 }, { "epoch": 1.1947463768115942, "grad_norm": 0.44608179429349926, "learning_rate": 7.538654163522319e-05, "loss": 0.8899, "step": 2638 }, { "epoch": 1.1951992753623188, "grad_norm": 0.3963519607150411, "learning_rate": 7.538064330947253e-05, "loss": 0.8885, "step": 2639 }, { "epoch": 1.1956521739130435, "grad_norm": 0.48441348106542115, "learning_rate": 7.53747414466629e-05, "loss": 0.8893, "step": 2640 }, { "epoch": 1.196105072463768, "grad_norm": 0.6371965904624619, "learning_rate": 7.53688360473843e-05, "loss": 0.8961, "step": 2641 }, { "epoch": 1.1965579710144927, "grad_norm": 0.8411880048141044, "learning_rate": 7.536292711222713e-05, "loss": 0.9031, "step": 2642 }, { "epoch": 1.1970108695652173, "grad_norm": 0.8943310012229242, "learning_rate": 7.53570146417821e-05, "loss": 0.8841, "step": 2643 }, { "epoch": 1.197463768115942, "grad_norm": 0.8334939299542817, "learning_rate": 7.535109863664028e-05, "loss": 0.9213, "step": 2644 }, { "epoch": 1.1979166666666667, "grad_norm": 0.7735434424025392, "learning_rate": 7.534517909739312e-05, "loss": 0.8929, "step": 2645 }, { "epoch": 1.1983695652173914, "grad_norm": 0.7479326657856369, "learning_rate": 7.533925602463237e-05, "loss": 0.8991, "step": 2646 }, { "epoch": 1.198822463768116, "grad_norm": 0.6732522989618391, "learning_rate": 7.533332941895023e-05, "loss": 0.8829, "step": 2647 }, { "epoch": 1.1992753623188406, "grad_norm": 0.5467060164290147, "learning_rate": 7.532739928093913e-05, "loss": 0.9043, "step": 2648 }, { "epoch": 1.1997282608695652, "grad_norm": 0.5272757013502672, "learning_rate": 7.532146561119196e-05, "loss": 0.8902, "step": 2649 }, { "epoch": 1.2001811594202898, "grad_norm": 0.4595450639214968, "learning_rate": 7.53155284103019e-05, "loss": 0.9075, "step": 2650 }, { "epoch": 1.2006340579710144, "grad_norm": 0.3761348273316996, "learning_rate": 7.530958767886249e-05, "loss": 0.8893, "step": 2651 }, { "epoch": 1.2010869565217392, "grad_norm": 0.3696009368444218, "learning_rate": 7.530364341746765e-05, "loss": 0.8843, "step": 2652 }, { "epoch": 1.2015398550724639, "grad_norm": 0.3027625906692911, "learning_rate": 7.529769562671164e-05, "loss": 0.8993, "step": 2653 }, { "epoch": 1.2019927536231885, "grad_norm": 0.270532320041284, "learning_rate": 7.529174430718906e-05, "loss": 0.8932, "step": 2654 }, { "epoch": 1.202445652173913, "grad_norm": 0.3620051926742045, "learning_rate": 7.528578945949488e-05, "loss": 0.8999, "step": 2655 }, { "epoch": 1.2028985507246377, "grad_norm": 0.34271672998969593, "learning_rate": 7.527983108422442e-05, "loss": 0.881, "step": 2656 }, { "epoch": 1.2033514492753623, "grad_norm": 0.29891217252360824, "learning_rate": 7.527386918197333e-05, "loss": 0.8832, "step": 2657 }, { "epoch": 1.203804347826087, "grad_norm": 0.3953478032978135, "learning_rate": 7.526790375333765e-05, "loss": 0.8768, "step": 2658 }, { "epoch": 1.2042572463768115, "grad_norm": 0.491896469758624, "learning_rate": 7.526193479891374e-05, "loss": 0.8889, "step": 2659 }, { "epoch": 1.2047101449275361, "grad_norm": 0.4768395856993518, "learning_rate": 7.525596231929833e-05, "loss": 0.9105, "step": 2660 }, { "epoch": 1.2051630434782608, "grad_norm": 0.4016577406529652, "learning_rate": 7.524998631508849e-05, "loss": 0.8919, "step": 2661 }, { "epoch": 1.2056159420289856, "grad_norm": 0.47033624267812973, "learning_rate": 7.524400678688167e-05, "loss": 0.9132, "step": 2662 }, { "epoch": 1.2060688405797102, "grad_norm": 0.3444655507780638, "learning_rate": 7.523802373527565e-05, "loss": 0.9085, "step": 2663 }, { "epoch": 1.2065217391304348, "grad_norm": 0.37926132044140437, "learning_rate": 7.523203716086855e-05, "loss": 0.902, "step": 2664 }, { "epoch": 1.2069746376811594, "grad_norm": 0.5085683095990452, "learning_rate": 7.522604706425886e-05, "loss": 0.8788, "step": 2665 }, { "epoch": 1.207427536231884, "grad_norm": 0.5740719441248844, "learning_rate": 7.522005344604543e-05, "loss": 0.9131, "step": 2666 }, { "epoch": 1.2078804347826086, "grad_norm": 0.6381182501829882, "learning_rate": 7.521405630682745e-05, "loss": 0.8933, "step": 2667 }, { "epoch": 1.2083333333333333, "grad_norm": 0.6753527636537167, "learning_rate": 7.520805564720444e-05, "loss": 0.8777, "step": 2668 }, { "epoch": 1.208786231884058, "grad_norm": 0.7635691104632795, "learning_rate": 7.520205146777635e-05, "loss": 0.9111, "step": 2669 }, { "epoch": 1.2092391304347827, "grad_norm": 0.899551003313181, "learning_rate": 7.519604376914336e-05, "loss": 0.8859, "step": 2670 }, { "epoch": 1.2096920289855073, "grad_norm": 1.1168488080310939, "learning_rate": 7.519003255190612e-05, "loss": 0.9101, "step": 2671 }, { "epoch": 1.210144927536232, "grad_norm": 0.8351751709780072, "learning_rate": 7.518401781666555e-05, "loss": 0.8963, "step": 2672 }, { "epoch": 1.2105978260869565, "grad_norm": 0.636299439383499, "learning_rate": 7.517799956402297e-05, "loss": 0.8762, "step": 2673 }, { "epoch": 1.2110507246376812, "grad_norm": 0.5061573777586261, "learning_rate": 7.517197779458003e-05, "loss": 0.8899, "step": 2674 }, { "epoch": 1.2115036231884058, "grad_norm": 0.37845525806501984, "learning_rate": 7.516595250893874e-05, "loss": 0.8866, "step": 2675 }, { "epoch": 1.2119565217391304, "grad_norm": 0.3385861067709102, "learning_rate": 7.515992370770145e-05, "loss": 0.8901, "step": 2676 }, { "epoch": 1.212409420289855, "grad_norm": 0.36172507700161494, "learning_rate": 7.515389139147087e-05, "loss": 0.9004, "step": 2677 }, { "epoch": 1.2128623188405796, "grad_norm": 0.4337283118248273, "learning_rate": 7.514785556085006e-05, "loss": 0.8734, "step": 2678 }, { "epoch": 1.2133152173913044, "grad_norm": 0.45410352751606864, "learning_rate": 7.514181621644243e-05, "loss": 0.9039, "step": 2679 }, { "epoch": 1.213768115942029, "grad_norm": 0.473990367603988, "learning_rate": 7.513577335885175e-05, "loss": 0.8768, "step": 2680 }, { "epoch": 1.2142210144927537, "grad_norm": 0.5588328185911741, "learning_rate": 7.512972698868214e-05, "loss": 0.887, "step": 2681 }, { "epoch": 1.2146739130434783, "grad_norm": 0.6937645750957683, "learning_rate": 7.512367710653804e-05, "loss": 0.9006, "step": 2682 }, { "epoch": 1.2151268115942029, "grad_norm": 0.8234033642737338, "learning_rate": 7.511762371302428e-05, "loss": 0.906, "step": 2683 }, { "epoch": 1.2155797101449275, "grad_norm": 1.0636627511971115, "learning_rate": 7.511156680874604e-05, "loss": 0.9112, "step": 2684 }, { "epoch": 1.216032608695652, "grad_norm": 1.1216518164708607, "learning_rate": 7.510550639430883e-05, "loss": 0.8961, "step": 2685 }, { "epoch": 1.2164855072463767, "grad_norm": 0.8335443312043864, "learning_rate": 7.50994424703185e-05, "loss": 0.8803, "step": 2686 }, { "epoch": 1.2169384057971016, "grad_norm": 0.5536784849003535, "learning_rate": 7.50933750373813e-05, "loss": 0.8773, "step": 2687 }, { "epoch": 1.2173913043478262, "grad_norm": 0.4116388318807075, "learning_rate": 7.508730409610377e-05, "loss": 0.9036, "step": 2688 }, { "epoch": 1.2178442028985508, "grad_norm": 0.47447899866334403, "learning_rate": 7.508122964709287e-05, "loss": 0.8596, "step": 2689 }, { "epoch": 1.2182971014492754, "grad_norm": 0.530932523378576, "learning_rate": 7.507515169095584e-05, "loss": 0.8636, "step": 2690 }, { "epoch": 1.21875, "grad_norm": 0.6395876219722701, "learning_rate": 7.506907022830032e-05, "loss": 0.8821, "step": 2691 }, { "epoch": 1.2192028985507246, "grad_norm": 0.710922052622373, "learning_rate": 7.506298525973427e-05, "loss": 0.8977, "step": 2692 }, { "epoch": 1.2196557971014492, "grad_norm": 0.7929558920814882, "learning_rate": 7.505689678586603e-05, "loss": 0.8826, "step": 2693 }, { "epoch": 1.2201086956521738, "grad_norm": 0.854609251637631, "learning_rate": 7.505080480730426e-05, "loss": 0.8916, "step": 2694 }, { "epoch": 1.2205615942028984, "grad_norm": 0.801746676296219, "learning_rate": 7.504470932465799e-05, "loss": 0.8902, "step": 2695 }, { "epoch": 1.2210144927536233, "grad_norm": 0.6087577650061925, "learning_rate": 7.50386103385366e-05, "loss": 0.8892, "step": 2696 }, { "epoch": 1.221467391304348, "grad_norm": 0.49895219021512804, "learning_rate": 7.50325078495498e-05, "loss": 0.8953, "step": 2697 }, { "epoch": 1.2219202898550725, "grad_norm": 0.4629739512463667, "learning_rate": 7.50264018583077e-05, "loss": 0.9045, "step": 2698 }, { "epoch": 1.2223731884057971, "grad_norm": 0.473466980207827, "learning_rate": 7.502029236542067e-05, "loss": 0.8643, "step": 2699 }, { "epoch": 1.2228260869565217, "grad_norm": 0.5179428891401592, "learning_rate": 7.501417937149954e-05, "loss": 0.9048, "step": 2700 }, { "epoch": 1.2232789855072463, "grad_norm": 0.5518366204209547, "learning_rate": 7.500806287715541e-05, "loss": 0.9222, "step": 2701 }, { "epoch": 1.223731884057971, "grad_norm": 0.45892591722962245, "learning_rate": 7.500194288299975e-05, "loss": 0.8837, "step": 2702 }, { "epoch": 1.2241847826086956, "grad_norm": 0.4483287208391819, "learning_rate": 7.499581938964442e-05, "loss": 0.898, "step": 2703 }, { "epoch": 1.2246376811594204, "grad_norm": 0.4293687654937929, "learning_rate": 7.498969239770155e-05, "loss": 0.8908, "step": 2704 }, { "epoch": 1.225090579710145, "grad_norm": 0.41728449549733765, "learning_rate": 7.498356190778369e-05, "loss": 0.9017, "step": 2705 }, { "epoch": 1.2255434782608696, "grad_norm": 0.3675307608345987, "learning_rate": 7.497742792050369e-05, "loss": 0.8694, "step": 2706 }, { "epoch": 1.2259963768115942, "grad_norm": 0.41671844721965945, "learning_rate": 7.497129043647482e-05, "loss": 0.9026, "step": 2707 }, { "epoch": 1.2264492753623188, "grad_norm": 0.38129384411245676, "learning_rate": 7.496514945631062e-05, "loss": 0.8828, "step": 2708 }, { "epoch": 1.2269021739130435, "grad_norm": 0.4074353199209213, "learning_rate": 7.495900498062501e-05, "loss": 0.8917, "step": 2709 }, { "epoch": 1.227355072463768, "grad_norm": 0.4346956383534201, "learning_rate": 7.49528570100323e-05, "loss": 0.9031, "step": 2710 }, { "epoch": 1.2278079710144927, "grad_norm": 0.4434553123198333, "learning_rate": 7.494670554514707e-05, "loss": 0.8936, "step": 2711 }, { "epoch": 1.2282608695652173, "grad_norm": 0.508496752545578, "learning_rate": 7.494055058658431e-05, "loss": 0.8969, "step": 2712 }, { "epoch": 1.228713768115942, "grad_norm": 0.607552735568504, "learning_rate": 7.493439213495935e-05, "loss": 0.865, "step": 2713 }, { "epoch": 1.2291666666666667, "grad_norm": 0.7328526945977025, "learning_rate": 7.492823019088785e-05, "loss": 0.8828, "step": 2714 }, { "epoch": 1.2296195652173914, "grad_norm": 0.7630676243386687, "learning_rate": 7.492206475498581e-05, "loss": 0.8843, "step": 2715 }, { "epoch": 1.230072463768116, "grad_norm": 0.6976938152092119, "learning_rate": 7.491589582786963e-05, "loss": 0.9052, "step": 2716 }, { "epoch": 1.2305253623188406, "grad_norm": 0.6477044541115795, "learning_rate": 7.490972341015601e-05, "loss": 0.9035, "step": 2717 }, { "epoch": 1.2309782608695652, "grad_norm": 0.5244322574537736, "learning_rate": 7.490354750246203e-05, "loss": 0.884, "step": 2718 }, { "epoch": 1.2314311594202898, "grad_norm": 0.35738193546351005, "learning_rate": 7.489736810540508e-05, "loss": 0.9079, "step": 2719 }, { "epoch": 1.2318840579710144, "grad_norm": 0.42762296514288045, "learning_rate": 7.489118521960296e-05, "loss": 0.8937, "step": 2720 }, { "epoch": 1.2323369565217392, "grad_norm": 0.6647709600155086, "learning_rate": 7.488499884567376e-05, "loss": 0.8976, "step": 2721 }, { "epoch": 1.2327898550724639, "grad_norm": 0.49580632463262075, "learning_rate": 7.487880898423595e-05, "loss": 0.8846, "step": 2722 }, { "epoch": 1.2332427536231885, "grad_norm": 0.4261354435026187, "learning_rate": 7.487261563590831e-05, "loss": 0.913, "step": 2723 }, { "epoch": 1.233695652173913, "grad_norm": 0.43699204311068324, "learning_rate": 7.486641880131006e-05, "loss": 0.8865, "step": 2724 }, { "epoch": 1.2341485507246377, "grad_norm": 0.6766546927958415, "learning_rate": 7.486021848106065e-05, "loss": 0.8696, "step": 2725 }, { "epoch": 1.2346014492753623, "grad_norm": 0.9581126013438866, "learning_rate": 7.485401467577996e-05, "loss": 0.9091, "step": 2726 }, { "epoch": 1.235054347826087, "grad_norm": 0.9975831193653213, "learning_rate": 7.48478073860882e-05, "loss": 0.9017, "step": 2727 }, { "epoch": 1.2355072463768115, "grad_norm": 0.9581256371129862, "learning_rate": 7.48415966126059e-05, "loss": 0.903, "step": 2728 }, { "epoch": 1.2359601449275361, "grad_norm": 0.991839486816409, "learning_rate": 7.4835382355954e-05, "loss": 0.8999, "step": 2729 }, { "epoch": 1.2364130434782608, "grad_norm": 1.0130796524412775, "learning_rate": 7.48291646167537e-05, "loss": 0.8612, "step": 2730 }, { "epoch": 1.2368659420289856, "grad_norm": 0.7643086749561641, "learning_rate": 7.482294339562663e-05, "loss": 0.8814, "step": 2731 }, { "epoch": 1.2373188405797102, "grad_norm": 0.3678559001738137, "learning_rate": 7.481671869319473e-05, "loss": 0.8895, "step": 2732 }, { "epoch": 1.2377717391304348, "grad_norm": 0.48941232619848196, "learning_rate": 7.481049051008028e-05, "loss": 0.906, "step": 2733 }, { "epoch": 1.2382246376811594, "grad_norm": 0.7914216637417568, "learning_rate": 7.480425884690594e-05, "loss": 0.9127, "step": 2734 }, { "epoch": 1.238677536231884, "grad_norm": 0.9878672611479448, "learning_rate": 7.47980237042947e-05, "loss": 0.9004, "step": 2735 }, { "epoch": 1.2391304347826086, "grad_norm": 1.1953445919719081, "learning_rate": 7.479178508286987e-05, "loss": 0.9443, "step": 2736 }, { "epoch": 1.2395833333333333, "grad_norm": 0.5948071298610634, "learning_rate": 7.478554298325517e-05, "loss": 0.8897, "step": 2737 }, { "epoch": 1.240036231884058, "grad_norm": 0.4563578079156692, "learning_rate": 7.47792974060746e-05, "loss": 0.9058, "step": 2738 }, { "epoch": 1.2404891304347827, "grad_norm": 0.5837596015408731, "learning_rate": 7.477304835195257e-05, "loss": 0.9182, "step": 2739 }, { "epoch": 1.2409420289855073, "grad_norm": 0.7645794855035182, "learning_rate": 7.476679582151379e-05, "loss": 0.8852, "step": 2740 }, { "epoch": 1.241394927536232, "grad_norm": 0.9373475985992099, "learning_rate": 7.476053981538332e-05, "loss": 0.9133, "step": 2741 }, { "epoch": 1.2418478260869565, "grad_norm": 0.8594248604648865, "learning_rate": 7.475428033418663e-05, "loss": 0.8969, "step": 2742 }, { "epoch": 1.2423007246376812, "grad_norm": 0.6367358643700288, "learning_rate": 7.474801737854944e-05, "loss": 0.8985, "step": 2743 }, { "epoch": 1.2427536231884058, "grad_norm": 0.5818447585206878, "learning_rate": 7.47417509490979e-05, "loss": 0.8807, "step": 2744 }, { "epoch": 1.2432065217391304, "grad_norm": 0.5540338629209884, "learning_rate": 7.473548104645847e-05, "loss": 0.904, "step": 2745 }, { "epoch": 1.243659420289855, "grad_norm": 0.48137637249789017, "learning_rate": 7.472920767125795e-05, "loss": 0.866, "step": 2746 }, { "epoch": 1.2441123188405796, "grad_norm": 0.43398119074597125, "learning_rate": 7.47229308241235e-05, "loss": 0.8903, "step": 2747 }, { "epoch": 1.2445652173913044, "grad_norm": 0.4783907349687721, "learning_rate": 7.471665050568265e-05, "loss": 0.8871, "step": 2748 }, { "epoch": 1.245018115942029, "grad_norm": 0.4784886074018755, "learning_rate": 7.471036671656322e-05, "loss": 0.8952, "step": 2749 }, { "epoch": 1.2454710144927537, "grad_norm": 0.5206126210533608, "learning_rate": 7.470407945739343e-05, "loss": 0.9327, "step": 2750 }, { "epoch": 1.2459239130434783, "grad_norm": 0.5200352940799347, "learning_rate": 7.469778872880183e-05, "loss": 0.8766, "step": 2751 }, { "epoch": 1.2463768115942029, "grad_norm": 0.42623135371794896, "learning_rate": 7.469149453141729e-05, "loss": 0.9086, "step": 2752 }, { "epoch": 1.2468297101449275, "grad_norm": 0.4789648780424578, "learning_rate": 7.468519686586909e-05, "loss": 0.8994, "step": 2753 }, { "epoch": 1.247282608695652, "grad_norm": 0.6176176034380523, "learning_rate": 7.467889573278678e-05, "loss": 0.8873, "step": 2754 }, { "epoch": 1.2477355072463767, "grad_norm": 0.6275865529387482, "learning_rate": 7.467259113280032e-05, "loss": 0.8941, "step": 2755 }, { "epoch": 1.2481884057971016, "grad_norm": 0.45514284262339344, "learning_rate": 7.466628306653997e-05, "loss": 0.9019, "step": 2756 }, { "epoch": 1.2486413043478262, "grad_norm": 0.4242512691518988, "learning_rate": 7.465997153463638e-05, "loss": 0.9123, "step": 2757 }, { "epoch": 1.2490942028985508, "grad_norm": 0.48027483835833906, "learning_rate": 7.46536565377205e-05, "loss": 0.8798, "step": 2758 }, { "epoch": 1.2495471014492754, "grad_norm": 0.4250578818534876, "learning_rate": 7.464733807642367e-05, "loss": 0.9001, "step": 2759 }, { "epoch": 1.25, "grad_norm": 0.31426359362592576, "learning_rate": 7.464101615137756e-05, "loss": 0.8906, "step": 2760 }, { "epoch": 1.2504528985507246, "grad_norm": 0.41095600552916045, "learning_rate": 7.463469076321416e-05, "loss": 0.8787, "step": 2761 }, { "epoch": 1.2509057971014492, "grad_norm": 0.4438302330687449, "learning_rate": 7.462836191256584e-05, "loss": 0.8999, "step": 2762 }, { "epoch": 1.2513586956521738, "grad_norm": 0.40075173965142236, "learning_rate": 7.462202960006531e-05, "loss": 0.9082, "step": 2763 }, { "epoch": 1.2518115942028984, "grad_norm": 0.41927573344278035, "learning_rate": 7.461569382634561e-05, "loss": 0.8972, "step": 2764 }, { "epoch": 1.252264492753623, "grad_norm": 0.4504945494133159, "learning_rate": 7.460935459204015e-05, "loss": 0.869, "step": 2765 }, { "epoch": 1.252717391304348, "grad_norm": 0.363545124688572, "learning_rate": 7.460301189778266e-05, "loss": 0.876, "step": 2766 }, { "epoch": 1.2531702898550725, "grad_norm": 0.3054400539543748, "learning_rate": 7.459666574420726e-05, "loss": 0.8785, "step": 2767 }, { "epoch": 1.2536231884057971, "grad_norm": 0.39758259673407903, "learning_rate": 7.459031613194834e-05, "loss": 0.9262, "step": 2768 }, { "epoch": 1.2540760869565217, "grad_norm": 0.3723233453882747, "learning_rate": 7.458396306164072e-05, "loss": 0.8807, "step": 2769 }, { "epoch": 1.2545289855072463, "grad_norm": 0.37258793883934743, "learning_rate": 7.45776065339195e-05, "loss": 0.8983, "step": 2770 }, { "epoch": 1.254981884057971, "grad_norm": 0.5254459130172944, "learning_rate": 7.457124654942016e-05, "loss": 0.9085, "step": 2771 }, { "epoch": 1.2554347826086958, "grad_norm": 0.488096410845411, "learning_rate": 7.456488310877853e-05, "loss": 0.8955, "step": 2772 }, { "epoch": 1.2558876811594204, "grad_norm": 0.3159550794821847, "learning_rate": 7.455851621263075e-05, "loss": 0.8948, "step": 2773 }, { "epoch": 1.256340579710145, "grad_norm": 0.32704608968951104, "learning_rate": 7.455214586161336e-05, "loss": 0.8736, "step": 2774 }, { "epoch": 1.2567934782608696, "grad_norm": 0.4088515213800953, "learning_rate": 7.454577205636318e-05, "loss": 0.8857, "step": 2775 }, { "epoch": 1.2572463768115942, "grad_norm": 0.5453146104442786, "learning_rate": 7.453939479751742e-05, "loss": 0.9012, "step": 2776 }, { "epoch": 1.2576992753623188, "grad_norm": 0.6444721486456141, "learning_rate": 7.453301408571363e-05, "loss": 0.9014, "step": 2777 }, { "epoch": 1.2581521739130435, "grad_norm": 1.149755203106174, "learning_rate": 7.452662992158971e-05, "loss": 0.9088, "step": 2778 }, { "epoch": 1.258605072463768, "grad_norm": 0.5937408761204725, "learning_rate": 7.452024230578389e-05, "loss": 0.8876, "step": 2779 }, { "epoch": 1.2590579710144927, "grad_norm": 0.6026801768503398, "learning_rate": 7.451385123893474e-05, "loss": 0.8928, "step": 2780 }, { "epoch": 1.2595108695652173, "grad_norm": 0.6389912489909558, "learning_rate": 7.450745672168118e-05, "loss": 0.9041, "step": 2781 }, { "epoch": 1.259963768115942, "grad_norm": 0.7200746203896417, "learning_rate": 7.45010587546625e-05, "loss": 0.8802, "step": 2782 }, { "epoch": 1.2604166666666667, "grad_norm": 0.8509256776296615, "learning_rate": 7.44946573385183e-05, "loss": 0.8983, "step": 2783 }, { "epoch": 1.2608695652173914, "grad_norm": 0.9651412446690646, "learning_rate": 7.448825247388854e-05, "loss": 0.8967, "step": 2784 }, { "epoch": 1.261322463768116, "grad_norm": 1.0302979265091807, "learning_rate": 7.448184416141355e-05, "loss": 0.912, "step": 2785 }, { "epoch": 1.2617753623188406, "grad_norm": 0.9729069955426134, "learning_rate": 7.447543240173392e-05, "loss": 0.8912, "step": 2786 }, { "epoch": 1.2622282608695652, "grad_norm": 0.8631737609624508, "learning_rate": 7.446901719549073e-05, "loss": 0.8906, "step": 2787 }, { "epoch": 1.2626811594202898, "grad_norm": 0.6899277728639498, "learning_rate": 7.446259854332523e-05, "loss": 0.9097, "step": 2788 }, { "epoch": 1.2631340579710144, "grad_norm": 0.47390461206828044, "learning_rate": 7.445617644587915e-05, "loss": 0.9018, "step": 2789 }, { "epoch": 1.2635869565217392, "grad_norm": 0.3877605561224879, "learning_rate": 7.444975090379453e-05, "loss": 0.8956, "step": 2790 }, { "epoch": 1.2640398550724639, "grad_norm": 0.45100960527811085, "learning_rate": 7.44433219177137e-05, "loss": 0.9053, "step": 2791 }, { "epoch": 1.2644927536231885, "grad_norm": 0.5535059871384657, "learning_rate": 7.443688948827941e-05, "loss": 0.8838, "step": 2792 }, { "epoch": 1.264945652173913, "grad_norm": 0.7780671542341792, "learning_rate": 7.44304536161347e-05, "loss": 0.8835, "step": 2793 }, { "epoch": 1.2653985507246377, "grad_norm": 0.928638391731949, "learning_rate": 7.442401430192299e-05, "loss": 0.8956, "step": 2794 }, { "epoch": 1.2658514492753623, "grad_norm": 0.9963325954598409, "learning_rate": 7.441757154628802e-05, "loss": 0.9012, "step": 2795 }, { "epoch": 1.266304347826087, "grad_norm": 0.8580218962021013, "learning_rate": 7.441112534987389e-05, "loss": 0.9057, "step": 2796 }, { "epoch": 1.2667572463768115, "grad_norm": 0.6923653400066643, "learning_rate": 7.440467571332502e-05, "loss": 0.8901, "step": 2797 }, { "epoch": 1.2672101449275361, "grad_norm": 0.6430570314106571, "learning_rate": 7.43982226372862e-05, "loss": 0.9024, "step": 2798 }, { "epoch": 1.2676630434782608, "grad_norm": 0.5097930184275737, "learning_rate": 7.439176612240255e-05, "loss": 0.9063, "step": 2799 }, { "epoch": 1.2681159420289856, "grad_norm": 0.3380963065826477, "learning_rate": 7.438530616931955e-05, "loss": 0.8976, "step": 2800 }, { "epoch": 1.2685688405797102, "grad_norm": 0.28117537040027707, "learning_rate": 7.4378842778683e-05, "loss": 0.8974, "step": 2801 }, { "epoch": 1.2690217391304348, "grad_norm": 0.35405412801347946, "learning_rate": 7.437237595113906e-05, "loss": 0.8942, "step": 2802 }, { "epoch": 1.2694746376811594, "grad_norm": 0.47167070277469497, "learning_rate": 7.436590568733422e-05, "loss": 0.8971, "step": 2803 }, { "epoch": 1.269927536231884, "grad_norm": 0.5365060178879755, "learning_rate": 7.435943198791535e-05, "loss": 0.8896, "step": 2804 }, { "epoch": 1.2703804347826086, "grad_norm": 0.5277569547309011, "learning_rate": 7.43529548535296e-05, "loss": 0.8922, "step": 2805 }, { "epoch": 1.2708333333333333, "grad_norm": 0.6336662980916348, "learning_rate": 7.434647428482453e-05, "loss": 0.8904, "step": 2806 }, { "epoch": 1.271286231884058, "grad_norm": 0.6978496539250317, "learning_rate": 7.4339990282448e-05, "loss": 0.8902, "step": 2807 }, { "epoch": 1.2717391304347827, "grad_norm": 0.660886987153646, "learning_rate": 7.433350284704822e-05, "loss": 0.9016, "step": 2808 }, { "epoch": 1.2721920289855073, "grad_norm": 0.5579563089137887, "learning_rate": 7.432701197927374e-05, "loss": 0.8917, "step": 2809 }, { "epoch": 1.272644927536232, "grad_norm": 0.4853700516781166, "learning_rate": 7.43205176797735e-05, "loss": 0.884, "step": 2810 }, { "epoch": 1.2730978260869565, "grad_norm": 0.4406924824411318, "learning_rate": 7.431401994919671e-05, "loss": 0.8917, "step": 2811 }, { "epoch": 1.2735507246376812, "grad_norm": 0.4291535772155829, "learning_rate": 7.430751878819297e-05, "loss": 0.9123, "step": 2812 }, { "epoch": 1.2740036231884058, "grad_norm": 0.4477822371761897, "learning_rate": 7.430101419741221e-05, "loss": 0.8835, "step": 2813 }, { "epoch": 1.2744565217391304, "grad_norm": 0.4132907987074286, "learning_rate": 7.429450617750471e-05, "loss": 0.8946, "step": 2814 }, { "epoch": 1.274909420289855, "grad_norm": 0.3506029946740727, "learning_rate": 7.428799472912108e-05, "loss": 0.8854, "step": 2815 }, { "epoch": 1.2753623188405796, "grad_norm": 0.35142808495317124, "learning_rate": 7.428147985291228e-05, "loss": 0.888, "step": 2816 }, { "epoch": 1.2758152173913042, "grad_norm": 0.32106860687156763, "learning_rate": 7.427496154952963e-05, "loss": 0.8827, "step": 2817 }, { "epoch": 1.276268115942029, "grad_norm": 0.3470820508718856, "learning_rate": 7.426843981962475e-05, "loss": 0.8936, "step": 2818 }, { "epoch": 1.2767210144927537, "grad_norm": 0.391756852532758, "learning_rate": 7.426191466384964e-05, "loss": 0.872, "step": 2819 }, { "epoch": 1.2771739130434783, "grad_norm": 0.4109665361252229, "learning_rate": 7.425538608285662e-05, "loss": 0.9092, "step": 2820 }, { "epoch": 1.2776268115942029, "grad_norm": 0.3295608929937743, "learning_rate": 7.424885407729838e-05, "loss": 0.8892, "step": 2821 }, { "epoch": 1.2780797101449275, "grad_norm": 0.30143458908984894, "learning_rate": 7.424231864782793e-05, "loss": 0.9095, "step": 2822 }, { "epoch": 1.278532608695652, "grad_norm": 0.31994894579565547, "learning_rate": 7.423577979509862e-05, "loss": 0.8865, "step": 2823 }, { "epoch": 1.278985507246377, "grad_norm": 0.4813786127705964, "learning_rate": 7.422923751976415e-05, "loss": 0.8968, "step": 2824 }, { "epoch": 1.2794384057971016, "grad_norm": 0.6358989823025261, "learning_rate": 7.422269182247856e-05, "loss": 0.877, "step": 2825 }, { "epoch": 1.2798913043478262, "grad_norm": 0.7491615823240447, "learning_rate": 7.421614270389623e-05, "loss": 0.9028, "step": 2826 }, { "epoch": 1.2803442028985508, "grad_norm": 0.8954906437319622, "learning_rate": 7.420959016467193e-05, "loss": 0.9013, "step": 2827 }, { "epoch": 1.2807971014492754, "grad_norm": 1.0656541132958934, "learning_rate": 7.420303420546067e-05, "loss": 0.8808, "step": 2828 }, { "epoch": 1.28125, "grad_norm": 0.9237926577066639, "learning_rate": 7.419647482691788e-05, "loss": 0.892, "step": 2829 }, { "epoch": 1.2817028985507246, "grad_norm": 0.794213378802678, "learning_rate": 7.41899120296993e-05, "loss": 0.9023, "step": 2830 }, { "epoch": 1.2821557971014492, "grad_norm": 0.638149850464519, "learning_rate": 7.418334581446106e-05, "loss": 0.8742, "step": 2831 }, { "epoch": 1.2826086956521738, "grad_norm": 0.42406555880481667, "learning_rate": 7.417677618185955e-05, "loss": 0.8841, "step": 2832 }, { "epoch": 1.2830615942028984, "grad_norm": 0.2393073328689181, "learning_rate": 7.417020313255158e-05, "loss": 0.8697, "step": 2833 }, { "epoch": 1.283514492753623, "grad_norm": 0.2960081713478775, "learning_rate": 7.416362666719425e-05, "loss": 0.8961, "step": 2834 }, { "epoch": 1.283967391304348, "grad_norm": 0.3883574625958972, "learning_rate": 7.415704678644501e-05, "loss": 0.8969, "step": 2835 }, { "epoch": 1.2844202898550725, "grad_norm": 0.45826273076798446, "learning_rate": 7.41504634909617e-05, "loss": 0.8821, "step": 2836 }, { "epoch": 1.2848731884057971, "grad_norm": 0.5336833901314922, "learning_rate": 7.414387678140244e-05, "loss": 0.8946, "step": 2837 }, { "epoch": 1.2853260869565217, "grad_norm": 0.5978532942217856, "learning_rate": 7.41372866584257e-05, "loss": 0.8979, "step": 2838 }, { "epoch": 1.2857789855072463, "grad_norm": 0.6976468908812087, "learning_rate": 7.413069312269033e-05, "loss": 0.8821, "step": 2839 }, { "epoch": 1.286231884057971, "grad_norm": 0.6936996623407288, "learning_rate": 7.412409617485547e-05, "loss": 0.8844, "step": 2840 }, { "epoch": 1.2866847826086958, "grad_norm": 0.695580983725279, "learning_rate": 7.411749581558064e-05, "loss": 0.8779, "step": 2841 }, { "epoch": 1.2871376811594204, "grad_norm": 0.664616348516922, "learning_rate": 7.41108920455257e-05, "loss": 0.901, "step": 2842 }, { "epoch": 1.287590579710145, "grad_norm": 0.6147819454866269, "learning_rate": 7.410428486535083e-05, "loss": 0.8826, "step": 2843 }, { "epoch": 1.2880434782608696, "grad_norm": 0.5922476820034552, "learning_rate": 7.409767427571655e-05, "loss": 0.8739, "step": 2844 }, { "epoch": 1.2884963768115942, "grad_norm": 0.6292713154303994, "learning_rate": 7.409106027728374e-05, "loss": 0.9001, "step": 2845 }, { "epoch": 1.2889492753623188, "grad_norm": 0.6227211087261502, "learning_rate": 7.408444287071362e-05, "loss": 0.8683, "step": 2846 }, { "epoch": 1.2894021739130435, "grad_norm": 0.6370485353487277, "learning_rate": 7.407782205666774e-05, "loss": 0.8945, "step": 2847 }, { "epoch": 1.289855072463768, "grad_norm": 0.6544851328737583, "learning_rate": 7.407119783580797e-05, "loss": 0.8865, "step": 2848 }, { "epoch": 1.2903079710144927, "grad_norm": 0.7031309469869814, "learning_rate": 7.406457020879658e-05, "loss": 0.9025, "step": 2849 }, { "epoch": 1.2907608695652173, "grad_norm": 0.7212194996186632, "learning_rate": 7.405793917629611e-05, "loss": 0.8736, "step": 2850 }, { "epoch": 1.291213768115942, "grad_norm": 0.7434898302887506, "learning_rate": 7.40513047389695e-05, "loss": 0.8999, "step": 2851 }, { "epoch": 1.2916666666666667, "grad_norm": 0.7150434403026765, "learning_rate": 7.404466689747999e-05, "loss": 0.8804, "step": 2852 }, { "epoch": 1.2921195652173914, "grad_norm": 0.6725256329492177, "learning_rate": 7.403802565249118e-05, "loss": 0.8747, "step": 2853 }, { "epoch": 1.292572463768116, "grad_norm": 0.6372843835478433, "learning_rate": 7.403138100466702e-05, "loss": 0.871, "step": 2854 }, { "epoch": 1.2930253623188406, "grad_norm": 0.6423262193615347, "learning_rate": 7.402473295467178e-05, "loss": 0.8845, "step": 2855 }, { "epoch": 1.2934782608695652, "grad_norm": 0.6412338999106608, "learning_rate": 7.401808150317005e-05, "loss": 0.8884, "step": 2856 }, { "epoch": 1.2939311594202898, "grad_norm": 0.5983818400696211, "learning_rate": 7.401142665082682e-05, "loss": 0.8789, "step": 2857 }, { "epoch": 1.2943840579710144, "grad_norm": 0.5302959367843835, "learning_rate": 7.400476839830736e-05, "loss": 0.9193, "step": 2858 }, { "epoch": 1.2948369565217392, "grad_norm": 0.45233712959797245, "learning_rate": 7.399810674627732e-05, "loss": 0.891, "step": 2859 }, { "epoch": 1.2952898550724639, "grad_norm": 0.3954844466923572, "learning_rate": 7.399144169540268e-05, "loss": 0.9023, "step": 2860 }, { "epoch": 1.2957427536231885, "grad_norm": 0.46861175224484675, "learning_rate": 7.398477324634975e-05, "loss": 0.8649, "step": 2861 }, { "epoch": 1.296195652173913, "grad_norm": 0.44843772919263253, "learning_rate": 7.397810139978518e-05, "loss": 0.885, "step": 2862 }, { "epoch": 1.2966485507246377, "grad_norm": 0.3192267171840374, "learning_rate": 7.397142615637598e-05, "loss": 0.8669, "step": 2863 }, { "epoch": 1.2971014492753623, "grad_norm": 0.3563471649093596, "learning_rate": 7.396474751678947e-05, "loss": 0.8828, "step": 2864 }, { "epoch": 1.297554347826087, "grad_norm": 0.3836275009942196, "learning_rate": 7.395806548169332e-05, "loss": 0.8874, "step": 2865 }, { "epoch": 1.2980072463768115, "grad_norm": 0.49594398094723613, "learning_rate": 7.395138005175555e-05, "loss": 0.8913, "step": 2866 }, { "epoch": 1.2984601449275361, "grad_norm": 0.6670607437122544, "learning_rate": 7.394469122764454e-05, "loss": 0.8864, "step": 2867 }, { "epoch": 1.2989130434782608, "grad_norm": 0.8015963445147191, "learning_rate": 7.393799901002894e-05, "loss": 0.872, "step": 2868 }, { "epoch": 1.2993659420289856, "grad_norm": 0.8941888393236056, "learning_rate": 7.39313033995778e-05, "loss": 0.8786, "step": 2869 }, { "epoch": 1.2998188405797102, "grad_norm": 0.9073965916520789, "learning_rate": 7.392460439696049e-05, "loss": 0.9015, "step": 2870 }, { "epoch": 1.3002717391304348, "grad_norm": 0.8864258576178426, "learning_rate": 7.391790200284673e-05, "loss": 0.8915, "step": 2871 }, { "epoch": 1.3007246376811594, "grad_norm": 0.7804354772930685, "learning_rate": 7.391119621790655e-05, "loss": 0.9057, "step": 2872 }, { "epoch": 1.301177536231884, "grad_norm": 0.6645540868857259, "learning_rate": 7.390448704281035e-05, "loss": 0.8852, "step": 2873 }, { "epoch": 1.3016304347826086, "grad_norm": 0.47258642745465634, "learning_rate": 7.389777447822885e-05, "loss": 0.8781, "step": 2874 }, { "epoch": 1.3020833333333333, "grad_norm": 0.3001439672458163, "learning_rate": 7.389105852483312e-05, "loss": 0.9105, "step": 2875 }, { "epoch": 1.302536231884058, "grad_norm": 0.4987703720971828, "learning_rate": 7.388433918329457e-05, "loss": 0.8816, "step": 2876 }, { "epoch": 1.3029891304347827, "grad_norm": 0.532455253623664, "learning_rate": 7.387761645428494e-05, "loss": 0.8782, "step": 2877 }, { "epoch": 1.3034420289855073, "grad_norm": 0.6258891773834379, "learning_rate": 7.387089033847631e-05, "loss": 0.896, "step": 2878 }, { "epoch": 1.303894927536232, "grad_norm": 0.7834806762363963, "learning_rate": 7.386416083654109e-05, "loss": 0.8885, "step": 2879 }, { "epoch": 1.3043478260869565, "grad_norm": 0.8474080248011211, "learning_rate": 7.385742794915205e-05, "loss": 0.8817, "step": 2880 }, { "epoch": 1.3048007246376812, "grad_norm": 0.7705120130416062, "learning_rate": 7.385069167698228e-05, "loss": 0.8733, "step": 2881 }, { "epoch": 1.3052536231884058, "grad_norm": 0.6474904816917524, "learning_rate": 7.384395202070523e-05, "loss": 0.885, "step": 2882 }, { "epoch": 1.3057065217391304, "grad_norm": 0.5104691849844898, "learning_rate": 7.383720898099468e-05, "loss": 0.9027, "step": 2883 }, { "epoch": 1.306159420289855, "grad_norm": 0.5614076071730398, "learning_rate": 7.38304625585247e-05, "loss": 0.9041, "step": 2884 }, { "epoch": 1.3066123188405796, "grad_norm": 0.5345453079382195, "learning_rate": 7.382371275396979e-05, "loss": 0.8851, "step": 2885 }, { "epoch": 1.3070652173913042, "grad_norm": 0.40711889620703023, "learning_rate": 7.381695956800473e-05, "loss": 0.8935, "step": 2886 }, { "epoch": 1.307518115942029, "grad_norm": 0.302788539306079, "learning_rate": 7.381020300130461e-05, "loss": 0.8772, "step": 2887 }, { "epoch": 1.3079710144927537, "grad_norm": 0.3513181562063532, "learning_rate": 7.380344305454493e-05, "loss": 0.8761, "step": 2888 }, { "epoch": 1.3084239130434783, "grad_norm": 0.44099457602199477, "learning_rate": 7.379667972840148e-05, "loss": 0.8695, "step": 2889 }, { "epoch": 1.3088768115942029, "grad_norm": 0.4765701868271292, "learning_rate": 7.378991302355038e-05, "loss": 0.872, "step": 2890 }, { "epoch": 1.3093297101449275, "grad_norm": 0.4061093435789149, "learning_rate": 7.378314294066817e-05, "loss": 0.8927, "step": 2891 }, { "epoch": 1.309782608695652, "grad_norm": 0.444547439637185, "learning_rate": 7.37763694804316e-05, "loss": 0.8969, "step": 2892 }, { "epoch": 1.310235507246377, "grad_norm": 0.5203114912523962, "learning_rate": 7.376959264351784e-05, "loss": 0.8972, "step": 2893 }, { "epoch": 1.3106884057971016, "grad_norm": 0.6151144209378233, "learning_rate": 7.376281243060443e-05, "loss": 0.8803, "step": 2894 }, { "epoch": 1.3111413043478262, "grad_norm": 0.8174409438703965, "learning_rate": 7.375602884236913e-05, "loss": 0.9083, "step": 2895 }, { "epoch": 1.3115942028985508, "grad_norm": 1.0220024497994664, "learning_rate": 7.374924187949014e-05, "loss": 0.8759, "step": 2896 }, { "epoch": 1.3120471014492754, "grad_norm": 1.0564554957130938, "learning_rate": 7.374245154264597e-05, "loss": 0.8765, "step": 2897 }, { "epoch": 1.3125, "grad_norm": 0.8550863513483059, "learning_rate": 7.373565783251544e-05, "loss": 0.887, "step": 2898 }, { "epoch": 1.3129528985507246, "grad_norm": 0.6214946497793886, "learning_rate": 7.372886074977774e-05, "loss": 0.8765, "step": 2899 }, { "epoch": 1.3134057971014492, "grad_norm": 0.4157473933747755, "learning_rate": 7.372206029511236e-05, "loss": 0.8725, "step": 2900 }, { "epoch": 1.3138586956521738, "grad_norm": 0.2639388405728049, "learning_rate": 7.371525646919921e-05, "loss": 0.9088, "step": 2901 }, { "epoch": 1.3143115942028984, "grad_norm": 0.40908177676430124, "learning_rate": 7.370844927271844e-05, "loss": 0.8908, "step": 2902 }, { "epoch": 1.314764492753623, "grad_norm": 0.6275524601385761, "learning_rate": 7.370163870635056e-05, "loss": 0.8929, "step": 2903 }, { "epoch": 1.315217391304348, "grad_norm": 0.8120583244837499, "learning_rate": 7.369482477077647e-05, "loss": 0.8772, "step": 2904 }, { "epoch": 1.3156702898550725, "grad_norm": 0.8560422792040255, "learning_rate": 7.368800746667734e-05, "loss": 0.8932, "step": 2905 }, { "epoch": 1.3161231884057971, "grad_norm": 0.7641674597539604, "learning_rate": 7.368118679473475e-05, "loss": 0.8841, "step": 2906 }, { "epoch": 1.3165760869565217, "grad_norm": 0.7024755959991477, "learning_rate": 7.367436275563052e-05, "loss": 0.912, "step": 2907 }, { "epoch": 1.3170289855072463, "grad_norm": 0.5891300370037915, "learning_rate": 7.366753535004687e-05, "loss": 0.8961, "step": 2908 }, { "epoch": 1.317481884057971, "grad_norm": 0.5277194894493809, "learning_rate": 7.366070457866639e-05, "loss": 0.885, "step": 2909 }, { "epoch": 1.3179347826086958, "grad_norm": 0.5130813129010281, "learning_rate": 7.365387044217192e-05, "loss": 0.8877, "step": 2910 }, { "epoch": 1.3183876811594204, "grad_norm": 0.6139382011759147, "learning_rate": 7.364703294124669e-05, "loss": 0.8978, "step": 2911 }, { "epoch": 1.318840579710145, "grad_norm": 0.674569032099142, "learning_rate": 7.364019207657427e-05, "loss": 0.8963, "step": 2912 }, { "epoch": 1.3192934782608696, "grad_norm": 0.7085174805787585, "learning_rate": 7.363334784883852e-05, "loss": 0.9005, "step": 2913 }, { "epoch": 1.3197463768115942, "grad_norm": 0.6885656897863652, "learning_rate": 7.36265002587237e-05, "loss": 0.8941, "step": 2914 }, { "epoch": 1.3201992753623188, "grad_norm": 0.6991629565752614, "learning_rate": 7.361964930691438e-05, "loss": 0.8838, "step": 2915 }, { "epoch": 1.3206521739130435, "grad_norm": 0.6976119952172317, "learning_rate": 7.361279499409543e-05, "loss": 0.9069, "step": 2916 }, { "epoch": 1.321105072463768, "grad_norm": 0.7051375316572982, "learning_rate": 7.36059373209521e-05, "loss": 0.8848, "step": 2917 }, { "epoch": 1.3215579710144927, "grad_norm": 0.6850675224787396, "learning_rate": 7.359907628816995e-05, "loss": 0.8987, "step": 2918 }, { "epoch": 1.3220108695652173, "grad_norm": 0.6807453470206042, "learning_rate": 7.359221189643491e-05, "loss": 0.8783, "step": 2919 }, { "epoch": 1.322463768115942, "grad_norm": 0.631477040311845, "learning_rate": 7.358534414643323e-05, "loss": 0.9107, "step": 2920 }, { "epoch": 1.3229166666666667, "grad_norm": 0.5203670900406336, "learning_rate": 7.357847303885146e-05, "loss": 0.8975, "step": 2921 }, { "epoch": 1.3233695652173914, "grad_norm": 0.43563331719147275, "learning_rate": 7.357159857437654e-05, "loss": 0.8914, "step": 2922 }, { "epoch": 1.323822463768116, "grad_norm": 0.39819799574196085, "learning_rate": 7.35647207536957e-05, "loss": 0.9126, "step": 2923 }, { "epoch": 1.3242753623188406, "grad_norm": 0.4546619382977849, "learning_rate": 7.355783957749656e-05, "loss": 0.9016, "step": 2924 }, { "epoch": 1.3247282608695652, "grad_norm": 0.4774277531568301, "learning_rate": 7.355095504646699e-05, "loss": 0.8767, "step": 2925 }, { "epoch": 1.3251811594202898, "grad_norm": 0.40067282024762346, "learning_rate": 7.35440671612953e-05, "loss": 0.8922, "step": 2926 }, { "epoch": 1.3256340579710144, "grad_norm": 0.4405808579991879, "learning_rate": 7.353717592267005e-05, "loss": 0.9127, "step": 2927 }, { "epoch": 1.3260869565217392, "grad_norm": 0.5444713091613987, "learning_rate": 7.353028133128018e-05, "loss": 0.8869, "step": 2928 }, { "epoch": 1.3265398550724639, "grad_norm": 0.5661815347729329, "learning_rate": 7.352338338781495e-05, "loss": 0.9245, "step": 2929 }, { "epoch": 1.3269927536231885, "grad_norm": 0.6241438780192516, "learning_rate": 7.351648209296397e-05, "loss": 0.91, "step": 2930 }, { "epoch": 1.327445652173913, "grad_norm": 0.723266620493994, "learning_rate": 7.350957744741716e-05, "loss": 0.8826, "step": 2931 }, { "epoch": 1.3278985507246377, "grad_norm": 0.8786821638677074, "learning_rate": 7.350266945186478e-05, "loss": 0.8796, "step": 2932 }, { "epoch": 1.3283514492753623, "grad_norm": 1.0142212755751123, "learning_rate": 7.349575810699745e-05, "loss": 0.8752, "step": 2933 }, { "epoch": 1.328804347826087, "grad_norm": 1.0081520591131814, "learning_rate": 7.348884341350612e-05, "loss": 0.9069, "step": 2934 }, { "epoch": 1.3292572463768115, "grad_norm": 0.8745453593471451, "learning_rate": 7.348192537208204e-05, "loss": 0.9055, "step": 2935 }, { "epoch": 1.3297101449275361, "grad_norm": 0.7743284723632471, "learning_rate": 7.347500398341682e-05, "loss": 0.8887, "step": 2936 }, { "epoch": 1.3301630434782608, "grad_norm": 0.6109574775434886, "learning_rate": 7.346807924820241e-05, "loss": 0.9093, "step": 2937 }, { "epoch": 1.3306159420289856, "grad_norm": 0.415817945276889, "learning_rate": 7.346115116713108e-05, "loss": 0.8986, "step": 2938 }, { "epoch": 1.3310688405797102, "grad_norm": 0.3266382496755978, "learning_rate": 7.345421974089543e-05, "loss": 0.9016, "step": 2939 }, { "epoch": 1.3315217391304348, "grad_norm": 0.3434350766182101, "learning_rate": 7.344728497018844e-05, "loss": 0.8956, "step": 2940 }, { "epoch": 1.3319746376811594, "grad_norm": 0.36515622291693756, "learning_rate": 7.344034685570335e-05, "loss": 0.8831, "step": 2941 }, { "epoch": 1.332427536231884, "grad_norm": 0.44933945391014457, "learning_rate": 7.343340539813382e-05, "loss": 0.906, "step": 2942 }, { "epoch": 1.3328804347826086, "grad_norm": 0.4590743596330593, "learning_rate": 7.342646059817376e-05, "loss": 0.8711, "step": 2943 }, { "epoch": 1.3333333333333333, "grad_norm": 0.40849727593927737, "learning_rate": 7.341951245651747e-05, "loss": 0.8741, "step": 2944 }, { "epoch": 1.333786231884058, "grad_norm": 0.4215778385794867, "learning_rate": 7.341256097385955e-05, "loss": 0.8584, "step": 2945 }, { "epoch": 1.3342391304347827, "grad_norm": 0.5339399580666161, "learning_rate": 7.340560615089496e-05, "loss": 0.8955, "step": 2946 }, { "epoch": 1.3346920289855073, "grad_norm": 0.6428335250064175, "learning_rate": 7.3398647988319e-05, "loss": 0.8925, "step": 2947 }, { "epoch": 1.335144927536232, "grad_norm": 0.7233160877279641, "learning_rate": 7.339168648682728e-05, "loss": 0.9028, "step": 2948 }, { "epoch": 1.3355978260869565, "grad_norm": 0.6953578924418993, "learning_rate": 7.338472164711574e-05, "loss": 0.9158, "step": 2949 }, { "epoch": 1.3360507246376812, "grad_norm": 0.6296538077416728, "learning_rate": 7.337775346988068e-05, "loss": 0.894, "step": 2950 }, { "epoch": 1.3365036231884058, "grad_norm": 0.534141783614168, "learning_rate": 7.33707819558187e-05, "loss": 0.8872, "step": 2951 }, { "epoch": 1.3369565217391304, "grad_norm": 0.4846168039434347, "learning_rate": 7.336380710562678e-05, "loss": 0.8935, "step": 2952 }, { "epoch": 1.337409420289855, "grad_norm": 0.47637420386492785, "learning_rate": 7.335682892000218e-05, "loss": 0.8977, "step": 2953 }, { "epoch": 1.3378623188405796, "grad_norm": 0.4814979074629972, "learning_rate": 7.334984739964255e-05, "loss": 0.9434, "step": 2954 }, { "epoch": 1.3383152173913042, "grad_norm": 0.4037464443884316, "learning_rate": 7.334286254524582e-05, "loss": 0.8907, "step": 2955 }, { "epoch": 1.338768115942029, "grad_norm": 0.3164195891480404, "learning_rate": 7.333587435751029e-05, "loss": 0.8808, "step": 2956 }, { "epoch": 1.3392210144927537, "grad_norm": 0.3240266929815799, "learning_rate": 7.332888283713457e-05, "loss": 0.9155, "step": 2957 }, { "epoch": 1.3396739130434783, "grad_norm": 0.31859352133929486, "learning_rate": 7.332188798481763e-05, "loss": 0.8747, "step": 2958 }, { "epoch": 1.3401268115942029, "grad_norm": 0.3103934868410287, "learning_rate": 7.331488980125873e-05, "loss": 0.8794, "step": 2959 }, { "epoch": 1.3405797101449275, "grad_norm": 0.2625760498806908, "learning_rate": 7.330788828715752e-05, "loss": 0.8938, "step": 2960 }, { "epoch": 1.341032608695652, "grad_norm": 0.3115168551939088, "learning_rate": 7.330088344321392e-05, "loss": 0.8987, "step": 2961 }, { "epoch": 1.341485507246377, "grad_norm": 0.37449863813850426, "learning_rate": 7.329387527012824e-05, "loss": 0.9158, "step": 2962 }, { "epoch": 1.3419384057971016, "grad_norm": 0.3546577115869493, "learning_rate": 7.32868637686011e-05, "loss": 0.9011, "step": 2963 }, { "epoch": 1.3423913043478262, "grad_norm": 0.39739566111322355, "learning_rate": 7.327984893933342e-05, "loss": 0.9074, "step": 2964 }, { "epoch": 1.3428442028985508, "grad_norm": 0.38604845773758045, "learning_rate": 7.327283078302652e-05, "loss": 0.8664, "step": 2965 }, { "epoch": 1.3432971014492754, "grad_norm": 0.31652229873088256, "learning_rate": 7.326580930038201e-05, "loss": 0.915, "step": 2966 }, { "epoch": 1.34375, "grad_norm": 0.2815978910708371, "learning_rate": 7.325878449210182e-05, "loss": 0.8884, "step": 2967 }, { "epoch": 1.3442028985507246, "grad_norm": 0.3759943480538137, "learning_rate": 7.325175635888823e-05, "loss": 0.8672, "step": 2968 }, { "epoch": 1.3446557971014492, "grad_norm": 0.45962748142200704, "learning_rate": 7.324472490144388e-05, "loss": 0.888, "step": 2969 }, { "epoch": 1.3451086956521738, "grad_norm": 0.43804312290616126, "learning_rate": 7.323769012047169e-05, "loss": 0.8986, "step": 2970 }, { "epoch": 1.3455615942028984, "grad_norm": 0.48524362444193997, "learning_rate": 7.323065201667496e-05, "loss": 0.8915, "step": 2971 }, { "epoch": 1.346014492753623, "grad_norm": 0.6141348506873779, "learning_rate": 7.322361059075729e-05, "loss": 0.8759, "step": 2972 }, { "epoch": 1.346467391304348, "grad_norm": 0.7334011392739032, "learning_rate": 7.321656584342261e-05, "loss": 0.8778, "step": 2973 }, { "epoch": 1.3469202898550725, "grad_norm": 0.7870834764520993, "learning_rate": 7.320951777537522e-05, "loss": 0.8871, "step": 2974 }, { "epoch": 1.3473731884057971, "grad_norm": 0.8270863096617043, "learning_rate": 7.32024663873197e-05, "loss": 0.8773, "step": 2975 }, { "epoch": 1.3478260869565217, "grad_norm": 0.8015701624043552, "learning_rate": 7.319541167996102e-05, "loss": 0.8887, "step": 2976 }, { "epoch": 1.3482789855072463, "grad_norm": 0.6780857844939351, "learning_rate": 7.318835365400442e-05, "loss": 0.9006, "step": 2977 }, { "epoch": 1.348731884057971, "grad_norm": 0.4605489221889336, "learning_rate": 7.318129231015551e-05, "loss": 0.9008, "step": 2978 }, { "epoch": 1.3491847826086958, "grad_norm": 0.384367309230792, "learning_rate": 7.317422764912023e-05, "loss": 0.8846, "step": 2979 }, { "epoch": 1.3496376811594204, "grad_norm": 0.32699464523168276, "learning_rate": 7.316715967160484e-05, "loss": 0.9047, "step": 2980 }, { "epoch": 1.350090579710145, "grad_norm": 0.26159823584836006, "learning_rate": 7.316008837831595e-05, "loss": 0.8683, "step": 2981 }, { "epoch": 1.3505434782608696, "grad_norm": 0.38309382375091816, "learning_rate": 7.315301376996047e-05, "loss": 0.9045, "step": 2982 }, { "epoch": 1.3509963768115942, "grad_norm": 0.5289218341169305, "learning_rate": 7.314593584724566e-05, "loss": 0.8911, "step": 2983 }, { "epoch": 1.3514492753623188, "grad_norm": 0.5152735445806926, "learning_rate": 7.313885461087912e-05, "loss": 0.8844, "step": 2984 }, { "epoch": 1.3519021739130435, "grad_norm": 0.4743985698400954, "learning_rate": 7.31317700615688e-05, "loss": 0.8907, "step": 2985 }, { "epoch": 1.352355072463768, "grad_norm": 0.5449820461015717, "learning_rate": 7.31246822000229e-05, "loss": 0.8844, "step": 2986 }, { "epoch": 1.3528079710144927, "grad_norm": 0.6354795785440761, "learning_rate": 7.311759102695004e-05, "loss": 0.8763, "step": 2987 }, { "epoch": 1.3532608695652173, "grad_norm": 0.6666579923654812, "learning_rate": 7.311049654305911e-05, "loss": 0.9017, "step": 2988 }, { "epoch": 1.353713768115942, "grad_norm": 0.6027230101237628, "learning_rate": 7.310339874905939e-05, "loss": 0.8864, "step": 2989 }, { "epoch": 1.3541666666666667, "grad_norm": 0.5698765004407815, "learning_rate": 7.309629764566042e-05, "loss": 0.8958, "step": 2990 }, { "epoch": 1.3546195652173914, "grad_norm": 0.6013713297232776, "learning_rate": 7.308919323357215e-05, "loss": 0.8942, "step": 2991 }, { "epoch": 1.355072463768116, "grad_norm": 0.7028818001712304, "learning_rate": 7.30820855135048e-05, "loss": 0.8878, "step": 2992 }, { "epoch": 1.3555253623188406, "grad_norm": 0.700402002210149, "learning_rate": 7.307497448616891e-05, "loss": 0.913, "step": 2993 }, { "epoch": 1.3559782608695652, "grad_norm": 0.6329771068339519, "learning_rate": 7.306786015227543e-05, "loss": 0.8906, "step": 2994 }, { "epoch": 1.3564311594202898, "grad_norm": 0.6070039836401205, "learning_rate": 7.306074251253558e-05, "loss": 0.9211, "step": 2995 }, { "epoch": 1.3568840579710144, "grad_norm": 0.5857466921139564, "learning_rate": 7.305362156766089e-05, "loss": 0.8873, "step": 2996 }, { "epoch": 1.3573369565217392, "grad_norm": 0.6053530942496111, "learning_rate": 7.304649731836328e-05, "loss": 0.887, "step": 2997 }, { "epoch": 1.3577898550724639, "grad_norm": 0.6276424384182343, "learning_rate": 7.303936976535498e-05, "loss": 0.8966, "step": 2998 }, { "epoch": 1.3582427536231885, "grad_norm": 0.5840957044101708, "learning_rate": 7.303223890934851e-05, "loss": 0.8851, "step": 2999 }, { "epoch": 1.358695652173913, "grad_norm": 0.622146247419218, "learning_rate": 7.302510475105679e-05, "loss": 0.8942, "step": 3000 }, { "epoch": 1.3591485507246377, "grad_norm": 0.7110784418208911, "learning_rate": 7.301796729119301e-05, "loss": 0.8951, "step": 3001 }, { "epoch": 1.3596014492753623, "grad_norm": 1.164010525809893, "learning_rate": 7.30108265304707e-05, "loss": 0.8985, "step": 3002 }, { "epoch": 1.360054347826087, "grad_norm": 0.382921744939772, "learning_rate": 7.300368246960377e-05, "loss": 0.8722, "step": 3003 }, { "epoch": 1.3605072463768115, "grad_norm": 0.47496092135042944, "learning_rate": 7.29965351093064e-05, "loss": 0.8996, "step": 3004 }, { "epoch": 1.3609601449275361, "grad_norm": 0.6871999730425521, "learning_rate": 7.298938445029314e-05, "loss": 0.8857, "step": 3005 }, { "epoch": 1.3614130434782608, "grad_norm": 0.8803341031004254, "learning_rate": 7.298223049327881e-05, "loss": 0.9104, "step": 3006 }, { "epoch": 1.3618659420289856, "grad_norm": 1.0489382350206586, "learning_rate": 7.297507323897867e-05, "loss": 0.8882, "step": 3007 }, { "epoch": 1.3623188405797102, "grad_norm": 0.9001450715186236, "learning_rate": 7.296791268810818e-05, "loss": 0.8961, "step": 3008 }, { "epoch": 1.3627717391304348, "grad_norm": 0.6968751710849226, "learning_rate": 7.296074884138322e-05, "loss": 0.8763, "step": 3009 }, { "epoch": 1.3632246376811594, "grad_norm": 0.43456339991284143, "learning_rate": 7.295358169951998e-05, "loss": 0.9078, "step": 3010 }, { "epoch": 1.363677536231884, "grad_norm": 0.48102678412903255, "learning_rate": 7.294641126323494e-05, "loss": 0.8875, "step": 3011 }, { "epoch": 1.3641304347826086, "grad_norm": 0.6214758276121017, "learning_rate": 7.293923753324496e-05, "loss": 0.8925, "step": 3012 }, { "epoch": 1.3645833333333333, "grad_norm": 0.7871720455620557, "learning_rate": 7.293206051026722e-05, "loss": 0.8853, "step": 3013 }, { "epoch": 1.365036231884058, "grad_norm": 0.9350755099367345, "learning_rate": 7.292488019501919e-05, "loss": 0.8731, "step": 3014 }, { "epoch": 1.3654891304347827, "grad_norm": 0.911116848677372, "learning_rate": 7.291769658821873e-05, "loss": 0.8976, "step": 3015 }, { "epoch": 1.3659420289855073, "grad_norm": 0.8343642370573963, "learning_rate": 7.291050969058397e-05, "loss": 0.889, "step": 3016 }, { "epoch": 1.366394927536232, "grad_norm": 0.7382970539261865, "learning_rate": 7.290331950283341e-05, "loss": 0.8922, "step": 3017 }, { "epoch": 1.3668478260869565, "grad_norm": 0.5947843336669999, "learning_rate": 7.289612602568586e-05, "loss": 0.8816, "step": 3018 }, { "epoch": 1.3673007246376812, "grad_norm": 0.4116747456722887, "learning_rate": 7.288892925986045e-05, "loss": 0.9002, "step": 3019 }, { "epoch": 1.3677536231884058, "grad_norm": 0.3314061060653942, "learning_rate": 7.288172920607668e-05, "loss": 0.8923, "step": 3020 }, { "epoch": 1.3682065217391304, "grad_norm": 0.37749075170532714, "learning_rate": 7.287452586505434e-05, "loss": 0.8841, "step": 3021 }, { "epoch": 1.368659420289855, "grad_norm": 0.5508964564036557, "learning_rate": 7.286731923751353e-05, "loss": 0.8799, "step": 3022 }, { "epoch": 1.3691123188405796, "grad_norm": 0.6936062148888457, "learning_rate": 7.286010932417474e-05, "loss": 0.8762, "step": 3023 }, { "epoch": 1.3695652173913042, "grad_norm": 0.8723969495471948, "learning_rate": 7.285289612575876e-05, "loss": 0.8791, "step": 3024 }, { "epoch": 1.370018115942029, "grad_norm": 1.0857554997781682, "learning_rate": 7.284567964298669e-05, "loss": 0.8917, "step": 3025 }, { "epoch": 1.3704710144927537, "grad_norm": 0.8177301168867975, "learning_rate": 7.283845987657998e-05, "loss": 0.9006, "step": 3026 }, { "epoch": 1.3709239130434783, "grad_norm": 0.5899337474100553, "learning_rate": 7.28312368272604e-05, "loss": 0.8879, "step": 3027 }, { "epoch": 1.3713768115942029, "grad_norm": 0.5624041520020165, "learning_rate": 7.282401049575005e-05, "loss": 0.8848, "step": 3028 }, { "epoch": 1.3718297101449275, "grad_norm": 0.6800134982418563, "learning_rate": 7.281678088277134e-05, "loss": 0.8848, "step": 3029 }, { "epoch": 1.372282608695652, "grad_norm": 0.7079729692962854, "learning_rate": 7.280954798904708e-05, "loss": 0.9064, "step": 3030 }, { "epoch": 1.372735507246377, "grad_norm": 0.5388486441202281, "learning_rate": 7.280231181530029e-05, "loss": 0.8979, "step": 3031 }, { "epoch": 1.3731884057971016, "grad_norm": 0.45779893720108894, "learning_rate": 7.279507236225441e-05, "loss": 0.8829, "step": 3032 }, { "epoch": 1.3736413043478262, "grad_norm": 0.48402864577061794, "learning_rate": 7.278782963063317e-05, "loss": 0.8845, "step": 3033 }, { "epoch": 1.3740942028985508, "grad_norm": 0.5539600757260337, "learning_rate": 7.278058362116065e-05, "loss": 0.9117, "step": 3034 }, { "epoch": 1.3745471014492754, "grad_norm": 0.5813205696890681, "learning_rate": 7.277333433456126e-05, "loss": 0.8855, "step": 3035 }, { "epoch": 1.375, "grad_norm": 0.531195224158702, "learning_rate": 7.276608177155968e-05, "loss": 0.895, "step": 3036 }, { "epoch": 1.3754528985507246, "grad_norm": 0.4738971657259408, "learning_rate": 7.275882593288099e-05, "loss": 0.8749, "step": 3037 }, { "epoch": 1.3759057971014492, "grad_norm": 1.2392947526113907, "learning_rate": 7.275156681925056e-05, "loss": 0.9212, "step": 3038 }, { "epoch": 1.3763586956521738, "grad_norm": 0.4102480499695737, "learning_rate": 7.27443044313941e-05, "loss": 0.8737, "step": 3039 }, { "epoch": 1.3768115942028984, "grad_norm": 0.5143838117895696, "learning_rate": 7.273703877003763e-05, "loss": 0.9095, "step": 3040 }, { "epoch": 1.377264492753623, "grad_norm": 0.6103697575010043, "learning_rate": 7.272976983590752e-05, "loss": 0.8827, "step": 3041 }, { "epoch": 1.377717391304348, "grad_norm": 0.7083599436818669, "learning_rate": 7.272249762973046e-05, "loss": 0.9004, "step": 3042 }, { "epoch": 1.3781702898550725, "grad_norm": 0.8360903961281264, "learning_rate": 7.271522215223344e-05, "loss": 0.8836, "step": 3043 }, { "epoch": 1.3786231884057971, "grad_norm": 0.8316653289439071, "learning_rate": 7.270794340414385e-05, "loss": 0.8904, "step": 3044 }, { "epoch": 1.3790760869565217, "grad_norm": 0.7203955089515389, "learning_rate": 7.270066138618931e-05, "loss": 0.8833, "step": 3045 }, { "epoch": 1.3795289855072463, "grad_norm": 0.5275189974192279, "learning_rate": 7.269337609909783e-05, "loss": 0.885, "step": 3046 }, { "epoch": 1.379981884057971, "grad_norm": 0.4401739748531849, "learning_rate": 7.268608754359774e-05, "loss": 0.8862, "step": 3047 }, { "epoch": 1.3804347826086958, "grad_norm": 0.3672554779553054, "learning_rate": 7.267879572041768e-05, "loss": 0.9065, "step": 3048 }, { "epoch": 1.3808876811594204, "grad_norm": 0.3281774406673306, "learning_rate": 7.267150063028664e-05, "loss": 0.8895, "step": 3049 }, { "epoch": 1.381340579710145, "grad_norm": 0.2943965891972056, "learning_rate": 7.26642022739339e-05, "loss": 0.8916, "step": 3050 }, { "epoch": 1.3817934782608696, "grad_norm": 0.35095004984165185, "learning_rate": 7.26569006520891e-05, "loss": 0.8973, "step": 3051 }, { "epoch": 1.3822463768115942, "grad_norm": 0.46213212866495296, "learning_rate": 7.264959576548218e-05, "loss": 0.8969, "step": 3052 }, { "epoch": 1.3826992753623188, "grad_norm": 0.5515444147667338, "learning_rate": 7.264228761484344e-05, "loss": 0.8744, "step": 3053 }, { "epoch": 1.3831521739130435, "grad_norm": 0.5993831102420728, "learning_rate": 7.263497620090349e-05, "loss": 0.8884, "step": 3054 }, { "epoch": 1.383605072463768, "grad_norm": 0.5965423775588453, "learning_rate": 7.262766152439326e-05, "loss": 0.8886, "step": 3055 }, { "epoch": 1.3840579710144927, "grad_norm": 0.6625409818306046, "learning_rate": 7.262034358604398e-05, "loss": 0.8831, "step": 3056 }, { "epoch": 1.3845108695652173, "grad_norm": 0.7356386080324481, "learning_rate": 7.261302238658727e-05, "loss": 0.9005, "step": 3057 }, { "epoch": 1.384963768115942, "grad_norm": 0.7852213745610195, "learning_rate": 7.260569792675504e-05, "loss": 0.8881, "step": 3058 }, { "epoch": 1.3854166666666667, "grad_norm": 0.7667996949494054, "learning_rate": 7.259837020727953e-05, "loss": 0.9061, "step": 3059 }, { "epoch": 1.3858695652173914, "grad_norm": 0.6717744573166633, "learning_rate": 7.259103922889328e-05, "loss": 0.8905, "step": 3060 }, { "epoch": 1.386322463768116, "grad_norm": 0.5195394302563745, "learning_rate": 7.258370499232918e-05, "loss": 0.8812, "step": 3061 }, { "epoch": 1.3867753623188406, "grad_norm": 0.4360083268641877, "learning_rate": 7.257636749832048e-05, "loss": 0.8624, "step": 3062 }, { "epoch": 1.3872282608695652, "grad_norm": 0.39758844857065173, "learning_rate": 7.256902674760069e-05, "loss": 0.9056, "step": 3063 }, { "epoch": 1.3876811594202898, "grad_norm": 0.3526115482405765, "learning_rate": 7.256168274090369e-05, "loss": 0.8877, "step": 3064 }, { "epoch": 1.3881340579710144, "grad_norm": 0.344070120079196, "learning_rate": 7.255433547896367e-05, "loss": 0.8977, "step": 3065 }, { "epoch": 1.3885869565217392, "grad_norm": 0.3523290285409416, "learning_rate": 7.254698496251513e-05, "loss": 0.8854, "step": 3066 }, { "epoch": 1.3890398550724639, "grad_norm": 0.32554290729860985, "learning_rate": 7.253963119229295e-05, "loss": 0.9004, "step": 3067 }, { "epoch": 1.3894927536231885, "grad_norm": 0.29598933450366527, "learning_rate": 7.253227416903226e-05, "loss": 0.8937, "step": 3068 }, { "epoch": 1.389945652173913, "grad_norm": 0.31601534379176416, "learning_rate": 7.252491389346858e-05, "loss": 0.8708, "step": 3069 }, { "epoch": 1.3903985507246377, "grad_norm": 0.3730344556530963, "learning_rate": 7.251755036633772e-05, "loss": 0.8738, "step": 3070 }, { "epoch": 1.3908514492753623, "grad_norm": 0.34561338240168465, "learning_rate": 7.25101835883758e-05, "loss": 0.9098, "step": 3071 }, { "epoch": 1.391304347826087, "grad_norm": 0.3229718219198772, "learning_rate": 7.250281356031932e-05, "loss": 0.8976, "step": 3072 }, { "epoch": 1.3917572463768115, "grad_norm": 0.2830875569901027, "learning_rate": 7.249544028290508e-05, "loss": 0.8904, "step": 3073 }, { "epoch": 1.3922101449275361, "grad_norm": 0.3558786005686538, "learning_rate": 7.248806375687016e-05, "loss": 0.9005, "step": 3074 }, { "epoch": 1.3926630434782608, "grad_norm": 0.40603621168169285, "learning_rate": 7.248068398295203e-05, "loss": 0.9041, "step": 3075 }, { "epoch": 1.3931159420289856, "grad_norm": 0.4188494786944694, "learning_rate": 7.247330096188846e-05, "loss": 0.933, "step": 3076 }, { "epoch": 1.3935688405797102, "grad_norm": 0.46937987160675415, "learning_rate": 7.246591469441754e-05, "loss": 0.8786, "step": 3077 }, { "epoch": 1.3940217391304348, "grad_norm": 0.5899411250763644, "learning_rate": 7.245852518127766e-05, "loss": 0.8964, "step": 3078 }, { "epoch": 1.3944746376811594, "grad_norm": 0.7022686007097576, "learning_rate": 7.24511324232076e-05, "loss": 0.8826, "step": 3079 }, { "epoch": 1.394927536231884, "grad_norm": 0.7581581119590639, "learning_rate": 7.244373642094641e-05, "loss": 0.9074, "step": 3080 }, { "epoch": 1.3953804347826086, "grad_norm": 0.7940998101565234, "learning_rate": 7.243633717523348e-05, "loss": 0.8832, "step": 3081 }, { "epoch": 1.3958333333333333, "grad_norm": 0.8890901089681339, "learning_rate": 7.242893468680849e-05, "loss": 0.8921, "step": 3082 }, { "epoch": 1.396286231884058, "grad_norm": 0.8443012390555873, "learning_rate": 7.242152895641154e-05, "loss": 0.8952, "step": 3083 }, { "epoch": 1.3967391304347827, "grad_norm": 0.7659027322680546, "learning_rate": 7.241411998478296e-05, "loss": 0.8947, "step": 3084 }, { "epoch": 1.3971920289855073, "grad_norm": 0.666759007821645, "learning_rate": 7.240670777266345e-05, "loss": 0.8872, "step": 3085 }, { "epoch": 1.397644927536232, "grad_norm": 0.5020635706812135, "learning_rate": 7.239929232079399e-05, "loss": 0.8875, "step": 3086 }, { "epoch": 1.3980978260869565, "grad_norm": 0.32243450517359157, "learning_rate": 7.239187362991595e-05, "loss": 0.9113, "step": 3087 }, { "epoch": 1.3985507246376812, "grad_norm": 0.34565358203207336, "learning_rate": 7.238445170077096e-05, "loss": 0.876, "step": 3088 }, { "epoch": 1.3990036231884058, "grad_norm": 0.3261256173510322, "learning_rate": 7.237702653410104e-05, "loss": 0.9018, "step": 3089 }, { "epoch": 1.3994565217391304, "grad_norm": 0.449798125297462, "learning_rate": 7.236959813064846e-05, "loss": 0.8801, "step": 3090 }, { "epoch": 1.399909420289855, "grad_norm": 0.5820221928535245, "learning_rate": 7.236216649115586e-05, "loss": 0.8773, "step": 3091 }, { "epoch": 1.4003623188405796, "grad_norm": 0.6842329594696188, "learning_rate": 7.235473161636618e-05, "loss": 0.8999, "step": 3092 }, { "epoch": 1.4008152173913042, "grad_norm": 0.7899472561148081, "learning_rate": 7.234729350702273e-05, "loss": 0.8827, "step": 3093 }, { "epoch": 1.401268115942029, "grad_norm": 0.7907827845708147, "learning_rate": 7.233985216386908e-05, "loss": 0.8846, "step": 3094 }, { "epoch": 1.4017210144927537, "grad_norm": 0.7416268382085259, "learning_rate": 7.233240758764916e-05, "loss": 0.8986, "step": 3095 }, { "epoch": 1.4021739130434783, "grad_norm": 0.5984847555190689, "learning_rate": 7.232495977910722e-05, "loss": 0.893, "step": 3096 }, { "epoch": 1.4026268115942029, "grad_norm": 0.416243613884527, "learning_rate": 7.231750873898783e-05, "loss": 0.8975, "step": 3097 }, { "epoch": 1.4030797101449275, "grad_norm": 0.2781073385866679, "learning_rate": 7.231005446803588e-05, "loss": 0.9009, "step": 3098 }, { "epoch": 1.403532608695652, "grad_norm": 0.27040429053854714, "learning_rate": 7.230259696699659e-05, "loss": 0.8998, "step": 3099 }, { "epoch": 1.403985507246377, "grad_norm": 0.3271945515091191, "learning_rate": 7.229513623661549e-05, "loss": 0.901, "step": 3100 }, { "epoch": 1.4044384057971016, "grad_norm": 0.40679209390780974, "learning_rate": 7.228767227763843e-05, "loss": 0.8859, "step": 3101 }, { "epoch": 1.4048913043478262, "grad_norm": 0.4746056649579459, "learning_rate": 7.228020509081162e-05, "loss": 0.8725, "step": 3102 }, { "epoch": 1.4053442028985508, "grad_norm": 0.6164341792626948, "learning_rate": 7.227273467688156e-05, "loss": 0.9082, "step": 3103 }, { "epoch": 1.4057971014492754, "grad_norm": 0.746725859325264, "learning_rate": 7.226526103659508e-05, "loss": 0.8607, "step": 3104 }, { "epoch": 1.40625, "grad_norm": 0.7218115274379501, "learning_rate": 7.225778417069932e-05, "loss": 0.8899, "step": 3105 }, { "epoch": 1.4067028985507246, "grad_norm": 0.6351480394690473, "learning_rate": 7.225030407994175e-05, "loss": 0.9132, "step": 3106 }, { "epoch": 1.4071557971014492, "grad_norm": 0.5296611005558516, "learning_rate": 7.224282076507017e-05, "loss": 0.8866, "step": 3107 }, { "epoch": 1.4076086956521738, "grad_norm": 0.45270935365251114, "learning_rate": 7.223533422683272e-05, "loss": 0.8767, "step": 3108 }, { "epoch": 1.4080615942028984, "grad_norm": 0.4532449464108838, "learning_rate": 7.222784446597783e-05, "loss": 0.8981, "step": 3109 }, { "epoch": 1.408514492753623, "grad_norm": 0.3863291918230121, "learning_rate": 7.222035148325424e-05, "loss": 0.8807, "step": 3110 }, { "epoch": 1.408967391304348, "grad_norm": 0.2996495641060439, "learning_rate": 7.221285527941106e-05, "loss": 0.902, "step": 3111 }, { "epoch": 1.4094202898550725, "grad_norm": 0.3047014775815713, "learning_rate": 7.22053558551977e-05, "loss": 0.8991, "step": 3112 }, { "epoch": 1.4098731884057971, "grad_norm": 0.27908037986483464, "learning_rate": 7.219785321136388e-05, "loss": 0.9047, "step": 3113 }, { "epoch": 1.4103260869565217, "grad_norm": 0.2454594217452982, "learning_rate": 7.219034734865965e-05, "loss": 0.8747, "step": 3114 }, { "epoch": 1.4107789855072463, "grad_norm": 0.2314077539914407, "learning_rate": 7.218283826783539e-05, "loss": 0.8848, "step": 3115 }, { "epoch": 1.411231884057971, "grad_norm": 0.2626994252701673, "learning_rate": 7.217532596964179e-05, "loss": 0.8883, "step": 3116 }, { "epoch": 1.4116847826086958, "grad_norm": 0.24235173919277161, "learning_rate": 7.216781045482986e-05, "loss": 0.8766, "step": 3117 }, { "epoch": 1.4121376811594204, "grad_norm": 0.3123964887715365, "learning_rate": 7.216029172415095e-05, "loss": 0.8995, "step": 3118 }, { "epoch": 1.412590579710145, "grad_norm": 0.32904054201281957, "learning_rate": 7.215276977835672e-05, "loss": 0.8883, "step": 3119 }, { "epoch": 1.4130434782608696, "grad_norm": 0.34058817152388543, "learning_rate": 7.214524461819913e-05, "loss": 0.9181, "step": 3120 }, { "epoch": 1.4134963768115942, "grad_norm": 0.40332163043128155, "learning_rate": 7.21377162444305e-05, "loss": 0.8781, "step": 3121 }, { "epoch": 1.4139492753623188, "grad_norm": 0.5275760518544874, "learning_rate": 7.213018465780347e-05, "loss": 0.9106, "step": 3122 }, { "epoch": 1.4144021739130435, "grad_norm": 0.6823763457568416, "learning_rate": 7.212264985907095e-05, "loss": 0.8945, "step": 3123 }, { "epoch": 1.414855072463768, "grad_norm": 0.8795115800115716, "learning_rate": 7.211511184898622e-05, "loss": 0.908, "step": 3124 }, { "epoch": 1.4153079710144927, "grad_norm": 0.9989877041758456, "learning_rate": 7.210757062830286e-05, "loss": 0.9024, "step": 3125 }, { "epoch": 1.4157608695652173, "grad_norm": 1.0039151834077225, "learning_rate": 7.21000261977748e-05, "loss": 0.8878, "step": 3126 }, { "epoch": 1.416213768115942, "grad_norm": 0.8870993352988904, "learning_rate": 7.209247855815625e-05, "loss": 0.8887, "step": 3127 }, { "epoch": 1.4166666666666667, "grad_norm": 0.6345638879240386, "learning_rate": 7.208492771020176e-05, "loss": 0.9077, "step": 3128 }, { "epoch": 1.4171195652173914, "grad_norm": 0.3258714123909608, "learning_rate": 7.20773736546662e-05, "loss": 0.8794, "step": 3129 }, { "epoch": 1.417572463768116, "grad_norm": 0.33924564827579134, "learning_rate": 7.206981639230478e-05, "loss": 0.9004, "step": 3130 }, { "epoch": 1.4180253623188406, "grad_norm": 0.5628396636445495, "learning_rate": 7.206225592387298e-05, "loss": 0.8862, "step": 3131 }, { "epoch": 1.4184782608695652, "grad_norm": 0.7424658106446033, "learning_rate": 7.205469225012666e-05, "loss": 0.9077, "step": 3132 }, { "epoch": 1.4189311594202898, "grad_norm": 0.8805026339826061, "learning_rate": 7.204712537182196e-05, "loss": 0.9039, "step": 3133 }, { "epoch": 1.4193840579710144, "grad_norm": 0.9256304660646014, "learning_rate": 7.203955528971534e-05, "loss": 0.8744, "step": 3134 }, { "epoch": 1.4198369565217392, "grad_norm": 0.7569592973481042, "learning_rate": 7.203198200456362e-05, "loss": 0.8937, "step": 3135 }, { "epoch": 1.4202898550724639, "grad_norm": 0.5378268077323882, "learning_rate": 7.20244055171239e-05, "loss": 0.8789, "step": 3136 }, { "epoch": 1.4207427536231885, "grad_norm": 0.3989499482821305, "learning_rate": 7.20168258281536e-05, "loss": 0.8993, "step": 3137 }, { "epoch": 1.421195652173913, "grad_norm": 0.3315044044774246, "learning_rate": 7.20092429384105e-05, "loss": 0.8923, "step": 3138 }, { "epoch": 1.4216485507246377, "grad_norm": 0.3965647522050676, "learning_rate": 7.200165684865266e-05, "loss": 0.8925, "step": 3139 }, { "epoch": 1.4221014492753623, "grad_norm": 0.5017933011602898, "learning_rate": 7.199406755963847e-05, "loss": 0.905, "step": 3140 }, { "epoch": 1.422554347826087, "grad_norm": 0.5956854536823823, "learning_rate": 7.198647507212664e-05, "loss": 0.8797, "step": 3141 }, { "epoch": 1.4230072463768115, "grad_norm": 0.6488114531447674, "learning_rate": 7.197887938687622e-05, "loss": 0.8924, "step": 3142 }, { "epoch": 1.4234601449275361, "grad_norm": 0.6763793364422596, "learning_rate": 7.197128050464655e-05, "loss": 0.881, "step": 3143 }, { "epoch": 1.4239130434782608, "grad_norm": 0.6661116759133398, "learning_rate": 7.196367842619732e-05, "loss": 0.9035, "step": 3144 }, { "epoch": 1.4243659420289856, "grad_norm": 0.6283094123905043, "learning_rate": 7.19560731522885e-05, "loss": 0.8912, "step": 3145 }, { "epoch": 1.4248188405797102, "grad_norm": 0.5348808965291191, "learning_rate": 7.194846468368038e-05, "loss": 0.8991, "step": 3146 }, { "epoch": 1.4252717391304348, "grad_norm": 0.47292509496448737, "learning_rate": 7.194085302113365e-05, "loss": 0.8901, "step": 3147 }, { "epoch": 1.4257246376811594, "grad_norm": 0.36184919070078453, "learning_rate": 7.193323816540922e-05, "loss": 0.894, "step": 3148 }, { "epoch": 1.426177536231884, "grad_norm": 0.3268608353187552, "learning_rate": 7.192562011726835e-05, "loss": 0.8904, "step": 3149 }, { "epoch": 1.4266304347826086, "grad_norm": 0.2741259385255917, "learning_rate": 7.191799887747267e-05, "loss": 0.8698, "step": 3150 }, { "epoch": 1.4270833333333333, "grad_norm": 0.34794638043840287, "learning_rate": 7.191037444678407e-05, "loss": 0.9069, "step": 3151 }, { "epoch": 1.427536231884058, "grad_norm": 0.5012755989593098, "learning_rate": 7.190274682596475e-05, "loss": 0.8806, "step": 3152 }, { "epoch": 1.4279891304347827, "grad_norm": 0.6429207070073101, "learning_rate": 7.189511601577728e-05, "loss": 0.8878, "step": 3153 }, { "epoch": 1.4284420289855073, "grad_norm": 0.6974352989616319, "learning_rate": 7.188748201698452e-05, "loss": 0.8859, "step": 3154 }, { "epoch": 1.428894927536232, "grad_norm": 0.6650741660180434, "learning_rate": 7.187984483034966e-05, "loss": 0.8883, "step": 3155 }, { "epoch": 1.4293478260869565, "grad_norm": 0.6208071489642172, "learning_rate": 7.187220445663618e-05, "loss": 0.8803, "step": 3156 }, { "epoch": 1.4298007246376812, "grad_norm": 0.5656649421553878, "learning_rate": 7.186456089660793e-05, "loss": 0.8869, "step": 3157 }, { "epoch": 1.4302536231884058, "grad_norm": 0.5616661435827365, "learning_rate": 7.185691415102903e-05, "loss": 0.8963, "step": 3158 }, { "epoch": 1.4307065217391304, "grad_norm": 0.6606578596776993, "learning_rate": 7.184926422066394e-05, "loss": 0.8949, "step": 3159 }, { "epoch": 1.431159420289855, "grad_norm": 0.7492840535911616, "learning_rate": 7.184161110627743e-05, "loss": 0.9045, "step": 3160 }, { "epoch": 1.4316123188405796, "grad_norm": 0.7495643437507751, "learning_rate": 7.183395480863461e-05, "loss": 0.8806, "step": 3161 }, { "epoch": 1.4320652173913042, "grad_norm": 0.7056898394794312, "learning_rate": 7.182629532850087e-05, "loss": 0.8937, "step": 3162 }, { "epoch": 1.432518115942029, "grad_norm": 0.6540363954810178, "learning_rate": 7.181863266664197e-05, "loss": 0.892, "step": 3163 }, { "epoch": 1.4329710144927537, "grad_norm": 0.7387863537869959, "learning_rate": 7.181096682382393e-05, "loss": 0.8774, "step": 3164 }, { "epoch": 1.4334239130434783, "grad_norm": 0.7907955015057659, "learning_rate": 7.180329780081313e-05, "loss": 0.8935, "step": 3165 }, { "epoch": 1.4338768115942029, "grad_norm": 0.7239337850230673, "learning_rate": 7.179562559837626e-05, "loss": 0.8932, "step": 3166 }, { "epoch": 1.4343297101449275, "grad_norm": 0.5361393662034681, "learning_rate": 7.17879502172803e-05, "loss": 0.8948, "step": 3167 }, { "epoch": 1.434782608695652, "grad_norm": 0.417633335796596, "learning_rate": 7.178027165829259e-05, "loss": 0.9132, "step": 3168 }, { "epoch": 1.435235507246377, "grad_norm": 0.3113444165695234, "learning_rate": 7.177258992218077e-05, "loss": 0.8877, "step": 3169 }, { "epoch": 1.4356884057971016, "grad_norm": 0.2906799393232747, "learning_rate": 7.176490500971277e-05, "loss": 0.8727, "step": 3170 }, { "epoch": 1.4361413043478262, "grad_norm": 0.3863643352464737, "learning_rate": 7.17572169216569e-05, "loss": 0.878, "step": 3171 }, { "epoch": 1.4365942028985508, "grad_norm": 0.5668116607793534, "learning_rate": 7.174952565878171e-05, "loss": 0.871, "step": 3172 }, { "epoch": 1.4370471014492754, "grad_norm": 0.6378501600190273, "learning_rate": 7.174183122185615e-05, "loss": 0.8853, "step": 3173 }, { "epoch": 1.4375, "grad_norm": 0.5591730321166928, "learning_rate": 7.173413361164941e-05, "loss": 0.8949, "step": 3174 }, { "epoch": 1.4379528985507246, "grad_norm": 0.46580456455440644, "learning_rate": 7.172643282893106e-05, "loss": 0.8948, "step": 3175 }, { "epoch": 1.4384057971014492, "grad_norm": 0.39133893070628045, "learning_rate": 7.171872887447093e-05, "loss": 0.863, "step": 3176 }, { "epoch": 1.4388586956521738, "grad_norm": 0.4773488086959425, "learning_rate": 7.171102174903922e-05, "loss": 0.8854, "step": 3177 }, { "epoch": 1.4393115942028984, "grad_norm": 0.5647431388385573, "learning_rate": 7.17033114534064e-05, "loss": 0.8862, "step": 3178 }, { "epoch": 1.439764492753623, "grad_norm": 0.6931295981171761, "learning_rate": 7.169559798834332e-05, "loss": 0.887, "step": 3179 }, { "epoch": 1.440217391304348, "grad_norm": 0.7159524626494298, "learning_rate": 7.168788135462106e-05, "loss": 0.8851, "step": 3180 }, { "epoch": 1.4406702898550725, "grad_norm": 0.5864519799157542, "learning_rate": 7.16801615530111e-05, "loss": 0.8941, "step": 3181 }, { "epoch": 1.4411231884057971, "grad_norm": 0.48995130681699534, "learning_rate": 7.16724385842852e-05, "loss": 0.8811, "step": 3182 }, { "epoch": 1.4415760869565217, "grad_norm": 0.3314895737151583, "learning_rate": 7.16647124492154e-05, "loss": 0.8991, "step": 3183 }, { "epoch": 1.4420289855072463, "grad_norm": 0.3417053961642744, "learning_rate": 7.165698314857413e-05, "loss": 0.903, "step": 3184 }, { "epoch": 1.442481884057971, "grad_norm": 0.4641663372021114, "learning_rate": 7.164925068313409e-05, "loss": 0.8875, "step": 3185 }, { "epoch": 1.4429347826086958, "grad_norm": 0.5165540910054369, "learning_rate": 7.164151505366831e-05, "loss": 0.8775, "step": 3186 }, { "epoch": 1.4433876811594204, "grad_norm": 0.5430210788854697, "learning_rate": 7.163377626095013e-05, "loss": 0.908, "step": 3187 }, { "epoch": 1.443840579710145, "grad_norm": 0.7081710552630922, "learning_rate": 7.16260343057532e-05, "loss": 0.9011, "step": 3188 }, { "epoch": 1.4442934782608696, "grad_norm": 0.8894727594496843, "learning_rate": 7.161828918885151e-05, "loss": 0.9059, "step": 3189 }, { "epoch": 1.4447463768115942, "grad_norm": 0.8255904955127419, "learning_rate": 7.161054091101936e-05, "loss": 0.882, "step": 3190 }, { "epoch": 1.4451992753623188, "grad_norm": 0.7453667549091368, "learning_rate": 7.160278947303132e-05, "loss": 0.9266, "step": 3191 }, { "epoch": 1.4456521739130435, "grad_norm": 0.6901866980090706, "learning_rate": 7.159503487566236e-05, "loss": 0.9052, "step": 3192 }, { "epoch": 1.446105072463768, "grad_norm": 0.6697397188992659, "learning_rate": 7.158727711968768e-05, "loss": 0.8853, "step": 3193 }, { "epoch": 1.4465579710144927, "grad_norm": 0.6467811925594062, "learning_rate": 7.157951620588286e-05, "loss": 0.8916, "step": 3194 }, { "epoch": 1.4470108695652173, "grad_norm": 0.574735298187606, "learning_rate": 7.157175213502376e-05, "loss": 0.8861, "step": 3195 }, { "epoch": 1.447463768115942, "grad_norm": 0.5151594120504962, "learning_rate": 7.156398490788656e-05, "loss": 0.8806, "step": 3196 }, { "epoch": 1.4479166666666667, "grad_norm": 0.4343735978285113, "learning_rate": 7.155621452524779e-05, "loss": 0.8897, "step": 3197 }, { "epoch": 1.4483695652173914, "grad_norm": 0.3563884307535028, "learning_rate": 7.154844098788424e-05, "loss": 0.8994, "step": 3198 }, { "epoch": 1.448822463768116, "grad_norm": 0.36437795725981725, "learning_rate": 7.154066429657304e-05, "loss": 0.8899, "step": 3199 }, { "epoch": 1.4492753623188406, "grad_norm": 0.41138492906840146, "learning_rate": 7.153288445209166e-05, "loss": 0.8855, "step": 3200 }, { "epoch": 1.4497282608695652, "grad_norm": 0.47584145682097034, "learning_rate": 7.152510145521786e-05, "loss": 0.8863, "step": 3201 }, { "epoch": 1.4501811594202898, "grad_norm": 0.5444977927076304, "learning_rate": 7.151731530672971e-05, "loss": 0.8847, "step": 3202 }, { "epoch": 1.4506340579710144, "grad_norm": 0.5843214565106498, "learning_rate": 7.15095260074056e-05, "loss": 0.8953, "step": 3203 }, { "epoch": 1.4510869565217392, "grad_norm": 0.6286583971799171, "learning_rate": 7.150173355802425e-05, "loss": 0.9168, "step": 3204 }, { "epoch": 1.4515398550724639, "grad_norm": 0.6071862618573268, "learning_rate": 7.149393795936468e-05, "loss": 0.8595, "step": 3205 }, { "epoch": 1.4519927536231885, "grad_norm": 0.5395512974124431, "learning_rate": 7.148613921220622e-05, "loss": 0.8747, "step": 3206 }, { "epoch": 1.452445652173913, "grad_norm": 0.48077159573124506, "learning_rate": 7.147833731732852e-05, "loss": 0.8832, "step": 3207 }, { "epoch": 1.4528985507246377, "grad_norm": 0.5059125576927522, "learning_rate": 7.147053227551157e-05, "loss": 0.8865, "step": 3208 }, { "epoch": 1.4533514492753623, "grad_norm": 0.46734471883865775, "learning_rate": 7.146272408753563e-05, "loss": 0.9076, "step": 3209 }, { "epoch": 1.453804347826087, "grad_norm": 0.3817233808661143, "learning_rate": 7.14549127541813e-05, "loss": 0.8858, "step": 3210 }, { "epoch": 1.4542572463768115, "grad_norm": 0.40778237036315207, "learning_rate": 7.144709827622952e-05, "loss": 0.9039, "step": 3211 }, { "epoch": 1.4547101449275361, "grad_norm": 0.4825304574496048, "learning_rate": 7.143928065446148e-05, "loss": 0.8884, "step": 3212 }, { "epoch": 1.4551630434782608, "grad_norm": 0.3786440197050443, "learning_rate": 7.143145988965872e-05, "loss": 0.8726, "step": 3213 }, { "epoch": 1.4556159420289856, "grad_norm": 0.3280763244003997, "learning_rate": 7.142363598260313e-05, "loss": 0.8712, "step": 3214 }, { "epoch": 1.4560688405797102, "grad_norm": 0.3731141109829373, "learning_rate": 7.141580893407683e-05, "loss": 0.8897, "step": 3215 }, { "epoch": 1.4565217391304348, "grad_norm": 0.3474368105804016, "learning_rate": 7.140797874486237e-05, "loss": 0.8858, "step": 3216 }, { "epoch": 1.4569746376811594, "grad_norm": 0.31386974600798156, "learning_rate": 7.140014541574246e-05, "loss": 0.9057, "step": 3217 }, { "epoch": 1.457427536231884, "grad_norm": 0.3451368935915898, "learning_rate": 7.139230894750028e-05, "loss": 0.8718, "step": 3218 }, { "epoch": 1.4578804347826086, "grad_norm": 0.3447795409470079, "learning_rate": 7.138446934091923e-05, "loss": 0.8616, "step": 3219 }, { "epoch": 1.4583333333333333, "grad_norm": 0.3401009378619367, "learning_rate": 7.137662659678303e-05, "loss": 0.8952, "step": 3220 }, { "epoch": 1.458786231884058, "grad_norm": 0.33004736018663644, "learning_rate": 7.136878071587576e-05, "loss": 0.8727, "step": 3221 }, { "epoch": 1.4592391304347827, "grad_norm": 0.316153623953174, "learning_rate": 7.136093169898178e-05, "loss": 0.8846, "step": 3222 }, { "epoch": 1.4596920289855073, "grad_norm": 0.345073671718519, "learning_rate": 7.135307954688575e-05, "loss": 0.8695, "step": 3223 }, { "epoch": 1.460144927536232, "grad_norm": 0.38208161137920504, "learning_rate": 7.134522426037268e-05, "loss": 0.8892, "step": 3224 }, { "epoch": 1.4605978260869565, "grad_norm": 0.438767022132904, "learning_rate": 7.133736584022787e-05, "loss": 0.8788, "step": 3225 }, { "epoch": 1.4610507246376812, "grad_norm": 0.4523706906735224, "learning_rate": 7.132950428723695e-05, "loss": 0.924, "step": 3226 }, { "epoch": 1.4615036231884058, "grad_norm": 0.5578097267931513, "learning_rate": 7.132163960218582e-05, "loss": 0.8802, "step": 3227 }, { "epoch": 1.4619565217391304, "grad_norm": 0.717806494791795, "learning_rate": 7.131377178586075e-05, "loss": 0.8762, "step": 3228 }, { "epoch": 1.462409420289855, "grad_norm": 0.7819339746822049, "learning_rate": 7.13059008390483e-05, "loss": 0.9001, "step": 3229 }, { "epoch": 1.4628623188405796, "grad_norm": 0.8417040621136092, "learning_rate": 7.129802676253534e-05, "loss": 0.8895, "step": 3230 }, { "epoch": 1.4633152173913042, "grad_norm": 0.95466291939544, "learning_rate": 7.129014955710902e-05, "loss": 0.8802, "step": 3231 }, { "epoch": 1.463768115942029, "grad_norm": 1.025605769807455, "learning_rate": 7.128226922355691e-05, "loss": 0.9025, "step": 3232 }, { "epoch": 1.4642210144927537, "grad_norm": 0.8455880391202348, "learning_rate": 7.127438576266673e-05, "loss": 0.8974, "step": 3233 }, { "epoch": 1.4646739130434783, "grad_norm": 0.5747105439417753, "learning_rate": 7.126649917522667e-05, "loss": 0.9044, "step": 3234 }, { "epoch": 1.4651268115942029, "grad_norm": 0.3720790846415079, "learning_rate": 7.125860946202516e-05, "loss": 0.8817, "step": 3235 }, { "epoch": 1.4655797101449275, "grad_norm": 0.2932904374351766, "learning_rate": 7.12507166238509e-05, "loss": 0.8893, "step": 3236 }, { "epoch": 1.466032608695652, "grad_norm": 0.4140545956023074, "learning_rate": 7.124282066149298e-05, "loss": 0.8866, "step": 3237 }, { "epoch": 1.466485507246377, "grad_norm": 0.5528604853793853, "learning_rate": 7.123492157574077e-05, "loss": 0.8935, "step": 3238 }, { "epoch": 1.4669384057971016, "grad_norm": 0.6246103287184153, "learning_rate": 7.122701936738397e-05, "loss": 0.8926, "step": 3239 }, { "epoch": 1.4673913043478262, "grad_norm": 0.6427118039400099, "learning_rate": 7.121911403721254e-05, "loss": 0.8969, "step": 3240 }, { "epoch": 1.4678442028985508, "grad_norm": 0.6729179592438717, "learning_rate": 7.121120558601683e-05, "loss": 0.8789, "step": 3241 }, { "epoch": 1.4682971014492754, "grad_norm": 0.6560814734343599, "learning_rate": 7.120329401458742e-05, "loss": 0.8851, "step": 3242 }, { "epoch": 1.46875, "grad_norm": 0.6335694038597942, "learning_rate": 7.119537932371527e-05, "loss": 0.8718, "step": 3243 }, { "epoch": 1.4692028985507246, "grad_norm": 0.5848682658431905, "learning_rate": 7.118746151419162e-05, "loss": 0.8688, "step": 3244 }, { "epoch": 1.4696557971014492, "grad_norm": 0.5141479932017106, "learning_rate": 7.117954058680803e-05, "loss": 0.8761, "step": 3245 }, { "epoch": 1.4701086956521738, "grad_norm": 0.44324558950497656, "learning_rate": 7.117161654235634e-05, "loss": 0.8983, "step": 3246 }, { "epoch": 1.4705615942028984, "grad_norm": 0.3524388771879365, "learning_rate": 7.116368938162875e-05, "loss": 0.8728, "step": 3247 }, { "epoch": 1.471014492753623, "grad_norm": 0.2975878258881075, "learning_rate": 7.115575910541776e-05, "loss": 0.883, "step": 3248 }, { "epoch": 1.471467391304348, "grad_norm": 0.3705547817889322, "learning_rate": 7.114782571451616e-05, "loss": 0.9058, "step": 3249 }, { "epoch": 1.4719202898550725, "grad_norm": 0.4863702802256972, "learning_rate": 7.113988920971706e-05, "loss": 0.8935, "step": 3250 }, { "epoch": 1.4723731884057971, "grad_norm": 0.6216987420000235, "learning_rate": 7.11319495918139e-05, "loss": 0.9066, "step": 3251 }, { "epoch": 1.4728260869565217, "grad_norm": 0.7647429874818142, "learning_rate": 7.112400686160039e-05, "loss": 0.8851, "step": 3252 }, { "epoch": 1.4732789855072463, "grad_norm": 0.820171025333598, "learning_rate": 7.11160610198706e-05, "loss": 0.8867, "step": 3253 }, { "epoch": 1.473731884057971, "grad_norm": 0.8465399930423448, "learning_rate": 7.110811206741889e-05, "loss": 0.8887, "step": 3254 }, { "epoch": 1.4741847826086958, "grad_norm": 0.8467708211198953, "learning_rate": 7.11001600050399e-05, "loss": 0.878, "step": 3255 }, { "epoch": 1.4746376811594204, "grad_norm": 0.7925820691939559, "learning_rate": 7.109220483352867e-05, "loss": 0.8893, "step": 3256 }, { "epoch": 1.475090579710145, "grad_norm": 0.6894647508177223, "learning_rate": 7.108424655368043e-05, "loss": 0.8948, "step": 3257 }, { "epoch": 1.4755434782608696, "grad_norm": 0.5898780323262315, "learning_rate": 7.107628516629079e-05, "loss": 0.879, "step": 3258 }, { "epoch": 1.4759963768115942, "grad_norm": 0.40395628023403507, "learning_rate": 7.106832067215569e-05, "loss": 0.889, "step": 3259 }, { "epoch": 1.4764492753623188, "grad_norm": 0.2481934000352548, "learning_rate": 7.106035307207135e-05, "loss": 0.9031, "step": 3260 }, { "epoch": 1.4769021739130435, "grad_norm": 0.3723445360717585, "learning_rate": 7.105238236683429e-05, "loss": 0.8752, "step": 3261 }, { "epoch": 1.477355072463768, "grad_norm": 0.41105208654345815, "learning_rate": 7.104440855724136e-05, "loss": 0.8861, "step": 3262 }, { "epoch": 1.4778079710144927, "grad_norm": 0.4135926480737131, "learning_rate": 7.103643164408971e-05, "loss": 0.8851, "step": 3263 }, { "epoch": 1.4782608695652173, "grad_norm": 0.44565881767150517, "learning_rate": 7.10284516281768e-05, "loss": 0.916, "step": 3264 }, { "epoch": 1.478713768115942, "grad_norm": 0.4929682786944962, "learning_rate": 7.102046851030041e-05, "loss": 0.8819, "step": 3265 }, { "epoch": 1.4791666666666667, "grad_norm": 0.5447914317529897, "learning_rate": 7.101248229125864e-05, "loss": 0.901, "step": 3266 }, { "epoch": 1.4796195652173914, "grad_norm": 0.5781626266944695, "learning_rate": 7.100449297184986e-05, "loss": 0.8807, "step": 3267 }, { "epoch": 1.480072463768116, "grad_norm": 0.5305685245115201, "learning_rate": 7.09965005528728e-05, "loss": 0.8884, "step": 3268 }, { "epoch": 1.4805253623188406, "grad_norm": 0.5024274354461111, "learning_rate": 7.098850503512645e-05, "loss": 0.8984, "step": 3269 }, { "epoch": 1.4809782608695652, "grad_norm": 0.5212672857189518, "learning_rate": 7.098050641941015e-05, "loss": 0.8808, "step": 3270 }, { "epoch": 1.4814311594202898, "grad_norm": 0.5218912200463502, "learning_rate": 7.097250470652353e-05, "loss": 0.8784, "step": 3271 }, { "epoch": 1.4818840579710144, "grad_norm": 0.5657635597976381, "learning_rate": 7.096449989726653e-05, "loss": 0.9047, "step": 3272 }, { "epoch": 1.4823369565217392, "grad_norm": 0.7438069421148162, "learning_rate": 7.095649199243943e-05, "loss": 0.9056, "step": 3273 }, { "epoch": 1.4827898550724639, "grad_norm": 0.9035841064218627, "learning_rate": 7.094848099284275e-05, "loss": 0.8665, "step": 3274 }, { "epoch": 1.4832427536231885, "grad_norm": 0.987241677279701, "learning_rate": 7.094046689927739e-05, "loss": 0.8795, "step": 3275 }, { "epoch": 1.483695652173913, "grad_norm": 0.9860936901465714, "learning_rate": 7.093244971254452e-05, "loss": 0.9072, "step": 3276 }, { "epoch": 1.4841485507246377, "grad_norm": 0.9408553430661575, "learning_rate": 7.092442943344563e-05, "loss": 0.9001, "step": 3277 }, { "epoch": 1.4846014492753623, "grad_norm": 0.8525849764492125, "learning_rate": 7.091640606278253e-05, "loss": 0.9009, "step": 3278 }, { "epoch": 1.485054347826087, "grad_norm": 0.7637596248642889, "learning_rate": 7.090837960135734e-05, "loss": 0.9172, "step": 3279 }, { "epoch": 1.4855072463768115, "grad_norm": 0.6937588005974165, "learning_rate": 7.090035004997245e-05, "loss": 0.8602, "step": 3280 }, { "epoch": 1.4859601449275361, "grad_norm": 0.6962553058837092, "learning_rate": 7.089231740943059e-05, "loss": 0.9005, "step": 3281 }, { "epoch": 1.4864130434782608, "grad_norm": 0.8354519174054056, "learning_rate": 7.088428168053481e-05, "loss": 0.8772, "step": 3282 }, { "epoch": 1.4868659420289856, "grad_norm": 0.9027919752516833, "learning_rate": 7.087624286408845e-05, "loss": 0.8962, "step": 3283 }, { "epoch": 1.4873188405797102, "grad_norm": 0.7576026539538896, "learning_rate": 7.086820096089518e-05, "loss": 0.886, "step": 3284 }, { "epoch": 1.4877717391304348, "grad_norm": 0.7255232035253862, "learning_rate": 7.086015597175893e-05, "loss": 0.8729, "step": 3285 }, { "epoch": 1.4882246376811594, "grad_norm": 0.7683147927945462, "learning_rate": 7.085210789748399e-05, "loss": 0.908, "step": 3286 }, { "epoch": 1.488677536231884, "grad_norm": 0.7787613256143958, "learning_rate": 7.084405673887494e-05, "loss": 0.8804, "step": 3287 }, { "epoch": 1.4891304347826086, "grad_norm": 0.6532326608529695, "learning_rate": 7.083600249673665e-05, "loss": 0.9008, "step": 3288 }, { "epoch": 1.4895833333333333, "grad_norm": 0.5202815120762776, "learning_rate": 7.082794517187432e-05, "loss": 0.8799, "step": 3289 }, { "epoch": 1.490036231884058, "grad_norm": 0.4169912141306032, "learning_rate": 7.081988476509348e-05, "loss": 0.8943, "step": 3290 }, { "epoch": 1.4904891304347827, "grad_norm": 0.33569281216968216, "learning_rate": 7.081182127719992e-05, "loss": 0.8773, "step": 3291 }, { "epoch": 1.4909420289855073, "grad_norm": 0.4187937009892978, "learning_rate": 7.080375470899974e-05, "loss": 0.8844, "step": 3292 }, { "epoch": 1.491394927536232, "grad_norm": 0.5284759562766796, "learning_rate": 7.079568506129941e-05, "loss": 0.9064, "step": 3293 }, { "epoch": 1.4918478260869565, "grad_norm": 0.6697302907384538, "learning_rate": 7.078761233490564e-05, "loss": 0.913, "step": 3294 }, { "epoch": 1.4923007246376812, "grad_norm": 0.793174224109817, "learning_rate": 7.077953653062548e-05, "loss": 0.8994, "step": 3295 }, { "epoch": 1.4927536231884058, "grad_norm": 0.7739806194641289, "learning_rate": 7.077145764926628e-05, "loss": 0.8904, "step": 3296 }, { "epoch": 1.4932065217391304, "grad_norm": 0.6019145002657037, "learning_rate": 7.076337569163568e-05, "loss": 0.883, "step": 3297 }, { "epoch": 1.493659420289855, "grad_norm": 0.45098296330060567, "learning_rate": 7.075529065854168e-05, "loss": 0.8899, "step": 3298 }, { "epoch": 1.4941123188405796, "grad_norm": 0.3960212003590651, "learning_rate": 7.074720255079252e-05, "loss": 0.8679, "step": 3299 }, { "epoch": 1.4945652173913042, "grad_norm": 0.3655017710878583, "learning_rate": 7.073911136919683e-05, "loss": 0.9168, "step": 3300 }, { "epoch": 1.495018115942029, "grad_norm": 0.4759096613807915, "learning_rate": 7.073101711456343e-05, "loss": 0.8887, "step": 3301 }, { "epoch": 1.4954710144927537, "grad_norm": 0.6315018737283784, "learning_rate": 7.072291978770157e-05, "loss": 0.8875, "step": 3302 }, { "epoch": 1.4959239130434783, "grad_norm": 0.6429226420097129, "learning_rate": 7.071481938942074e-05, "loss": 0.872, "step": 3303 }, { "epoch": 1.4963768115942029, "grad_norm": 0.5532372407269748, "learning_rate": 7.070671592053073e-05, "loss": 0.8666, "step": 3304 }, { "epoch": 1.4968297101449275, "grad_norm": 0.39142059053040135, "learning_rate": 7.069860938184168e-05, "loss": 0.8922, "step": 3305 }, { "epoch": 1.497282608695652, "grad_norm": 0.3356448469059604, "learning_rate": 7.069049977416399e-05, "loss": 0.9276, "step": 3306 }, { "epoch": 1.497735507246377, "grad_norm": 0.3319254550681614, "learning_rate": 7.06823870983084e-05, "loss": 0.8809, "step": 3307 }, { "epoch": 1.4981884057971016, "grad_norm": 0.3403591424292937, "learning_rate": 7.067427135508598e-05, "loss": 0.8964, "step": 3308 }, { "epoch": 1.4986413043478262, "grad_norm": 0.4387192297583605, "learning_rate": 7.066615254530803e-05, "loss": 0.8863, "step": 3309 }, { "epoch": 1.4990942028985508, "grad_norm": 0.41096463750664125, "learning_rate": 7.065803066978621e-05, "loss": 0.8819, "step": 3310 }, { "epoch": 1.4995471014492754, "grad_norm": 0.37183379002577843, "learning_rate": 7.06499057293325e-05, "loss": 0.8619, "step": 3311 }, { "epoch": 1.5, "grad_norm": 0.4069440169658705, "learning_rate": 7.064177772475912e-05, "loss": 0.8931, "step": 3312 }, { "epoch": 1.5004528985507246, "grad_norm": 0.38550386409018966, "learning_rate": 7.063364665687869e-05, "loss": 0.8849, "step": 3313 }, { "epoch": 1.5009057971014492, "grad_norm": 0.4048457093949384, "learning_rate": 7.062551252650405e-05, "loss": 0.9228, "step": 3314 }, { "epoch": 1.5013586956521738, "grad_norm": 0.4664665072762579, "learning_rate": 7.061737533444838e-05, "loss": 0.895, "step": 3315 }, { "epoch": 1.5018115942028984, "grad_norm": 0.46807787829388325, "learning_rate": 7.06092350815252e-05, "loss": 0.8949, "step": 3316 }, { "epoch": 1.502264492753623, "grad_norm": 0.48975372453500676, "learning_rate": 7.060109176854827e-05, "loss": 0.8861, "step": 3317 }, { "epoch": 1.5027173913043477, "grad_norm": 0.531949438300426, "learning_rate": 7.059294539633171e-05, "loss": 0.8927, "step": 3318 }, { "epoch": 1.5031702898550725, "grad_norm": 0.4925217679805419, "learning_rate": 7.058479596568993e-05, "loss": 0.8893, "step": 3319 }, { "epoch": 1.5036231884057971, "grad_norm": 0.4686495511040049, "learning_rate": 7.057664347743762e-05, "loss": 0.8895, "step": 3320 }, { "epoch": 1.5040760869565217, "grad_norm": 0.4608472909199961, "learning_rate": 7.056848793238981e-05, "loss": 0.9044, "step": 3321 }, { "epoch": 1.5045289855072463, "grad_norm": 0.45968810013096567, "learning_rate": 7.056032933136183e-05, "loss": 0.8654, "step": 3322 }, { "epoch": 1.5049818840579712, "grad_norm": 0.44479763892935187, "learning_rate": 7.055216767516928e-05, "loss": 0.9004, "step": 3323 }, { "epoch": 1.5054347826086958, "grad_norm": 0.4972291878721193, "learning_rate": 7.054400296462813e-05, "loss": 0.9008, "step": 3324 }, { "epoch": 1.5058876811594204, "grad_norm": 0.5558116499286458, "learning_rate": 7.053583520055459e-05, "loss": 0.8751, "step": 3325 }, { "epoch": 1.506340579710145, "grad_norm": 0.5572660176259708, "learning_rate": 7.052766438376522e-05, "loss": 0.9066, "step": 3326 }, { "epoch": 1.5067934782608696, "grad_norm": 0.5287016382881004, "learning_rate": 7.051949051507689e-05, "loss": 0.8806, "step": 3327 }, { "epoch": 1.5072463768115942, "grad_norm": 0.5453923620409072, "learning_rate": 7.051131359530671e-05, "loss": 0.903, "step": 3328 }, { "epoch": 1.5076992753623188, "grad_norm": 0.5246564106305326, "learning_rate": 7.050313362527216e-05, "loss": 0.8885, "step": 3329 }, { "epoch": 1.5081521739130435, "grad_norm": 0.4551281421639407, "learning_rate": 7.0494950605791e-05, "loss": 0.9002, "step": 3330 }, { "epoch": 1.508605072463768, "grad_norm": 0.4302887419662433, "learning_rate": 7.048676453768131e-05, "loss": 0.9125, "step": 3331 }, { "epoch": 1.5090579710144927, "grad_norm": 0.41596888325514964, "learning_rate": 7.047857542176145e-05, "loss": 0.8827, "step": 3332 }, { "epoch": 1.5095108695652173, "grad_norm": 0.41557187586881117, "learning_rate": 7.047038325885012e-05, "loss": 0.8932, "step": 3333 }, { "epoch": 1.509963768115942, "grad_norm": 0.3873891599948017, "learning_rate": 7.04621880497663e-05, "loss": 0.8721, "step": 3334 }, { "epoch": 1.5104166666666665, "grad_norm": 0.36662811908315307, "learning_rate": 7.045398979532925e-05, "loss": 0.9009, "step": 3335 }, { "epoch": 1.5108695652173914, "grad_norm": 0.4378252615806763, "learning_rate": 7.04457884963586e-05, "loss": 0.8841, "step": 3336 }, { "epoch": 1.511322463768116, "grad_norm": 0.47654717706392796, "learning_rate": 7.043758415367422e-05, "loss": 0.8881, "step": 3337 }, { "epoch": 1.5117753623188406, "grad_norm": 0.5203274609372227, "learning_rate": 7.042937676809631e-05, "loss": 0.8764, "step": 3338 }, { "epoch": 1.5122282608695652, "grad_norm": 0.5452756217136757, "learning_rate": 7.04211663404454e-05, "loss": 0.8907, "step": 3339 }, { "epoch": 1.5126811594202898, "grad_norm": 0.5987768427600002, "learning_rate": 7.041295287154228e-05, "loss": 0.8889, "step": 3340 }, { "epoch": 1.5131340579710146, "grad_norm": 0.6271573526493316, "learning_rate": 7.040473636220807e-05, "loss": 0.8744, "step": 3341 }, { "epoch": 1.5135869565217392, "grad_norm": 0.596401963922181, "learning_rate": 7.039651681326418e-05, "loss": 0.8874, "step": 3342 }, { "epoch": 1.5140398550724639, "grad_norm": 0.5006625948331096, "learning_rate": 7.038829422553234e-05, "loss": 0.8863, "step": 3343 }, { "epoch": 1.5144927536231885, "grad_norm": 0.5215996282030188, "learning_rate": 7.038006859983455e-05, "loss": 0.9008, "step": 3344 }, { "epoch": 1.514945652173913, "grad_norm": 0.5449323985656742, "learning_rate": 7.03718399369932e-05, "loss": 0.8861, "step": 3345 }, { "epoch": 1.5153985507246377, "grad_norm": 0.5121712959300164, "learning_rate": 7.036360823783085e-05, "loss": 0.8911, "step": 3346 }, { "epoch": 1.5158514492753623, "grad_norm": 0.3703584194705018, "learning_rate": 7.035537350317049e-05, "loss": 0.8881, "step": 3347 }, { "epoch": 1.516304347826087, "grad_norm": 0.3211553163168923, "learning_rate": 7.034713573383533e-05, "loss": 0.8879, "step": 3348 }, { "epoch": 1.5167572463768115, "grad_norm": 0.4417272446984872, "learning_rate": 7.033889493064893e-05, "loss": 0.9215, "step": 3349 }, { "epoch": 1.5172101449275361, "grad_norm": 0.4936028509440296, "learning_rate": 7.033065109443512e-05, "loss": 0.8812, "step": 3350 }, { "epoch": 1.5176630434782608, "grad_norm": 0.4661927031383706, "learning_rate": 7.032240422601805e-05, "loss": 0.8946, "step": 3351 }, { "epoch": 1.5181159420289854, "grad_norm": 0.5429742901281898, "learning_rate": 7.031415432622219e-05, "loss": 0.886, "step": 3352 }, { "epoch": 1.5185688405797102, "grad_norm": 0.6398344708478726, "learning_rate": 7.030590139587228e-05, "loss": 0.9004, "step": 3353 }, { "epoch": 1.5190217391304348, "grad_norm": 0.7444129121336636, "learning_rate": 7.029764543579339e-05, "loss": 0.8848, "step": 3354 }, { "epoch": 1.5194746376811594, "grad_norm": 0.82970856060134, "learning_rate": 7.028938644681088e-05, "loss": 0.8815, "step": 3355 }, { "epoch": 1.519927536231884, "grad_norm": 0.8465415879721574, "learning_rate": 7.028112442975038e-05, "loss": 0.8729, "step": 3356 }, { "epoch": 1.5203804347826086, "grad_norm": 0.7755904914671474, "learning_rate": 7.027285938543791e-05, "loss": 0.8685, "step": 3357 }, { "epoch": 1.5208333333333335, "grad_norm": 0.6047792556879085, "learning_rate": 7.026459131469972e-05, "loss": 0.888, "step": 3358 }, { "epoch": 1.521286231884058, "grad_norm": 0.39038803691837937, "learning_rate": 7.025632021836238e-05, "loss": 0.8911, "step": 3359 }, { "epoch": 1.5217391304347827, "grad_norm": 0.2802451488563246, "learning_rate": 7.024804609725275e-05, "loss": 0.898, "step": 3360 }, { "epoch": 1.5221920289855073, "grad_norm": 0.3605870812667753, "learning_rate": 7.023976895219802e-05, "loss": 0.8998, "step": 3361 }, { "epoch": 1.522644927536232, "grad_norm": 0.43831026403548795, "learning_rate": 7.023148878402566e-05, "loss": 0.8944, "step": 3362 }, { "epoch": 1.5230978260869565, "grad_norm": 0.5937132926281063, "learning_rate": 7.022320559356348e-05, "loss": 0.8824, "step": 3363 }, { "epoch": 1.5235507246376812, "grad_norm": 0.6682425806032631, "learning_rate": 7.021491938163953e-05, "loss": 0.9075, "step": 3364 }, { "epoch": 1.5240036231884058, "grad_norm": 0.6766027612137169, "learning_rate": 7.020663014908222e-05, "loss": 0.9003, "step": 3365 }, { "epoch": 1.5244565217391304, "grad_norm": 0.5935831096717897, "learning_rate": 7.019833789672023e-05, "loss": 0.8767, "step": 3366 }, { "epoch": 1.524909420289855, "grad_norm": 0.5436102463535987, "learning_rate": 7.019004262538254e-05, "loss": 0.879, "step": 3367 }, { "epoch": 1.5253623188405796, "grad_norm": 0.5391257345193337, "learning_rate": 7.018174433589844e-05, "loss": 0.8996, "step": 3368 }, { "epoch": 1.5258152173913042, "grad_norm": 0.5052168717278754, "learning_rate": 7.017344302909755e-05, "loss": 0.8797, "step": 3369 }, { "epoch": 1.5262681159420288, "grad_norm": 0.41083072427190476, "learning_rate": 7.016513870580973e-05, "loss": 0.8875, "step": 3370 }, { "epoch": 1.5267210144927537, "grad_norm": 0.35798472837860607, "learning_rate": 7.01568313668652e-05, "loss": 0.8951, "step": 3371 }, { "epoch": 1.5271739130434783, "grad_norm": 0.33607315530863785, "learning_rate": 7.014852101309445e-05, "loss": 0.8866, "step": 3372 }, { "epoch": 1.5276268115942029, "grad_norm": 0.3370955000922167, "learning_rate": 7.014020764532828e-05, "loss": 0.8981, "step": 3373 }, { "epoch": 1.5280797101449275, "grad_norm": 0.5031198871596354, "learning_rate": 7.013189126439778e-05, "loss": 0.8948, "step": 3374 }, { "epoch": 1.5285326086956523, "grad_norm": 0.6231406319614813, "learning_rate": 7.012357187113438e-05, "loss": 0.8679, "step": 3375 }, { "epoch": 1.528985507246377, "grad_norm": 0.6595495347246497, "learning_rate": 7.011524946636976e-05, "loss": 0.8802, "step": 3376 }, { "epoch": 1.5294384057971016, "grad_norm": 0.6520139574815272, "learning_rate": 7.010692405093593e-05, "loss": 0.8852, "step": 3377 }, { "epoch": 1.5298913043478262, "grad_norm": 0.654549890559206, "learning_rate": 7.009859562566518e-05, "loss": 0.892, "step": 3378 }, { "epoch": 1.5303442028985508, "grad_norm": 0.6143130613422727, "learning_rate": 7.009026419139015e-05, "loss": 0.8804, "step": 3379 }, { "epoch": 1.5307971014492754, "grad_norm": 0.5427570068226762, "learning_rate": 7.008192974894371e-05, "loss": 0.8852, "step": 3380 }, { "epoch": 1.53125, "grad_norm": 0.5154463210874624, "learning_rate": 7.00735922991591e-05, "loss": 0.8921, "step": 3381 }, { "epoch": 1.5317028985507246, "grad_norm": 0.5286543489193187, "learning_rate": 7.00652518428698e-05, "loss": 0.8752, "step": 3382 }, { "epoch": 1.5321557971014492, "grad_norm": 0.45807520698511517, "learning_rate": 7.005690838090964e-05, "loss": 0.8881, "step": 3383 }, { "epoch": 1.5326086956521738, "grad_norm": 0.4055260774681594, "learning_rate": 7.004856191411271e-05, "loss": 0.8744, "step": 3384 }, { "epoch": 1.5330615942028984, "grad_norm": 0.37180053763656046, "learning_rate": 7.004021244331344e-05, "loss": 0.8816, "step": 3385 }, { "epoch": 1.533514492753623, "grad_norm": 0.3676675478342172, "learning_rate": 7.003185996934653e-05, "loss": 0.8857, "step": 3386 }, { "epoch": 1.5339673913043477, "grad_norm": 0.4026107609330641, "learning_rate": 7.002350449304698e-05, "loss": 0.8582, "step": 3387 }, { "epoch": 1.5344202898550725, "grad_norm": 0.33941959941946076, "learning_rate": 7.001514601525012e-05, "loss": 0.8759, "step": 3388 }, { "epoch": 1.5348731884057971, "grad_norm": 0.30762658017049244, "learning_rate": 7.000678453679154e-05, "loss": 0.8806, "step": 3389 }, { "epoch": 1.5353260869565217, "grad_norm": 0.3344037348065886, "learning_rate": 6.999842005850715e-05, "loss": 0.9099, "step": 3390 }, { "epoch": 1.5357789855072463, "grad_norm": 0.38707831448203184, "learning_rate": 6.999005258123319e-05, "loss": 0.8902, "step": 3391 }, { "epoch": 1.5362318840579712, "grad_norm": 0.47526053346762015, "learning_rate": 6.998168210580614e-05, "loss": 0.8844, "step": 3392 }, { "epoch": 1.5366847826086958, "grad_norm": 0.45959316785761883, "learning_rate": 6.997330863306282e-05, "loss": 0.8906, "step": 3393 }, { "epoch": 1.5371376811594204, "grad_norm": 0.5086693322040807, "learning_rate": 6.996493216384034e-05, "loss": 0.9013, "step": 3394 }, { "epoch": 1.537590579710145, "grad_norm": 0.5992829433549804, "learning_rate": 6.995655269897611e-05, "loss": 0.8937, "step": 3395 }, { "epoch": 1.5380434782608696, "grad_norm": 0.5707161628891597, "learning_rate": 6.994817023930784e-05, "loss": 0.8798, "step": 3396 }, { "epoch": 1.5384963768115942, "grad_norm": 0.46068439720584753, "learning_rate": 6.993978478567352e-05, "loss": 0.9135, "step": 3397 }, { "epoch": 1.5389492753623188, "grad_norm": 0.510834902315046, "learning_rate": 6.99313963389115e-05, "loss": 0.8881, "step": 3398 }, { "epoch": 1.5394021739130435, "grad_norm": 0.5460795335177546, "learning_rate": 6.992300489986033e-05, "loss": 0.8925, "step": 3399 }, { "epoch": 1.539855072463768, "grad_norm": 0.5780438439133219, "learning_rate": 6.991461046935895e-05, "loss": 0.8941, "step": 3400 }, { "epoch": 1.5403079710144927, "grad_norm": 0.6537080896909043, "learning_rate": 6.990621304824656e-05, "loss": 0.8764, "step": 3401 }, { "epoch": 1.5407608695652173, "grad_norm": 0.8144510869011344, "learning_rate": 6.989781263736266e-05, "loss": 0.8927, "step": 3402 }, { "epoch": 1.541213768115942, "grad_norm": 0.914006029736203, "learning_rate": 6.988940923754706e-05, "loss": 0.9143, "step": 3403 }, { "epoch": 1.5416666666666665, "grad_norm": 0.9046257516140853, "learning_rate": 6.988100284963985e-05, "loss": 0.9065, "step": 3404 }, { "epoch": 1.5421195652173914, "grad_norm": 0.8703048657688788, "learning_rate": 6.987259347448143e-05, "loss": 0.8967, "step": 3405 }, { "epoch": 1.542572463768116, "grad_norm": 0.815997926900649, "learning_rate": 6.986418111291252e-05, "loss": 0.8813, "step": 3406 }, { "epoch": 1.5430253623188406, "grad_norm": 0.678577494201753, "learning_rate": 6.985576576577406e-05, "loss": 0.891, "step": 3407 }, { "epoch": 1.5434782608695652, "grad_norm": 0.6547378303947747, "learning_rate": 6.984734743390743e-05, "loss": 0.8988, "step": 3408 }, { "epoch": 1.5439311594202898, "grad_norm": 0.5115814786841584, "learning_rate": 6.983892611815416e-05, "loss": 0.8933, "step": 3409 }, { "epoch": 1.5443840579710146, "grad_norm": 0.3281723743689916, "learning_rate": 6.983050181935616e-05, "loss": 0.8852, "step": 3410 }, { "epoch": 1.5448369565217392, "grad_norm": 0.3438650210813816, "learning_rate": 6.982207453835563e-05, "loss": 0.8868, "step": 3411 }, { "epoch": 1.5452898550724639, "grad_norm": 0.38733893118530915, "learning_rate": 6.981364427599505e-05, "loss": 0.8849, "step": 3412 }, { "epoch": 1.5457427536231885, "grad_norm": 0.4667099766515002, "learning_rate": 6.98052110331172e-05, "loss": 0.8979, "step": 3413 }, { "epoch": 1.546195652173913, "grad_norm": 0.4794407433200265, "learning_rate": 6.979677481056519e-05, "loss": 0.8948, "step": 3414 }, { "epoch": 1.5466485507246377, "grad_norm": 0.47248089563807155, "learning_rate": 6.978833560918236e-05, "loss": 0.883, "step": 3415 }, { "epoch": 1.5471014492753623, "grad_norm": 0.41374541827900796, "learning_rate": 6.977989342981244e-05, "loss": 0.9001, "step": 3416 }, { "epoch": 1.547554347826087, "grad_norm": 0.3707215366014839, "learning_rate": 6.977144827329937e-05, "loss": 0.8927, "step": 3417 }, { "epoch": 1.5480072463768115, "grad_norm": 0.3296064725722886, "learning_rate": 6.976300014048745e-05, "loss": 0.8887, "step": 3418 }, { "epoch": 1.5484601449275361, "grad_norm": 0.3164336363996876, "learning_rate": 6.975454903222124e-05, "loss": 0.9081, "step": 3419 }, { "epoch": 1.5489130434782608, "grad_norm": 0.3191973462724632, "learning_rate": 6.97460949493456e-05, "loss": 0.8919, "step": 3420 }, { "epoch": 1.5493659420289854, "grad_norm": 0.32173215905245106, "learning_rate": 6.973763789270573e-05, "loss": 0.8899, "step": 3421 }, { "epoch": 1.5498188405797102, "grad_norm": 0.3533951150315436, "learning_rate": 6.972917786314706e-05, "loss": 0.8878, "step": 3422 }, { "epoch": 1.5502717391304348, "grad_norm": 0.3607824638925038, "learning_rate": 6.972071486151536e-05, "loss": 0.868, "step": 3423 }, { "epoch": 1.5507246376811594, "grad_norm": 0.38678082442619166, "learning_rate": 6.971224888865673e-05, "loss": 0.8835, "step": 3424 }, { "epoch": 1.551177536231884, "grad_norm": 0.33582658693708023, "learning_rate": 6.970377994541749e-05, "loss": 0.8953, "step": 3425 }, { "epoch": 1.5516304347826086, "grad_norm": 0.4676533135320974, "learning_rate": 6.969530803264428e-05, "loss": 0.8899, "step": 3426 }, { "epoch": 1.5520833333333335, "grad_norm": 0.5823074167393198, "learning_rate": 6.968683315118407e-05, "loss": 0.9204, "step": 3427 }, { "epoch": 1.552536231884058, "grad_norm": 0.6168342381728186, "learning_rate": 6.967835530188413e-05, "loss": 0.8818, "step": 3428 }, { "epoch": 1.5529891304347827, "grad_norm": 0.6439890173729893, "learning_rate": 6.966987448559196e-05, "loss": 0.905, "step": 3429 }, { "epoch": 1.5534420289855073, "grad_norm": 0.7284549242719442, "learning_rate": 6.966139070315543e-05, "loss": 0.8792, "step": 3430 }, { "epoch": 1.553894927536232, "grad_norm": 0.7865719589199713, "learning_rate": 6.965290395542267e-05, "loss": 0.9196, "step": 3431 }, { "epoch": 1.5543478260869565, "grad_norm": 0.8027284747658376, "learning_rate": 6.964441424324211e-05, "loss": 0.8886, "step": 3432 }, { "epoch": 1.5548007246376812, "grad_norm": 0.6716632224267342, "learning_rate": 6.963592156746249e-05, "loss": 0.86, "step": 3433 }, { "epoch": 1.5552536231884058, "grad_norm": 0.44486217029422015, "learning_rate": 6.96274259289328e-05, "loss": 0.8882, "step": 3434 }, { "epoch": 1.5557065217391304, "grad_norm": 0.28504112087836164, "learning_rate": 6.961892732850242e-05, "loss": 0.8961, "step": 3435 }, { "epoch": 1.556159420289855, "grad_norm": 0.3878054034720017, "learning_rate": 6.961042576702094e-05, "loss": 0.9014, "step": 3436 }, { "epoch": 1.5566123188405796, "grad_norm": 0.5673301173455687, "learning_rate": 6.960192124533826e-05, "loss": 0.8942, "step": 3437 }, { "epoch": 1.5570652173913042, "grad_norm": 0.7258574644299396, "learning_rate": 6.959341376430461e-05, "loss": 0.8747, "step": 3438 }, { "epoch": 1.5575181159420288, "grad_norm": 0.8261009895961586, "learning_rate": 6.95849033247705e-05, "loss": 0.8956, "step": 3439 }, { "epoch": 1.5579710144927537, "grad_norm": 0.8716904127589129, "learning_rate": 6.95763899275867e-05, "loss": 0.8878, "step": 3440 }, { "epoch": 1.5584239130434783, "grad_norm": 0.871829656132776, "learning_rate": 6.956787357360436e-05, "loss": 0.8893, "step": 3441 }, { "epoch": 1.5588768115942029, "grad_norm": 0.7650083061478152, "learning_rate": 6.955935426367482e-05, "loss": 0.8992, "step": 3442 }, { "epoch": 1.5593297101449275, "grad_norm": 0.6047753229890579, "learning_rate": 6.955083199864981e-05, "loss": 0.8822, "step": 3443 }, { "epoch": 1.5597826086956523, "grad_norm": 0.41208593337883515, "learning_rate": 6.95423067793813e-05, "loss": 0.8817, "step": 3444 }, { "epoch": 1.560235507246377, "grad_norm": 0.32945517540000957, "learning_rate": 6.953377860672156e-05, "loss": 0.8928, "step": 3445 }, { "epoch": 1.5606884057971016, "grad_norm": 0.2782704777802139, "learning_rate": 6.952524748152316e-05, "loss": 0.8858, "step": 3446 }, { "epoch": 1.5611413043478262, "grad_norm": 0.31155830352553265, "learning_rate": 6.951671340463901e-05, "loss": 0.8664, "step": 3447 }, { "epoch": 1.5615942028985508, "grad_norm": 0.4584666195186209, "learning_rate": 6.950817637692222e-05, "loss": 0.8649, "step": 3448 }, { "epoch": 1.5620471014492754, "grad_norm": 0.5333189090170045, "learning_rate": 6.949963639922629e-05, "loss": 0.8695, "step": 3449 }, { "epoch": 1.5625, "grad_norm": 0.5328843884145783, "learning_rate": 6.949109347240496e-05, "loss": 0.8781, "step": 3450 }, { "epoch": 1.5629528985507246, "grad_norm": 0.5254611372370579, "learning_rate": 6.94825475973123e-05, "loss": 0.8766, "step": 3451 }, { "epoch": 1.5634057971014492, "grad_norm": 0.4859368837145494, "learning_rate": 6.947399877480263e-05, "loss": 0.8892, "step": 3452 }, { "epoch": 1.5638586956521738, "grad_norm": 0.43878442002489804, "learning_rate": 6.946544700573058e-05, "loss": 0.8765, "step": 3453 }, { "epoch": 1.5643115942028984, "grad_norm": 0.43914038280738027, "learning_rate": 6.945689229095111e-05, "loss": 0.8746, "step": 3454 }, { "epoch": 1.564764492753623, "grad_norm": 0.5162192716169203, "learning_rate": 6.944833463131944e-05, "loss": 0.9028, "step": 3455 }, { "epoch": 1.5652173913043477, "grad_norm": 0.5518622275455761, "learning_rate": 6.94397740276911e-05, "loss": 0.9169, "step": 3456 }, { "epoch": 1.5656702898550725, "grad_norm": 0.474106218180923, "learning_rate": 6.943121048092188e-05, "loss": 0.881, "step": 3457 }, { "epoch": 1.5661231884057971, "grad_norm": 0.3556742036541498, "learning_rate": 6.942264399186793e-05, "loss": 0.8803, "step": 3458 }, { "epoch": 1.5665760869565217, "grad_norm": 0.4038072327834157, "learning_rate": 6.941407456138563e-05, "loss": 0.9018, "step": 3459 }, { "epoch": 1.5670289855072463, "grad_norm": 0.4360354993763063, "learning_rate": 6.940550219033168e-05, "loss": 0.903, "step": 3460 }, { "epoch": 1.5674818840579712, "grad_norm": 0.42242743866491833, "learning_rate": 6.939692687956309e-05, "loss": 0.9096, "step": 3461 }, { "epoch": 1.5679347826086958, "grad_norm": 0.502454222357513, "learning_rate": 6.938834862993713e-05, "loss": 0.8879, "step": 3462 }, { "epoch": 1.5683876811594204, "grad_norm": 0.6017251241806909, "learning_rate": 6.937976744231139e-05, "loss": 0.8811, "step": 3463 }, { "epoch": 1.568840579710145, "grad_norm": 0.7930529418983621, "learning_rate": 6.937118331754373e-05, "loss": 0.9085, "step": 3464 }, { "epoch": 1.5692934782608696, "grad_norm": 1.0530272489746129, "learning_rate": 6.936259625649235e-05, "loss": 0.8832, "step": 3465 }, { "epoch": 1.5697463768115942, "grad_norm": 1.0539504243827036, "learning_rate": 6.935400626001567e-05, "loss": 0.8963, "step": 3466 }, { "epoch": 1.5701992753623188, "grad_norm": 0.7638561329519977, "learning_rate": 6.93454133289725e-05, "loss": 0.9018, "step": 3467 }, { "epoch": 1.5706521739130435, "grad_norm": 0.4967807280099056, "learning_rate": 6.933681746422183e-05, "loss": 0.8873, "step": 3468 }, { "epoch": 1.571105072463768, "grad_norm": 0.3336414160159657, "learning_rate": 6.932821866662304e-05, "loss": 0.871, "step": 3469 }, { "epoch": 1.5715579710144927, "grad_norm": 0.4232900796621546, "learning_rate": 6.931961693703576e-05, "loss": 0.8727, "step": 3470 }, { "epoch": 1.5720108695652173, "grad_norm": 0.6859743645135552, "learning_rate": 6.931101227631991e-05, "loss": 0.8809, "step": 3471 }, { "epoch": 1.572463768115942, "grad_norm": 0.9010970395898683, "learning_rate": 6.93024046853357e-05, "loss": 0.885, "step": 3472 }, { "epoch": 1.5729166666666665, "grad_norm": 0.9958034729868097, "learning_rate": 6.929379416494369e-05, "loss": 0.887, "step": 3473 }, { "epoch": 1.5733695652173914, "grad_norm": 0.943765991907619, "learning_rate": 6.928518071600463e-05, "loss": 0.8805, "step": 3474 }, { "epoch": 1.573822463768116, "grad_norm": 0.7651419660949252, "learning_rate": 6.927656433937968e-05, "loss": 0.8798, "step": 3475 }, { "epoch": 1.5742753623188406, "grad_norm": 0.4678054677982523, "learning_rate": 6.926794503593017e-05, "loss": 0.867, "step": 3476 }, { "epoch": 1.5747282608695652, "grad_norm": 0.31309624055976665, "learning_rate": 6.925932280651783e-05, "loss": 0.8879, "step": 3477 }, { "epoch": 1.5751811594202898, "grad_norm": 0.3759147021788123, "learning_rate": 6.925069765200461e-05, "loss": 0.8832, "step": 3478 }, { "epoch": 1.5756340579710146, "grad_norm": 0.5814437653847302, "learning_rate": 6.924206957325281e-05, "loss": 0.8951, "step": 3479 }, { "epoch": 1.5760869565217392, "grad_norm": 0.7546709253711281, "learning_rate": 6.923343857112497e-05, "loss": 0.8983, "step": 3480 }, { "epoch": 1.5765398550724639, "grad_norm": 0.7757110902319493, "learning_rate": 6.922480464648395e-05, "loss": 0.8807, "step": 3481 }, { "epoch": 1.5769927536231885, "grad_norm": 0.6845797055730578, "learning_rate": 6.921616780019291e-05, "loss": 0.907, "step": 3482 }, { "epoch": 1.577445652173913, "grad_norm": 0.5255256228877072, "learning_rate": 6.920752803311527e-05, "loss": 0.9025, "step": 3483 }, { "epoch": 1.5778985507246377, "grad_norm": 0.40466640253691255, "learning_rate": 6.919888534611476e-05, "loss": 0.8836, "step": 3484 }, { "epoch": 1.5783514492753623, "grad_norm": 0.291839335361975, "learning_rate": 6.919023974005542e-05, "loss": 0.8962, "step": 3485 }, { "epoch": 1.578804347826087, "grad_norm": 0.3025692907932476, "learning_rate": 6.918159121580155e-05, "loss": 0.8793, "step": 3486 }, { "epoch": 1.5792572463768115, "grad_norm": 0.4779135974653199, "learning_rate": 6.917293977421778e-05, "loss": 0.9041, "step": 3487 }, { "epoch": 1.5797101449275361, "grad_norm": 0.656002314344528, "learning_rate": 6.916428541616899e-05, "loss": 0.9046, "step": 3488 }, { "epoch": 1.5801630434782608, "grad_norm": 0.7630411370042991, "learning_rate": 6.915562814252036e-05, "loss": 0.8795, "step": 3489 }, { "epoch": 1.5806159420289854, "grad_norm": 0.7477443256863407, "learning_rate": 6.91469679541374e-05, "loss": 0.8858, "step": 3490 }, { "epoch": 1.5810688405797102, "grad_norm": 0.6180352500708423, "learning_rate": 6.913830485188584e-05, "loss": 0.9032, "step": 3491 }, { "epoch": 1.5815217391304348, "grad_norm": 0.4808443608711631, "learning_rate": 6.91296388366318e-05, "loss": 0.8932, "step": 3492 }, { "epoch": 1.5819746376811594, "grad_norm": 0.3501387011156843, "learning_rate": 6.912096990924159e-05, "loss": 0.8788, "step": 3493 }, { "epoch": 1.582427536231884, "grad_norm": 0.2319890650259296, "learning_rate": 6.911229807058187e-05, "loss": 0.8909, "step": 3494 }, { "epoch": 1.5828804347826086, "grad_norm": 0.3408480605609763, "learning_rate": 6.91036233215196e-05, "loss": 0.9111, "step": 3495 }, { "epoch": 1.5833333333333335, "grad_norm": 0.423125805653411, "learning_rate": 6.909494566292195e-05, "loss": 0.8853, "step": 3496 }, { "epoch": 1.583786231884058, "grad_norm": 0.4788881173237839, "learning_rate": 6.908626509565651e-05, "loss": 0.8846, "step": 3497 }, { "epoch": 1.5842391304347827, "grad_norm": 0.48173765600021556, "learning_rate": 6.907758162059103e-05, "loss": 0.8806, "step": 3498 }, { "epoch": 1.5846920289855073, "grad_norm": 0.34846894672485246, "learning_rate": 6.906889523859366e-05, "loss": 0.8702, "step": 3499 }, { "epoch": 1.585144927536232, "grad_norm": 0.2820207532508777, "learning_rate": 6.906020595053276e-05, "loss": 0.877, "step": 3500 }, { "epoch": 1.5855978260869565, "grad_norm": 0.3162178424630541, "learning_rate": 6.905151375727704e-05, "loss": 0.8789, "step": 3501 }, { "epoch": 1.5860507246376812, "grad_norm": 0.3405747702359585, "learning_rate": 6.904281865969543e-05, "loss": 0.8881, "step": 3502 }, { "epoch": 1.5865036231884058, "grad_norm": 0.4114909215086068, "learning_rate": 6.903412065865722e-05, "loss": 0.8894, "step": 3503 }, { "epoch": 1.5869565217391304, "grad_norm": 0.4949425624802807, "learning_rate": 6.902541975503196e-05, "loss": 0.8922, "step": 3504 }, { "epoch": 1.587409420289855, "grad_norm": 0.47962091449080707, "learning_rate": 6.901671594968951e-05, "loss": 0.8851, "step": 3505 }, { "epoch": 1.5878623188405796, "grad_norm": 0.4210912142542904, "learning_rate": 6.900800924349997e-05, "loss": 0.897, "step": 3506 }, { "epoch": 1.5883152173913042, "grad_norm": 0.3293804042951442, "learning_rate": 6.899929963733378e-05, "loss": 0.8908, "step": 3507 }, { "epoch": 1.5887681159420288, "grad_norm": 0.3732819957250329, "learning_rate": 6.899058713206165e-05, "loss": 0.8921, "step": 3508 }, { "epoch": 1.5892210144927537, "grad_norm": 0.5713203699408389, "learning_rate": 6.898187172855458e-05, "loss": 0.8717, "step": 3509 }, { "epoch": 1.5896739130434783, "grad_norm": 0.6717057718200999, "learning_rate": 6.897315342768388e-05, "loss": 0.895, "step": 3510 }, { "epoch": 1.5901268115942029, "grad_norm": 0.6211464089401738, "learning_rate": 6.896443223032111e-05, "loss": 0.902, "step": 3511 }, { "epoch": 1.5905797101449275, "grad_norm": 0.5050295980300891, "learning_rate": 6.895570813733816e-05, "loss": 0.9036, "step": 3512 }, { "epoch": 1.5910326086956523, "grad_norm": 0.4889240889392168, "learning_rate": 6.894698114960718e-05, "loss": 0.8974, "step": 3513 }, { "epoch": 1.591485507246377, "grad_norm": 0.5682267558164661, "learning_rate": 6.893825126800062e-05, "loss": 0.8929, "step": 3514 }, { "epoch": 1.5919384057971016, "grad_norm": 0.634299314254502, "learning_rate": 6.892951849339124e-05, "loss": 0.8753, "step": 3515 }, { "epoch": 1.5923913043478262, "grad_norm": 0.5399183074700679, "learning_rate": 6.892078282665204e-05, "loss": 0.892, "step": 3516 }, { "epoch": 1.5928442028985508, "grad_norm": 0.4119486672237172, "learning_rate": 6.891204426865635e-05, "loss": 0.9016, "step": 3517 }, { "epoch": 1.5932971014492754, "grad_norm": 0.5004875330396056, "learning_rate": 6.890330282027778e-05, "loss": 0.8912, "step": 3518 }, { "epoch": 1.59375, "grad_norm": 0.5352926724716092, "learning_rate": 6.889455848239022e-05, "loss": 0.8934, "step": 3519 }, { "epoch": 1.5942028985507246, "grad_norm": 0.43477898355496736, "learning_rate": 6.888581125586788e-05, "loss": 0.8704, "step": 3520 }, { "epoch": 1.5946557971014492, "grad_norm": 0.4742433929727883, "learning_rate": 6.88770611415852e-05, "loss": 0.8789, "step": 3521 }, { "epoch": 1.5951086956521738, "grad_norm": 0.4646165962820182, "learning_rate": 6.886830814041697e-05, "loss": 0.8921, "step": 3522 }, { "epoch": 1.5955615942028984, "grad_norm": 0.35448716261584445, "learning_rate": 6.885955225323822e-05, "loss": 0.8971, "step": 3523 }, { "epoch": 1.596014492753623, "grad_norm": 0.40466563358662144, "learning_rate": 6.88507934809243e-05, "loss": 0.881, "step": 3524 }, { "epoch": 1.5964673913043477, "grad_norm": 0.5160222240745946, "learning_rate": 6.884203182435084e-05, "loss": 0.8816, "step": 3525 }, { "epoch": 1.5969202898550725, "grad_norm": 0.5091532667163663, "learning_rate": 6.883326728439375e-05, "loss": 0.9021, "step": 3526 }, { "epoch": 1.5973731884057971, "grad_norm": 0.5311127460472769, "learning_rate": 6.882449986192925e-05, "loss": 0.8685, "step": 3527 }, { "epoch": 1.5978260869565217, "grad_norm": 0.691849678519667, "learning_rate": 6.881572955783382e-05, "loss": 0.9051, "step": 3528 }, { "epoch": 1.5982789855072463, "grad_norm": 0.7803934759502074, "learning_rate": 6.880695637298424e-05, "loss": 0.8501, "step": 3529 }, { "epoch": 1.5987318840579712, "grad_norm": 0.7375077866253497, "learning_rate": 6.879818030825758e-05, "loss": 0.8975, "step": 3530 }, { "epoch": 1.5991847826086958, "grad_norm": 0.734508976786062, "learning_rate": 6.878940136453119e-05, "loss": 0.8902, "step": 3531 }, { "epoch": 1.5996376811594204, "grad_norm": 0.5902470825064686, "learning_rate": 6.878061954268274e-05, "loss": 0.8806, "step": 3532 }, { "epoch": 1.600090579710145, "grad_norm": 0.45150241015234294, "learning_rate": 6.877183484359016e-05, "loss": 0.9089, "step": 3533 }, { "epoch": 1.6005434782608696, "grad_norm": 0.44023742738558025, "learning_rate": 6.876304726813163e-05, "loss": 0.8898, "step": 3534 }, { "epoch": 1.6009963768115942, "grad_norm": 0.38906426717773, "learning_rate": 6.87542568171857e-05, "loss": 0.894, "step": 3535 }, { "epoch": 1.6014492753623188, "grad_norm": 0.2951508215681201, "learning_rate": 6.874546349163115e-05, "loss": 0.8917, "step": 3536 }, { "epoch": 1.6019021739130435, "grad_norm": 0.4990670666477802, "learning_rate": 6.873666729234707e-05, "loss": 0.8802, "step": 3537 }, { "epoch": 1.602355072463768, "grad_norm": 0.6766394492388116, "learning_rate": 6.872786822021281e-05, "loss": 0.8949, "step": 3538 }, { "epoch": 1.6028079710144927, "grad_norm": 0.7219689703355938, "learning_rate": 6.871906627610805e-05, "loss": 0.8838, "step": 3539 }, { "epoch": 1.6032608695652173, "grad_norm": 0.800008647832802, "learning_rate": 6.871026146091272e-05, "loss": 0.8812, "step": 3540 }, { "epoch": 1.603713768115942, "grad_norm": 0.7624386551644841, "learning_rate": 6.870145377550706e-05, "loss": 0.8845, "step": 3541 }, { "epoch": 1.6041666666666665, "grad_norm": 0.6449754107020813, "learning_rate": 6.869264322077158e-05, "loss": 0.9037, "step": 3542 }, { "epoch": 1.6046195652173914, "grad_norm": 0.489793087124558, "learning_rate": 6.86838297975871e-05, "loss": 0.8866, "step": 3543 }, { "epoch": 1.605072463768116, "grad_norm": 0.44200157548103575, "learning_rate": 6.867501350683469e-05, "loss": 0.8903, "step": 3544 }, { "epoch": 1.6055253623188406, "grad_norm": 0.4622220884402722, "learning_rate": 6.866619434939574e-05, "loss": 0.8856, "step": 3545 }, { "epoch": 1.6059782608695652, "grad_norm": 0.39145835855402566, "learning_rate": 6.865737232615193e-05, "loss": 0.8786, "step": 3546 }, { "epoch": 1.6064311594202898, "grad_norm": 0.3390395744231965, "learning_rate": 6.864854743798519e-05, "loss": 0.8915, "step": 3547 }, { "epoch": 1.6068840579710146, "grad_norm": 0.3490395358647507, "learning_rate": 6.863971968577776e-05, "loss": 0.8691, "step": 3548 }, { "epoch": 1.6073369565217392, "grad_norm": 0.3184038153336556, "learning_rate": 6.863088907041218e-05, "loss": 0.8839, "step": 3549 }, { "epoch": 1.6077898550724639, "grad_norm": 0.3378212678438521, "learning_rate": 6.862205559277124e-05, "loss": 0.8936, "step": 3550 }, { "epoch": 1.6082427536231885, "grad_norm": 0.4069673159659651, "learning_rate": 6.861321925373804e-05, "loss": 0.8914, "step": 3551 }, { "epoch": 1.608695652173913, "grad_norm": 0.34199157602988706, "learning_rate": 6.860438005419598e-05, "loss": 0.8989, "step": 3552 }, { "epoch": 1.6091485507246377, "grad_norm": 0.3315274831735942, "learning_rate": 6.859553799502871e-05, "loss": 0.8922, "step": 3553 }, { "epoch": 1.6096014492753623, "grad_norm": 0.3933133265948832, "learning_rate": 6.85866930771202e-05, "loss": 0.8864, "step": 3554 }, { "epoch": 1.610054347826087, "grad_norm": 0.3274486898518188, "learning_rate": 6.857784530135469e-05, "loss": 0.8774, "step": 3555 }, { "epoch": 1.6105072463768115, "grad_norm": 0.3137827249244487, "learning_rate": 6.85689946686167e-05, "loss": 0.8732, "step": 3556 }, { "epoch": 1.6109601449275361, "grad_norm": 0.33358100659346124, "learning_rate": 6.856014117979104e-05, "loss": 0.8976, "step": 3557 }, { "epoch": 1.6114130434782608, "grad_norm": 0.32570692633305587, "learning_rate": 6.85512848357628e-05, "loss": 0.9048, "step": 3558 }, { "epoch": 1.6118659420289854, "grad_norm": 0.38223340377249204, "learning_rate": 6.854242563741737e-05, "loss": 0.9058, "step": 3559 }, { "epoch": 1.6123188405797102, "grad_norm": 0.5221160250536164, "learning_rate": 6.853356358564043e-05, "loss": 0.9111, "step": 3560 }, { "epoch": 1.6127717391304348, "grad_norm": 0.6547128432119194, "learning_rate": 6.852469868131792e-05, "loss": 0.8893, "step": 3561 }, { "epoch": 1.6132246376811594, "grad_norm": 0.7810399373158747, "learning_rate": 6.851583092533608e-05, "loss": 0.8873, "step": 3562 }, { "epoch": 1.613677536231884, "grad_norm": 0.8766693067658348, "learning_rate": 6.850696031858143e-05, "loss": 0.8898, "step": 3563 }, { "epoch": 1.6141304347826086, "grad_norm": 0.8714173397889238, "learning_rate": 6.84980868619408e-05, "loss": 0.8946, "step": 3564 }, { "epoch": 1.6145833333333335, "grad_norm": 0.772967245649841, "learning_rate": 6.848921055630125e-05, "loss": 0.9, "step": 3565 }, { "epoch": 1.615036231884058, "grad_norm": 0.6208742860676963, "learning_rate": 6.848033140255019e-05, "loss": 0.8934, "step": 3566 }, { "epoch": 1.6154891304347827, "grad_norm": 0.5160583659444492, "learning_rate": 6.847144940157525e-05, "loss": 0.8811, "step": 3567 }, { "epoch": 1.6159420289855073, "grad_norm": 0.35661357474013644, "learning_rate": 6.84625645542644e-05, "loss": 0.8792, "step": 3568 }, { "epoch": 1.616394927536232, "grad_norm": 0.27575059204006963, "learning_rate": 6.84536768615059e-05, "loss": 0.8712, "step": 3569 }, { "epoch": 1.6168478260869565, "grad_norm": 0.4111192457207626, "learning_rate": 6.84447863241882e-05, "loss": 0.8907, "step": 3570 }, { "epoch": 1.6173007246376812, "grad_norm": 0.5606085879652605, "learning_rate": 6.843589294320014e-05, "loss": 0.8728, "step": 3571 }, { "epoch": 1.6177536231884058, "grad_norm": 0.6022499931946473, "learning_rate": 6.842699671943081e-05, "loss": 0.8978, "step": 3572 }, { "epoch": 1.6182065217391304, "grad_norm": 0.5321421276708745, "learning_rate": 6.841809765376957e-05, "loss": 0.8652, "step": 3573 }, { "epoch": 1.618659420289855, "grad_norm": 0.44302030684327104, "learning_rate": 6.840919574710608e-05, "loss": 0.8833, "step": 3574 }, { "epoch": 1.6191123188405796, "grad_norm": 0.33490290457349686, "learning_rate": 6.840029100033027e-05, "loss": 0.9071, "step": 3575 }, { "epoch": 1.6195652173913042, "grad_norm": 0.2809440998195494, "learning_rate": 6.839138341433236e-05, "loss": 0.8806, "step": 3576 }, { "epoch": 1.6200181159420288, "grad_norm": 0.28292227952887855, "learning_rate": 6.838247299000287e-05, "loss": 0.8961, "step": 3577 }, { "epoch": 1.6204710144927537, "grad_norm": 0.23190216807750652, "learning_rate": 6.837355972823257e-05, "loss": 0.8856, "step": 3578 }, { "epoch": 1.6209239130434783, "grad_norm": 0.2717517509249205, "learning_rate": 6.836464362991255e-05, "loss": 0.8965, "step": 3579 }, { "epoch": 1.6213768115942029, "grad_norm": 0.4162120387160152, "learning_rate": 6.835572469593415e-05, "loss": 0.883, "step": 3580 }, { "epoch": 1.6218297101449275, "grad_norm": 0.5084618096982045, "learning_rate": 6.834680292718902e-05, "loss": 0.8842, "step": 3581 }, { "epoch": 1.6222826086956523, "grad_norm": 0.47763772597049353, "learning_rate": 6.833787832456909e-05, "loss": 0.8981, "step": 3582 }, { "epoch": 1.622735507246377, "grad_norm": 0.5155791367379998, "learning_rate": 6.832895088896655e-05, "loss": 0.8747, "step": 3583 }, { "epoch": 1.6231884057971016, "grad_norm": 0.593507810028236, "learning_rate": 6.83200206212739e-05, "loss": 0.8922, "step": 3584 }, { "epoch": 1.6236413043478262, "grad_norm": 0.6343875874617785, "learning_rate": 6.831108752238392e-05, "loss": 0.8921, "step": 3585 }, { "epoch": 1.6240942028985508, "grad_norm": 0.6029273154400078, "learning_rate": 6.830215159318965e-05, "loss": 0.8762, "step": 3586 }, { "epoch": 1.6245471014492754, "grad_norm": 0.6085419468493898, "learning_rate": 6.829321283458445e-05, "loss": 0.8883, "step": 3587 }, { "epoch": 1.625, "grad_norm": 0.5228078757886636, "learning_rate": 6.828427124746191e-05, "loss": 0.8776, "step": 3588 }, { "epoch": 1.6254528985507246, "grad_norm": 0.4423338803648265, "learning_rate": 6.827532683271595e-05, "loss": 0.8736, "step": 3589 }, { "epoch": 1.6259057971014492, "grad_norm": 0.3929677476327635, "learning_rate": 6.826637959124078e-05, "loss": 0.878, "step": 3590 }, { "epoch": 1.6263586956521738, "grad_norm": 0.373473335739617, "learning_rate": 6.825742952393083e-05, "loss": 0.8976, "step": 3591 }, { "epoch": 1.6268115942028984, "grad_norm": 0.3933592095162357, "learning_rate": 6.824847663168089e-05, "loss": 0.8961, "step": 3592 }, { "epoch": 1.627264492753623, "grad_norm": 0.3666220134445041, "learning_rate": 6.823952091538598e-05, "loss": 0.864, "step": 3593 }, { "epoch": 1.6277173913043477, "grad_norm": 0.31188406388148016, "learning_rate": 6.82305623759414e-05, "loss": 0.8798, "step": 3594 }, { "epoch": 1.6281702898550725, "grad_norm": 0.3247518784163089, "learning_rate": 6.822160101424276e-05, "loss": 0.8916, "step": 3595 }, { "epoch": 1.6286231884057971, "grad_norm": 0.3597674769534177, "learning_rate": 6.821263683118596e-05, "loss": 0.8686, "step": 3596 }, { "epoch": 1.6290760869565217, "grad_norm": 0.37405547243855636, "learning_rate": 6.820366982766715e-05, "loss": 0.8859, "step": 3597 }, { "epoch": 1.6295289855072463, "grad_norm": 0.40452310360848037, "learning_rate": 6.819470000458276e-05, "loss": 0.8723, "step": 3598 }, { "epoch": 1.6299818840579712, "grad_norm": 0.3861359652034274, "learning_rate": 6.818572736282955e-05, "loss": 0.889, "step": 3599 }, { "epoch": 1.6304347826086958, "grad_norm": 0.42625191894482845, "learning_rate": 6.81767519033045e-05, "loss": 0.878, "step": 3600 }, { "epoch": 1.6308876811594204, "grad_norm": 0.45725742736668984, "learning_rate": 6.816777362690491e-05, "loss": 0.9006, "step": 3601 }, { "epoch": 1.631340579710145, "grad_norm": 0.47254755754476274, "learning_rate": 6.815879253452837e-05, "loss": 0.8938, "step": 3602 }, { "epoch": 1.6317934782608696, "grad_norm": 0.5476469373424353, "learning_rate": 6.81498086270727e-05, "loss": 0.8808, "step": 3603 }, { "epoch": 1.6322463768115942, "grad_norm": 0.6504113808471235, "learning_rate": 6.814082190543607e-05, "loss": 0.8959, "step": 3604 }, { "epoch": 1.6326992753623188, "grad_norm": 0.7620207532099692, "learning_rate": 6.813183237051687e-05, "loss": 0.8909, "step": 3605 }, { "epoch": 1.6331521739130435, "grad_norm": 0.8856255224748719, "learning_rate": 6.812284002321381e-05, "loss": 0.9045, "step": 3606 }, { "epoch": 1.633605072463768, "grad_norm": 0.9935917312042963, "learning_rate": 6.811384486442589e-05, "loss": 0.8882, "step": 3607 }, { "epoch": 1.6340579710144927, "grad_norm": 1.0224598746121638, "learning_rate": 6.810484689505232e-05, "loss": 0.8846, "step": 3608 }, { "epoch": 1.6345108695652173, "grad_norm": 0.8479123154104389, "learning_rate": 6.809584611599268e-05, "loss": 0.8936, "step": 3609 }, { "epoch": 1.634963768115942, "grad_norm": 0.7453955150888157, "learning_rate": 6.808684252814678e-05, "loss": 0.892, "step": 3610 }, { "epoch": 1.6354166666666665, "grad_norm": 0.6393522553690898, "learning_rate": 6.807783613241474e-05, "loss": 0.8822, "step": 3611 }, { "epoch": 1.6358695652173914, "grad_norm": 0.47189152138345336, "learning_rate": 6.80688269296969e-05, "loss": 0.8935, "step": 3612 }, { "epoch": 1.636322463768116, "grad_norm": 0.26436748319385284, "learning_rate": 6.805981492089398e-05, "loss": 0.8787, "step": 3613 }, { "epoch": 1.6367753623188406, "grad_norm": 0.27536712538620284, "learning_rate": 6.805080010690688e-05, "loss": 0.8868, "step": 3614 }, { "epoch": 1.6372282608695652, "grad_norm": 0.3876330895912719, "learning_rate": 6.804178248863685e-05, "loss": 0.8869, "step": 3615 }, { "epoch": 1.6376811594202898, "grad_norm": 0.4831830266354041, "learning_rate": 6.803276206698539e-05, "loss": 0.8993, "step": 3616 }, { "epoch": 1.6381340579710146, "grad_norm": 0.5339339597498296, "learning_rate": 6.802373884285428e-05, "loss": 0.9152, "step": 3617 }, { "epoch": 1.6385869565217392, "grad_norm": 0.4952986089028463, "learning_rate": 6.80147128171456e-05, "loss": 0.9, "step": 3618 }, { "epoch": 1.6390398550724639, "grad_norm": 0.5521544587670537, "learning_rate": 6.800568399076168e-05, "loss": 0.8928, "step": 3619 }, { "epoch": 1.6394927536231885, "grad_norm": 0.5949377042766674, "learning_rate": 6.799665236460514e-05, "loss": 0.8813, "step": 3620 }, { "epoch": 1.639945652173913, "grad_norm": 0.5900119543929973, "learning_rate": 6.798761793957893e-05, "loss": 0.8707, "step": 3621 }, { "epoch": 1.6403985507246377, "grad_norm": 0.5202447626339385, "learning_rate": 6.797858071658618e-05, "loss": 0.8942, "step": 3622 }, { "epoch": 1.6408514492753623, "grad_norm": 0.46408582360665934, "learning_rate": 6.79695406965304e-05, "loss": 0.8876, "step": 3623 }, { "epoch": 1.641304347826087, "grad_norm": 0.3997634173674193, "learning_rate": 6.796049788031528e-05, "loss": 0.8652, "step": 3624 }, { "epoch": 1.6417572463768115, "grad_norm": 0.3778729967239979, "learning_rate": 6.795145226884491e-05, "loss": 0.8888, "step": 3625 }, { "epoch": 1.6422101449275361, "grad_norm": 0.34240890321275486, "learning_rate": 6.794240386302356e-05, "loss": 0.8753, "step": 3626 }, { "epoch": 1.6426630434782608, "grad_norm": 0.29258217041449736, "learning_rate": 6.793335266375581e-05, "loss": 0.8764, "step": 3627 }, { "epoch": 1.6431159420289854, "grad_norm": 0.2348522242696148, "learning_rate": 6.792429867194653e-05, "loss": 0.8874, "step": 3628 }, { "epoch": 1.6435688405797102, "grad_norm": 0.250220960554489, "learning_rate": 6.791524188850086e-05, "loss": 0.8796, "step": 3629 }, { "epoch": 1.6440217391304348, "grad_norm": 0.230243501873386, "learning_rate": 6.790618231432422e-05, "loss": 0.9077, "step": 3630 }, { "epoch": 1.6444746376811594, "grad_norm": 0.25243324587430915, "learning_rate": 6.789711995032233e-05, "loss": 0.8818, "step": 3631 }, { "epoch": 1.644927536231884, "grad_norm": 0.28088862402631243, "learning_rate": 6.788805479740114e-05, "loss": 0.8977, "step": 3632 }, { "epoch": 1.6453804347826086, "grad_norm": 0.3539892456876423, "learning_rate": 6.787898685646692e-05, "loss": 0.9057, "step": 3633 }, { "epoch": 1.6458333333333335, "grad_norm": 0.41896888957774514, "learning_rate": 6.786991612842621e-05, "loss": 0.8768, "step": 3634 }, { "epoch": 1.646286231884058, "grad_norm": 0.5373484972107091, "learning_rate": 6.786084261418581e-05, "loss": 0.8898, "step": 3635 }, { "epoch": 1.6467391304347827, "grad_norm": 0.6499684315532217, "learning_rate": 6.785176631465284e-05, "loss": 0.893, "step": 3636 }, { "epoch": 1.6471920289855073, "grad_norm": 0.7254331061734004, "learning_rate": 6.784268723073463e-05, "loss": 0.8843, "step": 3637 }, { "epoch": 1.647644927536232, "grad_norm": 0.7537332140334264, "learning_rate": 6.78336053633389e-05, "loss": 0.8563, "step": 3638 }, { "epoch": 1.6480978260869565, "grad_norm": 0.7430886825085796, "learning_rate": 6.78245207133735e-05, "loss": 0.8886, "step": 3639 }, { "epoch": 1.6485507246376812, "grad_norm": 0.703671381853854, "learning_rate": 6.781543328174669e-05, "loss": 0.8914, "step": 3640 }, { "epoch": 1.6490036231884058, "grad_norm": 0.6511454661818256, "learning_rate": 6.780634306936692e-05, "loss": 0.8851, "step": 3641 }, { "epoch": 1.6494565217391304, "grad_norm": 0.588865197473307, "learning_rate": 6.7797250077143e-05, "loss": 0.8795, "step": 3642 }, { "epoch": 1.649909420289855, "grad_norm": 0.48746874308722143, "learning_rate": 6.77881543059839e-05, "loss": 0.9038, "step": 3643 }, { "epoch": 1.6503623188405796, "grad_norm": 0.3481134730435006, "learning_rate": 6.777905575679902e-05, "loss": 0.8823, "step": 3644 }, { "epoch": 1.6508152173913042, "grad_norm": 0.29546307513525993, "learning_rate": 6.776995443049789e-05, "loss": 0.8913, "step": 3645 }, { "epoch": 1.6512681159420288, "grad_norm": 0.32147440506925806, "learning_rate": 6.776085032799042e-05, "loss": 0.9038, "step": 3646 }, { "epoch": 1.6517210144927537, "grad_norm": 0.38743168033617537, "learning_rate": 6.775174345018676e-05, "loss": 0.9192, "step": 3647 }, { "epoch": 1.6521739130434783, "grad_norm": 0.44571670187710793, "learning_rate": 6.774263379799732e-05, "loss": 0.8809, "step": 3648 }, { "epoch": 1.6526268115942029, "grad_norm": 0.5194310231716187, "learning_rate": 6.773352137233282e-05, "loss": 0.8806, "step": 3649 }, { "epoch": 1.6530797101449275, "grad_norm": 0.6213271498432817, "learning_rate": 6.772440617410424e-05, "loss": 0.8998, "step": 3650 }, { "epoch": 1.6535326086956523, "grad_norm": 0.6415663693627601, "learning_rate": 6.771528820422285e-05, "loss": 0.8817, "step": 3651 }, { "epoch": 1.653985507246377, "grad_norm": 0.596493812465786, "learning_rate": 6.770616746360016e-05, "loss": 0.8812, "step": 3652 }, { "epoch": 1.6544384057971016, "grad_norm": 0.4736987749920116, "learning_rate": 6.769704395314803e-05, "loss": 0.8815, "step": 3653 }, { "epoch": 1.6548913043478262, "grad_norm": 0.44470486760229794, "learning_rate": 6.768791767377851e-05, "loss": 0.8872, "step": 3654 }, { "epoch": 1.6553442028985508, "grad_norm": 0.6364336587760363, "learning_rate": 6.767878862640398e-05, "loss": 0.8901, "step": 3655 }, { "epoch": 1.6557971014492754, "grad_norm": 0.7300274025341349, "learning_rate": 6.76696568119371e-05, "loss": 0.888, "step": 3656 }, { "epoch": 1.65625, "grad_norm": 0.6933036876523138, "learning_rate": 6.766052223129079e-05, "loss": 0.888, "step": 3657 }, { "epoch": 1.6567028985507246, "grad_norm": 0.586310426417269, "learning_rate": 6.765138488537823e-05, "loss": 0.8799, "step": 3658 }, { "epoch": 1.6571557971014492, "grad_norm": 0.5715321586589678, "learning_rate": 6.76422447751129e-05, "loss": 0.8992, "step": 3659 }, { "epoch": 1.6576086956521738, "grad_norm": 0.7176578774790804, "learning_rate": 6.763310190140856e-05, "loss": 0.8993, "step": 3660 }, { "epoch": 1.6580615942028984, "grad_norm": 0.6745080006958181, "learning_rate": 6.762395626517924e-05, "loss": 0.8764, "step": 3661 }, { "epoch": 1.658514492753623, "grad_norm": 0.5638805216485355, "learning_rate": 6.761480786733924e-05, "loss": 0.8805, "step": 3662 }, { "epoch": 1.6589673913043477, "grad_norm": 0.5463273357383294, "learning_rate": 6.760565670880313e-05, "loss": 0.8727, "step": 3663 }, { "epoch": 1.6594202898550725, "grad_norm": 0.611712663439509, "learning_rate": 6.759650279048577e-05, "loss": 0.8725, "step": 3664 }, { "epoch": 1.6598731884057971, "grad_norm": 0.5957168912848986, "learning_rate": 6.758734611330231e-05, "loss": 0.9065, "step": 3665 }, { "epoch": 1.6603260869565217, "grad_norm": 0.4330063925189238, "learning_rate": 6.757818667816815e-05, "loss": 0.8935, "step": 3666 }, { "epoch": 1.6607789855072463, "grad_norm": 0.4099054795802016, "learning_rate": 6.756902448599894e-05, "loss": 0.8632, "step": 3667 }, { "epoch": 1.6612318840579712, "grad_norm": 0.40916134859041636, "learning_rate": 6.755985953771068e-05, "loss": 0.8833, "step": 3668 }, { "epoch": 1.6616847826086958, "grad_norm": 0.3261909213966454, "learning_rate": 6.75506918342196e-05, "loss": 0.8899, "step": 3669 }, { "epoch": 1.6621376811594204, "grad_norm": 0.3195236714937347, "learning_rate": 6.754152137644218e-05, "loss": 0.8661, "step": 3670 }, { "epoch": 1.662590579710145, "grad_norm": 0.40998909863088434, "learning_rate": 6.753234816529522e-05, "loss": 0.8633, "step": 3671 }, { "epoch": 1.6630434782608696, "grad_norm": 0.4700496261742187, "learning_rate": 6.752317220169581e-05, "loss": 0.8902, "step": 3672 }, { "epoch": 1.6634963768115942, "grad_norm": 0.4556349475891146, "learning_rate": 6.751399348656124e-05, "loss": 0.9006, "step": 3673 }, { "epoch": 1.6639492753623188, "grad_norm": 0.4220522178432642, "learning_rate": 6.750481202080916e-05, "loss": 0.8881, "step": 3674 }, { "epoch": 1.6644021739130435, "grad_norm": 0.43020217187408494, "learning_rate": 6.749562780535743e-05, "loss": 0.8961, "step": 3675 }, { "epoch": 1.664855072463768, "grad_norm": 0.4464586900155846, "learning_rate": 6.748644084112421e-05, "loss": 0.8922, "step": 3676 }, { "epoch": 1.6653079710144927, "grad_norm": 0.4520998435417112, "learning_rate": 6.747725112902794e-05, "loss": 0.9047, "step": 3677 }, { "epoch": 1.6657608695652173, "grad_norm": 0.45216723447401114, "learning_rate": 6.746805866998735e-05, "loss": 0.8869, "step": 3678 }, { "epoch": 1.666213768115942, "grad_norm": 0.42482926401304033, "learning_rate": 6.74588634649214e-05, "loss": 0.8965, "step": 3679 }, { "epoch": 1.6666666666666665, "grad_norm": 0.4329245471550429, "learning_rate": 6.744966551474936e-05, "loss": 0.8895, "step": 3680 }, { "epoch": 1.6671195652173914, "grad_norm": 0.49983041304228865, "learning_rate": 6.744046482039075e-05, "loss": 0.8903, "step": 3681 }, { "epoch": 1.667572463768116, "grad_norm": 0.5341727838207756, "learning_rate": 6.743126138276539e-05, "loss": 0.8944, "step": 3682 }, { "epoch": 1.6680253623188406, "grad_norm": 0.4550043518983271, "learning_rate": 6.742205520279337e-05, "loss": 0.8793, "step": 3683 }, { "epoch": 1.6684782608695652, "grad_norm": 0.34195318895225085, "learning_rate": 6.741284628139504e-05, "loss": 0.8726, "step": 3684 }, { "epoch": 1.6689311594202898, "grad_norm": 0.2779852303138535, "learning_rate": 6.740363461949102e-05, "loss": 0.8793, "step": 3685 }, { "epoch": 1.6693840579710146, "grad_norm": 0.26231245804223385, "learning_rate": 6.739442021800223e-05, "loss": 0.8955, "step": 3686 }, { "epoch": 1.6698369565217392, "grad_norm": 0.34839860740869005, "learning_rate": 6.738520307784982e-05, "loss": 0.8759, "step": 3687 }, { "epoch": 1.6702898550724639, "grad_norm": 0.4540887345293808, "learning_rate": 6.737598319995527e-05, "loss": 0.8989, "step": 3688 }, { "epoch": 1.6707427536231885, "grad_norm": 0.5390511703569849, "learning_rate": 6.736676058524032e-05, "loss": 0.8887, "step": 3689 }, { "epoch": 1.671195652173913, "grad_norm": 0.6402719814975466, "learning_rate": 6.735753523462694e-05, "loss": 0.8933, "step": 3690 }, { "epoch": 1.6716485507246377, "grad_norm": 0.7965536336470309, "learning_rate": 6.73483071490374e-05, "loss": 0.8859, "step": 3691 }, { "epoch": 1.6721014492753623, "grad_norm": 0.8170516439744608, "learning_rate": 6.733907632939426e-05, "loss": 0.8844, "step": 3692 }, { "epoch": 1.672554347826087, "grad_norm": 0.7702843085943225, "learning_rate": 6.732984277662035e-05, "loss": 0.8821, "step": 3693 }, { "epoch": 1.6730072463768115, "grad_norm": 0.7020019790214195, "learning_rate": 6.732060649163872e-05, "loss": 0.8806, "step": 3694 }, { "epoch": 1.6734601449275361, "grad_norm": 0.6114244101507936, "learning_rate": 6.73113674753728e-05, "loss": 0.8776, "step": 3695 }, { "epoch": 1.6739130434782608, "grad_norm": 0.5063323035001203, "learning_rate": 6.730212572874618e-05, "loss": 0.8898, "step": 3696 }, { "epoch": 1.6743659420289854, "grad_norm": 0.4455571996089021, "learning_rate": 6.729288125268278e-05, "loss": 0.8808, "step": 3697 }, { "epoch": 1.6748188405797102, "grad_norm": 0.3866334096752379, "learning_rate": 6.72836340481068e-05, "loss": 0.9088, "step": 3698 }, { "epoch": 1.6752717391304348, "grad_norm": 0.3428613611194493, "learning_rate": 6.727438411594267e-05, "loss": 0.9037, "step": 3699 }, { "epoch": 1.6757246376811594, "grad_norm": 0.3596076831105522, "learning_rate": 6.726513145711517e-05, "loss": 0.8946, "step": 3700 }, { "epoch": 1.676177536231884, "grad_norm": 0.43957834124876705, "learning_rate": 6.725587607254925e-05, "loss": 0.9089, "step": 3701 }, { "epoch": 1.6766304347826086, "grad_norm": 0.34539978109844244, "learning_rate": 6.724661796317023e-05, "loss": 0.893, "step": 3702 }, { "epoch": 1.6770833333333335, "grad_norm": 0.32296902438798913, "learning_rate": 6.723735712990362e-05, "loss": 0.8901, "step": 3703 }, { "epoch": 1.677536231884058, "grad_norm": 0.4138942697215391, "learning_rate": 6.722809357367526e-05, "loss": 0.8859, "step": 3704 }, { "epoch": 1.6779891304347827, "grad_norm": 0.47066774255011473, "learning_rate": 6.721882729541125e-05, "loss": 0.8946, "step": 3705 }, { "epoch": 1.6784420289855073, "grad_norm": 0.44314145641220304, "learning_rate": 6.720955829603794e-05, "loss": 0.877, "step": 3706 }, { "epoch": 1.678894927536232, "grad_norm": 0.49656429303739874, "learning_rate": 6.720028657648195e-05, "loss": 0.8797, "step": 3707 }, { "epoch": 1.6793478260869565, "grad_norm": 0.5045099150942038, "learning_rate": 6.719101213767022e-05, "loss": 0.8904, "step": 3708 }, { "epoch": 1.6798007246376812, "grad_norm": 0.544021902824136, "learning_rate": 6.718173498052993e-05, "loss": 0.8968, "step": 3709 }, { "epoch": 1.6802536231884058, "grad_norm": 0.5344032196845476, "learning_rate": 6.717245510598851e-05, "loss": 0.9115, "step": 3710 }, { "epoch": 1.6807065217391304, "grad_norm": 0.4608970317915766, "learning_rate": 6.71631725149737e-05, "loss": 0.8852, "step": 3711 }, { "epoch": 1.681159420289855, "grad_norm": 0.39196014325324463, "learning_rate": 6.715388720841349e-05, "loss": 0.9098, "step": 3712 }, { "epoch": 1.6816123188405796, "grad_norm": 0.38930481139715145, "learning_rate": 6.714459918723613e-05, "loss": 0.8952, "step": 3713 }, { "epoch": 1.6820652173913042, "grad_norm": 0.41282884173295825, "learning_rate": 6.713530845237019e-05, "loss": 0.8861, "step": 3714 }, { "epoch": 1.6825181159420288, "grad_norm": 0.4637997337294865, "learning_rate": 6.712601500474447e-05, "loss": 0.8645, "step": 3715 }, { "epoch": 1.6829710144927537, "grad_norm": 0.4942661013118465, "learning_rate": 6.711671884528804e-05, "loss": 0.9048, "step": 3716 }, { "epoch": 1.6834239130434783, "grad_norm": 0.5636537895449015, "learning_rate": 6.710741997493025e-05, "loss": 0.8954, "step": 3717 }, { "epoch": 1.6838768115942029, "grad_norm": 0.6641307787415438, "learning_rate": 6.709811839460074e-05, "loss": 0.9024, "step": 3718 }, { "epoch": 1.6843297101449275, "grad_norm": 0.7947335009428582, "learning_rate": 6.708881410522939e-05, "loss": 0.8943, "step": 3719 }, { "epoch": 1.6847826086956523, "grad_norm": 0.8907530623871186, "learning_rate": 6.707950710774636e-05, "loss": 0.8885, "step": 3720 }, { "epoch": 1.685235507246377, "grad_norm": 0.936221188807138, "learning_rate": 6.70701974030821e-05, "loss": 0.8651, "step": 3721 }, { "epoch": 1.6856884057971016, "grad_norm": 0.8763603861992787, "learning_rate": 6.70608849921673e-05, "loss": 0.8893, "step": 3722 }, { "epoch": 1.6861413043478262, "grad_norm": 0.828370049186378, "learning_rate": 6.705156987593295e-05, "loss": 0.8909, "step": 3723 }, { "epoch": 1.6865942028985508, "grad_norm": 0.6720866285473451, "learning_rate": 6.70422520553103e-05, "loss": 0.865, "step": 3724 }, { "epoch": 1.6870471014492754, "grad_norm": 0.4740505686038839, "learning_rate": 6.703293153123087e-05, "loss": 0.8958, "step": 3725 }, { "epoch": 1.6875, "grad_norm": 0.36506620361661224, "learning_rate": 6.702360830462642e-05, "loss": 0.9053, "step": 3726 }, { "epoch": 1.6879528985507246, "grad_norm": 0.4271288185686102, "learning_rate": 6.701428237642904e-05, "loss": 0.8663, "step": 3727 }, { "epoch": 1.6884057971014492, "grad_norm": 0.5912582168856959, "learning_rate": 6.700495374757102e-05, "loss": 0.874, "step": 3728 }, { "epoch": 1.6888586956521738, "grad_norm": 0.6803700981308682, "learning_rate": 6.699562241898498e-05, "loss": 0.8875, "step": 3729 }, { "epoch": 1.6893115942028984, "grad_norm": 0.6837661781976093, "learning_rate": 6.698628839160381e-05, "loss": 0.885, "step": 3730 }, { "epoch": 1.689764492753623, "grad_norm": 0.6536613132382108, "learning_rate": 6.697695166636061e-05, "loss": 0.8882, "step": 3731 }, { "epoch": 1.6902173913043477, "grad_norm": 0.5769128697584215, "learning_rate": 6.696761224418882e-05, "loss": 0.8926, "step": 3732 }, { "epoch": 1.6906702898550725, "grad_norm": 0.4508742556223363, "learning_rate": 6.695827012602209e-05, "loss": 0.8896, "step": 3733 }, { "epoch": 1.6911231884057971, "grad_norm": 0.3768772301147291, "learning_rate": 6.694892531279439e-05, "loss": 0.8872, "step": 3734 }, { "epoch": 1.6915760869565217, "grad_norm": 0.32022758850909416, "learning_rate": 6.69395778054399e-05, "loss": 0.8574, "step": 3735 }, { "epoch": 1.6920289855072463, "grad_norm": 0.28880514977730826, "learning_rate": 6.693022760489314e-05, "loss": 0.8999, "step": 3736 }, { "epoch": 1.6924818840579712, "grad_norm": 0.27715558858698713, "learning_rate": 6.692087471208884e-05, "loss": 0.8696, "step": 3737 }, { "epoch": 1.6929347826086958, "grad_norm": 0.31996796766556945, "learning_rate": 6.691151912796205e-05, "loss": 0.8898, "step": 3738 }, { "epoch": 1.6933876811594204, "grad_norm": 0.4294562409938239, "learning_rate": 6.690216085344804e-05, "loss": 0.894, "step": 3739 }, { "epoch": 1.693840579710145, "grad_norm": 0.43298050368472296, "learning_rate": 6.689279988948239e-05, "loss": 0.891, "step": 3740 }, { "epoch": 1.6942934782608696, "grad_norm": 0.5893229843290054, "learning_rate": 6.68834362370009e-05, "loss": 0.8902, "step": 3741 }, { "epoch": 1.6947463768115942, "grad_norm": 0.712140275606331, "learning_rate": 6.68740698969397e-05, "loss": 0.884, "step": 3742 }, { "epoch": 1.6951992753623188, "grad_norm": 0.734446142708722, "learning_rate": 6.686470087023514e-05, "loss": 0.8811, "step": 3743 }, { "epoch": 1.6956521739130435, "grad_norm": 0.6765866848639488, "learning_rate": 6.685532915782388e-05, "loss": 0.8814, "step": 3744 }, { "epoch": 1.696105072463768, "grad_norm": 0.6221857233754057, "learning_rate": 6.68459547606428e-05, "loss": 0.9161, "step": 3745 }, { "epoch": 1.6965579710144927, "grad_norm": 0.570456652765803, "learning_rate": 6.683657767962908e-05, "loss": 0.8915, "step": 3746 }, { "epoch": 1.6970108695652173, "grad_norm": 0.5437135044087242, "learning_rate": 6.682719791572016e-05, "loss": 0.8893, "step": 3747 }, { "epoch": 1.697463768115942, "grad_norm": 0.48455879553782144, "learning_rate": 6.681781546985376e-05, "loss": 0.904, "step": 3748 }, { "epoch": 1.6979166666666665, "grad_norm": 0.40277592535383394, "learning_rate": 6.680843034296785e-05, "loss": 0.8891, "step": 3749 }, { "epoch": 1.6983695652173914, "grad_norm": 0.36547671512665725, "learning_rate": 6.679904253600066e-05, "loss": 0.8898, "step": 3750 }, { "epoch": 1.698822463768116, "grad_norm": 0.3716511237225048, "learning_rate": 6.678965204989073e-05, "loss": 0.8957, "step": 3751 }, { "epoch": 1.6992753623188406, "grad_norm": 0.289337741486804, "learning_rate": 6.678025888557684e-05, "loss": 0.8882, "step": 3752 }, { "epoch": 1.6997282608695652, "grad_norm": 0.311273699170471, "learning_rate": 6.677086304399803e-05, "loss": 0.884, "step": 3753 }, { "epoch": 1.7001811594202898, "grad_norm": 0.449479141590849, "learning_rate": 6.676146452609364e-05, "loss": 0.871, "step": 3754 }, { "epoch": 1.7006340579710146, "grad_norm": 0.6352300648762417, "learning_rate": 6.67520633328032e-05, "loss": 0.9046, "step": 3755 }, { "epoch": 1.7010869565217392, "grad_norm": 0.4912530442971865, "learning_rate": 6.67426594650666e-05, "loss": 0.8731, "step": 3756 }, { "epoch": 1.7015398550724639, "grad_norm": 0.32483501034778844, "learning_rate": 6.673325292382397e-05, "loss": 0.8902, "step": 3757 }, { "epoch": 1.7019927536231885, "grad_norm": 0.23426235533087403, "learning_rate": 6.672384371001569e-05, "loss": 0.8868, "step": 3758 }, { "epoch": 1.702445652173913, "grad_norm": 0.25940831404948206, "learning_rate": 6.671443182458238e-05, "loss": 0.8945, "step": 3759 }, { "epoch": 1.7028985507246377, "grad_norm": 0.308067462400355, "learning_rate": 6.670501726846501e-05, "loss": 0.8631, "step": 3760 }, { "epoch": 1.7033514492753623, "grad_norm": 0.38238866724868886, "learning_rate": 6.669560004260474e-05, "loss": 0.889, "step": 3761 }, { "epoch": 1.703804347826087, "grad_norm": 0.4426924159395077, "learning_rate": 6.668618014794302e-05, "loss": 0.8819, "step": 3762 }, { "epoch": 1.7042572463768115, "grad_norm": 0.44720842570041547, "learning_rate": 6.66767575854216e-05, "loss": 0.8806, "step": 3763 }, { "epoch": 1.7047101449275361, "grad_norm": 0.4507356935794973, "learning_rate": 6.666733235598245e-05, "loss": 0.8859, "step": 3764 }, { "epoch": 1.7051630434782608, "grad_norm": 0.5058027658905285, "learning_rate": 6.665790446056781e-05, "loss": 0.9022, "step": 3765 }, { "epoch": 1.7056159420289854, "grad_norm": 0.492514617113839, "learning_rate": 6.664847390012023e-05, "loss": 0.8689, "step": 3766 }, { "epoch": 1.7060688405797102, "grad_norm": 0.5105729465000927, "learning_rate": 6.663904067558247e-05, "loss": 0.9034, "step": 3767 }, { "epoch": 1.7065217391304348, "grad_norm": 0.45092736636956865, "learning_rate": 6.662960478789762e-05, "loss": 0.8933, "step": 3768 }, { "epoch": 1.7069746376811594, "grad_norm": 0.43887127124480135, "learning_rate": 6.662016623800899e-05, "loss": 0.8958, "step": 3769 }, { "epoch": 1.707427536231884, "grad_norm": 0.4770927321130078, "learning_rate": 6.661072502686014e-05, "loss": 0.8755, "step": 3770 }, { "epoch": 1.7078804347826086, "grad_norm": 0.4751436164962602, "learning_rate": 6.660128115539494e-05, "loss": 0.899, "step": 3771 }, { "epoch": 1.7083333333333335, "grad_norm": 0.7383510579898154, "learning_rate": 6.659183462455751e-05, "loss": 0.8983, "step": 3772 }, { "epoch": 1.708786231884058, "grad_norm": 0.3850739782636564, "learning_rate": 6.658238543529225e-05, "loss": 0.8981, "step": 3773 }, { "epoch": 1.7092391304347827, "grad_norm": 0.3351356811925701, "learning_rate": 6.657293358854376e-05, "loss": 0.8967, "step": 3774 }, { "epoch": 1.7096920289855073, "grad_norm": 0.3984116547613819, "learning_rate": 6.656347908525704e-05, "loss": 0.9023, "step": 3775 }, { "epoch": 1.710144927536232, "grad_norm": 0.589168818899834, "learning_rate": 6.655402192637719e-05, "loss": 0.902, "step": 3776 }, { "epoch": 1.7105978260869565, "grad_norm": 0.7681973544915057, "learning_rate": 6.654456211284971e-05, "loss": 0.9062, "step": 3777 }, { "epoch": 1.7110507246376812, "grad_norm": 0.9193066800856465, "learning_rate": 6.653509964562027e-05, "loss": 0.874, "step": 3778 }, { "epoch": 1.7115036231884058, "grad_norm": 0.9892699710914136, "learning_rate": 6.652563452563487e-05, "loss": 0.8891, "step": 3779 }, { "epoch": 1.7119565217391304, "grad_norm": 0.972335014437011, "learning_rate": 6.651616675383977e-05, "loss": 0.9076, "step": 3780 }, { "epoch": 1.712409420289855, "grad_norm": 0.9603585376734484, "learning_rate": 6.650669633118144e-05, "loss": 0.891, "step": 3781 }, { "epoch": 1.7128623188405796, "grad_norm": 0.8020308527637141, "learning_rate": 6.64972232586067e-05, "loss": 0.8754, "step": 3782 }, { "epoch": 1.7133152173913042, "grad_norm": 0.5262732267857766, "learning_rate": 6.648774753706254e-05, "loss": 0.893, "step": 3783 }, { "epoch": 1.7137681159420288, "grad_norm": 0.31484817189849107, "learning_rate": 6.64782691674963e-05, "loss": 0.9114, "step": 3784 }, { "epoch": 1.7142210144927537, "grad_norm": 0.362931214801955, "learning_rate": 6.646878815085552e-05, "loss": 0.8994, "step": 3785 }, { "epoch": 1.7146739130434783, "grad_norm": 0.5755801837619423, "learning_rate": 6.645930448808806e-05, "loss": 0.8941, "step": 3786 }, { "epoch": 1.7151268115942029, "grad_norm": 0.7785970034069141, "learning_rate": 6.644981818014199e-05, "loss": 0.9061, "step": 3787 }, { "epoch": 1.7155797101449275, "grad_norm": 0.9107437685256503, "learning_rate": 6.644032922796568e-05, "loss": 0.8922, "step": 3788 }, { "epoch": 1.7160326086956523, "grad_norm": 0.8874793470484384, "learning_rate": 6.643083763250776e-05, "loss": 0.8758, "step": 3789 }, { "epoch": 1.716485507246377, "grad_norm": 0.7548884177052809, "learning_rate": 6.642134339471713e-05, "loss": 0.8975, "step": 3790 }, { "epoch": 1.7169384057971016, "grad_norm": 0.5656236733278245, "learning_rate": 6.641184651554292e-05, "loss": 0.8907, "step": 3791 }, { "epoch": 1.7173913043478262, "grad_norm": 0.36537418067100014, "learning_rate": 6.640234699593457e-05, "loss": 0.9021, "step": 3792 }, { "epoch": 1.7178442028985508, "grad_norm": 0.32226796818910797, "learning_rate": 6.639284483684176e-05, "loss": 0.8651, "step": 3793 }, { "epoch": 1.7182971014492754, "grad_norm": 0.313162975862179, "learning_rate": 6.638334003921441e-05, "loss": 0.8846, "step": 3794 }, { "epoch": 1.71875, "grad_norm": 0.420435636250305, "learning_rate": 6.637383260400276e-05, "loss": 0.8707, "step": 3795 }, { "epoch": 1.7192028985507246, "grad_norm": 0.511447689940874, "learning_rate": 6.636432253215728e-05, "loss": 0.8797, "step": 3796 }, { "epoch": 1.7196557971014492, "grad_norm": 0.5312820076535464, "learning_rate": 6.635480982462868e-05, "loss": 0.9031, "step": 3797 }, { "epoch": 1.7201086956521738, "grad_norm": 0.49610447161334265, "learning_rate": 6.634529448236798e-05, "loss": 0.893, "step": 3798 }, { "epoch": 1.7205615942028984, "grad_norm": 0.40259253404869155, "learning_rate": 6.633577650632646e-05, "loss": 0.891, "step": 3799 }, { "epoch": 1.721014492753623, "grad_norm": 0.44230173224514946, "learning_rate": 6.632625589745561e-05, "loss": 0.886, "step": 3800 }, { "epoch": 1.7214673913043477, "grad_norm": 0.37563562065810535, "learning_rate": 6.631673265670726e-05, "loss": 0.8746, "step": 3801 }, { "epoch": 1.7219202898550725, "grad_norm": 0.3458848056939215, "learning_rate": 6.630720678503343e-05, "loss": 0.8624, "step": 3802 }, { "epoch": 1.7223731884057971, "grad_norm": 0.3130285068454116, "learning_rate": 6.629767828338645e-05, "loss": 0.894, "step": 3803 }, { "epoch": 1.7228260869565217, "grad_norm": 0.28954060381106717, "learning_rate": 6.628814715271891e-05, "loss": 0.8928, "step": 3804 }, { "epoch": 1.7232789855072463, "grad_norm": 0.3676513491124775, "learning_rate": 6.627861339398364e-05, "loss": 0.8972, "step": 3805 }, { "epoch": 1.7237318840579712, "grad_norm": 0.45239294371601413, "learning_rate": 6.626907700813373e-05, "loss": 0.8952, "step": 3806 }, { "epoch": 1.7241847826086958, "grad_norm": 0.45567942019577407, "learning_rate": 6.625953799612259e-05, "loss": 0.8784, "step": 3807 }, { "epoch": 1.7246376811594204, "grad_norm": 0.48659700281104856, "learning_rate": 6.62499963589038e-05, "loss": 0.8817, "step": 3808 }, { "epoch": 1.725090579710145, "grad_norm": 0.459869594072264, "learning_rate": 6.624045209743129e-05, "loss": 0.886, "step": 3809 }, { "epoch": 1.7255434782608696, "grad_norm": 0.5095623529156305, "learning_rate": 6.62309052126592e-05, "loss": 0.873, "step": 3810 }, { "epoch": 1.7259963768115942, "grad_norm": 0.5099857514521614, "learning_rate": 6.622135570554195e-05, "loss": 0.8766, "step": 3811 }, { "epoch": 1.7264492753623188, "grad_norm": 0.443668315839144, "learning_rate": 6.621180357703423e-05, "loss": 0.8848, "step": 3812 }, { "epoch": 1.7269021739130435, "grad_norm": 0.38566103639049837, "learning_rate": 6.620224882809096e-05, "loss": 0.9055, "step": 3813 }, { "epoch": 1.727355072463768, "grad_norm": 0.3118600654759906, "learning_rate": 6.619269145966735e-05, "loss": 0.8823, "step": 3814 }, { "epoch": 1.7278079710144927, "grad_norm": 0.3667270162387291, "learning_rate": 6.618313147271887e-05, "loss": 0.8854, "step": 3815 }, { "epoch": 1.7282608695652173, "grad_norm": 0.4175328104243543, "learning_rate": 6.617356886820125e-05, "loss": 0.8728, "step": 3816 }, { "epoch": 1.728713768115942, "grad_norm": 0.3643632508348047, "learning_rate": 6.616400364707048e-05, "loss": 0.9174, "step": 3817 }, { "epoch": 1.7291666666666665, "grad_norm": 0.36662449423762056, "learning_rate": 6.615443581028279e-05, "loss": 0.8922, "step": 3818 }, { "epoch": 1.7296195652173914, "grad_norm": 0.5735872365191491, "learning_rate": 6.614486535879472e-05, "loss": 0.9013, "step": 3819 }, { "epoch": 1.730072463768116, "grad_norm": 0.6119092368066701, "learning_rate": 6.613529229356303e-05, "loss": 0.8745, "step": 3820 }, { "epoch": 1.7305253623188406, "grad_norm": 0.47956518830418937, "learning_rate": 6.612571661554474e-05, "loss": 0.8839, "step": 3821 }, { "epoch": 1.7309782608695652, "grad_norm": 0.36601784815307004, "learning_rate": 6.611613832569717e-05, "loss": 0.8974, "step": 3822 }, { "epoch": 1.7314311594202898, "grad_norm": 0.38228569175484217, "learning_rate": 6.610655742497787e-05, "loss": 0.8949, "step": 3823 }, { "epoch": 1.7318840579710146, "grad_norm": 0.3587435038626096, "learning_rate": 6.609697391434465e-05, "loss": 0.8872, "step": 3824 }, { "epoch": 1.7323369565217392, "grad_norm": 0.296988194231477, "learning_rate": 6.60873877947556e-05, "loss": 0.8799, "step": 3825 }, { "epoch": 1.7327898550724639, "grad_norm": 0.364422584264065, "learning_rate": 6.607779906716905e-05, "loss": 0.8742, "step": 3826 }, { "epoch": 1.7332427536231885, "grad_norm": 0.4312662664416277, "learning_rate": 6.60682077325436e-05, "loss": 0.8749, "step": 3827 }, { "epoch": 1.733695652173913, "grad_norm": 0.32459276181569413, "learning_rate": 6.605861379183811e-05, "loss": 0.8697, "step": 3828 }, { "epoch": 1.7341485507246377, "grad_norm": 0.32945375695521345, "learning_rate": 6.604901724601171e-05, "loss": 0.8919, "step": 3829 }, { "epoch": 1.7346014492753623, "grad_norm": 0.4091555266392855, "learning_rate": 6.60394180960238e-05, "loss": 0.8816, "step": 3830 }, { "epoch": 1.735054347826087, "grad_norm": 0.4447448597905667, "learning_rate": 6.602981634283396e-05, "loss": 0.907, "step": 3831 }, { "epoch": 1.7355072463768115, "grad_norm": 0.43435989875141123, "learning_rate": 6.602021198740217e-05, "loss": 0.8818, "step": 3832 }, { "epoch": 1.7359601449275361, "grad_norm": 0.3778859936257223, "learning_rate": 6.601060503068853e-05, "loss": 0.8946, "step": 3833 }, { "epoch": 1.7364130434782608, "grad_norm": 0.354961436672549, "learning_rate": 6.600099547365348e-05, "loss": 0.8895, "step": 3834 }, { "epoch": 1.7368659420289854, "grad_norm": 0.35610606367592945, "learning_rate": 6.599138331725774e-05, "loss": 0.8843, "step": 3835 }, { "epoch": 1.7373188405797102, "grad_norm": 0.38630591890660676, "learning_rate": 6.59817685624622e-05, "loss": 0.8909, "step": 3836 }, { "epoch": 1.7377717391304348, "grad_norm": 0.38263217744243555, "learning_rate": 6.597215121022809e-05, "loss": 0.9002, "step": 3837 }, { "epoch": 1.7382246376811594, "grad_norm": 0.34857726087695184, "learning_rate": 6.596253126151688e-05, "loss": 0.8996, "step": 3838 }, { "epoch": 1.738677536231884, "grad_norm": 0.3986990515120951, "learning_rate": 6.595290871729027e-05, "loss": 0.8833, "step": 3839 }, { "epoch": 1.7391304347826086, "grad_norm": 0.4732464243321195, "learning_rate": 6.594328357851024e-05, "loss": 0.8875, "step": 3840 }, { "epoch": 1.7395833333333335, "grad_norm": 0.5684469093001615, "learning_rate": 6.593365584613906e-05, "loss": 0.8912, "step": 3841 }, { "epoch": 1.740036231884058, "grad_norm": 0.6616395368480343, "learning_rate": 6.59240255211392e-05, "loss": 0.8916, "step": 3842 }, { "epoch": 1.7404891304347827, "grad_norm": 0.7334493082259675, "learning_rate": 6.591439260447342e-05, "loss": 0.8945, "step": 3843 }, { "epoch": 1.7409420289855073, "grad_norm": 0.7796691773654347, "learning_rate": 6.590475709710476e-05, "loss": 0.9146, "step": 3844 }, { "epoch": 1.741394927536232, "grad_norm": 0.8648127968885893, "learning_rate": 6.589511899999647e-05, "loss": 0.8846, "step": 3845 }, { "epoch": 1.7418478260869565, "grad_norm": 0.8676013943245497, "learning_rate": 6.588547831411211e-05, "loss": 0.8922, "step": 3846 }, { "epoch": 1.7423007246376812, "grad_norm": 0.7477991975082916, "learning_rate": 6.587583504041545e-05, "loss": 0.9053, "step": 3847 }, { "epoch": 1.7427536231884058, "grad_norm": 0.5029253643679066, "learning_rate": 6.586618917987058e-05, "loss": 0.8677, "step": 3848 }, { "epoch": 1.7432065217391304, "grad_norm": 0.32848526197249345, "learning_rate": 6.585654073344177e-05, "loss": 0.9058, "step": 3849 }, { "epoch": 1.743659420289855, "grad_norm": 0.38276660211822094, "learning_rate": 6.584688970209361e-05, "loss": 0.8915, "step": 3850 }, { "epoch": 1.7441123188405796, "grad_norm": 0.5794406030964959, "learning_rate": 6.583723608679095e-05, "loss": 0.8802, "step": 3851 }, { "epoch": 1.7445652173913042, "grad_norm": 0.7831601805347601, "learning_rate": 6.582757988849884e-05, "loss": 0.8725, "step": 3852 }, { "epoch": 1.7450181159420288, "grad_norm": 0.9533667242619959, "learning_rate": 6.581792110818263e-05, "loss": 0.9009, "step": 3853 }, { "epoch": 1.7454710144927537, "grad_norm": 1.0085949151461246, "learning_rate": 6.580825974680794e-05, "loss": 0.8902, "step": 3854 }, { "epoch": 1.7459239130434783, "grad_norm": 0.8939710640159605, "learning_rate": 6.579859580534063e-05, "loss": 0.8859, "step": 3855 }, { "epoch": 1.7463768115942029, "grad_norm": 0.6457616093885118, "learning_rate": 6.578892928474682e-05, "loss": 0.8917, "step": 3856 }, { "epoch": 1.7468297101449275, "grad_norm": 0.3368416092350408, "learning_rate": 6.577926018599289e-05, "loss": 0.878, "step": 3857 }, { "epoch": 1.7472826086956523, "grad_norm": 0.28884125248881104, "learning_rate": 6.576958851004546e-05, "loss": 0.8818, "step": 3858 }, { "epoch": 1.747735507246377, "grad_norm": 0.49781833130369485, "learning_rate": 6.575991425787143e-05, "loss": 0.8885, "step": 3859 }, { "epoch": 1.7481884057971016, "grad_norm": 0.626376580141757, "learning_rate": 6.575023743043795e-05, "loss": 0.8733, "step": 3860 }, { "epoch": 1.7486413043478262, "grad_norm": 0.7257713117794738, "learning_rate": 6.574055802871244e-05, "loss": 0.8784, "step": 3861 }, { "epoch": 1.7490942028985508, "grad_norm": 0.757863720501457, "learning_rate": 6.573087605366255e-05, "loss": 0.8876, "step": 3862 }, { "epoch": 1.7495471014492754, "grad_norm": 0.6443574542794556, "learning_rate": 6.57211915062562e-05, "loss": 0.8908, "step": 3863 }, { "epoch": 1.75, "grad_norm": 0.41456955714722216, "learning_rate": 6.571150438746157e-05, "loss": 0.9158, "step": 3864 }, { "epoch": 1.7504528985507246, "grad_norm": 0.24511322553265574, "learning_rate": 6.570181469824712e-05, "loss": 0.8749, "step": 3865 }, { "epoch": 1.7509057971014492, "grad_norm": 0.2945928086879495, "learning_rate": 6.569212243958153e-05, "loss": 0.8753, "step": 3866 }, { "epoch": 1.7513586956521738, "grad_norm": 0.4261234360116771, "learning_rate": 6.568242761243373e-05, "loss": 0.8956, "step": 3867 }, { "epoch": 1.7518115942028984, "grad_norm": 0.5562865196434247, "learning_rate": 6.567273021777296e-05, "loss": 0.9012, "step": 3868 }, { "epoch": 1.752264492753623, "grad_norm": 0.6042215139516067, "learning_rate": 6.566303025656865e-05, "loss": 0.9022, "step": 3869 }, { "epoch": 1.7527173913043477, "grad_norm": 0.6240117542804884, "learning_rate": 6.565332772979055e-05, "loss": 0.9073, "step": 3870 }, { "epoch": 1.7531702898550725, "grad_norm": 0.6219303039198829, "learning_rate": 6.564362263840861e-05, "loss": 0.8903, "step": 3871 }, { "epoch": 1.7536231884057971, "grad_norm": 0.5388908828679535, "learning_rate": 6.563391498339312e-05, "loss": 0.9022, "step": 3872 }, { "epoch": 1.7540760869565217, "grad_norm": 0.4037351383710571, "learning_rate": 6.56242047657145e-05, "loss": 0.9005, "step": 3873 }, { "epoch": 1.7545289855072463, "grad_norm": 0.28032026881928385, "learning_rate": 6.561449198634352e-05, "loss": 0.9026, "step": 3874 }, { "epoch": 1.7549818840579712, "grad_norm": 0.2480887203712311, "learning_rate": 6.56047766462512e-05, "loss": 0.877, "step": 3875 }, { "epoch": 1.7554347826086958, "grad_norm": 0.2934291444239557, "learning_rate": 6.559505874640875e-05, "loss": 0.8888, "step": 3876 }, { "epoch": 1.7558876811594204, "grad_norm": 0.3793050209091224, "learning_rate": 6.558533828778776e-05, "loss": 0.8905, "step": 3877 }, { "epoch": 1.756340579710145, "grad_norm": 0.5317125404445538, "learning_rate": 6.557561527135994e-05, "loss": 0.8839, "step": 3878 }, { "epoch": 1.7567934782608696, "grad_norm": 0.6289387965578646, "learning_rate": 6.556588969809733e-05, "loss": 0.896, "step": 3879 }, { "epoch": 1.7572463768115942, "grad_norm": 0.608651693148754, "learning_rate": 6.555616156897222e-05, "loss": 0.8926, "step": 3880 }, { "epoch": 1.7576992753623188, "grad_norm": 0.5505465516407465, "learning_rate": 6.554643088495715e-05, "loss": 0.9225, "step": 3881 }, { "epoch": 1.7581521739130435, "grad_norm": 0.4539580461886618, "learning_rate": 6.55366976470249e-05, "loss": 0.8797, "step": 3882 }, { "epoch": 1.758605072463768, "grad_norm": 0.3907590774909615, "learning_rate": 6.552696185614852e-05, "loss": 0.8735, "step": 3883 }, { "epoch": 1.7590579710144927, "grad_norm": 0.3333240406886703, "learning_rate": 6.55172235133013e-05, "loss": 0.8892, "step": 3884 }, { "epoch": 1.7595108695652173, "grad_norm": 0.30539583684421734, "learning_rate": 6.55074826194568e-05, "loss": 0.9053, "step": 3885 }, { "epoch": 1.759963768115942, "grad_norm": 0.37706642502035226, "learning_rate": 6.549773917558886e-05, "loss": 0.896, "step": 3886 }, { "epoch": 1.7604166666666665, "grad_norm": 0.4628369156229589, "learning_rate": 6.548799318267154e-05, "loss": 0.8995, "step": 3887 }, { "epoch": 1.7608695652173914, "grad_norm": 0.5079211585204734, "learning_rate": 6.547824464167913e-05, "loss": 0.9043, "step": 3888 }, { "epoch": 1.761322463768116, "grad_norm": 0.474887194199462, "learning_rate": 6.546849355358624e-05, "loss": 0.8867, "step": 3889 }, { "epoch": 1.7617753623188406, "grad_norm": 0.4385400154821647, "learning_rate": 6.54587399193677e-05, "loss": 0.9064, "step": 3890 }, { "epoch": 1.7622282608695652, "grad_norm": 0.43566497612929755, "learning_rate": 6.544898373999857e-05, "loss": 0.8872, "step": 3891 }, { "epoch": 1.7626811594202898, "grad_norm": 0.4074455185083816, "learning_rate": 6.543922501645424e-05, "loss": 0.8928, "step": 3892 }, { "epoch": 1.7631340579710146, "grad_norm": 0.4194058123994107, "learning_rate": 6.542946374971024e-05, "loss": 0.873, "step": 3893 }, { "epoch": 1.7635869565217392, "grad_norm": 0.373040277272108, "learning_rate": 6.541969994074246e-05, "loss": 0.8749, "step": 3894 }, { "epoch": 1.7640398550724639, "grad_norm": 0.28998272511494955, "learning_rate": 6.5409933590527e-05, "loss": 0.8824, "step": 3895 }, { "epoch": 1.7644927536231885, "grad_norm": 0.3274889350681042, "learning_rate": 6.54001647000402e-05, "loss": 0.8853, "step": 3896 }, { "epoch": 1.764945652173913, "grad_norm": 0.330427782022199, "learning_rate": 6.539039327025871e-05, "loss": 0.8749, "step": 3897 }, { "epoch": 1.7653985507246377, "grad_norm": 0.3915311917753186, "learning_rate": 6.538061930215936e-05, "loss": 0.8884, "step": 3898 }, { "epoch": 1.7658514492753623, "grad_norm": 0.49342269893778534, "learning_rate": 6.537084279671927e-05, "loss": 0.8836, "step": 3899 }, { "epoch": 1.766304347826087, "grad_norm": 0.5328904649523796, "learning_rate": 6.536106375491583e-05, "loss": 0.8912, "step": 3900 }, { "epoch": 1.7667572463768115, "grad_norm": 0.5437363369705362, "learning_rate": 6.535128217772667e-05, "loss": 0.9012, "step": 3901 }, { "epoch": 1.7672101449275361, "grad_norm": 0.46916491026413065, "learning_rate": 6.534149806612962e-05, "loss": 0.8786, "step": 3902 }, { "epoch": 1.7676630434782608, "grad_norm": 0.3504979709379077, "learning_rate": 6.53317114211029e-05, "loss": 0.8731, "step": 3903 }, { "epoch": 1.7681159420289854, "grad_norm": 0.2644922566619548, "learning_rate": 6.532192224362482e-05, "loss": 0.8707, "step": 3904 }, { "epoch": 1.7685688405797102, "grad_norm": 0.27703364461510044, "learning_rate": 6.531213053467404e-05, "loss": 0.9046, "step": 3905 }, { "epoch": 1.7690217391304348, "grad_norm": 0.3436247865265645, "learning_rate": 6.530233629522948e-05, "loss": 0.8899, "step": 3906 }, { "epoch": 1.7694746376811594, "grad_norm": 0.38035543859167686, "learning_rate": 6.529253952627026e-05, "loss": 0.9025, "step": 3907 }, { "epoch": 1.769927536231884, "grad_norm": 0.46927481259597587, "learning_rate": 6.528274022877578e-05, "loss": 0.8907, "step": 3908 }, { "epoch": 1.7703804347826086, "grad_norm": 0.4838201624791859, "learning_rate": 6.527293840372569e-05, "loss": 0.9057, "step": 3909 }, { "epoch": 1.7708333333333335, "grad_norm": 0.43629892291419087, "learning_rate": 6.526313405209991e-05, "loss": 0.8756, "step": 3910 }, { "epoch": 1.771286231884058, "grad_norm": 0.3573551854374048, "learning_rate": 6.525332717487859e-05, "loss": 0.8629, "step": 3911 }, { "epoch": 1.7717391304347827, "grad_norm": 0.315190929150106, "learning_rate": 6.524351777304212e-05, "loss": 0.8658, "step": 3912 }, { "epoch": 1.7721920289855073, "grad_norm": 0.39328428246305913, "learning_rate": 6.523370584757118e-05, "loss": 0.8916, "step": 3913 }, { "epoch": 1.772644927536232, "grad_norm": 0.42116628512982124, "learning_rate": 6.522389139944668e-05, "loss": 0.8924, "step": 3914 }, { "epoch": 1.7730978260869565, "grad_norm": 0.39253244267361237, "learning_rate": 6.52140744296498e-05, "loss": 0.8825, "step": 3915 }, { "epoch": 1.7735507246376812, "grad_norm": 0.43289871660009993, "learning_rate": 6.520425493916194e-05, "loss": 0.9148, "step": 3916 }, { "epoch": 1.7740036231884058, "grad_norm": 0.5832451064714166, "learning_rate": 6.519443292896476e-05, "loss": 0.9048, "step": 3917 }, { "epoch": 1.7744565217391304, "grad_norm": 0.6513424078976435, "learning_rate": 6.51846084000402e-05, "loss": 0.8905, "step": 3918 }, { "epoch": 1.774909420289855, "grad_norm": 0.751736800074061, "learning_rate": 6.517478135337044e-05, "loss": 0.8669, "step": 3919 }, { "epoch": 1.7753623188405796, "grad_norm": 0.7897514308876916, "learning_rate": 6.516495178993788e-05, "loss": 0.896, "step": 3920 }, { "epoch": 1.7758152173913042, "grad_norm": 0.7496465053266349, "learning_rate": 6.515511971072523e-05, "loss": 0.8922, "step": 3921 }, { "epoch": 1.7762681159420288, "grad_norm": 0.5908818389215882, "learning_rate": 6.51452851167154e-05, "loss": 0.8795, "step": 3922 }, { "epoch": 1.7767210144927537, "grad_norm": 0.3734001928258197, "learning_rate": 6.513544800889155e-05, "loss": 0.8846, "step": 3923 }, { "epoch": 1.7771739130434783, "grad_norm": 0.23757577401665236, "learning_rate": 6.512560838823715e-05, "loss": 0.8884, "step": 3924 }, { "epoch": 1.7776268115942029, "grad_norm": 0.24691657338501508, "learning_rate": 6.511576625573585e-05, "loss": 0.8978, "step": 3925 }, { "epoch": 1.7780797101449275, "grad_norm": 0.3034376252221564, "learning_rate": 6.510592161237162e-05, "loss": 0.8632, "step": 3926 }, { "epoch": 1.7785326086956523, "grad_norm": 0.4201959378319555, "learning_rate": 6.50960744591286e-05, "loss": 0.9076, "step": 3927 }, { "epoch": 1.778985507246377, "grad_norm": 0.5216387066501867, "learning_rate": 6.508622479699126e-05, "loss": 0.8885, "step": 3928 }, { "epoch": 1.7794384057971016, "grad_norm": 0.539244510092113, "learning_rate": 6.507637262694427e-05, "loss": 0.8839, "step": 3929 }, { "epoch": 1.7798913043478262, "grad_norm": 0.5198226089505594, "learning_rate": 6.506651794997257e-05, "loss": 0.8876, "step": 3930 }, { "epoch": 1.7803442028985508, "grad_norm": 0.4593263992418376, "learning_rate": 6.505666076706136e-05, "loss": 0.8874, "step": 3931 }, { "epoch": 1.7807971014492754, "grad_norm": 0.4083379283358627, "learning_rate": 6.504680107919606e-05, "loss": 0.8784, "step": 3932 }, { "epoch": 1.78125, "grad_norm": 0.393327707100185, "learning_rate": 6.503693888736238e-05, "loss": 0.892, "step": 3933 }, { "epoch": 1.7817028985507246, "grad_norm": 0.4116998122461958, "learning_rate": 6.502707419254622e-05, "loss": 0.8677, "step": 3934 }, { "epoch": 1.7821557971014492, "grad_norm": 0.4267185199877024, "learning_rate": 6.50172069957338e-05, "loss": 0.8753, "step": 3935 }, { "epoch": 1.7826086956521738, "grad_norm": 0.37499536059392297, "learning_rate": 6.500733729791156e-05, "loss": 0.8718, "step": 3936 }, { "epoch": 1.7830615942028984, "grad_norm": 0.3174441180638176, "learning_rate": 6.499746510006618e-05, "loss": 0.8939, "step": 3937 }, { "epoch": 1.783514492753623, "grad_norm": 0.3187504677205597, "learning_rate": 6.498759040318461e-05, "loss": 0.8893, "step": 3938 }, { "epoch": 1.7839673913043477, "grad_norm": 0.3612454101611358, "learning_rate": 6.497771320825402e-05, "loss": 0.8951, "step": 3939 }, { "epoch": 1.7844202898550725, "grad_norm": 0.39165888097712503, "learning_rate": 6.496783351626187e-05, "loss": 0.8882, "step": 3940 }, { "epoch": 1.7848731884057971, "grad_norm": 0.41554689141885337, "learning_rate": 6.495795132819585e-05, "loss": 0.8869, "step": 3941 }, { "epoch": 1.7853260869565217, "grad_norm": 0.420359486312801, "learning_rate": 6.494806664504387e-05, "loss": 0.8878, "step": 3942 }, { "epoch": 1.7857789855072463, "grad_norm": 0.3545926483785166, "learning_rate": 6.493817946779412e-05, "loss": 0.8681, "step": 3943 }, { "epoch": 1.7862318840579712, "grad_norm": 0.26526887460954823, "learning_rate": 6.492828979743507e-05, "loss": 0.8878, "step": 3944 }, { "epoch": 1.7866847826086958, "grad_norm": 0.2976277068866101, "learning_rate": 6.491839763495538e-05, "loss": 0.8692, "step": 3945 }, { "epoch": 1.7871376811594204, "grad_norm": 0.29211345945496314, "learning_rate": 6.490850298134399e-05, "loss": 0.8982, "step": 3946 }, { "epoch": 1.787590579710145, "grad_norm": 0.28846380501905294, "learning_rate": 6.48986058375901e-05, "loss": 0.8796, "step": 3947 }, { "epoch": 1.7880434782608696, "grad_norm": 0.26772846110499343, "learning_rate": 6.488870620468312e-05, "loss": 0.87, "step": 3948 }, { "epoch": 1.7884963768115942, "grad_norm": 0.3236276294903073, "learning_rate": 6.487880408361273e-05, "loss": 0.8846, "step": 3949 }, { "epoch": 1.7889492753623188, "grad_norm": 0.3869154688087481, "learning_rate": 6.486889947536888e-05, "loss": 0.91, "step": 3950 }, { "epoch": 1.7894021739130435, "grad_norm": 0.4266140457711818, "learning_rate": 6.485899238094175e-05, "loss": 0.888, "step": 3951 }, { "epoch": 1.789855072463768, "grad_norm": 0.47476272385696905, "learning_rate": 6.484908280132175e-05, "loss": 0.8772, "step": 3952 }, { "epoch": 1.7903079710144927, "grad_norm": 0.5036946472632191, "learning_rate": 6.483917073749957e-05, "loss": 0.8552, "step": 3953 }, { "epoch": 1.7907608695652173, "grad_norm": 0.5351271243500114, "learning_rate": 6.482925619046614e-05, "loss": 0.8662, "step": 3954 }, { "epoch": 1.791213768115942, "grad_norm": 0.48884161554433947, "learning_rate": 6.481933916121261e-05, "loss": 0.8782, "step": 3955 }, { "epoch": 1.7916666666666665, "grad_norm": 0.4206527303353511, "learning_rate": 6.480941965073041e-05, "loss": 0.8741, "step": 3956 }, { "epoch": 1.7921195652173914, "grad_norm": 0.4068737144245641, "learning_rate": 6.479949766001122e-05, "loss": 0.9273, "step": 3957 }, { "epoch": 1.792572463768116, "grad_norm": 0.4458650415824046, "learning_rate": 6.478957319004695e-05, "loss": 0.8976, "step": 3958 }, { "epoch": 1.7930253623188406, "grad_norm": 0.43619775186612025, "learning_rate": 6.477964624182975e-05, "loss": 0.8822, "step": 3959 }, { "epoch": 1.7934782608695652, "grad_norm": 0.45716922965671436, "learning_rate": 6.476971681635209e-05, "loss": 0.897, "step": 3960 }, { "epoch": 1.7939311594202898, "grad_norm": 0.47346138031104884, "learning_rate": 6.475978491460656e-05, "loss": 0.8726, "step": 3961 }, { "epoch": 1.7943840579710146, "grad_norm": 0.42451165018368914, "learning_rate": 6.47498505375861e-05, "loss": 0.8991, "step": 3962 }, { "epoch": 1.7948369565217392, "grad_norm": 0.44256775140757354, "learning_rate": 6.473991368628386e-05, "loss": 0.9102, "step": 3963 }, { "epoch": 1.7952898550724639, "grad_norm": 0.521453848674066, "learning_rate": 6.472997436169324e-05, "loss": 0.8948, "step": 3964 }, { "epoch": 1.7957427536231885, "grad_norm": 0.6587566580419221, "learning_rate": 6.472003256480791e-05, "loss": 0.8751, "step": 3965 }, { "epoch": 1.796195652173913, "grad_norm": 0.7874226468062187, "learning_rate": 6.471008829662175e-05, "loss": 0.9028, "step": 3966 }, { "epoch": 1.7966485507246377, "grad_norm": 0.8275949312777132, "learning_rate": 6.47001415581289e-05, "loss": 0.8758, "step": 3967 }, { "epoch": 1.7971014492753623, "grad_norm": 0.720641841654992, "learning_rate": 6.469019235032376e-05, "loss": 0.8926, "step": 3968 }, { "epoch": 1.797554347826087, "grad_norm": 0.5298562492774062, "learning_rate": 6.468024067420097e-05, "loss": 0.8735, "step": 3969 }, { "epoch": 1.7980072463768115, "grad_norm": 0.40660306539861896, "learning_rate": 6.46702865307554e-05, "loss": 0.8783, "step": 3970 }, { "epoch": 1.7984601449275361, "grad_norm": 0.28315738903729415, "learning_rate": 6.46603299209822e-05, "loss": 0.8704, "step": 3971 }, { "epoch": 1.7989130434782608, "grad_norm": 0.31461298252824155, "learning_rate": 6.465037084587673e-05, "loss": 0.876, "step": 3972 }, { "epoch": 1.7993659420289854, "grad_norm": 0.4132349766366711, "learning_rate": 6.464040930643464e-05, "loss": 0.8633, "step": 3973 }, { "epoch": 1.7998188405797102, "grad_norm": 0.5612489473944954, "learning_rate": 6.463044530365178e-05, "loss": 0.8963, "step": 3974 }, { "epoch": 1.8002717391304348, "grad_norm": 0.6774615451367889, "learning_rate": 6.462047883852427e-05, "loss": 0.8938, "step": 3975 }, { "epoch": 1.8007246376811594, "grad_norm": 0.6359056256178138, "learning_rate": 6.461050991204846e-05, "loss": 0.873, "step": 3976 }, { "epoch": 1.801177536231884, "grad_norm": 0.5120135765485834, "learning_rate": 6.4600538525221e-05, "loss": 0.8963, "step": 3977 }, { "epoch": 1.8016304347826086, "grad_norm": 0.38387125748514217, "learning_rate": 6.45905646790387e-05, "loss": 0.8807, "step": 3978 }, { "epoch": 1.8020833333333335, "grad_norm": 0.28375507782414905, "learning_rate": 6.458058837449871e-05, "loss": 0.8638, "step": 3979 }, { "epoch": 1.802536231884058, "grad_norm": 0.3173100788838732, "learning_rate": 6.457060961259833e-05, "loss": 0.8818, "step": 3980 }, { "epoch": 1.8029891304347827, "grad_norm": 0.37562859593815806, "learning_rate": 6.456062839433518e-05, "loss": 0.8969, "step": 3981 }, { "epoch": 1.8034420289855073, "grad_norm": 0.43133412958653106, "learning_rate": 6.455064472070709e-05, "loss": 0.8841, "step": 3982 }, { "epoch": 1.803894927536232, "grad_norm": 0.49503779355439453, "learning_rate": 6.454065859271213e-05, "loss": 0.8947, "step": 3983 }, { "epoch": 1.8043478260869565, "grad_norm": 0.5845567146025965, "learning_rate": 6.453067001134866e-05, "loss": 0.9128, "step": 3984 }, { "epoch": 1.8048007246376812, "grad_norm": 0.5709894271612723, "learning_rate": 6.452067897761522e-05, "loss": 0.8774, "step": 3985 }, { "epoch": 1.8052536231884058, "grad_norm": 0.5076060945251746, "learning_rate": 6.451068549251066e-05, "loss": 0.881, "step": 3986 }, { "epoch": 1.8057065217391304, "grad_norm": 0.38760307629684937, "learning_rate": 6.450068955703403e-05, "loss": 0.8862, "step": 3987 }, { "epoch": 1.806159420289855, "grad_norm": 0.33420633101787894, "learning_rate": 6.449069117218463e-05, "loss": 0.8693, "step": 3988 }, { "epoch": 1.8066123188405796, "grad_norm": 0.2620354438456043, "learning_rate": 6.448069033896203e-05, "loss": 0.892, "step": 3989 }, { "epoch": 1.8070652173913042, "grad_norm": 0.2865692444545845, "learning_rate": 6.447068705836602e-05, "loss": 0.8785, "step": 3990 }, { "epoch": 1.8075181159420288, "grad_norm": 0.2717575830978325, "learning_rate": 6.446068133139666e-05, "loss": 0.8697, "step": 3991 }, { "epoch": 1.8079710144927537, "grad_norm": 0.2876234012158772, "learning_rate": 6.445067315905421e-05, "loss": 0.8567, "step": 3992 }, { "epoch": 1.8084239130434783, "grad_norm": 0.3188502361145313, "learning_rate": 6.444066254233923e-05, "loss": 0.8882, "step": 3993 }, { "epoch": 1.8088768115942029, "grad_norm": 0.29648423527798323, "learning_rate": 6.443064948225248e-05, "loss": 0.901, "step": 3994 }, { "epoch": 1.8093297101449275, "grad_norm": 0.30734916978927673, "learning_rate": 6.4420633979795e-05, "loss": 0.8708, "step": 3995 }, { "epoch": 1.8097826086956523, "grad_norm": 0.3220050167962907, "learning_rate": 6.441061603596803e-05, "loss": 0.9015, "step": 3996 }, { "epoch": 1.810235507246377, "grad_norm": 0.335620076547473, "learning_rate": 6.44005956517731e-05, "loss": 0.886, "step": 3997 }, { "epoch": 1.8106884057971016, "grad_norm": 0.48558286634771947, "learning_rate": 6.439057282821195e-05, "loss": 0.8736, "step": 3998 }, { "epoch": 1.8111413043478262, "grad_norm": 0.6404905658144691, "learning_rate": 6.438054756628659e-05, "loss": 0.8857, "step": 3999 }, { "epoch": 1.8115942028985508, "grad_norm": 0.7757804643750206, "learning_rate": 6.437051986699924e-05, "loss": 0.8819, "step": 4000 }, { "epoch": 1.8120471014492754, "grad_norm": 0.9569338270768512, "learning_rate": 6.436048973135242e-05, "loss": 0.9129, "step": 4001 }, { "epoch": 1.8125, "grad_norm": 1.0425509258949355, "learning_rate": 6.435045716034883e-05, "loss": 0.876, "step": 4002 }, { "epoch": 1.8129528985507246, "grad_norm": 0.8363069489767507, "learning_rate": 6.434042215499146e-05, "loss": 0.8858, "step": 4003 }, { "epoch": 1.8134057971014492, "grad_norm": 0.5447806167900245, "learning_rate": 6.43303847162835e-05, "loss": 0.9064, "step": 4004 }, { "epoch": 1.8138586956521738, "grad_norm": 0.40002788042081877, "learning_rate": 6.432034484522843e-05, "loss": 0.8887, "step": 4005 }, { "epoch": 1.8143115942028984, "grad_norm": 0.4250124313593499, "learning_rate": 6.431030254282995e-05, "loss": 0.8939, "step": 4006 }, { "epoch": 1.814764492753623, "grad_norm": 0.4801986727585039, "learning_rate": 6.430025781009202e-05, "loss": 0.8999, "step": 4007 }, { "epoch": 1.8152173913043477, "grad_norm": 0.5214733827445333, "learning_rate": 6.429021064801879e-05, "loss": 0.8832, "step": 4008 }, { "epoch": 1.8156702898550725, "grad_norm": 0.5659472152828524, "learning_rate": 6.42801610576147e-05, "loss": 0.8783, "step": 4009 }, { "epoch": 1.8161231884057971, "grad_norm": 0.5985187037180544, "learning_rate": 6.427010903988445e-05, "loss": 0.8864, "step": 4010 }, { "epoch": 1.8165760869565217, "grad_norm": 0.595812800310808, "learning_rate": 6.426005459583295e-05, "loss": 0.8895, "step": 4011 }, { "epoch": 1.8170289855072463, "grad_norm": 0.5143628036763662, "learning_rate": 6.424999772646533e-05, "loss": 0.8845, "step": 4012 }, { "epoch": 1.8174818840579712, "grad_norm": 0.36595023427273743, "learning_rate": 6.423993843278702e-05, "loss": 0.8958, "step": 4013 }, { "epoch": 1.8179347826086958, "grad_norm": 0.354441127804359, "learning_rate": 6.422987671580364e-05, "loss": 0.8778, "step": 4014 }, { "epoch": 1.8183876811594204, "grad_norm": 0.3601110512879736, "learning_rate": 6.42198125765211e-05, "loss": 0.8866, "step": 4015 }, { "epoch": 1.818840579710145, "grad_norm": 0.32661696596535056, "learning_rate": 6.420974601594552e-05, "loss": 0.8718, "step": 4016 }, { "epoch": 1.8192934782608696, "grad_norm": 0.3247409330888699, "learning_rate": 6.419967703508326e-05, "loss": 0.8638, "step": 4017 }, { "epoch": 1.8197463768115942, "grad_norm": 0.40663369838664204, "learning_rate": 6.418960563494095e-05, "loss": 0.8987, "step": 4018 }, { "epoch": 1.8201992753623188, "grad_norm": 0.524264046634495, "learning_rate": 6.417953181652543e-05, "loss": 0.8832, "step": 4019 }, { "epoch": 1.8206521739130435, "grad_norm": 0.4720269831472797, "learning_rate": 6.416945558084379e-05, "loss": 0.8925, "step": 4020 }, { "epoch": 1.821105072463768, "grad_norm": 0.49761436731115294, "learning_rate": 6.415937692890337e-05, "loss": 0.8874, "step": 4021 }, { "epoch": 1.8215579710144927, "grad_norm": 0.49586851708817076, "learning_rate": 6.414929586171177e-05, "loss": 0.9034, "step": 4022 }, { "epoch": 1.8220108695652173, "grad_norm": 0.4921151815559854, "learning_rate": 6.41392123802768e-05, "loss": 0.8713, "step": 4023 }, { "epoch": 1.822463768115942, "grad_norm": 0.44545371000929046, "learning_rate": 6.41291264856065e-05, "loss": 0.9045, "step": 4024 }, { "epoch": 1.8229166666666665, "grad_norm": 0.45494561943433676, "learning_rate": 6.411903817870919e-05, "loss": 0.8839, "step": 4025 }, { "epoch": 1.8233695652173914, "grad_norm": 0.39216370976589454, "learning_rate": 6.410894746059342e-05, "loss": 0.8781, "step": 4026 }, { "epoch": 1.823822463768116, "grad_norm": 0.37960732530030494, "learning_rate": 6.409885433226796e-05, "loss": 0.8757, "step": 4027 }, { "epoch": 1.8242753623188406, "grad_norm": 0.39025041672170524, "learning_rate": 6.408875879474186e-05, "loss": 0.8838, "step": 4028 }, { "epoch": 1.8247282608695652, "grad_norm": 0.4977756600224817, "learning_rate": 6.407866084902437e-05, "loss": 0.8843, "step": 4029 }, { "epoch": 1.8251811594202898, "grad_norm": 0.5724626377300363, "learning_rate": 6.406856049612499e-05, "loss": 0.8657, "step": 4030 }, { "epoch": 1.8256340579710146, "grad_norm": 0.5340711233108469, "learning_rate": 6.405845773705346e-05, "loss": 0.8933, "step": 4031 }, { "epoch": 1.8260869565217392, "grad_norm": 0.5086875869148381, "learning_rate": 6.404835257281981e-05, "loss": 0.8832, "step": 4032 }, { "epoch": 1.8265398550724639, "grad_norm": 0.5273706130818949, "learning_rate": 6.403824500443425e-05, "loss": 0.8878, "step": 4033 }, { "epoch": 1.8269927536231885, "grad_norm": 0.5389539405604055, "learning_rate": 6.402813503290724e-05, "loss": 0.9005, "step": 4034 }, { "epoch": 1.827445652173913, "grad_norm": 0.44757354584939857, "learning_rate": 6.401802265924949e-05, "loss": 0.8767, "step": 4035 }, { "epoch": 1.8278985507246377, "grad_norm": 0.3254101260952603, "learning_rate": 6.400790788447195e-05, "loss": 0.9003, "step": 4036 }, { "epoch": 1.8283514492753623, "grad_norm": 0.3838237058887761, "learning_rate": 6.399779070958581e-05, "loss": 0.8788, "step": 4037 }, { "epoch": 1.828804347826087, "grad_norm": 0.28034778387910725, "learning_rate": 6.398767113560253e-05, "loss": 0.879, "step": 4038 }, { "epoch": 1.8292572463768115, "grad_norm": 0.27828433651937357, "learning_rate": 6.397754916353373e-05, "loss": 0.8958, "step": 4039 }, { "epoch": 1.8297101449275361, "grad_norm": 0.36625016652978787, "learning_rate": 6.396742479439136e-05, "loss": 0.8997, "step": 4040 }, { "epoch": 1.8301630434782608, "grad_norm": 0.36654873422731205, "learning_rate": 6.395729802918754e-05, "loss": 0.8857, "step": 4041 }, { "epoch": 1.8306159420289854, "grad_norm": 0.3928525964843142, "learning_rate": 6.394716886893468e-05, "loss": 0.9013, "step": 4042 }, { "epoch": 1.8310688405797102, "grad_norm": 0.5203332996076627, "learning_rate": 6.39370373146454e-05, "loss": 0.8804, "step": 4043 }, { "epoch": 1.8315217391304348, "grad_norm": 0.5716177015864637, "learning_rate": 6.392690336733257e-05, "loss": 0.8906, "step": 4044 }, { "epoch": 1.8319746376811594, "grad_norm": 0.5696352615941238, "learning_rate": 6.391676702800929e-05, "loss": 0.8947, "step": 4045 }, { "epoch": 1.832427536231884, "grad_norm": 0.5349971027210692, "learning_rate": 6.390662829768894e-05, "loss": 0.8889, "step": 4046 }, { "epoch": 1.8328804347826086, "grad_norm": 0.4738122385001586, "learning_rate": 6.389648717738504e-05, "loss": 0.8932, "step": 4047 }, { "epoch": 1.8333333333333335, "grad_norm": 0.25969011711233375, "learning_rate": 6.388634366811146e-05, "loss": 0.873, "step": 4048 }, { "epoch": 1.833786231884058, "grad_norm": 0.3844845170772719, "learning_rate": 6.387619777088224e-05, "loss": 0.883, "step": 4049 }, { "epoch": 1.8342391304347827, "grad_norm": 0.3938968499526256, "learning_rate": 6.386604948671171e-05, "loss": 0.8878, "step": 4050 }, { "epoch": 1.8346920289855073, "grad_norm": 0.40456851001297056, "learning_rate": 6.385589881661439e-05, "loss": 0.8685, "step": 4051 }, { "epoch": 1.835144927536232, "grad_norm": 0.5034061898820381, "learning_rate": 6.384574576160507e-05, "loss": 0.884, "step": 4052 }, { "epoch": 1.8355978260869565, "grad_norm": 0.5472105777731303, "learning_rate": 6.383559032269874e-05, "loss": 0.8854, "step": 4053 }, { "epoch": 1.8360507246376812, "grad_norm": 0.5348967510703359, "learning_rate": 6.38254325009107e-05, "loss": 0.8958, "step": 4054 }, { "epoch": 1.8365036231884058, "grad_norm": 0.4744974632062329, "learning_rate": 6.38152722972564e-05, "loss": 0.877, "step": 4055 }, { "epoch": 1.8369565217391304, "grad_norm": 0.39877312769750217, "learning_rate": 6.38051097127516e-05, "loss": 0.8975, "step": 4056 }, { "epoch": 1.837409420289855, "grad_norm": 0.30438426969590976, "learning_rate": 6.379494474841225e-05, "loss": 0.9023, "step": 4057 }, { "epoch": 1.8378623188405796, "grad_norm": 0.30353579172458084, "learning_rate": 6.378477740525456e-05, "loss": 0.8939, "step": 4058 }, { "epoch": 1.8383152173913042, "grad_norm": 0.26098143770079685, "learning_rate": 6.377460768429502e-05, "loss": 0.875, "step": 4059 }, { "epoch": 1.8387681159420288, "grad_norm": 0.3008904378925055, "learning_rate": 6.376443558655024e-05, "loss": 0.8843, "step": 4060 }, { "epoch": 1.8392210144927537, "grad_norm": 0.3955636886417234, "learning_rate": 6.375426111303718e-05, "loss": 0.8845, "step": 4061 }, { "epoch": 1.8396739130434783, "grad_norm": 0.4092704592105104, "learning_rate": 6.374408426477301e-05, "loss": 0.8923, "step": 4062 }, { "epoch": 1.8401268115942029, "grad_norm": 0.49211880359615007, "learning_rate": 6.373390504277509e-05, "loss": 0.8927, "step": 4063 }, { "epoch": 1.8405797101449275, "grad_norm": 0.6276951955429657, "learning_rate": 6.372372344806109e-05, "loss": 0.8764, "step": 4064 }, { "epoch": 1.8410326086956523, "grad_norm": 0.7125137331591633, "learning_rate": 6.371353948164885e-05, "loss": 0.8905, "step": 4065 }, { "epoch": 1.841485507246377, "grad_norm": 0.7007442073438735, "learning_rate": 6.370335314455651e-05, "loss": 0.884, "step": 4066 }, { "epoch": 1.8419384057971016, "grad_norm": 0.6279429084990855, "learning_rate": 6.369316443780238e-05, "loss": 0.8951, "step": 4067 }, { "epoch": 1.8423913043478262, "grad_norm": 0.5054086691597123, "learning_rate": 6.368297336240506e-05, "loss": 0.8841, "step": 4068 }, { "epoch": 1.8428442028985508, "grad_norm": 0.40152233843859414, "learning_rate": 6.367277991938337e-05, "loss": 0.8937, "step": 4069 }, { "epoch": 1.8432971014492754, "grad_norm": 0.3782222622283842, "learning_rate": 6.366258410975635e-05, "loss": 0.8839, "step": 4070 }, { "epoch": 1.84375, "grad_norm": 0.3910895423982326, "learning_rate": 6.365238593454331e-05, "loss": 0.908, "step": 4071 }, { "epoch": 1.8442028985507246, "grad_norm": 0.3933582749822391, "learning_rate": 6.364218539476375e-05, "loss": 0.8862, "step": 4072 }, { "epoch": 1.8446557971014492, "grad_norm": 0.4893994731943934, "learning_rate": 6.363198249143746e-05, "loss": 0.9016, "step": 4073 }, { "epoch": 1.8451086956521738, "grad_norm": 0.6468381437825601, "learning_rate": 6.362177722558444e-05, "loss": 0.8817, "step": 4074 }, { "epoch": 1.8455615942028984, "grad_norm": 0.6842486535438017, "learning_rate": 6.361156959822492e-05, "loss": 0.8815, "step": 4075 }, { "epoch": 1.846014492753623, "grad_norm": 0.6496456276945853, "learning_rate": 6.360135961037936e-05, "loss": 0.8928, "step": 4076 }, { "epoch": 1.8464673913043477, "grad_norm": 0.550111976391273, "learning_rate": 6.359114726306847e-05, "loss": 0.8955, "step": 4077 }, { "epoch": 1.8469202898550725, "grad_norm": 0.4797343861896901, "learning_rate": 6.358093255731323e-05, "loss": 0.8866, "step": 4078 }, { "epoch": 1.8473731884057971, "grad_norm": 0.4284818335277306, "learning_rate": 6.357071549413478e-05, "loss": 0.8846, "step": 4079 }, { "epoch": 1.8478260869565217, "grad_norm": 0.43308709283966396, "learning_rate": 6.356049607455456e-05, "loss": 0.8931, "step": 4080 }, { "epoch": 1.8482789855072463, "grad_norm": 0.34085020816026235, "learning_rate": 6.355027429959421e-05, "loss": 0.8736, "step": 4081 }, { "epoch": 1.8487318840579712, "grad_norm": 0.28192661754365467, "learning_rate": 6.354005017027563e-05, "loss": 0.8947, "step": 4082 }, { "epoch": 1.8491847826086958, "grad_norm": 0.3429267885060807, "learning_rate": 6.352982368762091e-05, "loss": 0.8969, "step": 4083 }, { "epoch": 1.8496376811594204, "grad_norm": 0.2976788022149553, "learning_rate": 6.351959485265245e-05, "loss": 0.8861, "step": 4084 }, { "epoch": 1.850090579710145, "grad_norm": 0.35747006081838234, "learning_rate": 6.350936366639281e-05, "loss": 0.8934, "step": 4085 }, { "epoch": 1.8505434782608696, "grad_norm": 0.3777780767094651, "learning_rate": 6.349913012986482e-05, "loss": 0.8686, "step": 4086 }, { "epoch": 1.8509963768115942, "grad_norm": 0.4327459641072318, "learning_rate": 6.348889424409158e-05, "loss": 0.8799, "step": 4087 }, { "epoch": 1.8514492753623188, "grad_norm": 0.48127753631378756, "learning_rate": 6.347865601009635e-05, "loss": 0.8931, "step": 4088 }, { "epoch": 1.8519021739130435, "grad_norm": 0.5296842454266386, "learning_rate": 6.346841542890266e-05, "loss": 0.8967, "step": 4089 }, { "epoch": 1.852355072463768, "grad_norm": 0.6111359899705296, "learning_rate": 6.34581725015343e-05, "loss": 0.8932, "step": 4090 }, { "epoch": 1.8528079710144927, "grad_norm": 0.4723869052054687, "learning_rate": 6.344792722901527e-05, "loss": 0.8931, "step": 4091 }, { "epoch": 1.8532608695652173, "grad_norm": 0.40607084253631026, "learning_rate": 6.34376796123698e-05, "loss": 0.8767, "step": 4092 }, { "epoch": 1.853713768115942, "grad_norm": 0.40619694501290987, "learning_rate": 6.342742965262235e-05, "loss": 0.8732, "step": 4093 }, { "epoch": 1.8541666666666665, "grad_norm": 0.3645670392754444, "learning_rate": 6.341717735079763e-05, "loss": 0.8774, "step": 4094 }, { "epoch": 1.8546195652173914, "grad_norm": 0.27927266812917717, "learning_rate": 6.340692270792059e-05, "loss": 0.9207, "step": 4095 }, { "epoch": 1.855072463768116, "grad_norm": 0.261137512095884, "learning_rate": 6.339666572501639e-05, "loss": 0.8918, "step": 4096 }, { "epoch": 1.8555253623188406, "grad_norm": 0.2717542774340151, "learning_rate": 6.338640640311043e-05, "loss": 0.8738, "step": 4097 }, { "epoch": 1.8559782608695652, "grad_norm": 0.23922455349262112, "learning_rate": 6.33761447432284e-05, "loss": 0.8955, "step": 4098 }, { "epoch": 1.8564311594202898, "grad_norm": 0.3281663501419657, "learning_rate": 6.336588074639612e-05, "loss": 0.8951, "step": 4099 }, { "epoch": 1.8568840579710146, "grad_norm": 0.3903429807058965, "learning_rate": 6.33556144136397e-05, "loss": 0.8904, "step": 4100 }, { "epoch": 1.8573369565217392, "grad_norm": 0.3304606432380709, "learning_rate": 6.334534574598551e-05, "loss": 0.8881, "step": 4101 }, { "epoch": 1.8577898550724639, "grad_norm": 0.33076851690119735, "learning_rate": 6.333507474446011e-05, "loss": 0.8797, "step": 4102 }, { "epoch": 1.8582427536231885, "grad_norm": 0.2866674330079984, "learning_rate": 6.332480141009031e-05, "loss": 0.873, "step": 4103 }, { "epoch": 1.858695652173913, "grad_norm": 0.25878835311573056, "learning_rate": 6.331452574390314e-05, "loss": 0.8988, "step": 4104 }, { "epoch": 1.8591485507246377, "grad_norm": 0.2555425235301531, "learning_rate": 6.330424774692591e-05, "loss": 0.8936, "step": 4105 }, { "epoch": 1.8596014492753623, "grad_norm": 0.3365085462413698, "learning_rate": 6.329396742018609e-05, "loss": 0.8769, "step": 4106 }, { "epoch": 1.860054347826087, "grad_norm": 0.4250660548866355, "learning_rate": 6.328368476471143e-05, "loss": 0.9087, "step": 4107 }, { "epoch": 1.8605072463768115, "grad_norm": 0.44067068920285696, "learning_rate": 6.327339978152992e-05, "loss": 0.8588, "step": 4108 }, { "epoch": 1.8609601449275361, "grad_norm": 0.5809086686374848, "learning_rate": 6.326311247166975e-05, "loss": 0.8784, "step": 4109 }, { "epoch": 1.8614130434782608, "grad_norm": 0.6638506298610602, "learning_rate": 6.325282283615937e-05, "loss": 0.8798, "step": 4110 }, { "epoch": 1.8618659420289854, "grad_norm": 0.6037320378359602, "learning_rate": 6.324253087602743e-05, "loss": 0.9048, "step": 4111 }, { "epoch": 1.8623188405797102, "grad_norm": 0.5133877140744222, "learning_rate": 6.323223659230284e-05, "loss": 0.9022, "step": 4112 }, { "epoch": 1.8627717391304348, "grad_norm": 0.554294648774425, "learning_rate": 6.322193998601476e-05, "loss": 0.874, "step": 4113 }, { "epoch": 1.8632246376811594, "grad_norm": 0.6684794579672274, "learning_rate": 6.321164105819253e-05, "loss": 0.8962, "step": 4114 }, { "epoch": 1.863677536231884, "grad_norm": 0.7899286544312276, "learning_rate": 6.320133980986579e-05, "loss": 0.863, "step": 4115 }, { "epoch": 1.8641304347826086, "grad_norm": 0.7905338550209018, "learning_rate": 6.319103624206431e-05, "loss": 0.8842, "step": 4116 }, { "epoch": 1.8645833333333335, "grad_norm": 0.6757707691024534, "learning_rate": 6.318073035581821e-05, "loss": 0.8767, "step": 4117 }, { "epoch": 1.865036231884058, "grad_norm": 0.5938135268847162, "learning_rate": 6.317042215215775e-05, "loss": 0.8605, "step": 4118 }, { "epoch": 1.8654891304347827, "grad_norm": 0.6026563344079467, "learning_rate": 6.316011163211346e-05, "loss": 0.8801, "step": 4119 }, { "epoch": 1.8659420289855073, "grad_norm": 0.6074562817017525, "learning_rate": 6.314979879671614e-05, "loss": 0.8818, "step": 4120 }, { "epoch": 1.866394927536232, "grad_norm": 0.6138210238596694, "learning_rate": 6.313948364699671e-05, "loss": 0.8728, "step": 4121 }, { "epoch": 1.8668478260869565, "grad_norm": 0.5755352301546584, "learning_rate": 6.312916618398646e-05, "loss": 0.8875, "step": 4122 }, { "epoch": 1.8673007246376812, "grad_norm": 0.4465544940769136, "learning_rate": 6.311884640871682e-05, "loss": 0.889, "step": 4123 }, { "epoch": 1.8677536231884058, "grad_norm": 0.3177430322522961, "learning_rate": 6.310852432221946e-05, "loss": 0.8925, "step": 4124 }, { "epoch": 1.8682065217391304, "grad_norm": 0.2600851278274218, "learning_rate": 6.30981999255263e-05, "loss": 0.8908, "step": 4125 }, { "epoch": 1.868659420289855, "grad_norm": 0.30135576226796923, "learning_rate": 6.30878732196695e-05, "loss": 0.8795, "step": 4126 }, { "epoch": 1.8691123188405796, "grad_norm": 0.38142021780037155, "learning_rate": 6.307754420568143e-05, "loss": 0.8993, "step": 4127 }, { "epoch": 1.8695652173913042, "grad_norm": 0.4867117307820342, "learning_rate": 6.30672128845947e-05, "loss": 0.883, "step": 4128 }, { "epoch": 1.8700181159420288, "grad_norm": 0.5119770963433116, "learning_rate": 6.305687925744214e-05, "loss": 0.9049, "step": 4129 }, { "epoch": 1.8704710144927537, "grad_norm": 0.554171798737661, "learning_rate": 6.304654332525681e-05, "loss": 0.8897, "step": 4130 }, { "epoch": 1.8709239130434783, "grad_norm": 0.616633046046237, "learning_rate": 6.303620508907202e-05, "loss": 0.8796, "step": 4131 }, { "epoch": 1.8713768115942029, "grad_norm": 0.715985795518362, "learning_rate": 6.302586454992133e-05, "loss": 0.8711, "step": 4132 }, { "epoch": 1.8718297101449275, "grad_norm": 0.7138307230719856, "learning_rate": 6.301552170883845e-05, "loss": 0.8744, "step": 4133 }, { "epoch": 1.8722826086956523, "grad_norm": 0.5864784550681401, "learning_rate": 6.30051765668574e-05, "loss": 0.8933, "step": 4134 }, { "epoch": 1.872735507246377, "grad_norm": 0.4668817340823443, "learning_rate": 6.299482912501241e-05, "loss": 0.8773, "step": 4135 }, { "epoch": 1.8731884057971016, "grad_norm": 0.3884549655962828, "learning_rate": 6.29844793843379e-05, "loss": 0.8942, "step": 4136 }, { "epoch": 1.8736413043478262, "grad_norm": 0.352643179758855, "learning_rate": 6.297412734586856e-05, "loss": 0.8935, "step": 4137 }, { "epoch": 1.8740942028985508, "grad_norm": 0.28157508626371025, "learning_rate": 6.296377301063932e-05, "loss": 0.9079, "step": 4138 }, { "epoch": 1.8745471014492754, "grad_norm": 0.2630584487353468, "learning_rate": 6.29534163796853e-05, "loss": 0.8955, "step": 4139 }, { "epoch": 1.875, "grad_norm": 0.33304304614428726, "learning_rate": 6.294305745404185e-05, "loss": 0.8943, "step": 4140 }, { "epoch": 1.8754528985507246, "grad_norm": 0.39671630434162675, "learning_rate": 6.293269623474461e-05, "loss": 0.872, "step": 4141 }, { "epoch": 1.8759057971014492, "grad_norm": 0.3900931962599451, "learning_rate": 6.292233272282938e-05, "loss": 0.8702, "step": 4142 }, { "epoch": 1.8763586956521738, "grad_norm": 0.33565578789582967, "learning_rate": 6.291196691933222e-05, "loss": 0.8989, "step": 4143 }, { "epoch": 1.8768115942028984, "grad_norm": 0.33911773664447853, "learning_rate": 6.290159882528942e-05, "loss": 0.8636, "step": 4144 }, { "epoch": 1.877264492753623, "grad_norm": 0.35528187532229594, "learning_rate": 6.28912284417375e-05, "loss": 0.8893, "step": 4145 }, { "epoch": 1.8777173913043477, "grad_norm": 0.3759494518576048, "learning_rate": 6.28808557697132e-05, "loss": 0.9215, "step": 4146 }, { "epoch": 1.8781702898550725, "grad_norm": 0.38376951512103974, "learning_rate": 6.28704808102535e-05, "loss": 0.9075, "step": 4147 }, { "epoch": 1.8786231884057971, "grad_norm": 0.4146888853536762, "learning_rate": 6.286010356439558e-05, "loss": 0.872, "step": 4148 }, { "epoch": 1.8790760869565217, "grad_norm": 0.528116466943682, "learning_rate": 6.284972403317687e-05, "loss": 0.8808, "step": 4149 }, { "epoch": 1.8795289855072463, "grad_norm": 0.7281689000081362, "learning_rate": 6.283934221763506e-05, "loss": 0.9017, "step": 4150 }, { "epoch": 1.8799818840579712, "grad_norm": 0.8415361583670091, "learning_rate": 6.282895811880801e-05, "loss": 0.9061, "step": 4151 }, { "epoch": 1.8804347826086958, "grad_norm": 0.8966083078299103, "learning_rate": 6.281857173773384e-05, "loss": 0.8955, "step": 4152 }, { "epoch": 1.8808876811594204, "grad_norm": 0.9066352326871726, "learning_rate": 6.280818307545089e-05, "loss": 0.8874, "step": 4153 }, { "epoch": 1.881340579710145, "grad_norm": 0.8240119259104491, "learning_rate": 6.279779213299775e-05, "loss": 0.8899, "step": 4154 }, { "epoch": 1.8817934782608696, "grad_norm": 0.649256430077586, "learning_rate": 6.278739891141319e-05, "loss": 0.9112, "step": 4155 }, { "epoch": 1.8822463768115942, "grad_norm": 0.4970323790852866, "learning_rate": 6.277700341173625e-05, "loss": 0.8599, "step": 4156 }, { "epoch": 1.8826992753623188, "grad_norm": 0.3183491869372036, "learning_rate": 6.276660563500619e-05, "loss": 0.8851, "step": 4157 }, { "epoch": 1.8831521739130435, "grad_norm": 0.3691061772072866, "learning_rate": 6.27562055822625e-05, "loss": 0.8856, "step": 4158 }, { "epoch": 1.883605072463768, "grad_norm": 0.5977274327582941, "learning_rate": 6.274580325454485e-05, "loss": 0.8831, "step": 4159 }, { "epoch": 1.8840579710144927, "grad_norm": 0.8118446556977922, "learning_rate": 6.273539865289322e-05, "loss": 0.8853, "step": 4160 }, { "epoch": 1.8845108695652173, "grad_norm": 0.8375482545473004, "learning_rate": 6.272499177834777e-05, "loss": 0.9242, "step": 4161 }, { "epoch": 1.884963768115942, "grad_norm": 0.6761945136282568, "learning_rate": 6.271458263194888e-05, "loss": 0.8952, "step": 4162 }, { "epoch": 1.8854166666666665, "grad_norm": 0.5360451618592742, "learning_rate": 6.270417121473716e-05, "loss": 0.8928, "step": 4163 }, { "epoch": 1.8858695652173914, "grad_norm": 0.4523614376787487, "learning_rate": 6.269375752775346e-05, "loss": 0.8947, "step": 4164 }, { "epoch": 1.886322463768116, "grad_norm": 0.4205387760530702, "learning_rate": 6.268334157203886e-05, "loss": 0.8913, "step": 4165 }, { "epoch": 1.8867753623188406, "grad_norm": 0.5195872830325992, "learning_rate": 6.267292334863467e-05, "loss": 0.8741, "step": 4166 }, { "epoch": 1.8872282608695652, "grad_norm": 0.6570780716549556, "learning_rate": 6.26625028585824e-05, "loss": 0.8793, "step": 4167 }, { "epoch": 1.8876811594202898, "grad_norm": 0.6991034211768642, "learning_rate": 6.26520801029238e-05, "loss": 0.8979, "step": 4168 }, { "epoch": 1.8881340579710146, "grad_norm": 0.7157979114282582, "learning_rate": 6.264165508270087e-05, "loss": 0.8921, "step": 4169 }, { "epoch": 1.8885869565217392, "grad_norm": 0.6629302763933529, "learning_rate": 6.26312277989558e-05, "loss": 0.8713, "step": 4170 }, { "epoch": 1.8890398550724639, "grad_norm": 0.5055744891698108, "learning_rate": 6.262079825273101e-05, "loss": 0.8886, "step": 4171 }, { "epoch": 1.8894927536231885, "grad_norm": 0.37691976665469473, "learning_rate": 6.261036644506918e-05, "loss": 0.8896, "step": 4172 }, { "epoch": 1.889945652173913, "grad_norm": 0.31805136839262454, "learning_rate": 6.25999323770132e-05, "loss": 0.8873, "step": 4173 }, { "epoch": 1.8903985507246377, "grad_norm": 0.43877692098041443, "learning_rate": 6.258949604960616e-05, "loss": 0.8829, "step": 4174 }, { "epoch": 1.8908514492753623, "grad_norm": 0.5441695030861406, "learning_rate": 6.25790574638914e-05, "loss": 0.8694, "step": 4175 }, { "epoch": 1.891304347826087, "grad_norm": 0.5701258097328049, "learning_rate": 6.25686166209125e-05, "loss": 0.8772, "step": 4176 }, { "epoch": 1.8917572463768115, "grad_norm": 0.5545919451498332, "learning_rate": 6.25581735217132e-05, "loss": 0.87, "step": 4177 }, { "epoch": 1.8922101449275361, "grad_norm": 0.48302373272428456, "learning_rate": 6.254772816733758e-05, "loss": 0.8892, "step": 4178 }, { "epoch": 1.8926630434782608, "grad_norm": 0.460787055418829, "learning_rate": 6.253728055882983e-05, "loss": 0.9077, "step": 4179 }, { "epoch": 1.8931159420289854, "grad_norm": 0.48653476041942695, "learning_rate": 6.252683069723443e-05, "loss": 0.8808, "step": 4180 }, { "epoch": 1.8935688405797102, "grad_norm": 0.46679007997980787, "learning_rate": 6.251637858359607e-05, "loss": 0.8929, "step": 4181 }, { "epoch": 1.8940217391304348, "grad_norm": 0.3192584710214685, "learning_rate": 6.250592421895966e-05, "loss": 0.8939, "step": 4182 }, { "epoch": 1.8944746376811594, "grad_norm": 0.3186092857364115, "learning_rate": 6.249546760437035e-05, "loss": 0.8858, "step": 4183 }, { "epoch": 1.894927536231884, "grad_norm": 0.4577935435385762, "learning_rate": 6.248500874087349e-05, "loss": 0.8824, "step": 4184 }, { "epoch": 1.8953804347826086, "grad_norm": 0.46238349912573745, "learning_rate": 6.247454762951468e-05, "loss": 0.8975, "step": 4185 }, { "epoch": 1.8958333333333335, "grad_norm": 0.476964864884021, "learning_rate": 6.246408427133972e-05, "loss": 0.8875, "step": 4186 }, { "epoch": 1.896286231884058, "grad_norm": 0.4746383026373845, "learning_rate": 6.245361866739468e-05, "loss": 0.8801, "step": 4187 }, { "epoch": 1.8967391304347827, "grad_norm": 0.3802277378927663, "learning_rate": 6.244315081872577e-05, "loss": 0.8857, "step": 4188 }, { "epoch": 1.8971920289855073, "grad_norm": 0.32421525852277266, "learning_rate": 6.243268072637952e-05, "loss": 0.9127, "step": 4189 }, { "epoch": 1.897644927536232, "grad_norm": 0.37392863542627264, "learning_rate": 6.242220839140264e-05, "loss": 0.8857, "step": 4190 }, { "epoch": 1.8980978260869565, "grad_norm": 0.2767887628847923, "learning_rate": 6.241173381484204e-05, "loss": 0.8899, "step": 4191 }, { "epoch": 1.8985507246376812, "grad_norm": 0.3136391680094673, "learning_rate": 6.240125699774492e-05, "loss": 0.8977, "step": 4192 }, { "epoch": 1.8990036231884058, "grad_norm": 0.4450342159890434, "learning_rate": 6.239077794115863e-05, "loss": 0.8914, "step": 4193 }, { "epoch": 1.8994565217391304, "grad_norm": 0.3881345287211543, "learning_rate": 6.238029664613078e-05, "loss": 0.9056, "step": 4194 }, { "epoch": 1.899909420289855, "grad_norm": 0.31629202942843765, "learning_rate": 6.23698131137092e-05, "loss": 0.8962, "step": 4195 }, { "epoch": 1.9003623188405796, "grad_norm": 0.3403990245700415, "learning_rate": 6.235932734494198e-05, "loss": 0.8859, "step": 4196 }, { "epoch": 1.9008152173913042, "grad_norm": 0.3376980990304271, "learning_rate": 6.234883934087737e-05, "loss": 0.8793, "step": 4197 }, { "epoch": 1.9012681159420288, "grad_norm": 0.3267666823059989, "learning_rate": 6.233834910256386e-05, "loss": 0.8754, "step": 4198 }, { "epoch": 1.9017210144927537, "grad_norm": 0.3358586659440953, "learning_rate": 6.23278566310502e-05, "loss": 0.8972, "step": 4199 }, { "epoch": 1.9021739130434783, "grad_norm": 0.34153124907233284, "learning_rate": 6.231736192738535e-05, "loss": 0.8879, "step": 4200 }, { "epoch": 1.9026268115942029, "grad_norm": 0.3017097994967698, "learning_rate": 6.230686499261844e-05, "loss": 0.8862, "step": 4201 }, { "epoch": 1.9030797101449275, "grad_norm": 0.3219893647028006, "learning_rate": 6.22963658277989e-05, "loss": 0.8918, "step": 4202 }, { "epoch": 1.9035326086956523, "grad_norm": 0.27319833933678805, "learning_rate": 6.228586443397632e-05, "loss": 0.8832, "step": 4203 }, { "epoch": 1.903985507246377, "grad_norm": 0.2886397719364271, "learning_rate": 6.227536081220057e-05, "loss": 0.8819, "step": 4204 }, { "epoch": 1.9044384057971016, "grad_norm": 0.3414329777170869, "learning_rate": 6.22648549635217e-05, "loss": 0.8826, "step": 4205 }, { "epoch": 1.9048913043478262, "grad_norm": 0.3709179137944656, "learning_rate": 6.225434688899e-05, "loss": 0.8755, "step": 4206 }, { "epoch": 1.9053442028985508, "grad_norm": 0.37433711072633996, "learning_rate": 6.224383658965597e-05, "loss": 0.9114, "step": 4207 }, { "epoch": 1.9057971014492754, "grad_norm": 0.43345215182739083, "learning_rate": 6.223332406657034e-05, "loss": 0.8965, "step": 4208 }, { "epoch": 1.90625, "grad_norm": 0.5085193144799993, "learning_rate": 6.22228093207841e-05, "loss": 0.8992, "step": 4209 }, { "epoch": 1.9067028985507246, "grad_norm": 0.591362068399513, "learning_rate": 6.221229235334838e-05, "loss": 0.871, "step": 4210 }, { "epoch": 1.9071557971014492, "grad_norm": 0.5854278425144093, "learning_rate": 6.220177316531459e-05, "loss": 0.8853, "step": 4211 }, { "epoch": 1.9076086956521738, "grad_norm": 0.4968446208578798, "learning_rate": 6.219125175773436e-05, "loss": 0.8807, "step": 4212 }, { "epoch": 1.9080615942028984, "grad_norm": 0.4145340740026646, "learning_rate": 6.218072813165952e-05, "loss": 0.9064, "step": 4213 }, { "epoch": 1.908514492753623, "grad_norm": 0.3788782763525862, "learning_rate": 6.217020228814215e-05, "loss": 0.8905, "step": 4214 }, { "epoch": 1.9089673913043477, "grad_norm": 0.33039418632319867, "learning_rate": 6.215967422823452e-05, "loss": 0.8939, "step": 4215 }, { "epoch": 1.9094202898550725, "grad_norm": 0.28420394092714224, "learning_rate": 6.214914395298914e-05, "loss": 0.8894, "step": 4216 }, { "epoch": 1.9098731884057971, "grad_norm": 0.2685968888809134, "learning_rate": 6.213861146345876e-05, "loss": 0.9026, "step": 4217 }, { "epoch": 1.9103260869565217, "grad_norm": 0.25060686475361377, "learning_rate": 6.21280767606963e-05, "loss": 0.8816, "step": 4218 }, { "epoch": 1.9107789855072463, "grad_norm": 0.23748438925793133, "learning_rate": 6.211753984575493e-05, "loss": 0.8946, "step": 4219 }, { "epoch": 1.9112318840579712, "grad_norm": 0.21574333635394916, "learning_rate": 6.210700071968806e-05, "loss": 0.9026, "step": 4220 }, { "epoch": 1.9116847826086958, "grad_norm": 0.21230458970020824, "learning_rate": 6.20964593835493e-05, "loss": 0.8836, "step": 4221 }, { "epoch": 1.9121376811594204, "grad_norm": 0.25041725610561966, "learning_rate": 6.208591583839248e-05, "loss": 0.906, "step": 4222 }, { "epoch": 1.912590579710145, "grad_norm": 0.24137407437788297, "learning_rate": 6.207537008527165e-05, "loss": 0.8758, "step": 4223 }, { "epoch": 1.9130434782608696, "grad_norm": 0.21980390737951375, "learning_rate": 6.206482212524109e-05, "loss": 0.8619, "step": 4224 }, { "epoch": 1.9134963768115942, "grad_norm": 0.22499349149923223, "learning_rate": 6.20542719593553e-05, "loss": 0.8806, "step": 4225 }, { "epoch": 1.9139492753623188, "grad_norm": 0.2541976394413538, "learning_rate": 6.204371958866901e-05, "loss": 0.892, "step": 4226 }, { "epoch": 1.9144021739130435, "grad_norm": 0.28737913867934306, "learning_rate": 6.203316501423712e-05, "loss": 0.8855, "step": 4227 }, { "epoch": 1.914855072463768, "grad_norm": 0.25920215821262366, "learning_rate": 6.202260823711482e-05, "loss": 0.8652, "step": 4228 }, { "epoch": 1.9153079710144927, "grad_norm": 0.24744176416605387, "learning_rate": 6.201204925835747e-05, "loss": 0.9015, "step": 4229 }, { "epoch": 1.9157608695652173, "grad_norm": 0.32696798822906903, "learning_rate": 6.200148807902069e-05, "loss": 0.8804, "step": 4230 }, { "epoch": 1.916213768115942, "grad_norm": 0.3689971247096851, "learning_rate": 6.199092470016026e-05, "loss": 0.8905, "step": 4231 }, { "epoch": 1.9166666666666665, "grad_norm": 0.4826161943810626, "learning_rate": 6.198035912283225e-05, "loss": 0.876, "step": 4232 }, { "epoch": 1.9171195652173914, "grad_norm": 0.5599589249138318, "learning_rate": 6.19697913480929e-05, "loss": 0.8868, "step": 4233 }, { "epoch": 1.917572463768116, "grad_norm": 0.6943391693635846, "learning_rate": 6.195922137699872e-05, "loss": 0.8739, "step": 4234 }, { "epoch": 1.9180253623188406, "grad_norm": 0.7350618942094495, "learning_rate": 6.194864921060636e-05, "loss": 0.8895, "step": 4235 }, { "epoch": 1.9184782608695652, "grad_norm": 0.6616784398383084, "learning_rate": 6.193807484997275e-05, "loss": 0.8843, "step": 4236 }, { "epoch": 1.9189311594202898, "grad_norm": 0.5050277103542654, "learning_rate": 6.192749829615505e-05, "loss": 0.8775, "step": 4237 }, { "epoch": 1.9193840579710146, "grad_norm": 0.33765194020865774, "learning_rate": 6.191691955021059e-05, "loss": 0.8835, "step": 4238 }, { "epoch": 1.9198369565217392, "grad_norm": 0.20701458471956383, "learning_rate": 6.190633861319696e-05, "loss": 0.8911, "step": 4239 }, { "epoch": 1.9202898550724639, "grad_norm": 0.2214609402682648, "learning_rate": 6.189575548617194e-05, "loss": 0.876, "step": 4240 }, { "epoch": 1.9207427536231885, "grad_norm": 0.2768312212407701, "learning_rate": 6.188517017019355e-05, "loss": 0.8754, "step": 4241 }, { "epoch": 1.921195652173913, "grad_norm": 0.3617622807669982, "learning_rate": 6.187458266632002e-05, "loss": 0.8709, "step": 4242 }, { "epoch": 1.9216485507246377, "grad_norm": 0.3951253986063765, "learning_rate": 6.186399297560979e-05, "loss": 0.8831, "step": 4243 }, { "epoch": 1.9221014492753623, "grad_norm": 0.42478799853288385, "learning_rate": 6.185340109912155e-05, "loss": 0.8855, "step": 4244 }, { "epoch": 1.922554347826087, "grad_norm": 0.45245977931963083, "learning_rate": 6.184280703791415e-05, "loss": 0.9054, "step": 4245 }, { "epoch": 1.9230072463768115, "grad_norm": 0.460998629024106, "learning_rate": 6.183221079304673e-05, "loss": 0.8964, "step": 4246 }, { "epoch": 1.9234601449275361, "grad_norm": 0.4644374126903968, "learning_rate": 6.18216123655786e-05, "loss": 0.8967, "step": 4247 }, { "epoch": 1.9239130434782608, "grad_norm": 0.3997632493406689, "learning_rate": 6.181101175656932e-05, "loss": 0.8971, "step": 4248 }, { "epoch": 1.9243659420289854, "grad_norm": 0.35653597736297316, "learning_rate": 6.180040896707861e-05, "loss": 0.8903, "step": 4249 }, { "epoch": 1.9248188405797102, "grad_norm": 0.3266299907797819, "learning_rate": 6.178980399816647e-05, "loss": 0.8864, "step": 4250 }, { "epoch": 1.9252717391304348, "grad_norm": 0.3522714966893562, "learning_rate": 6.17791968508931e-05, "loss": 0.8897, "step": 4251 }, { "epoch": 1.9257246376811594, "grad_norm": 0.3363486844922753, "learning_rate": 6.17685875263189e-05, "loss": 0.8814, "step": 4252 }, { "epoch": 1.926177536231884, "grad_norm": 0.2814570959787273, "learning_rate": 6.175797602550452e-05, "loss": 0.8861, "step": 4253 }, { "epoch": 1.9266304347826086, "grad_norm": 0.2821646447920771, "learning_rate": 6.17473623495108e-05, "loss": 0.87, "step": 4254 }, { "epoch": 1.9270833333333335, "grad_norm": 0.2979033908590265, "learning_rate": 6.17367464993988e-05, "loss": 0.8912, "step": 4255 }, { "epoch": 1.927536231884058, "grad_norm": 0.24094283406521355, "learning_rate": 6.172612847622982e-05, "loss": 0.8787, "step": 4256 }, { "epoch": 1.9279891304347827, "grad_norm": 0.2771190622329758, "learning_rate": 6.171550828106534e-05, "loss": 0.8835, "step": 4257 }, { "epoch": 1.9284420289855073, "grad_norm": 0.3057015449137052, "learning_rate": 6.170488591496707e-05, "loss": 0.8881, "step": 4258 }, { "epoch": 1.928894927536232, "grad_norm": 0.28502357680232443, "learning_rate": 6.169426137899697e-05, "loss": 0.897, "step": 4259 }, { "epoch": 1.9293478260869565, "grad_norm": 0.3682248101317461, "learning_rate": 6.168363467421719e-05, "loss": 0.8737, "step": 4260 }, { "epoch": 1.9298007246376812, "grad_norm": 0.3579495997811797, "learning_rate": 6.167300580169009e-05, "loss": 0.8801, "step": 4261 }, { "epoch": 1.9302536231884058, "grad_norm": 0.3159713341349206, "learning_rate": 6.166237476247826e-05, "loss": 0.888, "step": 4262 }, { "epoch": 1.9307065217391304, "grad_norm": 0.2627151529870103, "learning_rate": 6.16517415576445e-05, "loss": 0.8884, "step": 4263 }, { "epoch": 1.931159420289855, "grad_norm": 0.3340495490157807, "learning_rate": 6.164110618825181e-05, "loss": 0.8844, "step": 4264 }, { "epoch": 1.9316123188405796, "grad_norm": 0.3532775597206331, "learning_rate": 6.163046865536345e-05, "loss": 0.9021, "step": 4265 }, { "epoch": 1.9320652173913042, "grad_norm": 0.31563830983550106, "learning_rate": 6.161982896004286e-05, "loss": 0.8955, "step": 4266 }, { "epoch": 1.9325181159420288, "grad_norm": 0.3027444659274207, "learning_rate": 6.16091871033537e-05, "loss": 0.8848, "step": 4267 }, { "epoch": 1.9329710144927537, "grad_norm": 0.29832809138617533, "learning_rate": 6.159854308635988e-05, "loss": 0.8795, "step": 4268 }, { "epoch": 1.9334239130434783, "grad_norm": 0.2863921731794351, "learning_rate": 6.158789691012548e-05, "loss": 0.8745, "step": 4269 }, { "epoch": 1.9338768115942029, "grad_norm": 0.31313527629896976, "learning_rate": 6.15772485757148e-05, "loss": 0.878, "step": 4270 }, { "epoch": 1.9343297101449275, "grad_norm": 0.36271198543063454, "learning_rate": 6.156659808419241e-05, "loss": 0.8836, "step": 4271 }, { "epoch": 1.9347826086956523, "grad_norm": 0.3911455739113859, "learning_rate": 6.155594543662304e-05, "loss": 0.8799, "step": 4272 }, { "epoch": 1.935235507246377, "grad_norm": 0.4768299441344383, "learning_rate": 6.154529063407161e-05, "loss": 0.8764, "step": 4273 }, { "epoch": 1.9356884057971016, "grad_norm": 0.46903393908095103, "learning_rate": 6.153463367760336e-05, "loss": 0.8857, "step": 4274 }, { "epoch": 1.9361413043478262, "grad_norm": 0.4012068773779218, "learning_rate": 6.152397456828366e-05, "loss": 0.8832, "step": 4275 }, { "epoch": 1.9365942028985508, "grad_norm": 0.3984051871685291, "learning_rate": 6.15133133071781e-05, "loss": 0.8667, "step": 4276 }, { "epoch": 1.9370471014492754, "grad_norm": 0.4559764948882655, "learning_rate": 6.150264989535252e-05, "loss": 0.8842, "step": 4277 }, { "epoch": 1.9375, "grad_norm": 0.408154346447958, "learning_rate": 6.149198433387297e-05, "loss": 0.866, "step": 4278 }, { "epoch": 1.9379528985507246, "grad_norm": 0.3747300185417019, "learning_rate": 6.148131662380567e-05, "loss": 0.897, "step": 4279 }, { "epoch": 1.9384057971014492, "grad_norm": 0.385753666016061, "learning_rate": 6.147064676621713e-05, "loss": 0.8844, "step": 4280 }, { "epoch": 1.9388586956521738, "grad_norm": 0.3950155022557531, "learning_rate": 6.145997476217398e-05, "loss": 0.9044, "step": 4281 }, { "epoch": 1.9393115942028984, "grad_norm": 0.3778622771374154, "learning_rate": 6.144930061274317e-05, "loss": 0.8716, "step": 4282 }, { "epoch": 1.939764492753623, "grad_norm": 0.34541515425490193, "learning_rate": 6.143862431899177e-05, "loss": 0.9068, "step": 4283 }, { "epoch": 1.9402173913043477, "grad_norm": 0.26997108732635616, "learning_rate": 6.142794588198715e-05, "loss": 0.9087, "step": 4284 }, { "epoch": 1.9406702898550725, "grad_norm": 0.2916709192786286, "learning_rate": 6.14172653027968e-05, "loss": 0.9104, "step": 4285 }, { "epoch": 1.9411231884057971, "grad_norm": 0.29902013677622186, "learning_rate": 6.14065825824885e-05, "loss": 0.9068, "step": 4286 }, { "epoch": 1.9415760869565217, "grad_norm": 0.25469591276005354, "learning_rate": 6.139589772213024e-05, "loss": 0.8795, "step": 4287 }, { "epoch": 1.9420289855072463, "grad_norm": 0.35799672246125924, "learning_rate": 6.138521072279017e-05, "loss": 0.903, "step": 4288 }, { "epoch": 1.9424818840579712, "grad_norm": 0.3516040409834551, "learning_rate": 6.137452158553668e-05, "loss": 0.882, "step": 4289 }, { "epoch": 1.9429347826086958, "grad_norm": 0.3409485584268516, "learning_rate": 6.136383031143842e-05, "loss": 0.8956, "step": 4290 }, { "epoch": 1.9433876811594204, "grad_norm": 0.4396910307206737, "learning_rate": 6.135313690156418e-05, "loss": 0.8773, "step": 4291 }, { "epoch": 1.943840579710145, "grad_norm": 0.5075863654322251, "learning_rate": 6.1342441356983e-05, "loss": 0.8894, "step": 4292 }, { "epoch": 1.9442934782608696, "grad_norm": 0.47211854470973147, "learning_rate": 6.133174367876415e-05, "loss": 0.886, "step": 4293 }, { "epoch": 1.9447463768115942, "grad_norm": 0.5150678406717892, "learning_rate": 6.132104386797707e-05, "loss": 0.8829, "step": 4294 }, { "epoch": 1.9451992753623188, "grad_norm": 0.7145976596486417, "learning_rate": 6.131034192569145e-05, "loss": 0.8911, "step": 4295 }, { "epoch": 1.9456521739130435, "grad_norm": 0.8594897659175421, "learning_rate": 6.12996378529772e-05, "loss": 0.8794, "step": 4296 }, { "epoch": 1.946105072463768, "grad_norm": 0.8557778396689205, "learning_rate": 6.128893165090437e-05, "loss": 0.8746, "step": 4297 }, { "epoch": 1.9465579710144927, "grad_norm": 0.7604207108981597, "learning_rate": 6.127822332054332e-05, "loss": 0.8969, "step": 4298 }, { "epoch": 1.9470108695652173, "grad_norm": 0.6581994150446808, "learning_rate": 6.126751286296456e-05, "loss": 0.8793, "step": 4299 }, { "epoch": 1.947463768115942, "grad_norm": 0.514788420001674, "learning_rate": 6.125680027923884e-05, "loss": 0.8926, "step": 4300 }, { "epoch": 1.9479166666666665, "grad_norm": 0.4062255774027685, "learning_rate": 6.124608557043713e-05, "loss": 0.8685, "step": 4301 }, { "epoch": 1.9483695652173914, "grad_norm": 0.392190795522421, "learning_rate": 6.123536873763056e-05, "loss": 0.9078, "step": 4302 }, { "epoch": 1.948822463768116, "grad_norm": 0.3748747533130713, "learning_rate": 6.122464978189053e-05, "loss": 0.8831, "step": 4303 }, { "epoch": 1.9492753623188406, "grad_norm": 0.5749896315474362, "learning_rate": 6.121392870428864e-05, "loss": 0.876, "step": 4304 }, { "epoch": 1.9497282608695652, "grad_norm": 0.7217907512536449, "learning_rate": 6.120320550589667e-05, "loss": 0.8795, "step": 4305 }, { "epoch": 1.9501811594202898, "grad_norm": 0.7252129473230162, "learning_rate": 6.119248018778665e-05, "loss": 0.8842, "step": 4306 }, { "epoch": 1.9506340579710146, "grad_norm": 0.6505260410308168, "learning_rate": 6.118175275103079e-05, "loss": 0.9073, "step": 4307 }, { "epoch": 1.9510869565217392, "grad_norm": 0.5419854707731775, "learning_rate": 6.117102319670156e-05, "loss": 0.8885, "step": 4308 }, { "epoch": 1.9515398550724639, "grad_norm": 0.4703258286230629, "learning_rate": 6.116029152587159e-05, "loss": 0.8997, "step": 4309 }, { "epoch": 1.9519927536231885, "grad_norm": 0.3331652791743112, "learning_rate": 6.114955773961374e-05, "loss": 0.8894, "step": 4310 }, { "epoch": 1.952445652173913, "grad_norm": 0.28841277878054483, "learning_rate": 6.113882183900108e-05, "loss": 0.8608, "step": 4311 }, { "epoch": 1.9528985507246377, "grad_norm": 0.23813173630824805, "learning_rate": 6.112808382510692e-05, "loss": 0.8727, "step": 4312 }, { "epoch": 1.9533514492753623, "grad_norm": 0.23130244930036833, "learning_rate": 6.111734369900474e-05, "loss": 0.8925, "step": 4313 }, { "epoch": 1.953804347826087, "grad_norm": 0.3022803083664877, "learning_rate": 6.110660146176823e-05, "loss": 0.8763, "step": 4314 }, { "epoch": 1.9542572463768115, "grad_norm": 0.3540796333676115, "learning_rate": 6.109585711447134e-05, "loss": 0.8946, "step": 4315 }, { "epoch": 1.9547101449275361, "grad_norm": 0.3894732421752882, "learning_rate": 6.108511065818818e-05, "loss": 0.8617, "step": 4316 }, { "epoch": 1.9551630434782608, "grad_norm": 0.41826862421631544, "learning_rate": 6.107436209399309e-05, "loss": 0.8788, "step": 4317 }, { "epoch": 1.9556159420289854, "grad_norm": 0.3936510235128157, "learning_rate": 6.106361142296064e-05, "loss": 0.8837, "step": 4318 }, { "epoch": 1.9560688405797102, "grad_norm": 0.32567458763374024, "learning_rate": 6.105285864616556e-05, "loss": 0.8962, "step": 4319 }, { "epoch": 1.9565217391304348, "grad_norm": 0.3570929366724528, "learning_rate": 6.104210376468284e-05, "loss": 0.8868, "step": 4320 }, { "epoch": 1.9569746376811594, "grad_norm": 0.34416034836471193, "learning_rate": 6.1031346779587677e-05, "loss": 0.8763, "step": 4321 }, { "epoch": 1.957427536231884, "grad_norm": 0.2807060255184943, "learning_rate": 6.1020587691955446e-05, "loss": 0.8891, "step": 4322 }, { "epoch": 1.9578804347826086, "grad_norm": 0.21552267708238557, "learning_rate": 6.100982650286176e-05, "loss": 0.89, "step": 4323 }, { "epoch": 1.9583333333333335, "grad_norm": 0.22089934734987496, "learning_rate": 6.099906321338241e-05, "loss": 0.8829, "step": 4324 }, { "epoch": 1.958786231884058, "grad_norm": 0.24950946857582504, "learning_rate": 6.0988297824593444e-05, "loss": 0.8747, "step": 4325 }, { "epoch": 1.9592391304347827, "grad_norm": 0.22061204921251898, "learning_rate": 6.097753033757109e-05, "loss": 0.8987, "step": 4326 }, { "epoch": 1.9596920289855073, "grad_norm": 0.278922104932681, "learning_rate": 6.096676075339178e-05, "loss": 0.8959, "step": 4327 }, { "epoch": 1.960144927536232, "grad_norm": 0.27656298807548185, "learning_rate": 6.095598907313217e-05, "loss": 0.9043, "step": 4328 }, { "epoch": 1.9605978260869565, "grad_norm": 0.2088716035735608, "learning_rate": 6.094521529786912e-05, "loss": 0.8761, "step": 4329 }, { "epoch": 1.9610507246376812, "grad_norm": 0.2423418656596856, "learning_rate": 6.0934439428679726e-05, "loss": 0.8843, "step": 4330 }, { "epoch": 1.9615036231884058, "grad_norm": 0.29241025154583633, "learning_rate": 6.092366146664123e-05, "loss": 0.8846, "step": 4331 }, { "epoch": 1.9619565217391304, "grad_norm": 0.2605968325688786, "learning_rate": 6.0912881412831145e-05, "loss": 0.899, "step": 4332 }, { "epoch": 1.962409420289855, "grad_norm": 0.2759339886852092, "learning_rate": 6.0902099268327174e-05, "loss": 0.8987, "step": 4333 }, { "epoch": 1.9628623188405796, "grad_norm": 0.30385008639262634, "learning_rate": 6.089131503420721e-05, "loss": 0.8994, "step": 4334 }, { "epoch": 1.9633152173913042, "grad_norm": 0.30551671904432265, "learning_rate": 6.088052871154938e-05, "loss": 0.8881, "step": 4335 }, { "epoch": 1.9637681159420288, "grad_norm": 0.33717197273203436, "learning_rate": 6.0869740301432e-05, "loss": 0.8867, "step": 4336 }, { "epoch": 1.9642210144927537, "grad_norm": 0.399582169496217, "learning_rate": 6.0858949804933616e-05, "loss": 0.8789, "step": 4337 }, { "epoch": 1.9646739130434783, "grad_norm": 0.40917594772819404, "learning_rate": 6.084815722313297e-05, "loss": 0.9044, "step": 4338 }, { "epoch": 1.9651268115942029, "grad_norm": 0.47350072646175495, "learning_rate": 6.0837362557109e-05, "loss": 0.9006, "step": 4339 }, { "epoch": 1.9655797101449275, "grad_norm": 0.5523911264034516, "learning_rate": 6.082656580794089e-05, "loss": 0.8917, "step": 4340 }, { "epoch": 1.9660326086956523, "grad_norm": 0.642137268775581, "learning_rate": 6.081576697670798e-05, "loss": 0.8986, "step": 4341 }, { "epoch": 1.966485507246377, "grad_norm": 0.6824594458750142, "learning_rate": 6.080496606448986e-05, "loss": 0.8826, "step": 4342 }, { "epoch": 1.9669384057971016, "grad_norm": 0.693956641824655, "learning_rate": 6.079416307236632e-05, "loss": 0.8765, "step": 4343 }, { "epoch": 1.9673913043478262, "grad_norm": 0.6967347780936307, "learning_rate": 6.078335800141735e-05, "loss": 0.8999, "step": 4344 }, { "epoch": 1.9678442028985508, "grad_norm": 0.674223278471952, "learning_rate": 6.0772550852723136e-05, "loss": 0.8693, "step": 4345 }, { "epoch": 1.9682971014492754, "grad_norm": 0.5519182242857846, "learning_rate": 6.0761741627364104e-05, "loss": 0.8876, "step": 4346 }, { "epoch": 1.96875, "grad_norm": 0.37837080495100484, "learning_rate": 6.075093032642087e-05, "loss": 0.893, "step": 4347 }, { "epoch": 1.9692028985507246, "grad_norm": 0.22566521378882415, "learning_rate": 6.074011695097423e-05, "loss": 0.9112, "step": 4348 }, { "epoch": 1.9696557971014492, "grad_norm": 0.288942206451578, "learning_rate": 6.072930150210525e-05, "loss": 0.8934, "step": 4349 }, { "epoch": 1.9701086956521738, "grad_norm": 0.44679009640120076, "learning_rate": 6.071848398089513e-05, "loss": 0.8537, "step": 4350 }, { "epoch": 1.9705615942028984, "grad_norm": 0.5103956733112448, "learning_rate": 6.070766438842535e-05, "loss": 0.9015, "step": 4351 }, { "epoch": 1.971014492753623, "grad_norm": 0.6019938492002544, "learning_rate": 6.069684272577755e-05, "loss": 0.87, "step": 4352 }, { "epoch": 1.9714673913043477, "grad_norm": 0.7164808412207561, "learning_rate": 6.0686018994033574e-05, "loss": 0.9038, "step": 4353 }, { "epoch": 1.9719202898550725, "grad_norm": 0.8623910135052486, "learning_rate": 6.06751931942755e-05, "loss": 0.9024, "step": 4354 }, { "epoch": 1.9723731884057971, "grad_norm": 0.9566174383580558, "learning_rate": 6.0664365327585614e-05, "loss": 0.9005, "step": 4355 }, { "epoch": 1.9728260869565217, "grad_norm": 1.0235983339687644, "learning_rate": 6.065353539504637e-05, "loss": 0.9039, "step": 4356 }, { "epoch": 1.9732789855072463, "grad_norm": 0.9781253548512963, "learning_rate": 6.0642703397740464e-05, "loss": 0.8788, "step": 4357 }, { "epoch": 1.9737318840579712, "grad_norm": 0.9572414986234964, "learning_rate": 6.063186933675079e-05, "loss": 0.9018, "step": 4358 }, { "epoch": 1.9741847826086958, "grad_norm": 0.7917455298724705, "learning_rate": 6.062103321316045e-05, "loss": 0.8919, "step": 4359 }, { "epoch": 1.9746376811594204, "grad_norm": 0.5417655772123078, "learning_rate": 6.061019502805273e-05, "loss": 0.8819, "step": 4360 }, { "epoch": 1.975090579710145, "grad_norm": 0.3497092904438662, "learning_rate": 6.059935478251117e-05, "loss": 0.8784, "step": 4361 }, { "epoch": 1.9755434782608696, "grad_norm": 0.25066001628837664, "learning_rate": 6.0588512477619453e-05, "loss": 0.8919, "step": 4362 }, { "epoch": 1.9759963768115942, "grad_norm": 0.33828890308240495, "learning_rate": 6.0577668114461534e-05, "loss": 0.8923, "step": 4363 }, { "epoch": 1.9764492753623188, "grad_norm": 0.4241669490635184, "learning_rate": 6.0566821694121506e-05, "loss": 0.8909, "step": 4364 }, { "epoch": 1.9769021739130435, "grad_norm": 0.47259159266023093, "learning_rate": 6.055597321768375e-05, "loss": 0.8732, "step": 4365 }, { "epoch": 1.977355072463768, "grad_norm": 0.578140380268246, "learning_rate": 6.054512268623275e-05, "loss": 0.8864, "step": 4366 }, { "epoch": 1.9778079710144927, "grad_norm": 0.6940836205423114, "learning_rate": 6.0534270100853284e-05, "loss": 0.8835, "step": 4367 }, { "epoch": 1.9782608695652173, "grad_norm": 0.6872955734615932, "learning_rate": 6.052341546263032e-05, "loss": 0.8926, "step": 4368 }, { "epoch": 1.978713768115942, "grad_norm": 0.665504614361685, "learning_rate": 6.051255877264897e-05, "loss": 0.8669, "step": 4369 }, { "epoch": 1.9791666666666665, "grad_norm": 0.6598137248299748, "learning_rate": 6.0501700031994613e-05, "loss": 0.887, "step": 4370 }, { "epoch": 1.9796195652173914, "grad_norm": 0.5962124882185632, "learning_rate": 6.049083924175282e-05, "loss": 0.8754, "step": 4371 }, { "epoch": 1.980072463768116, "grad_norm": 0.4693375345414046, "learning_rate": 6.047997640300936e-05, "loss": 0.8955, "step": 4372 }, { "epoch": 1.9805253623188406, "grad_norm": 0.45044401564338954, "learning_rate": 6.0469111516850204e-05, "loss": 0.8886, "step": 4373 }, { "epoch": 1.9809782608695652, "grad_norm": 0.38765221491588636, "learning_rate": 6.045824458436153e-05, "loss": 0.8772, "step": 4374 }, { "epoch": 1.9814311594202898, "grad_norm": 0.286954641341792, "learning_rate": 6.044737560662972e-05, "loss": 0.8786, "step": 4375 }, { "epoch": 1.9818840579710146, "grad_norm": 0.34850478527911966, "learning_rate": 6.043650458474139e-05, "loss": 0.8902, "step": 4376 }, { "epoch": 1.9823369565217392, "grad_norm": 0.39299109821180067, "learning_rate": 6.0425631519783304e-05, "loss": 0.9076, "step": 4377 }, { "epoch": 1.9827898550724639, "grad_norm": 0.3862530308122043, "learning_rate": 6.041475641284246e-05, "loss": 0.8896, "step": 4378 }, { "epoch": 1.9832427536231885, "grad_norm": 0.36320087998563266, "learning_rate": 6.0403879265006064e-05, "loss": 0.911, "step": 4379 }, { "epoch": 1.983695652173913, "grad_norm": 0.34955526529784725, "learning_rate": 6.039300007736154e-05, "loss": 0.8669, "step": 4380 }, { "epoch": 1.9841485507246377, "grad_norm": 0.35873224367839796, "learning_rate": 6.038211885099648e-05, "loss": 0.8514, "step": 4381 }, { "epoch": 1.9846014492753623, "grad_norm": 0.37440442474204655, "learning_rate": 6.037123558699868e-05, "loss": 0.8778, "step": 4382 }, { "epoch": 1.985054347826087, "grad_norm": 0.4431020285046377, "learning_rate": 6.03603502864562e-05, "loss": 0.8711, "step": 4383 }, { "epoch": 1.9855072463768115, "grad_norm": 0.5798913876158011, "learning_rate": 6.0349462950457234e-05, "loss": 0.887, "step": 4384 }, { "epoch": 1.9859601449275361, "grad_norm": 0.6219930665759231, "learning_rate": 6.0338573580090214e-05, "loss": 0.8975, "step": 4385 }, { "epoch": 1.9864130434782608, "grad_norm": 0.6650979249713982, "learning_rate": 6.032768217644375e-05, "loss": 0.8958, "step": 4386 }, { "epoch": 1.9868659420289854, "grad_norm": 0.7575880391689698, "learning_rate": 6.031678874060669e-05, "loss": 0.865, "step": 4387 }, { "epoch": 1.9873188405797102, "grad_norm": 0.8174099526519374, "learning_rate": 6.030589327366807e-05, "loss": 0.8668, "step": 4388 }, { "epoch": 1.9877717391304348, "grad_norm": 0.753953278860636, "learning_rate": 6.029499577671713e-05, "loss": 0.8906, "step": 4389 }, { "epoch": 1.9882246376811594, "grad_norm": 0.6129566054911076, "learning_rate": 6.0284096250843295e-05, "loss": 0.9001, "step": 4390 }, { "epoch": 1.988677536231884, "grad_norm": 0.46490801755835187, "learning_rate": 6.027319469713621e-05, "loss": 0.8711, "step": 4391 }, { "epoch": 1.9891304347826086, "grad_norm": 0.3816182346560127, "learning_rate": 6.026229111668572e-05, "loss": 0.9003, "step": 4392 }, { "epoch": 1.9895833333333335, "grad_norm": 0.3565164803639658, "learning_rate": 6.02513855105819e-05, "loss": 0.9003, "step": 4393 }, { "epoch": 1.990036231884058, "grad_norm": 0.4521545612664858, "learning_rate": 6.024047787991496e-05, "loss": 0.8767, "step": 4394 }, { "epoch": 1.9904891304347827, "grad_norm": 0.5380027799768349, "learning_rate": 6.022956822577538e-05, "loss": 0.8719, "step": 4395 }, { "epoch": 1.9909420289855073, "grad_norm": 0.633521861512697, "learning_rate": 6.021865654925381e-05, "loss": 0.9023, "step": 4396 }, { "epoch": 1.991394927536232, "grad_norm": 0.7407232017779355, "learning_rate": 6.020774285144111e-05, "loss": 0.8906, "step": 4397 }, { "epoch": 1.9918478260869565, "grad_norm": 0.8622742664397508, "learning_rate": 6.019682713342833e-05, "loss": 0.9209, "step": 4398 }, { "epoch": 1.9923007246376812, "grad_norm": 0.8662866463749613, "learning_rate": 6.0185909396306727e-05, "loss": 0.8791, "step": 4399 }, { "epoch": 1.9927536231884058, "grad_norm": 0.8070418429854276, "learning_rate": 6.017498964116778e-05, "loss": 0.8928, "step": 4400 }, { "epoch": 1.9932065217391304, "grad_norm": 0.7176282515141317, "learning_rate": 6.016406786910315e-05, "loss": 0.9127, "step": 4401 }, { "epoch": 1.993659420289855, "grad_norm": 0.5532407992242645, "learning_rate": 6.0153144081204715e-05, "loss": 0.9053, "step": 4402 }, { "epoch": 1.9941123188405796, "grad_norm": 0.3915056343090548, "learning_rate": 6.0142218278564516e-05, "loss": 0.8775, "step": 4403 }, { "epoch": 1.9945652173913042, "grad_norm": 0.324786689101745, "learning_rate": 6.0131290462274835e-05, "loss": 0.8941, "step": 4404 }, { "epoch": 1.9950181159420288, "grad_norm": 0.294346393090784, "learning_rate": 6.012036063342815e-05, "loss": 0.9081, "step": 4405 }, { "epoch": 1.9954710144927537, "grad_norm": 0.3882477954676458, "learning_rate": 6.010942879311714e-05, "loss": 0.8925, "step": 4406 }, { "epoch": 1.9959239130434783, "grad_norm": 0.47968067368485406, "learning_rate": 6.009849494243465e-05, "loss": 0.8848, "step": 4407 }, { "epoch": 1.9963768115942029, "grad_norm": 0.5253649887546336, "learning_rate": 6.008755908247378e-05, "loss": 0.9022, "step": 4408 }, { "epoch": 1.9968297101449275, "grad_norm": 0.5309268126926205, "learning_rate": 6.00766212143278e-05, "loss": 0.8897, "step": 4409 }, { "epoch": 1.9972826086956523, "grad_norm": 1.5497997071143919, "learning_rate": 6.006568133909018e-05, "loss": 0.9071, "step": 4410 }, { "epoch": 1.997735507246377, "grad_norm": 0.3363168295401611, "learning_rate": 6.00547394578546e-05, "loss": 0.907, "step": 4411 }, { "epoch": 1.9981884057971016, "grad_norm": 0.6798761885727157, "learning_rate": 6.004379557171493e-05, "loss": 0.8886, "step": 4412 }, { "epoch": 1.9986413043478262, "grad_norm": 0.7590913250134742, "learning_rate": 6.0032849681765265e-05, "loss": 0.8917, "step": 4413 }, { "epoch": 1.9990942028985508, "grad_norm": 0.7073350598658242, "learning_rate": 6.002190178909986e-05, "loss": 0.8869, "step": 4414 }, { "epoch": 1.9995471014492754, "grad_norm": 0.6295102352581702, "learning_rate": 6.0010951894813227e-05, "loss": 0.9036, "step": 4415 }, { "epoch": 2.0, "grad_norm": 0.5457084690641719, "learning_rate": 6.000000000000001e-05, "loss": 0.8675, "step": 4416 }, { "epoch": 2.0004528985507246, "grad_norm": 0.4947739824657384, "learning_rate": 5.99890461057551e-05, "loss": 0.8858, "step": 4417 }, { "epoch": 2.0009057971014492, "grad_norm": 0.32448406869820023, "learning_rate": 5.9978090213173575e-05, "loss": 0.8728, "step": 4418 }, { "epoch": 2.001358695652174, "grad_norm": 0.3873664518470588, "learning_rate": 5.996713232335071e-05, "loss": 0.8774, "step": 4419 }, { "epoch": 2.0018115942028984, "grad_norm": 0.534134157002732, "learning_rate": 5.9956172437381993e-05, "loss": 0.8742, "step": 4420 }, { "epoch": 2.002264492753623, "grad_norm": 0.5721233363052168, "learning_rate": 5.994521055636309e-05, "loss": 0.8531, "step": 4421 }, { "epoch": 2.0027173913043477, "grad_norm": 0.5306641757026073, "learning_rate": 5.993424668138988e-05, "loss": 0.8897, "step": 4422 }, { "epoch": 2.0031702898550723, "grad_norm": 0.4536081130457519, "learning_rate": 5.992328081355846e-05, "loss": 0.9016, "step": 4423 }, { "epoch": 2.003623188405797, "grad_norm": 0.4274549008579753, "learning_rate": 5.991231295396507e-05, "loss": 0.8585, "step": 4424 }, { "epoch": 2.004076086956522, "grad_norm": 0.4350370729870853, "learning_rate": 5.990134310370619e-05, "loss": 0.8772, "step": 4425 }, { "epoch": 2.0045289855072466, "grad_norm": 0.43222013192198033, "learning_rate": 5.989037126387852e-05, "loss": 0.8921, "step": 4426 }, { "epoch": 2.004981884057971, "grad_norm": 0.32532766522514756, "learning_rate": 5.9879397435578914e-05, "loss": 0.8771, "step": 4427 }, { "epoch": 2.005434782608696, "grad_norm": 0.2947005690763087, "learning_rate": 5.986842161990443e-05, "loss": 0.8656, "step": 4428 }, { "epoch": 2.0058876811594204, "grad_norm": 0.424533001743201, "learning_rate": 5.9857443817952366e-05, "loss": 0.8951, "step": 4429 }, { "epoch": 2.006340579710145, "grad_norm": 0.3678255276439724, "learning_rate": 5.9846464030820166e-05, "loss": 0.8584, "step": 4430 }, { "epoch": 2.0067934782608696, "grad_norm": 0.28303747163658555, "learning_rate": 5.9835482259605526e-05, "loss": 0.8673, "step": 4431 }, { "epoch": 2.0072463768115942, "grad_norm": 0.3283416356369443, "learning_rate": 5.9824498505406275e-05, "loss": 0.871, "step": 4432 }, { "epoch": 2.007699275362319, "grad_norm": 0.35352597294426785, "learning_rate": 5.9813512769320505e-05, "loss": 0.8644, "step": 4433 }, { "epoch": 2.0081521739130435, "grad_norm": 0.32712950524137596, "learning_rate": 5.980252505244646e-05, "loss": 0.8834, "step": 4434 }, { "epoch": 2.008605072463768, "grad_norm": 0.3120413966524635, "learning_rate": 5.979153535588261e-05, "loss": 0.8902, "step": 4435 }, { "epoch": 2.0090579710144927, "grad_norm": 0.35815952869858775, "learning_rate": 5.97805436807276e-05, "loss": 0.8737, "step": 4436 }, { "epoch": 2.0095108695652173, "grad_norm": 0.3709397499980713, "learning_rate": 5.97695500280803e-05, "loss": 0.8796, "step": 4437 }, { "epoch": 2.009963768115942, "grad_norm": 0.3196528156186191, "learning_rate": 5.975855439903975e-05, "loss": 0.8915, "step": 4438 }, { "epoch": 2.0104166666666665, "grad_norm": 0.2606191193800286, "learning_rate": 5.9747556794705213e-05, "loss": 0.8781, "step": 4439 }, { "epoch": 2.010869565217391, "grad_norm": 0.35459066339044876, "learning_rate": 5.973655721617613e-05, "loss": 0.8815, "step": 4440 }, { "epoch": 2.0113224637681157, "grad_norm": 0.4321464998244881, "learning_rate": 5.9725555664552147e-05, "loss": 0.8929, "step": 4441 }, { "epoch": 2.011775362318841, "grad_norm": 0.47480135032009735, "learning_rate": 5.9714552140933104e-05, "loss": 0.8704, "step": 4442 }, { "epoch": 2.0122282608695654, "grad_norm": 0.4829824049265868, "learning_rate": 5.970354664641904e-05, "loss": 0.8831, "step": 4443 }, { "epoch": 2.01268115942029, "grad_norm": 0.45456070093300793, "learning_rate": 5.9692539182110205e-05, "loss": 0.8757, "step": 4444 }, { "epoch": 2.0131340579710146, "grad_norm": 0.39411410755717236, "learning_rate": 5.968152974910702e-05, "loss": 0.8933, "step": 4445 }, { "epoch": 2.0135869565217392, "grad_norm": 0.3101565345002051, "learning_rate": 5.9670518348510114e-05, "loss": 0.8924, "step": 4446 }, { "epoch": 2.014039855072464, "grad_norm": 0.2740061694520637, "learning_rate": 5.965950498142031e-05, "loss": 0.8656, "step": 4447 }, { "epoch": 2.0144927536231885, "grad_norm": 0.2725940897108291, "learning_rate": 5.964848964893866e-05, "loss": 0.8655, "step": 4448 }, { "epoch": 2.014945652173913, "grad_norm": 0.29951480039261025, "learning_rate": 5.9637472352166345e-05, "loss": 0.8675, "step": 4449 }, { "epoch": 2.0153985507246377, "grad_norm": 0.3335295685235477, "learning_rate": 5.962645309220481e-05, "loss": 0.8756, "step": 4450 }, { "epoch": 2.0158514492753623, "grad_norm": 0.39233790845699545, "learning_rate": 5.961543187015565e-05, "loss": 0.8527, "step": 4451 }, { "epoch": 2.016304347826087, "grad_norm": 0.4102338226596593, "learning_rate": 5.96044086871207e-05, "loss": 0.8777, "step": 4452 }, { "epoch": 2.0167572463768115, "grad_norm": 0.4011083710097336, "learning_rate": 5.959338354420194e-05, "loss": 0.8773, "step": 4453 }, { "epoch": 2.017210144927536, "grad_norm": 0.38298271449987675, "learning_rate": 5.958235644250157e-05, "loss": 0.8745, "step": 4454 }, { "epoch": 2.0176630434782608, "grad_norm": 0.34165905714289757, "learning_rate": 5.9571327383122e-05, "loss": 0.8959, "step": 4455 }, { "epoch": 2.0181159420289854, "grad_norm": 0.26910838519093394, "learning_rate": 5.956029636716581e-05, "loss": 0.8759, "step": 4456 }, { "epoch": 2.01856884057971, "grad_norm": 0.26625868047025397, "learning_rate": 5.9549263395735814e-05, "loss": 0.8885, "step": 4457 }, { "epoch": 2.0190217391304346, "grad_norm": 0.27339703731768117, "learning_rate": 5.953822846993497e-05, "loss": 0.8813, "step": 4458 }, { "epoch": 2.019474637681159, "grad_norm": 0.31946343031598706, "learning_rate": 5.9527191590866464e-05, "loss": 0.8801, "step": 4459 }, { "epoch": 2.0199275362318843, "grad_norm": 0.3339979370220577, "learning_rate": 5.951615275963367e-05, "loss": 0.8749, "step": 4460 }, { "epoch": 2.020380434782609, "grad_norm": 0.3626363332992384, "learning_rate": 5.950511197734017e-05, "loss": 0.8552, "step": 4461 }, { "epoch": 2.0208333333333335, "grad_norm": 0.342864629208014, "learning_rate": 5.94940692450897e-05, "loss": 0.8849, "step": 4462 }, { "epoch": 2.021286231884058, "grad_norm": 0.341302566733858, "learning_rate": 5.948302456398624e-05, "loss": 0.8787, "step": 4463 }, { "epoch": 2.0217391304347827, "grad_norm": 0.3471039592429989, "learning_rate": 5.947197793513394e-05, "loss": 0.8704, "step": 4464 }, { "epoch": 2.0221920289855073, "grad_norm": 0.3226089747377838, "learning_rate": 5.946092935963716e-05, "loss": 0.8835, "step": 4465 }, { "epoch": 2.022644927536232, "grad_norm": 0.32182295812669737, "learning_rate": 5.944987883860043e-05, "loss": 0.8907, "step": 4466 }, { "epoch": 2.0230978260869565, "grad_norm": 0.29772263645756325, "learning_rate": 5.9438826373128485e-05, "loss": 0.8705, "step": 4467 }, { "epoch": 2.023550724637681, "grad_norm": 0.2685961837147486, "learning_rate": 5.942777196432627e-05, "loss": 0.8671, "step": 4468 }, { "epoch": 2.0240036231884058, "grad_norm": 0.2542212069584743, "learning_rate": 5.941671561329891e-05, "loss": 0.8756, "step": 4469 }, { "epoch": 2.0244565217391304, "grad_norm": 0.33376156900189113, "learning_rate": 5.940565732115172e-05, "loss": 0.8899, "step": 4470 }, { "epoch": 2.024909420289855, "grad_norm": 0.4588819707120372, "learning_rate": 5.939459708899021e-05, "loss": 0.88, "step": 4471 }, { "epoch": 2.0253623188405796, "grad_norm": 0.5038554842552639, "learning_rate": 5.9383534917920096e-05, "loss": 0.8844, "step": 4472 }, { "epoch": 2.025815217391304, "grad_norm": 0.49790504054884555, "learning_rate": 5.9372470809047286e-05, "loss": 0.8541, "step": 4473 }, { "epoch": 2.026268115942029, "grad_norm": 0.4902959373694062, "learning_rate": 5.9361404763477865e-05, "loss": 0.884, "step": 4474 }, { "epoch": 2.0267210144927534, "grad_norm": 0.3754038058043893, "learning_rate": 5.935033678231813e-05, "loss": 0.859, "step": 4475 }, { "epoch": 2.027173913043478, "grad_norm": 0.3488535297924424, "learning_rate": 5.933926686667456e-05, "loss": 0.8679, "step": 4476 }, { "epoch": 2.027626811594203, "grad_norm": 0.3101330648895172, "learning_rate": 5.932819501765384e-05, "loss": 0.8686, "step": 4477 }, { "epoch": 2.0280797101449277, "grad_norm": 0.3496466319013611, "learning_rate": 5.931712123636285e-05, "loss": 0.88, "step": 4478 }, { "epoch": 2.0285326086956523, "grad_norm": 0.2810904151822481, "learning_rate": 5.930604552390862e-05, "loss": 0.8705, "step": 4479 }, { "epoch": 2.028985507246377, "grad_norm": 0.30317472136527474, "learning_rate": 5.929496788139843e-05, "loss": 0.8855, "step": 4480 }, { "epoch": 2.0294384057971016, "grad_norm": 0.3456477103959638, "learning_rate": 5.928388830993973e-05, "loss": 0.8934, "step": 4481 }, { "epoch": 2.029891304347826, "grad_norm": 0.32776167192828826, "learning_rate": 5.927280681064016e-05, "loss": 0.8959, "step": 4482 }, { "epoch": 2.0303442028985508, "grad_norm": 0.2503517863821845, "learning_rate": 5.9261723384607546e-05, "loss": 0.866, "step": 4483 }, { "epoch": 2.0307971014492754, "grad_norm": 0.23442173305968694, "learning_rate": 5.9250638032949916e-05, "loss": 0.8712, "step": 4484 }, { "epoch": 2.03125, "grad_norm": 0.28309227363883266, "learning_rate": 5.923955075677551e-05, "loss": 0.8687, "step": 4485 }, { "epoch": 2.0317028985507246, "grad_norm": 0.38856429592228575, "learning_rate": 5.9228461557192724e-05, "loss": 0.8857, "step": 4486 }, { "epoch": 2.0321557971014492, "grad_norm": 0.461569221964781, "learning_rate": 5.921737043531016e-05, "loss": 0.8714, "step": 4487 }, { "epoch": 2.032608695652174, "grad_norm": 0.5197902585140958, "learning_rate": 5.920627739223663e-05, "loss": 0.8737, "step": 4488 }, { "epoch": 2.0330615942028984, "grad_norm": 0.4843117669401289, "learning_rate": 5.9195182429081103e-05, "loss": 0.8495, "step": 4489 }, { "epoch": 2.033514492753623, "grad_norm": 0.42884489557464156, "learning_rate": 5.918408554695278e-05, "loss": 0.891, "step": 4490 }, { "epoch": 2.0339673913043477, "grad_norm": 0.3001341491328616, "learning_rate": 5.9172986746961025e-05, "loss": 0.8643, "step": 4491 }, { "epoch": 2.0344202898550723, "grad_norm": 0.22148021913879573, "learning_rate": 5.91618860302154e-05, "loss": 0.8715, "step": 4492 }, { "epoch": 2.034873188405797, "grad_norm": 0.21056730474987648, "learning_rate": 5.915078339782566e-05, "loss": 0.8816, "step": 4493 }, { "epoch": 2.035326086956522, "grad_norm": 0.31420528837600803, "learning_rate": 5.913967885090177e-05, "loss": 0.8989, "step": 4494 }, { "epoch": 2.0357789855072466, "grad_norm": 0.441183839339466, "learning_rate": 5.9128572390553846e-05, "loss": 0.8716, "step": 4495 }, { "epoch": 2.036231884057971, "grad_norm": 0.4602848290032526, "learning_rate": 5.911746401789223e-05, "loss": 0.8592, "step": 4496 }, { "epoch": 2.036684782608696, "grad_norm": 0.3170694825718748, "learning_rate": 5.910635373402743e-05, "loss": 0.8963, "step": 4497 }, { "epoch": 2.0371376811594204, "grad_norm": 0.2886532237433918, "learning_rate": 5.909524154007019e-05, "loss": 0.8736, "step": 4498 }, { "epoch": 2.037590579710145, "grad_norm": 0.2653117537162811, "learning_rate": 5.908412743713139e-05, "loss": 0.8761, "step": 4499 }, { "epoch": 2.0380434782608696, "grad_norm": 0.22404043935655377, "learning_rate": 5.9073011426322116e-05, "loss": 0.8676, "step": 4500 }, { "epoch": 2.0384963768115942, "grad_norm": 0.2766128899033137, "learning_rate": 5.9061893508753676e-05, "loss": 0.8839, "step": 4501 }, { "epoch": 2.038949275362319, "grad_norm": 0.31121929720377295, "learning_rate": 5.905077368553754e-05, "loss": 0.8719, "step": 4502 }, { "epoch": 2.0394021739130435, "grad_norm": 0.3439769954617727, "learning_rate": 5.903965195778537e-05, "loss": 0.8889, "step": 4503 }, { "epoch": 2.039855072463768, "grad_norm": 0.32952585997331096, "learning_rate": 5.902852832660901e-05, "loss": 0.866, "step": 4504 }, { "epoch": 2.0403079710144927, "grad_norm": 0.2905435909101227, "learning_rate": 5.901740279312053e-05, "loss": 0.8774, "step": 4505 }, { "epoch": 2.0407608695652173, "grad_norm": 0.24647420463841055, "learning_rate": 5.9006275358432145e-05, "loss": 0.8726, "step": 4506 }, { "epoch": 2.041213768115942, "grad_norm": 0.2297426403308828, "learning_rate": 5.8995146023656315e-05, "loss": 0.8784, "step": 4507 }, { "epoch": 2.0416666666666665, "grad_norm": 0.24235981756024016, "learning_rate": 5.8984014789905625e-05, "loss": 0.8616, "step": 4508 }, { "epoch": 2.042119565217391, "grad_norm": 0.2991644139824753, "learning_rate": 5.897288165829289e-05, "loss": 0.8698, "step": 4509 }, { "epoch": 2.0425724637681157, "grad_norm": 0.3183178797539586, "learning_rate": 5.896174662993111e-05, "loss": 0.8817, "step": 4510 }, { "epoch": 2.043025362318841, "grad_norm": 0.29026851389332625, "learning_rate": 5.895060970593348e-05, "loss": 0.8747, "step": 4511 }, { "epoch": 2.0434782608695654, "grad_norm": 0.28412474955313843, "learning_rate": 5.8939470887413364e-05, "loss": 0.8707, "step": 4512 }, { "epoch": 2.04393115942029, "grad_norm": 0.29672724686404645, "learning_rate": 5.892833017548433e-05, "loss": 0.8698, "step": 4513 }, { "epoch": 2.0443840579710146, "grad_norm": 0.32307069898186297, "learning_rate": 5.891718757126013e-05, "loss": 0.8882, "step": 4514 }, { "epoch": 2.0448369565217392, "grad_norm": 0.35659022948333635, "learning_rate": 5.890604307585473e-05, "loss": 0.8708, "step": 4515 }, { "epoch": 2.045289855072464, "grad_norm": 0.34761053840855816, "learning_rate": 5.889489669038223e-05, "loss": 0.8738, "step": 4516 }, { "epoch": 2.0457427536231885, "grad_norm": 0.3676513182860784, "learning_rate": 5.888374841595697e-05, "loss": 0.8721, "step": 4517 }, { "epoch": 2.046195652173913, "grad_norm": 0.3431118759352165, "learning_rate": 5.887259825369345e-05, "loss": 0.8743, "step": 4518 }, { "epoch": 2.0466485507246377, "grad_norm": 0.2998604542815545, "learning_rate": 5.886144620470638e-05, "loss": 0.8691, "step": 4519 }, { "epoch": 2.0471014492753623, "grad_norm": 0.2802301083825554, "learning_rate": 5.885029227011066e-05, "loss": 0.8629, "step": 4520 }, { "epoch": 2.047554347826087, "grad_norm": 0.3300712081575446, "learning_rate": 5.8839136451021337e-05, "loss": 0.8696, "step": 4521 }, { "epoch": 2.0480072463768115, "grad_norm": 0.39860878950038453, "learning_rate": 5.8827978748553685e-05, "loss": 0.8829, "step": 4522 }, { "epoch": 2.048460144927536, "grad_norm": 0.4138795713364075, "learning_rate": 5.8816819163823166e-05, "loss": 0.8533, "step": 4523 }, { "epoch": 2.0489130434782608, "grad_norm": 0.3899278349895713, "learning_rate": 5.880565769794541e-05, "loss": 0.8581, "step": 4524 }, { "epoch": 2.0493659420289854, "grad_norm": 0.28507712219830694, "learning_rate": 5.879449435203627e-05, "loss": 0.8736, "step": 4525 }, { "epoch": 2.04981884057971, "grad_norm": 0.2634248672646501, "learning_rate": 5.878332912721174e-05, "loss": 0.8857, "step": 4526 }, { "epoch": 2.0502717391304346, "grad_norm": 0.33035726620261835, "learning_rate": 5.877216202458803e-05, "loss": 0.8669, "step": 4527 }, { "epoch": 2.050724637681159, "grad_norm": 0.23280990627641032, "learning_rate": 5.876099304528153e-05, "loss": 0.8787, "step": 4528 }, { "epoch": 2.0511775362318843, "grad_norm": 0.22127765824442044, "learning_rate": 5.874982219040884e-05, "loss": 0.8671, "step": 4529 }, { "epoch": 2.051630434782609, "grad_norm": 0.2832748939456172, "learning_rate": 5.87386494610867e-05, "loss": 0.8715, "step": 4530 }, { "epoch": 2.0520833333333335, "grad_norm": 0.2834725572747518, "learning_rate": 5.8727474858432085e-05, "loss": 0.8749, "step": 4531 }, { "epoch": 2.052536231884058, "grad_norm": 0.2637289490474925, "learning_rate": 5.871629838356214e-05, "loss": 0.8648, "step": 4532 }, { "epoch": 2.0529891304347827, "grad_norm": 0.2197024385728896, "learning_rate": 5.870512003759418e-05, "loss": 0.8764, "step": 4533 }, { "epoch": 2.0534420289855073, "grad_norm": 0.260680649257349, "learning_rate": 5.869393982164573e-05, "loss": 0.8806, "step": 4534 }, { "epoch": 2.053894927536232, "grad_norm": 0.28272753120854305, "learning_rate": 5.868275773683449e-05, "loss": 0.861, "step": 4535 }, { "epoch": 2.0543478260869565, "grad_norm": 0.3698597096099344, "learning_rate": 5.867157378427835e-05, "loss": 0.8779, "step": 4536 }, { "epoch": 2.054800724637681, "grad_norm": 0.48852733191481257, "learning_rate": 5.866038796509541e-05, "loss": 0.8666, "step": 4537 }, { "epoch": 2.0552536231884058, "grad_norm": 0.5200893664060111, "learning_rate": 5.8649200280403895e-05, "loss": 0.8656, "step": 4538 }, { "epoch": 2.0557065217391304, "grad_norm": 0.5520209870281232, "learning_rate": 5.863801073132228e-05, "loss": 0.8743, "step": 4539 }, { "epoch": 2.056159420289855, "grad_norm": 0.6226646978686365, "learning_rate": 5.86268193189692e-05, "loss": 0.8802, "step": 4540 }, { "epoch": 2.0566123188405796, "grad_norm": 0.6138072948495882, "learning_rate": 5.861562604446349e-05, "loss": 0.8693, "step": 4541 }, { "epoch": 2.057065217391304, "grad_norm": 0.5992510782450337, "learning_rate": 5.8604430908924126e-05, "loss": 0.8783, "step": 4542 }, { "epoch": 2.057518115942029, "grad_norm": 0.46033530152538793, "learning_rate": 5.8593233913470326e-05, "loss": 0.8549, "step": 4543 }, { "epoch": 2.0579710144927534, "grad_norm": 0.37619542882896767, "learning_rate": 5.858203505922147e-05, "loss": 0.8795, "step": 4544 }, { "epoch": 2.058423913043478, "grad_norm": 0.3167612417869657, "learning_rate": 5.857083434729711e-05, "loss": 0.8441, "step": 4545 }, { "epoch": 2.058876811594203, "grad_norm": 0.28929083397951916, "learning_rate": 5.8559631778817046e-05, "loss": 0.8473, "step": 4546 }, { "epoch": 2.0593297101449277, "grad_norm": 0.27992142316054963, "learning_rate": 5.854842735490116e-05, "loss": 0.8482, "step": 4547 }, { "epoch": 2.0597826086956523, "grad_norm": 0.3264941171302842, "learning_rate": 5.85372210766696e-05, "loss": 0.8772, "step": 4548 }, { "epoch": 2.060235507246377, "grad_norm": 0.4402539982927182, "learning_rate": 5.852601294524268e-05, "loss": 0.8619, "step": 4549 }, { "epoch": 2.0606884057971016, "grad_norm": 0.43224359004054536, "learning_rate": 5.851480296174089e-05, "loss": 0.9006, "step": 4550 }, { "epoch": 2.061141304347826, "grad_norm": 0.44193984266248854, "learning_rate": 5.850359112728489e-05, "loss": 0.8824, "step": 4551 }, { "epoch": 2.0615942028985508, "grad_norm": 0.47433721547813174, "learning_rate": 5.849237744299558e-05, "loss": 0.8782, "step": 4552 }, { "epoch": 2.0620471014492754, "grad_norm": 0.5568927895896525, "learning_rate": 5.848116190999399e-05, "loss": 0.8785, "step": 4553 }, { "epoch": 2.0625, "grad_norm": 0.5869846800212092, "learning_rate": 5.846994452940137e-05, "loss": 0.8791, "step": 4554 }, { "epoch": 2.0629528985507246, "grad_norm": 0.550378748136553, "learning_rate": 5.845872530233911e-05, "loss": 0.878, "step": 4555 }, { "epoch": 2.0634057971014492, "grad_norm": 0.4538617377266762, "learning_rate": 5.844750422992883e-05, "loss": 0.8696, "step": 4556 }, { "epoch": 2.063858695652174, "grad_norm": 0.38436831182955483, "learning_rate": 5.843628131329233e-05, "loss": 0.8747, "step": 4557 }, { "epoch": 2.0643115942028984, "grad_norm": 0.32093925463091444, "learning_rate": 5.842505655355158e-05, "loss": 0.9046, "step": 4558 }, { "epoch": 2.064764492753623, "grad_norm": 0.2857669614977709, "learning_rate": 5.841382995182871e-05, "loss": 0.8823, "step": 4559 }, { "epoch": 2.0652173913043477, "grad_norm": 0.23490346171241488, "learning_rate": 5.840260150924609e-05, "loss": 0.8715, "step": 4560 }, { "epoch": 2.0656702898550723, "grad_norm": 0.22652274039182937, "learning_rate": 5.839137122692624e-05, "loss": 0.8477, "step": 4561 }, { "epoch": 2.066123188405797, "grad_norm": 0.2537792775172766, "learning_rate": 5.838013910599186e-05, "loss": 0.8725, "step": 4562 }, { "epoch": 2.066576086956522, "grad_norm": 0.29426283475880083, "learning_rate": 5.836890514756585e-05, "loss": 0.8867, "step": 4563 }, { "epoch": 2.0670289855072466, "grad_norm": 0.3136092665440354, "learning_rate": 5.835766935277128e-05, "loss": 0.8557, "step": 4564 }, { "epoch": 2.067481884057971, "grad_norm": 0.3076528835357055, "learning_rate": 5.834643172273141e-05, "loss": 0.8807, "step": 4565 }, { "epoch": 2.067934782608696, "grad_norm": 0.3858500223826293, "learning_rate": 5.833519225856969e-05, "loss": 0.8824, "step": 4566 }, { "epoch": 2.0683876811594204, "grad_norm": 0.3947825942668338, "learning_rate": 5.8323950961409753e-05, "loss": 0.8913, "step": 4567 }, { "epoch": 2.068840579710145, "grad_norm": 0.40470278671375465, "learning_rate": 5.83127078323754e-05, "loss": 0.862, "step": 4568 }, { "epoch": 2.0692934782608696, "grad_norm": 0.3509157416887644, "learning_rate": 5.830146287259062e-05, "loss": 0.8927, "step": 4569 }, { "epoch": 2.0697463768115942, "grad_norm": 0.37148390192217756, "learning_rate": 5.82902160831796e-05, "loss": 0.8955, "step": 4570 }, { "epoch": 2.070199275362319, "grad_norm": 0.2995001345396057, "learning_rate": 5.8278967465266706e-05, "loss": 0.8711, "step": 4571 }, { "epoch": 2.0706521739130435, "grad_norm": 0.28134120675375135, "learning_rate": 5.826771701997645e-05, "loss": 0.8903, "step": 4572 }, { "epoch": 2.071105072463768, "grad_norm": 0.292615279976981, "learning_rate": 5.825646474843359e-05, "loss": 0.8744, "step": 4573 }, { "epoch": 2.0715579710144927, "grad_norm": 0.25420639094830205, "learning_rate": 5.824521065176301e-05, "loss": 0.884, "step": 4574 }, { "epoch": 2.0720108695652173, "grad_norm": 0.26064387309757786, "learning_rate": 5.8233954731089816e-05, "loss": 0.8684, "step": 4575 }, { "epoch": 2.072463768115942, "grad_norm": 0.2414180936501228, "learning_rate": 5.8222696987539264e-05, "loss": 0.8963, "step": 4576 }, { "epoch": 2.0729166666666665, "grad_norm": 0.18875068568377543, "learning_rate": 5.821143742223682e-05, "loss": 0.8847, "step": 4577 }, { "epoch": 2.073369565217391, "grad_norm": 0.1795063729606562, "learning_rate": 5.820017603630812e-05, "loss": 0.8779, "step": 4578 }, { "epoch": 2.0738224637681157, "grad_norm": 0.19827724376248096, "learning_rate": 5.818891283087899e-05, "loss": 0.8855, "step": 4579 }, { "epoch": 2.074275362318841, "grad_norm": 0.18561083525878735, "learning_rate": 5.8177647807075413e-05, "loss": 0.8752, "step": 4580 }, { "epoch": 2.0747282608695654, "grad_norm": 0.20967531085462218, "learning_rate": 5.816638096602358e-05, "loss": 0.8657, "step": 4581 }, { "epoch": 2.07518115942029, "grad_norm": 0.2315229781214519, "learning_rate": 5.815511230884985e-05, "loss": 0.8735, "step": 4582 }, { "epoch": 2.0756340579710146, "grad_norm": 0.2534413646432964, "learning_rate": 5.814384183668079e-05, "loss": 0.8706, "step": 4583 }, { "epoch": 2.0760869565217392, "grad_norm": 0.3572361464011894, "learning_rate": 5.81325695506431e-05, "loss": 0.8912, "step": 4584 }, { "epoch": 2.076539855072464, "grad_norm": 0.3914854974852823, "learning_rate": 5.812129545186369e-05, "loss": 0.8808, "step": 4585 }, { "epoch": 2.0769927536231885, "grad_norm": 0.3394579845404793, "learning_rate": 5.8110019541469663e-05, "loss": 0.8964, "step": 4586 }, { "epoch": 2.077445652173913, "grad_norm": 0.3547942777991634, "learning_rate": 5.8098741820588286e-05, "loss": 0.8766, "step": 4587 }, { "epoch": 2.0778985507246377, "grad_norm": 0.35392854591878475, "learning_rate": 5.808746229034702e-05, "loss": 0.8719, "step": 4588 }, { "epoch": 2.0783514492753623, "grad_norm": 0.3866135169734733, "learning_rate": 5.807618095187347e-05, "loss": 0.8772, "step": 4589 }, { "epoch": 2.078804347826087, "grad_norm": 0.3930517496136729, "learning_rate": 5.806489780629547e-05, "loss": 0.859, "step": 4590 }, { "epoch": 2.0792572463768115, "grad_norm": 0.3730809644607854, "learning_rate": 5.805361285474101e-05, "loss": 0.8758, "step": 4591 }, { "epoch": 2.079710144927536, "grad_norm": 0.3366009823391214, "learning_rate": 5.8042326098338285e-05, "loss": 0.8809, "step": 4592 }, { "epoch": 2.0801630434782608, "grad_norm": 0.35675197144713205, "learning_rate": 5.8031037538215606e-05, "loss": 0.8743, "step": 4593 }, { "epoch": 2.0806159420289854, "grad_norm": 0.38694696242038074, "learning_rate": 5.801974717550154e-05, "loss": 0.8782, "step": 4594 }, { "epoch": 2.08106884057971, "grad_norm": 0.4140831551454422, "learning_rate": 5.800845501132479e-05, "loss": 0.8719, "step": 4595 }, { "epoch": 2.0815217391304346, "grad_norm": 0.41426837528491117, "learning_rate": 5.799716104681426e-05, "loss": 0.8639, "step": 4596 }, { "epoch": 2.081974637681159, "grad_norm": 0.45308883848556775, "learning_rate": 5.798586528309901e-05, "loss": 0.8447, "step": 4597 }, { "epoch": 2.0824275362318843, "grad_norm": 0.6030012219537734, "learning_rate": 5.797456772130832e-05, "loss": 0.858, "step": 4598 }, { "epoch": 2.082880434782609, "grad_norm": 0.5130950428267984, "learning_rate": 5.7963268362571606e-05, "loss": 0.8649, "step": 4599 }, { "epoch": 2.0833333333333335, "grad_norm": 0.4750875083087678, "learning_rate": 5.79519672080185e-05, "loss": 0.8853, "step": 4600 }, { "epoch": 2.083786231884058, "grad_norm": 0.4160091908995181, "learning_rate": 5.7940664258778764e-05, "loss": 0.8999, "step": 4601 }, { "epoch": 2.0842391304347827, "grad_norm": 0.4382525980693855, "learning_rate": 5.792935951598241e-05, "loss": 0.8907, "step": 4602 }, { "epoch": 2.0846920289855073, "grad_norm": 0.4568499141871992, "learning_rate": 5.7918052980759554e-05, "loss": 0.8771, "step": 4603 }, { "epoch": 2.085144927536232, "grad_norm": 0.3452072800795893, "learning_rate": 5.790674465424057e-05, "loss": 0.8793, "step": 4604 }, { "epoch": 2.0855978260869565, "grad_norm": 0.5276721164690477, "learning_rate": 5.7895434537555936e-05, "loss": 0.8829, "step": 4605 }, { "epoch": 2.086050724637681, "grad_norm": 0.26206493158251426, "learning_rate": 5.7884122631836364e-05, "loss": 0.8749, "step": 4606 }, { "epoch": 2.0865036231884058, "grad_norm": 0.26822146950738013, "learning_rate": 5.78728089382127e-05, "loss": 0.8711, "step": 4607 }, { "epoch": 2.0869565217391304, "grad_norm": 0.3636881884062064, "learning_rate": 5.7861493457816005e-05, "loss": 0.8588, "step": 4608 }, { "epoch": 2.087409420289855, "grad_norm": 0.36761963386492996, "learning_rate": 5.7850176191777514e-05, "loss": 0.8718, "step": 4609 }, { "epoch": 2.0878623188405796, "grad_norm": 0.3425410645387502, "learning_rate": 5.783885714122862e-05, "loss": 0.879, "step": 4610 }, { "epoch": 2.088315217391304, "grad_norm": 0.37066454460382137, "learning_rate": 5.7827536307300904e-05, "loss": 0.8736, "step": 4611 }, { "epoch": 2.088768115942029, "grad_norm": 0.39792477334081844, "learning_rate": 5.781621369112613e-05, "loss": 0.8812, "step": 4612 }, { "epoch": 2.0892210144927534, "grad_norm": 0.29261594749199993, "learning_rate": 5.7804889293836254e-05, "loss": 0.885, "step": 4613 }, { "epoch": 2.089673913043478, "grad_norm": 0.25026828409648455, "learning_rate": 5.7793563116563376e-05, "loss": 0.8674, "step": 4614 }, { "epoch": 2.090126811594203, "grad_norm": 0.3272596563348139, "learning_rate": 5.778223516043979e-05, "loss": 0.8728, "step": 4615 }, { "epoch": 2.0905797101449277, "grad_norm": 0.42527187984753134, "learning_rate": 5.7770905426597986e-05, "loss": 0.8895, "step": 4616 }, { "epoch": 2.0910326086956523, "grad_norm": 0.5476702038926821, "learning_rate": 5.775957391617061e-05, "loss": 0.8649, "step": 4617 }, { "epoch": 2.091485507246377, "grad_norm": 0.5961164214975048, "learning_rate": 5.774824063029048e-05, "loss": 0.8795, "step": 4618 }, { "epoch": 2.0919384057971016, "grad_norm": 0.4991593760658182, "learning_rate": 5.77369055700906e-05, "loss": 0.877, "step": 4619 }, { "epoch": 2.092391304347826, "grad_norm": 0.4418141950895011, "learning_rate": 5.772556873670417e-05, "loss": 0.8757, "step": 4620 }, { "epoch": 2.0928442028985508, "grad_norm": 0.41945672775174514, "learning_rate": 5.771423013126455e-05, "loss": 0.876, "step": 4621 }, { "epoch": 2.0932971014492754, "grad_norm": 0.3662490034168141, "learning_rate": 5.7702889754905276e-05, "loss": 0.8869, "step": 4622 }, { "epoch": 2.09375, "grad_norm": 0.3110485895633489, "learning_rate": 5.7691547608760055e-05, "loss": 0.8944, "step": 4623 }, { "epoch": 2.0942028985507246, "grad_norm": 0.341042364738184, "learning_rate": 5.768020369396278e-05, "loss": 0.8633, "step": 4624 }, { "epoch": 2.0946557971014492, "grad_norm": 0.3987833385295322, "learning_rate": 5.766885801164754e-05, "loss": 0.8743, "step": 4625 }, { "epoch": 2.095108695652174, "grad_norm": 0.3822545101781682, "learning_rate": 5.7657510562948555e-05, "loss": 0.8824, "step": 4626 }, { "epoch": 2.0955615942028984, "grad_norm": 0.3007377976099677, "learning_rate": 5.764616134900026e-05, "loss": 0.8716, "step": 4627 }, { "epoch": 2.096014492753623, "grad_norm": 0.20460828467766146, "learning_rate": 5.763481037093725e-05, "loss": 0.8436, "step": 4628 }, { "epoch": 2.0964673913043477, "grad_norm": 0.2619519810456749, "learning_rate": 5.7623457629894314e-05, "loss": 0.8684, "step": 4629 }, { "epoch": 2.0969202898550723, "grad_norm": 0.35113991198923256, "learning_rate": 5.7612103127006375e-05, "loss": 0.8584, "step": 4630 }, { "epoch": 2.097373188405797, "grad_norm": 0.36997425177012, "learning_rate": 5.760074686340858e-05, "loss": 0.8542, "step": 4631 }, { "epoch": 2.097826086956522, "grad_norm": 0.3969462165342517, "learning_rate": 5.7589388840236235e-05, "loss": 0.8756, "step": 4632 }, { "epoch": 2.0982789855072466, "grad_norm": 0.43454772475550413, "learning_rate": 5.757802905862481e-05, "loss": 0.8879, "step": 4633 }, { "epoch": 2.098731884057971, "grad_norm": 0.41963446990736475, "learning_rate": 5.756666751970998e-05, "loss": 0.8671, "step": 4634 }, { "epoch": 2.099184782608696, "grad_norm": 0.37493110397717444, "learning_rate": 5.755530422462755e-05, "loss": 0.8938, "step": 4635 }, { "epoch": 2.0996376811594204, "grad_norm": 0.33049265870926087, "learning_rate": 5.754393917451354e-05, "loss": 0.8638, "step": 4636 }, { "epoch": 2.100090579710145, "grad_norm": 0.33151153241770376, "learning_rate": 5.7532572370504126e-05, "loss": 0.8743, "step": 4637 }, { "epoch": 2.1005434782608696, "grad_norm": 0.32413749879805726, "learning_rate": 5.752120381373568e-05, "loss": 0.8498, "step": 4638 }, { "epoch": 2.1009963768115942, "grad_norm": 0.3103543597212889, "learning_rate": 5.750983350534471e-05, "loss": 0.8847, "step": 4639 }, { "epoch": 2.101449275362319, "grad_norm": 0.27096189454740455, "learning_rate": 5.749846144646794e-05, "loss": 0.8687, "step": 4640 }, { "epoch": 2.1019021739130435, "grad_norm": 0.20266986197907602, "learning_rate": 5.7487087638242265e-05, "loss": 0.8799, "step": 4641 }, { "epoch": 2.102355072463768, "grad_norm": 0.20862354139052644, "learning_rate": 5.747571208180472e-05, "loss": 0.8808, "step": 4642 }, { "epoch": 2.1028079710144927, "grad_norm": 0.25287627832708753, "learning_rate": 5.7464334778292535e-05, "loss": 0.8671, "step": 4643 }, { "epoch": 2.1032608695652173, "grad_norm": 0.25178477801643917, "learning_rate": 5.745295572884315e-05, "loss": 0.8751, "step": 4644 }, { "epoch": 2.103713768115942, "grad_norm": 0.26840304670618664, "learning_rate": 5.74415749345941e-05, "loss": 0.8747, "step": 4645 }, { "epoch": 2.1041666666666665, "grad_norm": 0.3359136684893718, "learning_rate": 5.743019239668318e-05, "loss": 0.8578, "step": 4646 }, { "epoch": 2.104619565217391, "grad_norm": 0.36765504174507874, "learning_rate": 5.741880811624831e-05, "loss": 0.8818, "step": 4647 }, { "epoch": 2.1050724637681157, "grad_norm": 0.34732621186699103, "learning_rate": 5.740742209442758e-05, "loss": 0.8778, "step": 4648 }, { "epoch": 2.105525362318841, "grad_norm": 0.2672251381715893, "learning_rate": 5.739603433235927e-05, "loss": 0.8471, "step": 4649 }, { "epoch": 2.1059782608695654, "grad_norm": 0.27165164494888555, "learning_rate": 5.738464483118186e-05, "loss": 0.8759, "step": 4650 }, { "epoch": 2.10643115942029, "grad_norm": 0.2893942530972752, "learning_rate": 5.737325359203395e-05, "loss": 0.8768, "step": 4651 }, { "epoch": 2.1068840579710146, "grad_norm": 0.3101330387348875, "learning_rate": 5.736186061605435e-05, "loss": 0.8638, "step": 4652 }, { "epoch": 2.1073369565217392, "grad_norm": 0.298808896474537, "learning_rate": 5.735046590438202e-05, "loss": 0.874, "step": 4653 }, { "epoch": 2.107789855072464, "grad_norm": 0.29282868426787634, "learning_rate": 5.733906945815613e-05, "loss": 0.868, "step": 4654 }, { "epoch": 2.1082427536231885, "grad_norm": 0.3053528771174306, "learning_rate": 5.732767127851599e-05, "loss": 0.8545, "step": 4655 }, { "epoch": 2.108695652173913, "grad_norm": 0.3382797607705386, "learning_rate": 5.73162713666011e-05, "loss": 0.8714, "step": 4656 }, { "epoch": 2.1091485507246377, "grad_norm": 0.37573115324102196, "learning_rate": 5.730486972355112e-05, "loss": 0.8567, "step": 4657 }, { "epoch": 2.1096014492753623, "grad_norm": 0.34466304503318823, "learning_rate": 5.729346635050588e-05, "loss": 0.867, "step": 4658 }, { "epoch": 2.110054347826087, "grad_norm": 0.2376948583462052, "learning_rate": 5.728206124860541e-05, "loss": 0.9045, "step": 4659 }, { "epoch": 2.1105072463768115, "grad_norm": 0.32609022796817405, "learning_rate": 5.7270654418989904e-05, "loss": 0.8979, "step": 4660 }, { "epoch": 2.110960144927536, "grad_norm": 0.3121289981244467, "learning_rate": 5.725924586279969e-05, "loss": 0.8789, "step": 4661 }, { "epoch": 2.1114130434782608, "grad_norm": 0.23375142025508883, "learning_rate": 5.7247835581175326e-05, "loss": 0.887, "step": 4662 }, { "epoch": 2.1118659420289854, "grad_norm": 0.2902372778382258, "learning_rate": 5.7236423575257496e-05, "loss": 0.8793, "step": 4663 }, { "epoch": 2.11231884057971, "grad_norm": 0.27814117033756824, "learning_rate": 5.722500984618711e-05, "loss": 0.8724, "step": 4664 }, { "epoch": 2.1127717391304346, "grad_norm": 0.26982036986904606, "learning_rate": 5.721359439510518e-05, "loss": 0.8518, "step": 4665 }, { "epoch": 2.113224637681159, "grad_norm": 0.2708467980554112, "learning_rate": 5.720217722315294e-05, "loss": 0.89, "step": 4666 }, { "epoch": 2.1136775362318843, "grad_norm": 0.31852525898187106, "learning_rate": 5.719075833147179e-05, "loss": 0.8689, "step": 4667 }, { "epoch": 2.114130434782609, "grad_norm": 0.3837462523887715, "learning_rate": 5.717933772120329e-05, "loss": 0.8835, "step": 4668 }, { "epoch": 2.1145833333333335, "grad_norm": 0.3233568445202643, "learning_rate": 5.716791539348917e-05, "loss": 0.8668, "step": 4669 }, { "epoch": 2.115036231884058, "grad_norm": 0.3473525387616927, "learning_rate": 5.7156491349471336e-05, "loss": 0.8754, "step": 4670 }, { "epoch": 2.1154891304347827, "grad_norm": 0.42326543448418036, "learning_rate": 5.714506559029188e-05, "loss": 0.8892, "step": 4671 }, { "epoch": 2.1159420289855073, "grad_norm": 0.5004969897250938, "learning_rate": 5.713363811709306e-05, "loss": 0.8814, "step": 4672 }, { "epoch": 2.116394927536232, "grad_norm": 0.42098590353904614, "learning_rate": 5.7122208931017276e-05, "loss": 0.8675, "step": 4673 }, { "epoch": 2.1168478260869565, "grad_norm": 0.47527053814383485, "learning_rate": 5.711077803320713e-05, "loss": 0.8748, "step": 4674 }, { "epoch": 2.117300724637681, "grad_norm": 0.46571389446870864, "learning_rate": 5.7099345424805394e-05, "loss": 0.8565, "step": 4675 }, { "epoch": 2.1177536231884058, "grad_norm": 0.46874444070143517, "learning_rate": 5.7087911106955e-05, "loss": 0.8573, "step": 4676 }, { "epoch": 2.1182065217391304, "grad_norm": 0.48971895776078994, "learning_rate": 5.7076475080799055e-05, "loss": 0.8864, "step": 4677 }, { "epoch": 2.118659420289855, "grad_norm": 0.5352354675626955, "learning_rate": 5.706503734748084e-05, "loss": 0.8774, "step": 4678 }, { "epoch": 2.1191123188405796, "grad_norm": 0.47646117854049846, "learning_rate": 5.7053597908143797e-05, "loss": 0.8941, "step": 4679 }, { "epoch": 2.119565217391304, "grad_norm": 0.4494309910182222, "learning_rate": 5.704215676393154e-05, "loss": 0.8859, "step": 4680 }, { "epoch": 2.120018115942029, "grad_norm": 0.4251133988943027, "learning_rate": 5.703071391598787e-05, "loss": 0.8775, "step": 4681 }, { "epoch": 2.1204710144927534, "grad_norm": 0.4245737270298813, "learning_rate": 5.701926936545675e-05, "loss": 0.8619, "step": 4682 }, { "epoch": 2.120923913043478, "grad_norm": 0.37677016161408555, "learning_rate": 5.7007823113482286e-05, "loss": 0.8454, "step": 4683 }, { "epoch": 2.121376811594203, "grad_norm": 0.3177006700878052, "learning_rate": 5.699637516120881e-05, "loss": 0.8586, "step": 4684 }, { "epoch": 2.1218297101449277, "grad_norm": 0.284890552599405, "learning_rate": 5.6984925509780774e-05, "loss": 0.8907, "step": 4685 }, { "epoch": 2.1222826086956523, "grad_norm": 0.29061585131979933, "learning_rate": 5.697347416034281e-05, "loss": 0.8658, "step": 4686 }, { "epoch": 2.122735507246377, "grad_norm": 0.30267001075948274, "learning_rate": 5.696202111403974e-05, "loss": 0.8822, "step": 4687 }, { "epoch": 2.1231884057971016, "grad_norm": 0.24061336467353822, "learning_rate": 5.695056637201655e-05, "loss": 0.8943, "step": 4688 }, { "epoch": 2.123641304347826, "grad_norm": 0.2119746606236408, "learning_rate": 5.693910993541837e-05, "loss": 0.8534, "step": 4689 }, { "epoch": 2.1240942028985508, "grad_norm": 0.26771900449940844, "learning_rate": 5.692765180539053e-05, "loss": 0.8708, "step": 4690 }, { "epoch": 2.1245471014492754, "grad_norm": 0.3278008895727268, "learning_rate": 5.691619198307852e-05, "loss": 0.8856, "step": 4691 }, { "epoch": 2.125, "grad_norm": 0.31897278617956754, "learning_rate": 5.6904730469627985e-05, "loss": 0.8726, "step": 4692 }, { "epoch": 2.1254528985507246, "grad_norm": 0.24168579687955807, "learning_rate": 5.689326726618476e-05, "loss": 0.8689, "step": 4693 }, { "epoch": 2.1259057971014492, "grad_norm": 0.22492021458980532, "learning_rate": 5.688180237389484e-05, "loss": 0.885, "step": 4694 }, { "epoch": 2.126358695652174, "grad_norm": 0.3397933313455651, "learning_rate": 5.687033579390438e-05, "loss": 0.8774, "step": 4695 }, { "epoch": 2.1268115942028984, "grad_norm": 0.39319933555344133, "learning_rate": 5.6858867527359706e-05, "loss": 0.88, "step": 4696 }, { "epoch": 2.127264492753623, "grad_norm": 0.4428068871994456, "learning_rate": 5.684739757540735e-05, "loss": 0.8884, "step": 4697 }, { "epoch": 2.1277173913043477, "grad_norm": 0.5774716009867357, "learning_rate": 5.683592593919395e-05, "loss": 0.8819, "step": 4698 }, { "epoch": 2.1281702898550723, "grad_norm": 0.7689119358711755, "learning_rate": 5.682445261986636e-05, "loss": 0.8689, "step": 4699 }, { "epoch": 2.128623188405797, "grad_norm": 0.8959426128539515, "learning_rate": 5.681297761857158e-05, "loss": 0.8615, "step": 4700 }, { "epoch": 2.1290760869565215, "grad_norm": 0.8720050029070989, "learning_rate": 5.680150093645678e-05, "loss": 0.8785, "step": 4701 }, { "epoch": 2.1295289855072466, "grad_norm": 0.7213100575537778, "learning_rate": 5.6790022574669324e-05, "loss": 0.8789, "step": 4702 }, { "epoch": 2.129981884057971, "grad_norm": 0.5785196214513209, "learning_rate": 5.677854253435669e-05, "loss": 0.8744, "step": 4703 }, { "epoch": 2.130434782608696, "grad_norm": 0.4212932320656827, "learning_rate": 5.676706081666657e-05, "loss": 0.8917, "step": 4704 }, { "epoch": 2.1308876811594204, "grad_norm": 0.20454180212321735, "learning_rate": 5.6755577422746805e-05, "loss": 0.8858, "step": 4705 }, { "epoch": 2.131340579710145, "grad_norm": 0.3324216642312036, "learning_rate": 5.6744092353745444e-05, "loss": 0.8791, "step": 4706 }, { "epoch": 2.1317934782608696, "grad_norm": 0.5311222495422808, "learning_rate": 5.67326056108106e-05, "loss": 0.872, "step": 4707 }, { "epoch": 2.1322463768115942, "grad_norm": 0.686717907374942, "learning_rate": 5.672111719509068e-05, "loss": 0.8751, "step": 4708 }, { "epoch": 2.132699275362319, "grad_norm": 0.7850284535217064, "learning_rate": 5.670962710773417e-05, "loss": 0.8842, "step": 4709 }, { "epoch": 2.1331521739130435, "grad_norm": 0.7795003760454137, "learning_rate": 5.669813534988977e-05, "loss": 0.8842, "step": 4710 }, { "epoch": 2.133605072463768, "grad_norm": 0.6478636576222819, "learning_rate": 5.66866419227063e-05, "loss": 0.856, "step": 4711 }, { "epoch": 2.1340579710144927, "grad_norm": 0.4875938102598631, "learning_rate": 5.6675146827332805e-05, "loss": 0.8723, "step": 4712 }, { "epoch": 2.1345108695652173, "grad_norm": 0.2507724685057168, "learning_rate": 5.6663650064918456e-05, "loss": 0.8733, "step": 4713 }, { "epoch": 2.134963768115942, "grad_norm": 0.2910608644286923, "learning_rate": 5.665215163661261e-05, "loss": 0.8984, "step": 4714 }, { "epoch": 2.1354166666666665, "grad_norm": 0.3850514543085621, "learning_rate": 5.664065154356477e-05, "loss": 0.8868, "step": 4715 }, { "epoch": 2.135869565217391, "grad_norm": 0.5336400683686772, "learning_rate": 5.662914978692463e-05, "loss": 0.8921, "step": 4716 }, { "epoch": 2.1363224637681157, "grad_norm": 0.6520844835156953, "learning_rate": 5.6617646367842025e-05, "loss": 0.8645, "step": 4717 }, { "epoch": 2.136775362318841, "grad_norm": 0.6338338264425499, "learning_rate": 5.6606141287466996e-05, "loss": 0.8817, "step": 4718 }, { "epoch": 2.1372282608695654, "grad_norm": 0.5253170694570024, "learning_rate": 5.659463454694969e-05, "loss": 0.8977, "step": 4719 }, { "epoch": 2.13768115942029, "grad_norm": 0.41840351403938725, "learning_rate": 5.6583126147440486e-05, "loss": 0.8831, "step": 4720 }, { "epoch": 2.1381340579710146, "grad_norm": 0.33574135145660283, "learning_rate": 5.657161609008987e-05, "loss": 0.8582, "step": 4721 }, { "epoch": 2.1385869565217392, "grad_norm": 0.28106697044629364, "learning_rate": 5.656010437604854e-05, "loss": 0.8646, "step": 4722 }, { "epoch": 2.139039855072464, "grad_norm": 0.3535573202213538, "learning_rate": 5.654859100646734e-05, "loss": 0.8895, "step": 4723 }, { "epoch": 2.1394927536231885, "grad_norm": 0.35621150103818244, "learning_rate": 5.653707598249726e-05, "loss": 0.8733, "step": 4724 }, { "epoch": 2.139945652173913, "grad_norm": 0.27554214683794653, "learning_rate": 5.6525559305289506e-05, "loss": 0.8894, "step": 4725 }, { "epoch": 2.1403985507246377, "grad_norm": 0.358785254007161, "learning_rate": 5.651404097599539e-05, "loss": 0.8758, "step": 4726 }, { "epoch": 2.1408514492753623, "grad_norm": 0.3564001197896599, "learning_rate": 5.650252099576644e-05, "loss": 0.8716, "step": 4727 }, { "epoch": 2.141304347826087, "grad_norm": 0.17353104108226913, "learning_rate": 5.6490999365754304e-05, "loss": 0.8749, "step": 4728 }, { "epoch": 2.1417572463768115, "grad_norm": 0.26808180694405725, "learning_rate": 5.6479476087110835e-05, "loss": 0.8757, "step": 4729 }, { "epoch": 2.142210144927536, "grad_norm": 0.2833225312310797, "learning_rate": 5.646795116098801e-05, "loss": 0.893, "step": 4730 }, { "epoch": 2.1426630434782608, "grad_norm": 0.2496654503200682, "learning_rate": 5.645642458853804e-05, "loss": 0.8702, "step": 4731 }, { "epoch": 2.1431159420289854, "grad_norm": 0.2972207444150904, "learning_rate": 5.644489637091321e-05, "loss": 0.8835, "step": 4732 }, { "epoch": 2.14356884057971, "grad_norm": 0.3313079387259796, "learning_rate": 5.643336650926603e-05, "loss": 0.8866, "step": 4733 }, { "epoch": 2.1440217391304346, "grad_norm": 0.3615902897576942, "learning_rate": 5.642183500474916e-05, "loss": 0.8862, "step": 4734 }, { "epoch": 2.144474637681159, "grad_norm": 0.30197106828249665, "learning_rate": 5.6410301858515435e-05, "loss": 0.8795, "step": 4735 }, { "epoch": 2.1449275362318843, "grad_norm": 0.34076402409641987, "learning_rate": 5.639876707171782e-05, "loss": 0.887, "step": 4736 }, { "epoch": 2.145380434782609, "grad_norm": 0.3771948401864898, "learning_rate": 5.638723064550947e-05, "loss": 0.8454, "step": 4737 }, { "epoch": 2.1458333333333335, "grad_norm": 0.3616898692341585, "learning_rate": 5.6375692581043705e-05, "loss": 0.865, "step": 4738 }, { "epoch": 2.146286231884058, "grad_norm": 0.3549694375989574, "learning_rate": 5.636415287947401e-05, "loss": 0.8634, "step": 4739 }, { "epoch": 2.1467391304347827, "grad_norm": 0.3790991596380694, "learning_rate": 5.635261154195402e-05, "loss": 0.8797, "step": 4740 }, { "epoch": 2.1471920289855073, "grad_norm": 0.4257936523728185, "learning_rate": 5.634106856963753e-05, "loss": 0.8609, "step": 4741 }, { "epoch": 2.147644927536232, "grad_norm": 0.3850337609280332, "learning_rate": 5.6329523963678525e-05, "loss": 0.8662, "step": 4742 }, { "epoch": 2.1480978260869565, "grad_norm": 0.36018119593576603, "learning_rate": 5.631797772523113e-05, "loss": 0.8752, "step": 4743 }, { "epoch": 2.148550724637681, "grad_norm": 0.3499620008822903, "learning_rate": 5.630642985544967e-05, "loss": 0.8622, "step": 4744 }, { "epoch": 2.1490036231884058, "grad_norm": 0.3973085237031702, "learning_rate": 5.6294880355488555e-05, "loss": 0.8573, "step": 4745 }, { "epoch": 2.1494565217391304, "grad_norm": 0.33781920795042186, "learning_rate": 5.628332922650242e-05, "loss": 0.8581, "step": 4746 }, { "epoch": 2.149909420289855, "grad_norm": 0.23708732664105092, "learning_rate": 5.627177646964607e-05, "loss": 0.8622, "step": 4747 }, { "epoch": 2.1503623188405796, "grad_norm": 0.23703472244530396, "learning_rate": 5.626022208607445e-05, "loss": 0.8802, "step": 4748 }, { "epoch": 2.150815217391304, "grad_norm": 0.3454872935747737, "learning_rate": 5.624866607694266e-05, "loss": 0.8598, "step": 4749 }, { "epoch": 2.151268115942029, "grad_norm": 0.3719165797024765, "learning_rate": 5.623710844340596e-05, "loss": 0.8772, "step": 4750 }, { "epoch": 2.1517210144927534, "grad_norm": 0.3269556339963967, "learning_rate": 5.622554918661981e-05, "loss": 0.8814, "step": 4751 }, { "epoch": 2.1521739130434785, "grad_norm": 0.36004711621288027, "learning_rate": 5.621398830773979e-05, "loss": 0.8499, "step": 4752 }, { "epoch": 2.152626811594203, "grad_norm": 0.43753840979181946, "learning_rate": 5.620242580792167e-05, "loss": 0.8779, "step": 4753 }, { "epoch": 2.1530797101449277, "grad_norm": 0.43705624091025924, "learning_rate": 5.6190861688321366e-05, "loss": 0.8854, "step": 4754 }, { "epoch": 2.1535326086956523, "grad_norm": 0.4693284510920875, "learning_rate": 5.617929595009497e-05, "loss": 0.8425, "step": 4755 }, { "epoch": 2.153985507246377, "grad_norm": 0.5621308123564648, "learning_rate": 5.616772859439871e-05, "loss": 0.8972, "step": 4756 }, { "epoch": 2.1544384057971016, "grad_norm": 0.5204523506802088, "learning_rate": 5.6156159622389006e-05, "loss": 0.8726, "step": 4757 }, { "epoch": 2.154891304347826, "grad_norm": 0.4536963358530456, "learning_rate": 5.6144589035222416e-05, "loss": 0.8364, "step": 4758 }, { "epoch": 2.1553442028985508, "grad_norm": 0.4116926969795311, "learning_rate": 5.613301683405568e-05, "loss": 0.8795, "step": 4759 }, { "epoch": 2.1557971014492754, "grad_norm": 0.3350591966750769, "learning_rate": 5.612144302004569e-05, "loss": 0.8809, "step": 4760 }, { "epoch": 2.15625, "grad_norm": 0.252073769393055, "learning_rate": 5.61098675943495e-05, "loss": 0.8788, "step": 4761 }, { "epoch": 2.1567028985507246, "grad_norm": 0.29205910613714126, "learning_rate": 5.6098290558124304e-05, "loss": 0.8645, "step": 4762 }, { "epoch": 2.1571557971014492, "grad_norm": 0.31510223661780873, "learning_rate": 5.6086711912527496e-05, "loss": 0.883, "step": 4763 }, { "epoch": 2.157608695652174, "grad_norm": 0.3525135974945182, "learning_rate": 5.607513165871661e-05, "loss": 0.8661, "step": 4764 }, { "epoch": 2.1580615942028984, "grad_norm": 0.3819735508518765, "learning_rate": 5.606354979784934e-05, "loss": 0.8763, "step": 4765 }, { "epoch": 2.158514492753623, "grad_norm": 0.34219064135568006, "learning_rate": 5.605196633108354e-05, "loss": 0.8711, "step": 4766 }, { "epoch": 2.1589673913043477, "grad_norm": 0.24644208413130428, "learning_rate": 5.604038125957722e-05, "loss": 0.8847, "step": 4767 }, { "epoch": 2.1594202898550723, "grad_norm": 0.2758230749712281, "learning_rate": 5.6028794584488585e-05, "loss": 0.8889, "step": 4768 }, { "epoch": 2.159873188405797, "grad_norm": 0.2609079572731213, "learning_rate": 5.601720630697595e-05, "loss": 0.8682, "step": 4769 }, { "epoch": 2.1603260869565215, "grad_norm": 0.2511649897557706, "learning_rate": 5.6005616428197816e-05, "loss": 0.874, "step": 4770 }, { "epoch": 2.1607789855072466, "grad_norm": 0.3449311657980978, "learning_rate": 5.5994024949312845e-05, "loss": 0.8936, "step": 4771 }, { "epoch": 2.161231884057971, "grad_norm": 0.3689991281741085, "learning_rate": 5.5982431871479856e-05, "loss": 0.8649, "step": 4772 }, { "epoch": 2.161684782608696, "grad_norm": 0.4011058293935755, "learning_rate": 5.597083719585782e-05, "loss": 0.8902, "step": 4773 }, { "epoch": 2.1621376811594204, "grad_norm": 0.3686685313070326, "learning_rate": 5.595924092360589e-05, "loss": 0.8616, "step": 4774 }, { "epoch": 2.162590579710145, "grad_norm": 0.3920569564077262, "learning_rate": 5.5947643055883355e-05, "loss": 0.8591, "step": 4775 }, { "epoch": 2.1630434782608696, "grad_norm": 0.4368969549968523, "learning_rate": 5.593604359384967e-05, "loss": 0.8565, "step": 4776 }, { "epoch": 2.1634963768115942, "grad_norm": 0.5227699879635509, "learning_rate": 5.5924442538664454e-05, "loss": 0.8869, "step": 4777 }, { "epoch": 2.163949275362319, "grad_norm": 0.5769442528759293, "learning_rate": 5.5912839891487484e-05, "loss": 0.8759, "step": 4778 }, { "epoch": 2.1644021739130435, "grad_norm": 0.5492814563260201, "learning_rate": 5.590123565347869e-05, "loss": 0.8629, "step": 4779 }, { "epoch": 2.164855072463768, "grad_norm": 0.523952053537981, "learning_rate": 5.5889629825798164e-05, "loss": 0.8739, "step": 4780 }, { "epoch": 2.1653079710144927, "grad_norm": 0.5274018159702796, "learning_rate": 5.587802240960617e-05, "loss": 0.8777, "step": 4781 }, { "epoch": 2.1657608695652173, "grad_norm": 0.5008820831831622, "learning_rate": 5.5866413406063136e-05, "loss": 0.8892, "step": 4782 }, { "epoch": 2.166213768115942, "grad_norm": 0.4524309523144941, "learning_rate": 5.5854802816329587e-05, "loss": 0.888, "step": 4783 }, { "epoch": 2.1666666666666665, "grad_norm": 0.3671700150922473, "learning_rate": 5.584319064156628e-05, "loss": 0.8886, "step": 4784 }, { "epoch": 2.167119565217391, "grad_norm": 0.31939288095042234, "learning_rate": 5.5831576882934104e-05, "loss": 0.8559, "step": 4785 }, { "epoch": 2.1675724637681157, "grad_norm": 0.2631267681871071, "learning_rate": 5.5819961541594104e-05, "loss": 0.8597, "step": 4786 }, { "epoch": 2.168025362318841, "grad_norm": 0.24958595745788673, "learning_rate": 5.580834461870747e-05, "loss": 0.8623, "step": 4787 }, { "epoch": 2.1684782608695654, "grad_norm": 0.24963248194250903, "learning_rate": 5.579672611543557e-05, "loss": 0.8878, "step": 4788 }, { "epoch": 2.16893115942029, "grad_norm": 0.2509069926282382, "learning_rate": 5.578510603293994e-05, "loss": 0.8651, "step": 4789 }, { "epoch": 2.1693840579710146, "grad_norm": 0.18973499135375044, "learning_rate": 5.577348437238223e-05, "loss": 0.8652, "step": 4790 }, { "epoch": 2.1698369565217392, "grad_norm": 0.27492349882328504, "learning_rate": 5.576186113492432e-05, "loss": 0.8874, "step": 4791 }, { "epoch": 2.170289855072464, "grad_norm": 0.2818030033348966, "learning_rate": 5.575023632172816e-05, "loss": 0.8705, "step": 4792 }, { "epoch": 2.1707427536231885, "grad_norm": 0.27876999963365734, "learning_rate": 5.573860993395591e-05, "loss": 0.8437, "step": 4793 }, { "epoch": 2.171195652173913, "grad_norm": 0.2800547162678692, "learning_rate": 5.572698197276989e-05, "loss": 0.8816, "step": 4794 }, { "epoch": 2.1716485507246377, "grad_norm": 0.3256228361619434, "learning_rate": 5.571535243933257e-05, "loss": 0.8773, "step": 4795 }, { "epoch": 2.1721014492753623, "grad_norm": 0.3258906486340443, "learning_rate": 5.5703721334806555e-05, "loss": 0.9078, "step": 4796 }, { "epoch": 2.172554347826087, "grad_norm": 0.30364970878006464, "learning_rate": 5.569208866035462e-05, "loss": 0.8842, "step": 4797 }, { "epoch": 2.1730072463768115, "grad_norm": 0.44773172194395644, "learning_rate": 5.568045441713973e-05, "loss": 0.8749, "step": 4798 }, { "epoch": 2.173460144927536, "grad_norm": 0.49793390553307987, "learning_rate": 5.5668818606324975e-05, "loss": 0.878, "step": 4799 }, { "epoch": 2.1739130434782608, "grad_norm": 0.4949080284223779, "learning_rate": 5.5657181229073584e-05, "loss": 0.8561, "step": 4800 }, { "epoch": 2.1743659420289854, "grad_norm": 0.5287031212720228, "learning_rate": 5.5645542286548965e-05, "loss": 0.8792, "step": 4801 }, { "epoch": 2.17481884057971, "grad_norm": 0.5214705382728417, "learning_rate": 5.563390177991469e-05, "loss": 0.8904, "step": 4802 }, { "epoch": 2.1752717391304346, "grad_norm": 0.5787790149545392, "learning_rate": 5.56222597103345e-05, "loss": 0.8773, "step": 4803 }, { "epoch": 2.175724637681159, "grad_norm": 0.509324504467446, "learning_rate": 5.561061607897224e-05, "loss": 0.8595, "step": 4804 }, { "epoch": 2.1761775362318843, "grad_norm": 0.34586085322265064, "learning_rate": 5.5598970886991956e-05, "loss": 0.8776, "step": 4805 }, { "epoch": 2.176630434782609, "grad_norm": 0.2654679625673988, "learning_rate": 5.558732413555783e-05, "loss": 0.86, "step": 4806 }, { "epoch": 2.1770833333333335, "grad_norm": 0.2674115525705017, "learning_rate": 5.5575675825834215e-05, "loss": 0.8613, "step": 4807 }, { "epoch": 2.177536231884058, "grad_norm": 0.2502490760444557, "learning_rate": 5.55640259589856e-05, "loss": 0.8636, "step": 4808 }, { "epoch": 2.1779891304347827, "grad_norm": 0.254155161304018, "learning_rate": 5.555237453617665e-05, "loss": 0.8812, "step": 4809 }, { "epoch": 2.1784420289855073, "grad_norm": 0.34690987326619815, "learning_rate": 5.554072155857217e-05, "loss": 0.8632, "step": 4810 }, { "epoch": 2.178894927536232, "grad_norm": 0.3245073315748438, "learning_rate": 5.552906702733713e-05, "loss": 0.8731, "step": 4811 }, { "epoch": 2.1793478260869565, "grad_norm": 0.2611717764362986, "learning_rate": 5.5517410943636654e-05, "loss": 0.8714, "step": 4812 }, { "epoch": 2.179800724637681, "grad_norm": 0.23662816288179292, "learning_rate": 5.5505753308636015e-05, "loss": 0.8964, "step": 4813 }, { "epoch": 2.1802536231884058, "grad_norm": 0.2603583958633786, "learning_rate": 5.549409412350065e-05, "loss": 0.8632, "step": 4814 }, { "epoch": 2.1807065217391304, "grad_norm": 0.32889546804298264, "learning_rate": 5.548243338939613e-05, "loss": 0.8718, "step": 4815 }, { "epoch": 2.181159420289855, "grad_norm": 0.2818103394645501, "learning_rate": 5.547077110748823e-05, "loss": 0.8841, "step": 4816 }, { "epoch": 2.1816123188405796, "grad_norm": 0.24781417955573407, "learning_rate": 5.5459107278942815e-05, "loss": 0.8809, "step": 4817 }, { "epoch": 2.182065217391304, "grad_norm": 0.3154374619866268, "learning_rate": 5.544744190492594e-05, "loss": 0.8918, "step": 4818 }, { "epoch": 2.182518115942029, "grad_norm": 0.3783568951822401, "learning_rate": 5.5435774986603835e-05, "loss": 0.879, "step": 4819 }, { "epoch": 2.1829710144927534, "grad_norm": 0.3846548876318862, "learning_rate": 5.542410652514284e-05, "loss": 0.8563, "step": 4820 }, { "epoch": 2.1834239130434785, "grad_norm": 0.34964106462848843, "learning_rate": 5.5412436521709476e-05, "loss": 0.8754, "step": 4821 }, { "epoch": 2.183876811594203, "grad_norm": 0.34814978109926986, "learning_rate": 5.5400764977470405e-05, "loss": 0.8785, "step": 4822 }, { "epoch": 2.1843297101449277, "grad_norm": 0.3206191537379284, "learning_rate": 5.5389091893592454e-05, "loss": 0.8486, "step": 4823 }, { "epoch": 2.1847826086956523, "grad_norm": 0.32678522660635323, "learning_rate": 5.5377417271242605e-05, "loss": 0.8583, "step": 4824 }, { "epoch": 2.185235507246377, "grad_norm": 0.3425108750654235, "learning_rate": 5.536574111158798e-05, "loss": 0.8621, "step": 4825 }, { "epoch": 2.1856884057971016, "grad_norm": 0.2937842424772557, "learning_rate": 5.535406341579587e-05, "loss": 0.8739, "step": 4826 }, { "epoch": 2.186141304347826, "grad_norm": 0.2616232587722703, "learning_rate": 5.534238418503371e-05, "loss": 0.8735, "step": 4827 }, { "epoch": 2.1865942028985508, "grad_norm": 0.21320901610235843, "learning_rate": 5.533070342046909e-05, "loss": 0.8787, "step": 4828 }, { "epoch": 2.1870471014492754, "grad_norm": 0.24317462181086222, "learning_rate": 5.5319021123269756e-05, "loss": 0.8815, "step": 4829 }, { "epoch": 2.1875, "grad_norm": 0.26857185683636575, "learning_rate": 5.5307337294603595e-05, "loss": 0.884, "step": 4830 }, { "epoch": 2.1879528985507246, "grad_norm": 0.2552645603505968, "learning_rate": 5.529565193563868e-05, "loss": 0.8858, "step": 4831 }, { "epoch": 2.1884057971014492, "grad_norm": 0.304068147495728, "learning_rate": 5.528396504754319e-05, "loss": 0.8892, "step": 4832 }, { "epoch": 2.188858695652174, "grad_norm": 0.31862510663447435, "learning_rate": 5.5272276631485516e-05, "loss": 0.871, "step": 4833 }, { "epoch": 2.1893115942028984, "grad_norm": 0.31506105368874715, "learning_rate": 5.526058668863413e-05, "loss": 0.8729, "step": 4834 }, { "epoch": 2.189764492753623, "grad_norm": 0.27673194342121915, "learning_rate": 5.524889522015771e-05, "loss": 0.8815, "step": 4835 }, { "epoch": 2.1902173913043477, "grad_norm": 0.31836584780880595, "learning_rate": 5.523720222722508e-05, "loss": 0.8767, "step": 4836 }, { "epoch": 2.1906702898550723, "grad_norm": 0.2850083528276479, "learning_rate": 5.5225507711005196e-05, "loss": 0.8721, "step": 4837 }, { "epoch": 2.191123188405797, "grad_norm": 0.2448004708643071, "learning_rate": 5.521381167266718e-05, "loss": 0.8749, "step": 4838 }, { "epoch": 2.1915760869565215, "grad_norm": 0.2794172359876601, "learning_rate": 5.5202114113380296e-05, "loss": 0.8648, "step": 4839 }, { "epoch": 2.1920289855072466, "grad_norm": 0.38570694456310217, "learning_rate": 5.519041503431398e-05, "loss": 0.8718, "step": 4840 }, { "epoch": 2.192481884057971, "grad_norm": 0.4843296654155831, "learning_rate": 5.5178714436637806e-05, "loss": 0.8743, "step": 4841 }, { "epoch": 2.192934782608696, "grad_norm": 0.5071309060498818, "learning_rate": 5.516701232152148e-05, "loss": 0.8715, "step": 4842 }, { "epoch": 2.1933876811594204, "grad_norm": 0.5284642198149667, "learning_rate": 5.515530869013491e-05, "loss": 0.876, "step": 4843 }, { "epoch": 2.193840579710145, "grad_norm": 0.5577739983340685, "learning_rate": 5.5143603543648114e-05, "loss": 0.8752, "step": 4844 }, { "epoch": 2.1942934782608696, "grad_norm": 0.5399852006222997, "learning_rate": 5.5131896883231275e-05, "loss": 0.8708, "step": 4845 }, { "epoch": 2.1947463768115942, "grad_norm": 0.49731587081923084, "learning_rate": 5.512018871005472e-05, "loss": 0.896, "step": 4846 }, { "epoch": 2.195199275362319, "grad_norm": 0.4320218613696413, "learning_rate": 5.5108479025288946e-05, "loss": 0.8958, "step": 4847 }, { "epoch": 2.1956521739130435, "grad_norm": 0.35441638554494453, "learning_rate": 5.5096767830104574e-05, "loss": 0.8778, "step": 4848 }, { "epoch": 2.196105072463768, "grad_norm": 0.28813178435461534, "learning_rate": 5.508505512567242e-05, "loss": 0.8685, "step": 4849 }, { "epoch": 2.1965579710144927, "grad_norm": 0.2758697987977899, "learning_rate": 5.507334091316338e-05, "loss": 0.8769, "step": 4850 }, { "epoch": 2.1970108695652173, "grad_norm": 0.23187245594999595, "learning_rate": 5.506162519374857e-05, "loss": 0.8664, "step": 4851 }, { "epoch": 2.197463768115942, "grad_norm": 0.23059282190618047, "learning_rate": 5.504990796859922e-05, "loss": 0.8603, "step": 4852 }, { "epoch": 2.1979166666666665, "grad_norm": 0.29845253514747794, "learning_rate": 5.5038189238886724e-05, "loss": 0.9064, "step": 4853 }, { "epoch": 2.198369565217391, "grad_norm": 0.36894330915699275, "learning_rate": 5.502646900578263e-05, "loss": 0.8543, "step": 4854 }, { "epoch": 2.1988224637681157, "grad_norm": 0.36802336464437, "learning_rate": 5.5014747270458605e-05, "loss": 0.8819, "step": 4855 }, { "epoch": 2.199275362318841, "grad_norm": 0.3846859164472856, "learning_rate": 5.5003024034086504e-05, "loss": 0.8831, "step": 4856 }, { "epoch": 2.1997282608695654, "grad_norm": 0.3735089606202636, "learning_rate": 5.4991299297838306e-05, "loss": 0.8543, "step": 4857 }, { "epoch": 2.20018115942029, "grad_norm": 0.48265030927932984, "learning_rate": 5.497957306288618e-05, "loss": 0.9028, "step": 4858 }, { "epoch": 2.2006340579710146, "grad_norm": 0.42565058029348185, "learning_rate": 5.49678453304024e-05, "loss": 0.8692, "step": 4859 }, { "epoch": 2.2010869565217392, "grad_norm": 0.30676679251849, "learning_rate": 5.495611610155939e-05, "loss": 0.8618, "step": 4860 }, { "epoch": 2.201539855072464, "grad_norm": 0.22763770595880722, "learning_rate": 5.4944385377529755e-05, "loss": 0.8784, "step": 4861 }, { "epoch": 2.2019927536231885, "grad_norm": 0.24852735107166074, "learning_rate": 5.4932653159486236e-05, "loss": 0.8921, "step": 4862 }, { "epoch": 2.202445652173913, "grad_norm": 0.21802503650924274, "learning_rate": 5.492091944860171e-05, "loss": 0.8743, "step": 4863 }, { "epoch": 2.2028985507246377, "grad_norm": 0.20082779266614015, "learning_rate": 5.490918424604923e-05, "loss": 0.8901, "step": 4864 }, { "epoch": 2.2033514492753623, "grad_norm": 0.24161170224315365, "learning_rate": 5.4897447553001976e-05, "loss": 0.8684, "step": 4865 }, { "epoch": 2.203804347826087, "grad_norm": 0.31344222444629777, "learning_rate": 5.4885709370633286e-05, "loss": 0.8835, "step": 4866 }, { "epoch": 2.2042572463768115, "grad_norm": 0.355986986236126, "learning_rate": 5.487396970011664e-05, "loss": 0.8763, "step": 4867 }, { "epoch": 2.204710144927536, "grad_norm": 0.3383335772218244, "learning_rate": 5.4862228542625665e-05, "loss": 0.8791, "step": 4868 }, { "epoch": 2.2051630434782608, "grad_norm": 0.3071456211238503, "learning_rate": 5.4850485899334154e-05, "loss": 0.8745, "step": 4869 }, { "epoch": 2.2056159420289854, "grad_norm": 0.24300227990966913, "learning_rate": 5.483874177141605e-05, "loss": 0.8653, "step": 4870 }, { "epoch": 2.20606884057971, "grad_norm": 0.2532169460198357, "learning_rate": 5.4826996160045396e-05, "loss": 0.8862, "step": 4871 }, { "epoch": 2.2065217391304346, "grad_norm": 0.2455161996171346, "learning_rate": 5.481524906639645e-05, "loss": 0.873, "step": 4872 }, { "epoch": 2.206974637681159, "grad_norm": 0.20669683038752526, "learning_rate": 5.480350049164358e-05, "loss": 0.8747, "step": 4873 }, { "epoch": 2.2074275362318843, "grad_norm": 0.197356141766674, "learning_rate": 5.4791750436961295e-05, "loss": 0.8847, "step": 4874 }, { "epoch": 2.207880434782609, "grad_norm": 0.2341571576927417, "learning_rate": 5.47799989035243e-05, "loss": 0.8647, "step": 4875 }, { "epoch": 2.2083333333333335, "grad_norm": 0.22496942839686074, "learning_rate": 5.476824589250738e-05, "loss": 0.8966, "step": 4876 }, { "epoch": 2.208786231884058, "grad_norm": 0.21811776557898283, "learning_rate": 5.4756491405085516e-05, "loss": 0.8702, "step": 4877 }, { "epoch": 2.2092391304347827, "grad_norm": 0.2729684502896788, "learning_rate": 5.474473544243383e-05, "loss": 0.8736, "step": 4878 }, { "epoch": 2.2096920289855073, "grad_norm": 0.33905918325933443, "learning_rate": 5.4732978005727575e-05, "loss": 0.8729, "step": 4879 }, { "epoch": 2.210144927536232, "grad_norm": 0.299678858108706, "learning_rate": 5.472121909614216e-05, "loss": 0.8857, "step": 4880 }, { "epoch": 2.2105978260869565, "grad_norm": 0.2810446815996399, "learning_rate": 5.470945871485314e-05, "loss": 0.8764, "step": 4881 }, { "epoch": 2.211050724637681, "grad_norm": 0.30741351156940855, "learning_rate": 5.4697696863036224e-05, "loss": 0.8753, "step": 4882 }, { "epoch": 2.2115036231884058, "grad_norm": 0.3993434409768133, "learning_rate": 5.4685933541867266e-05, "loss": 0.8681, "step": 4883 }, { "epoch": 2.2119565217391304, "grad_norm": 0.4306104692748844, "learning_rate": 5.467416875252227e-05, "loss": 0.8817, "step": 4884 }, { "epoch": 2.212409420289855, "grad_norm": 0.38611583160377727, "learning_rate": 5.466240249617736e-05, "loss": 0.8878, "step": 4885 }, { "epoch": 2.2128623188405796, "grad_norm": 0.41886421462483586, "learning_rate": 5.465063477400883e-05, "loss": 0.8725, "step": 4886 }, { "epoch": 2.213315217391304, "grad_norm": 0.5408183591885639, "learning_rate": 5.4638865587193136e-05, "loss": 0.8695, "step": 4887 }, { "epoch": 2.213768115942029, "grad_norm": 0.67574962516574, "learning_rate": 5.462709493690686e-05, "loss": 0.8635, "step": 4888 }, { "epoch": 2.2142210144927534, "grad_norm": 0.6244964363426208, "learning_rate": 5.4615322824326697e-05, "loss": 0.8851, "step": 4889 }, { "epoch": 2.2146739130434785, "grad_norm": 0.5314266769855236, "learning_rate": 5.460354925062957e-05, "loss": 0.8792, "step": 4890 }, { "epoch": 2.215126811594203, "grad_norm": 0.47693911422927243, "learning_rate": 5.459177421699249e-05, "loss": 0.8839, "step": 4891 }, { "epoch": 2.2155797101449277, "grad_norm": 0.5066150979625095, "learning_rate": 5.457999772459261e-05, "loss": 0.8748, "step": 4892 }, { "epoch": 2.2160326086956523, "grad_norm": 0.5159746113057814, "learning_rate": 5.4568219774607255e-05, "loss": 0.8823, "step": 4893 }, { "epoch": 2.216485507246377, "grad_norm": 0.4819835403822278, "learning_rate": 5.455644036821388e-05, "loss": 0.8869, "step": 4894 }, { "epoch": 2.2169384057971016, "grad_norm": 0.47837704670385234, "learning_rate": 5.45446595065901e-05, "loss": 0.8751, "step": 4895 }, { "epoch": 2.217391304347826, "grad_norm": 0.5370570535973113, "learning_rate": 5.453287719091366e-05, "loss": 0.8759, "step": 4896 }, { "epoch": 2.2178442028985508, "grad_norm": 0.5739626498233966, "learning_rate": 5.452109342236245e-05, "loss": 0.8565, "step": 4897 }, { "epoch": 2.2182971014492754, "grad_norm": 0.539555798471366, "learning_rate": 5.4509308202114535e-05, "loss": 0.8932, "step": 4898 }, { "epoch": 2.21875, "grad_norm": 0.38017082241752975, "learning_rate": 5.4497521531348066e-05, "loss": 0.8737, "step": 4899 }, { "epoch": 2.2192028985507246, "grad_norm": 0.27104466627928403, "learning_rate": 5.448573341124142e-05, "loss": 0.8773, "step": 4900 }, { "epoch": 2.2196557971014492, "grad_norm": 0.2908572524704478, "learning_rate": 5.4473943842973035e-05, "loss": 0.8788, "step": 4901 }, { "epoch": 2.220108695652174, "grad_norm": 0.2784170106615751, "learning_rate": 5.446215282772155e-05, "loss": 0.8704, "step": 4902 }, { "epoch": 2.2205615942028984, "grad_norm": 0.2764718595546362, "learning_rate": 5.445036036666573e-05, "loss": 0.8682, "step": 4903 }, { "epoch": 2.221014492753623, "grad_norm": 0.3553639714105954, "learning_rate": 5.4438566460984486e-05, "loss": 0.8703, "step": 4904 }, { "epoch": 2.2214673913043477, "grad_norm": 0.44565522230891114, "learning_rate": 5.442677111185689e-05, "loss": 0.8941, "step": 4905 }, { "epoch": 2.2219202898550723, "grad_norm": 0.512137887970793, "learning_rate": 5.44149743204621e-05, "loss": 0.8655, "step": 4906 }, { "epoch": 2.222373188405797, "grad_norm": 0.4328875248042008, "learning_rate": 5.4403176087979485e-05, "loss": 0.8766, "step": 4907 }, { "epoch": 2.2228260869565215, "grad_norm": 0.35440895946051615, "learning_rate": 5.4391376415588554e-05, "loss": 0.8861, "step": 4908 }, { "epoch": 2.2232789855072466, "grad_norm": 0.29938491817352275, "learning_rate": 5.43795753044689e-05, "loss": 0.8854, "step": 4909 }, { "epoch": 2.223731884057971, "grad_norm": 0.2944726594558164, "learning_rate": 5.4367772755800324e-05, "loss": 0.867, "step": 4910 }, { "epoch": 2.224184782608696, "grad_norm": 0.2559055135030395, "learning_rate": 5.435596877076273e-05, "loss": 0.8597, "step": 4911 }, { "epoch": 2.2246376811594204, "grad_norm": 0.22955833664720982, "learning_rate": 5.4344163350536196e-05, "loss": 0.8806, "step": 4912 }, { "epoch": 2.225090579710145, "grad_norm": 0.25429199221792403, "learning_rate": 5.4332356496300925e-05, "loss": 0.8774, "step": 4913 }, { "epoch": 2.2255434782608696, "grad_norm": 0.2623430551555357, "learning_rate": 5.432054820923725e-05, "loss": 0.8703, "step": 4914 }, { "epoch": 2.2259963768115942, "grad_norm": 0.40727131819343204, "learning_rate": 5.430873849052569e-05, "loss": 0.8584, "step": 4915 }, { "epoch": 2.226449275362319, "grad_norm": 0.5059192009068745, "learning_rate": 5.4296927341346864e-05, "loss": 0.853, "step": 4916 }, { "epoch": 2.2269021739130435, "grad_norm": 0.47076332778994806, "learning_rate": 5.4285114762881564e-05, "loss": 0.8848, "step": 4917 }, { "epoch": 2.227355072463768, "grad_norm": 0.46454814094436864, "learning_rate": 5.427330075631069e-05, "loss": 0.8838, "step": 4918 }, { "epoch": 2.2278079710144927, "grad_norm": 0.4737690526136072, "learning_rate": 5.426148532281532e-05, "loss": 0.8645, "step": 4919 }, { "epoch": 2.2282608695652173, "grad_norm": 0.43041509555923463, "learning_rate": 5.424966846357668e-05, "loss": 0.8658, "step": 4920 }, { "epoch": 2.228713768115942, "grad_norm": 0.37022757794780214, "learning_rate": 5.42378501797761e-05, "loss": 0.8662, "step": 4921 }, { "epoch": 2.2291666666666665, "grad_norm": 0.4264291016208675, "learning_rate": 5.4226030472595075e-05, "loss": 0.9002, "step": 4922 }, { "epoch": 2.229619565217391, "grad_norm": 0.3906596292405371, "learning_rate": 5.4214209343215234e-05, "loss": 0.8663, "step": 4923 }, { "epoch": 2.2300724637681157, "grad_norm": 0.32023387542169135, "learning_rate": 5.420238679281836e-05, "loss": 0.8664, "step": 4924 }, { "epoch": 2.230525362318841, "grad_norm": 0.5514610853236528, "learning_rate": 5.419056282258637e-05, "loss": 0.8858, "step": 4925 }, { "epoch": 2.2309782608695654, "grad_norm": 0.5090619578698247, "learning_rate": 5.417873743370134e-05, "loss": 0.882, "step": 4926 }, { "epoch": 2.23143115942029, "grad_norm": 0.48811212319858877, "learning_rate": 5.416691062734545e-05, "loss": 0.8675, "step": 4927 }, { "epoch": 2.2318840579710146, "grad_norm": 0.6738870075809845, "learning_rate": 5.4155082404701056e-05, "loss": 0.8741, "step": 4928 }, { "epoch": 2.2323369565217392, "grad_norm": 0.7889281076641244, "learning_rate": 5.4143252766950653e-05, "loss": 0.8693, "step": 4929 }, { "epoch": 2.232789855072464, "grad_norm": 0.7644984230533097, "learning_rate": 5.4131421715276856e-05, "loss": 0.8791, "step": 4930 }, { "epoch": 2.2332427536231885, "grad_norm": 0.6916018442028038, "learning_rate": 5.411958925086243e-05, "loss": 0.8845, "step": 4931 }, { "epoch": 2.233695652173913, "grad_norm": 0.5857271526777551, "learning_rate": 5.410775537489029e-05, "loss": 0.8528, "step": 4932 }, { "epoch": 2.2341485507246377, "grad_norm": 0.436244822472771, "learning_rate": 5.40959200885435e-05, "loss": 0.8666, "step": 4933 }, { "epoch": 2.2346014492753623, "grad_norm": 0.2974262740970286, "learning_rate": 5.4084083393005235e-05, "loss": 0.8903, "step": 4934 }, { "epoch": 2.235054347826087, "grad_norm": 0.26759931944778814, "learning_rate": 5.407224528945884e-05, "loss": 0.8676, "step": 4935 }, { "epoch": 2.2355072463768115, "grad_norm": 0.30884812776647674, "learning_rate": 5.406040577908778e-05, "loss": 0.8723, "step": 4936 }, { "epoch": 2.235960144927536, "grad_norm": 0.3474011823971214, "learning_rate": 5.4048564863075674e-05, "loss": 0.8763, "step": 4937 }, { "epoch": 2.2364130434782608, "grad_norm": 0.3796535523695248, "learning_rate": 5.403672254260628e-05, "loss": 0.865, "step": 4938 }, { "epoch": 2.2368659420289854, "grad_norm": 0.3274299999452318, "learning_rate": 5.402487881886349e-05, "loss": 0.8796, "step": 4939 }, { "epoch": 2.23731884057971, "grad_norm": 0.32190666727960293, "learning_rate": 5.4013033693031336e-05, "loss": 0.8815, "step": 4940 }, { "epoch": 2.2377717391304346, "grad_norm": 0.3884878294542894, "learning_rate": 5.4001187166294e-05, "loss": 0.8931, "step": 4941 }, { "epoch": 2.238224637681159, "grad_norm": 0.34538039197725434, "learning_rate": 5.3989339239835795e-05, "loss": 0.8642, "step": 4942 }, { "epoch": 2.2386775362318843, "grad_norm": 0.32930764119954087, "learning_rate": 5.397748991484119e-05, "loss": 0.8762, "step": 4943 }, { "epoch": 2.239130434782609, "grad_norm": 0.2898918908994444, "learning_rate": 5.3965639192494746e-05, "loss": 0.8504, "step": 4944 }, { "epoch": 2.2395833333333335, "grad_norm": 0.2561627630682915, "learning_rate": 5.3953787073981236e-05, "loss": 0.8671, "step": 4945 }, { "epoch": 2.240036231884058, "grad_norm": 0.41340857884035376, "learning_rate": 5.394193356048551e-05, "loss": 0.8807, "step": 4946 }, { "epoch": 2.2404891304347827, "grad_norm": 0.23988858421942405, "learning_rate": 5.393007865319262e-05, "loss": 0.862, "step": 4947 }, { "epoch": 2.2409420289855073, "grad_norm": 0.21274722538768556, "learning_rate": 5.391822235328767e-05, "loss": 0.8821, "step": 4948 }, { "epoch": 2.241394927536232, "grad_norm": 0.21877547140023423, "learning_rate": 5.390636466195598e-05, "loss": 0.864, "step": 4949 }, { "epoch": 2.2418478260869565, "grad_norm": 0.21417993041597883, "learning_rate": 5.389450558038297e-05, "loss": 0.8884, "step": 4950 }, { "epoch": 2.242300724637681, "grad_norm": 0.2260379190279925, "learning_rate": 5.3882645109754234e-05, "loss": 0.8542, "step": 4951 }, { "epoch": 2.2427536231884058, "grad_norm": 0.22776756448719582, "learning_rate": 5.3870783251255464e-05, "loss": 0.8485, "step": 4952 }, { "epoch": 2.2432065217391304, "grad_norm": 0.2126641961453412, "learning_rate": 5.385892000607251e-05, "loss": 0.8636, "step": 4953 }, { "epoch": 2.243659420289855, "grad_norm": 0.24215522196972367, "learning_rate": 5.384705537539136e-05, "loss": 0.8816, "step": 4954 }, { "epoch": 2.2441123188405796, "grad_norm": 0.18866373283973634, "learning_rate": 5.3835189360398146e-05, "loss": 0.8716, "step": 4955 }, { "epoch": 2.244565217391304, "grad_norm": 0.24223807673639755, "learning_rate": 5.3823321962279126e-05, "loss": 0.8618, "step": 4956 }, { "epoch": 2.245018115942029, "grad_norm": 0.24873389060146264, "learning_rate": 5.3811453182220696e-05, "loss": 0.8663, "step": 4957 }, { "epoch": 2.2454710144927534, "grad_norm": 0.2631591280293886, "learning_rate": 5.3799583021409414e-05, "loss": 0.8825, "step": 4958 }, { "epoch": 2.2459239130434785, "grad_norm": 0.24850006838320943, "learning_rate": 5.378771148103195e-05, "loss": 0.8677, "step": 4959 }, { "epoch": 2.246376811594203, "grad_norm": 0.20151533047273198, "learning_rate": 5.377583856227512e-05, "loss": 0.8692, "step": 4960 }, { "epoch": 2.2468297101449277, "grad_norm": 0.2963502891758927, "learning_rate": 5.376396426632588e-05, "loss": 0.8669, "step": 4961 }, { "epoch": 2.2472826086956523, "grad_norm": 0.2756954850486591, "learning_rate": 5.3752088594371324e-05, "loss": 0.8888, "step": 4962 }, { "epoch": 2.247735507246377, "grad_norm": 0.2459651510077647, "learning_rate": 5.374021154759866e-05, "loss": 0.8866, "step": 4963 }, { "epoch": 2.2481884057971016, "grad_norm": 0.25188107264171017, "learning_rate": 5.372833312719531e-05, "loss": 0.8726, "step": 4964 }, { "epoch": 2.248641304347826, "grad_norm": 0.2956179978282856, "learning_rate": 5.3716453334348724e-05, "loss": 0.8755, "step": 4965 }, { "epoch": 2.2490942028985508, "grad_norm": 0.28114844744745354, "learning_rate": 5.3704572170246545e-05, "loss": 0.8723, "step": 4966 }, { "epoch": 2.2495471014492754, "grad_norm": 0.2906231952127366, "learning_rate": 5.369268963607659e-05, "loss": 0.8544, "step": 4967 }, { "epoch": 2.25, "grad_norm": 0.3368300635277098, "learning_rate": 5.368080573302676e-05, "loss": 0.8907, "step": 4968 }, { "epoch": 2.2504528985507246, "grad_norm": 0.3513922645938684, "learning_rate": 5.366892046228509e-05, "loss": 0.8703, "step": 4969 }, { "epoch": 2.2509057971014492, "grad_norm": 0.3560011918530087, "learning_rate": 5.365703382503977e-05, "loss": 0.8596, "step": 4970 }, { "epoch": 2.251358695652174, "grad_norm": 0.3106275756730232, "learning_rate": 5.364514582247916e-05, "loss": 0.8844, "step": 4971 }, { "epoch": 2.2518115942028984, "grad_norm": 0.26024904836315815, "learning_rate": 5.363325645579169e-05, "loss": 0.859, "step": 4972 }, { "epoch": 2.252264492753623, "grad_norm": 0.28711539774339523, "learning_rate": 5.3621365726165966e-05, "loss": 0.878, "step": 4973 }, { "epoch": 2.2527173913043477, "grad_norm": 0.36876770411449644, "learning_rate": 5.3609473634790724e-05, "loss": 0.8775, "step": 4974 }, { "epoch": 2.2531702898550723, "grad_norm": 0.32347573346694314, "learning_rate": 5.359758018285484e-05, "loss": 0.8712, "step": 4975 }, { "epoch": 2.253623188405797, "grad_norm": 0.28838773973933907, "learning_rate": 5.3585685371547326e-05, "loss": 0.8748, "step": 4976 }, { "epoch": 2.2540760869565215, "grad_norm": 0.3440951810082395, "learning_rate": 5.3573789202057306e-05, "loss": 0.877, "step": 4977 }, { "epoch": 2.2545289855072466, "grad_norm": 0.365313399055707, "learning_rate": 5.356189167557407e-05, "loss": 0.8806, "step": 4978 }, { "epoch": 2.254981884057971, "grad_norm": 0.3348325506422884, "learning_rate": 5.354999279328703e-05, "loss": 0.8832, "step": 4979 }, { "epoch": 2.255434782608696, "grad_norm": 0.3319121400076632, "learning_rate": 5.3538092556385754e-05, "loss": 0.8783, "step": 4980 }, { "epoch": 2.2558876811594204, "grad_norm": 0.3461826986593243, "learning_rate": 5.35261909660599e-05, "loss": 0.8854, "step": 4981 }, { "epoch": 2.256340579710145, "grad_norm": 0.330731447881999, "learning_rate": 5.35142880234993e-05, "loss": 0.8717, "step": 4982 }, { "epoch": 2.2567934782608696, "grad_norm": 0.2789054626110093, "learning_rate": 5.350238372989392e-05, "loss": 0.8851, "step": 4983 }, { "epoch": 2.2572463768115942, "grad_norm": 0.24570451883794528, "learning_rate": 5.3490478086433836e-05, "loss": 0.8708, "step": 4984 }, { "epoch": 2.257699275362319, "grad_norm": 0.24970548571461268, "learning_rate": 5.347857109430928e-05, "loss": 0.883, "step": 4985 }, { "epoch": 2.2581521739130435, "grad_norm": 0.24701225104378738, "learning_rate": 5.346666275471061e-05, "loss": 0.8851, "step": 4986 }, { "epoch": 2.258605072463768, "grad_norm": 0.32336640926670185, "learning_rate": 5.345475306882833e-05, "loss": 0.8694, "step": 4987 }, { "epoch": 2.2590579710144927, "grad_norm": 0.3936767817667446, "learning_rate": 5.344284203785305e-05, "loss": 0.8543, "step": 4988 }, { "epoch": 2.2595108695652173, "grad_norm": 0.4265076748874329, "learning_rate": 5.343092966297558e-05, "loss": 0.8888, "step": 4989 }, { "epoch": 2.259963768115942, "grad_norm": 0.5187849423503699, "learning_rate": 5.3419015945386766e-05, "loss": 0.8538, "step": 4990 }, { "epoch": 2.2604166666666665, "grad_norm": 0.6064149506339251, "learning_rate": 5.340710088627766e-05, "loss": 0.8714, "step": 4991 }, { "epoch": 2.260869565217391, "grad_norm": 0.6420413397081264, "learning_rate": 5.339518448683945e-05, "loss": 0.8686, "step": 4992 }, { "epoch": 2.261322463768116, "grad_norm": 0.6075719969519626, "learning_rate": 5.338326674826341e-05, "loss": 0.8644, "step": 4993 }, { "epoch": 2.261775362318841, "grad_norm": 0.5814730276083194, "learning_rate": 5.3371347671740994e-05, "loss": 0.8744, "step": 4994 }, { "epoch": 2.2622282608695654, "grad_norm": 0.5408166568405526, "learning_rate": 5.3359427258463745e-05, "loss": 0.8761, "step": 4995 }, { "epoch": 2.26268115942029, "grad_norm": 0.47516043056347557, "learning_rate": 5.33475055096234e-05, "loss": 0.8747, "step": 4996 }, { "epoch": 2.2631340579710146, "grad_norm": 0.41649506726993046, "learning_rate": 5.333558242641178e-05, "loss": 0.892, "step": 4997 }, { "epoch": 2.2635869565217392, "grad_norm": 0.33963255409063686, "learning_rate": 5.332365801002084e-05, "loss": 0.8529, "step": 4998 }, { "epoch": 2.264039855072464, "grad_norm": 0.30603541158993464, "learning_rate": 5.33117322616427e-05, "loss": 0.8796, "step": 4999 }, { "epoch": 2.2644927536231885, "grad_norm": 0.23326336613786433, "learning_rate": 5.3299805182469594e-05, "loss": 0.8688, "step": 5000 }, { "epoch": 2.264945652173913, "grad_norm": 0.24460755354936486, "learning_rate": 5.328787677369389e-05, "loss": 0.8614, "step": 5001 }, { "epoch": 2.2653985507246377, "grad_norm": 0.3356194996697718, "learning_rate": 5.327594703650809e-05, "loss": 0.8897, "step": 5002 }, { "epoch": 2.2658514492753623, "grad_norm": 0.44646747436608664, "learning_rate": 5.3264015972104814e-05, "loss": 0.8712, "step": 5003 }, { "epoch": 2.266304347826087, "grad_norm": 0.5263583927046814, "learning_rate": 5.325208358167685e-05, "loss": 0.8859, "step": 5004 }, { "epoch": 2.2667572463768115, "grad_norm": 0.5979620509998336, "learning_rate": 5.32401498664171e-05, "loss": 0.8775, "step": 5005 }, { "epoch": 2.267210144927536, "grad_norm": 0.5949434195971717, "learning_rate": 5.322821482751858e-05, "loss": 0.8632, "step": 5006 }, { "epoch": 2.2676630434782608, "grad_norm": 0.5290590827446677, "learning_rate": 5.3216278466174456e-05, "loss": 0.8682, "step": 5007 }, { "epoch": 2.2681159420289854, "grad_norm": 0.4326791794523088, "learning_rate": 5.320434078357803e-05, "loss": 0.8723, "step": 5008 }, { "epoch": 2.26856884057971, "grad_norm": 0.35301960230985413, "learning_rate": 5.3192401780922736e-05, "loss": 0.8561, "step": 5009 }, { "epoch": 2.2690217391304346, "grad_norm": 0.2927755504718816, "learning_rate": 5.318046145940214e-05, "loss": 0.8956, "step": 5010 }, { "epoch": 2.269474637681159, "grad_norm": 0.27989592634332877, "learning_rate": 5.316851982020992e-05, "loss": 0.8699, "step": 5011 }, { "epoch": 2.269927536231884, "grad_norm": 0.295385072759692, "learning_rate": 5.31565768645399e-05, "loss": 0.88, "step": 5012 }, { "epoch": 2.270380434782609, "grad_norm": 0.3143576680442234, "learning_rate": 5.3144632593586046e-05, "loss": 0.8845, "step": 5013 }, { "epoch": 2.2708333333333335, "grad_norm": 0.4389230122169771, "learning_rate": 5.3132687008542454e-05, "loss": 0.889, "step": 5014 }, { "epoch": 2.271286231884058, "grad_norm": 0.4664992910646554, "learning_rate": 5.312074011060331e-05, "loss": 0.8648, "step": 5015 }, { "epoch": 2.2717391304347827, "grad_norm": 0.4232906690313404, "learning_rate": 5.3108791900963e-05, "loss": 0.8613, "step": 5016 }, { "epoch": 2.2721920289855073, "grad_norm": 0.4117293961920844, "learning_rate": 5.309684238081599e-05, "loss": 0.8772, "step": 5017 }, { "epoch": 2.272644927536232, "grad_norm": 0.47695992886876487, "learning_rate": 5.3084891551356886e-05, "loss": 0.8695, "step": 5018 }, { "epoch": 2.2730978260869565, "grad_norm": 0.3579742405765378, "learning_rate": 5.307293941378045e-05, "loss": 0.8889, "step": 5019 }, { "epoch": 2.273550724637681, "grad_norm": 0.23030685401433412, "learning_rate": 5.306098596928154e-05, "loss": 0.8739, "step": 5020 }, { "epoch": 2.2740036231884058, "grad_norm": 0.21156878827054812, "learning_rate": 5.304903121905516e-05, "loss": 0.8641, "step": 5021 }, { "epoch": 2.2744565217391304, "grad_norm": 0.27066758791878215, "learning_rate": 5.303707516429645e-05, "loss": 0.8732, "step": 5022 }, { "epoch": 2.274909420289855, "grad_norm": 0.307796756384482, "learning_rate": 5.302511780620069e-05, "loss": 0.8814, "step": 5023 }, { "epoch": 2.2753623188405796, "grad_norm": 0.3383784250013058, "learning_rate": 5.301315914596325e-05, "loss": 0.8753, "step": 5024 }, { "epoch": 2.275815217391304, "grad_norm": 0.4099127565535007, "learning_rate": 5.3001199184779676e-05, "loss": 0.8736, "step": 5025 }, { "epoch": 2.276268115942029, "grad_norm": 0.4731896912451143, "learning_rate": 5.298923792384561e-05, "loss": 0.8761, "step": 5026 }, { "epoch": 2.2767210144927534, "grad_norm": 0.42407166747095315, "learning_rate": 5.297727536435686e-05, "loss": 0.8706, "step": 5027 }, { "epoch": 2.2771739130434785, "grad_norm": 0.3249703180825379, "learning_rate": 5.2965311507509314e-05, "loss": 0.8795, "step": 5028 }, { "epoch": 2.277626811594203, "grad_norm": 0.3156928512216398, "learning_rate": 5.2953346354499036e-05, "loss": 0.8933, "step": 5029 }, { "epoch": 2.2780797101449277, "grad_norm": 0.3061721443588992, "learning_rate": 5.294137990652219e-05, "loss": 0.8815, "step": 5030 }, { "epoch": 2.2785326086956523, "grad_norm": 0.2683365212379366, "learning_rate": 5.292941216477509e-05, "loss": 0.864, "step": 5031 }, { "epoch": 2.278985507246377, "grad_norm": 0.22305565002890496, "learning_rate": 5.2917443130454174e-05, "loss": 0.8646, "step": 5032 }, { "epoch": 2.2794384057971016, "grad_norm": 0.21434659976765522, "learning_rate": 5.2905472804755995e-05, "loss": 0.8791, "step": 5033 }, { "epoch": 2.279891304347826, "grad_norm": 0.2929605364645233, "learning_rate": 5.289350118887725e-05, "loss": 0.8744, "step": 5034 }, { "epoch": 2.2803442028985508, "grad_norm": 0.36772270878166186, "learning_rate": 5.2881528284014764e-05, "loss": 0.8742, "step": 5035 }, { "epoch": 2.2807971014492754, "grad_norm": 0.4295012087052114, "learning_rate": 5.286955409136548e-05, "loss": 0.8951, "step": 5036 }, { "epoch": 2.28125, "grad_norm": 0.471420658096777, "learning_rate": 5.2857578612126466e-05, "loss": 0.8804, "step": 5037 }, { "epoch": 2.2817028985507246, "grad_norm": 0.4961039117401751, "learning_rate": 5.2845601847494963e-05, "loss": 0.8644, "step": 5038 }, { "epoch": 2.2821557971014492, "grad_norm": 0.47232905360359567, "learning_rate": 5.283362379866828e-05, "loss": 0.8817, "step": 5039 }, { "epoch": 2.282608695652174, "grad_norm": 0.3835786441622787, "learning_rate": 5.282164446684389e-05, "loss": 0.8741, "step": 5040 }, { "epoch": 2.2830615942028984, "grad_norm": 0.2865508323034075, "learning_rate": 5.280966385321939e-05, "loss": 0.8811, "step": 5041 }, { "epoch": 2.283514492753623, "grad_norm": 0.24332953715231706, "learning_rate": 5.279768195899248e-05, "loss": 0.8801, "step": 5042 }, { "epoch": 2.2839673913043477, "grad_norm": 0.2222118665207234, "learning_rate": 5.278569878536104e-05, "loss": 0.8697, "step": 5043 }, { "epoch": 2.2844202898550723, "grad_norm": 0.2256778618564971, "learning_rate": 5.277371433352302e-05, "loss": 0.8731, "step": 5044 }, { "epoch": 2.284873188405797, "grad_norm": 0.2194087147049817, "learning_rate": 5.2761728604676557e-05, "loss": 0.8497, "step": 5045 }, { "epoch": 2.2853260869565215, "grad_norm": 0.2550056981486632, "learning_rate": 5.274974160001984e-05, "loss": 0.8752, "step": 5046 }, { "epoch": 2.2857789855072466, "grad_norm": 0.3420826276565244, "learning_rate": 5.273775332075126e-05, "loss": 0.8684, "step": 5047 }, { "epoch": 2.286231884057971, "grad_norm": 0.3082465202804256, "learning_rate": 5.27257637680693e-05, "loss": 0.8897, "step": 5048 }, { "epoch": 2.286684782608696, "grad_norm": 0.2762813339267251, "learning_rate": 5.271377294317256e-05, "loss": 0.8553, "step": 5049 }, { "epoch": 2.2871376811594204, "grad_norm": 0.3257434936920582, "learning_rate": 5.2701780847259806e-05, "loss": 0.8696, "step": 5050 }, { "epoch": 2.287590579710145, "grad_norm": 0.3437274702398149, "learning_rate": 5.2689787481529885e-05, "loss": 0.8647, "step": 5051 }, { "epoch": 2.2880434782608696, "grad_norm": 0.3341294825007939, "learning_rate": 5.267779284718181e-05, "loss": 0.9011, "step": 5052 }, { "epoch": 2.2884963768115942, "grad_norm": 0.37501405721368997, "learning_rate": 5.266579694541468e-05, "loss": 0.882, "step": 5053 }, { "epoch": 2.288949275362319, "grad_norm": 0.4303517893005848, "learning_rate": 5.265379977742776e-05, "loss": 0.8664, "step": 5054 }, { "epoch": 2.2894021739130435, "grad_norm": 0.46103135432730863, "learning_rate": 5.264180134442043e-05, "loss": 0.8579, "step": 5055 }, { "epoch": 2.289855072463768, "grad_norm": 0.3969714393118422, "learning_rate": 5.262980164759219e-05, "loss": 0.876, "step": 5056 }, { "epoch": 2.2903079710144927, "grad_norm": 0.40239649708267033, "learning_rate": 5.261780068814267e-05, "loss": 0.8973, "step": 5057 }, { "epoch": 2.2907608695652173, "grad_norm": 0.41465107478861435, "learning_rate": 5.2605798467271605e-05, "loss": 0.8731, "step": 5058 }, { "epoch": 2.291213768115942, "grad_norm": 0.37478887201687, "learning_rate": 5.25937949861789e-05, "loss": 0.8698, "step": 5059 }, { "epoch": 2.2916666666666665, "grad_norm": 0.26341627649539484, "learning_rate": 5.258179024606455e-05, "loss": 0.8626, "step": 5060 }, { "epoch": 2.292119565217391, "grad_norm": 0.2730976741725638, "learning_rate": 5.25697842481287e-05, "loss": 0.8661, "step": 5061 }, { "epoch": 2.292572463768116, "grad_norm": 0.2377056598441287, "learning_rate": 5.255777699357159e-05, "loss": 0.8722, "step": 5062 }, { "epoch": 2.293025362318841, "grad_norm": 0.24532285049356423, "learning_rate": 5.254576848359362e-05, "loss": 0.8753, "step": 5063 }, { "epoch": 2.2934782608695654, "grad_norm": 0.22436291375021322, "learning_rate": 5.2533758719395294e-05, "loss": 0.885, "step": 5064 }, { "epoch": 2.29393115942029, "grad_norm": 0.2854190503292372, "learning_rate": 5.2521747702177246e-05, "loss": 0.9033, "step": 5065 }, { "epoch": 2.2943840579710146, "grad_norm": 0.37269027787952785, "learning_rate": 5.2509735433140244e-05, "loss": 0.8747, "step": 5066 }, { "epoch": 2.2948369565217392, "grad_norm": 0.3913142447955704, "learning_rate": 5.2497721913485163e-05, "loss": 0.8724, "step": 5067 }, { "epoch": 2.295289855072464, "grad_norm": 0.4104810484151395, "learning_rate": 5.248570714441301e-05, "loss": 0.8692, "step": 5068 }, { "epoch": 2.2957427536231885, "grad_norm": 0.3742110075726837, "learning_rate": 5.247369112712495e-05, "loss": 0.8792, "step": 5069 }, { "epoch": 2.296195652173913, "grad_norm": 0.33827794541258965, "learning_rate": 5.2461673862822196e-05, "loss": 0.8689, "step": 5070 }, { "epoch": 2.2966485507246377, "grad_norm": 0.3790626645938668, "learning_rate": 5.244965535270617e-05, "loss": 0.8947, "step": 5071 }, { "epoch": 2.2971014492753623, "grad_norm": 0.32425104911889496, "learning_rate": 5.243763559797836e-05, "loss": 0.8859, "step": 5072 }, { "epoch": 2.297554347826087, "grad_norm": 0.2919642064316824, "learning_rate": 5.2425614599840436e-05, "loss": 0.8815, "step": 5073 }, { "epoch": 2.2980072463768115, "grad_norm": 0.39273118692546216, "learning_rate": 5.2413592359494105e-05, "loss": 0.8725, "step": 5074 }, { "epoch": 2.298460144927536, "grad_norm": 0.34785867323891795, "learning_rate": 5.240156887814129e-05, "loss": 0.8778, "step": 5075 }, { "epoch": 2.2989130434782608, "grad_norm": 0.4494939095595602, "learning_rate": 5.2389544156983975e-05, "loss": 0.8702, "step": 5076 }, { "epoch": 2.2993659420289854, "grad_norm": 0.6109248631306649, "learning_rate": 5.2377518197224304e-05, "loss": 0.839, "step": 5077 }, { "epoch": 2.29981884057971, "grad_norm": 0.6410809895190698, "learning_rate": 5.236549100006452e-05, "loss": 0.8654, "step": 5078 }, { "epoch": 2.3002717391304346, "grad_norm": 0.5437107500714529, "learning_rate": 5.235346256670701e-05, "loss": 0.8692, "step": 5079 }, { "epoch": 2.300724637681159, "grad_norm": 0.4653967539484604, "learning_rate": 5.234143289835426e-05, "loss": 0.8559, "step": 5080 }, { "epoch": 2.301177536231884, "grad_norm": 0.3981397946562516, "learning_rate": 5.232940199620891e-05, "loss": 0.8711, "step": 5081 }, { "epoch": 2.301630434782609, "grad_norm": 0.30784609747617997, "learning_rate": 5.231736986147372e-05, "loss": 0.8936, "step": 5082 }, { "epoch": 2.3020833333333335, "grad_norm": 0.32049889507473855, "learning_rate": 5.2305336495351536e-05, "loss": 0.8668, "step": 5083 }, { "epoch": 2.302536231884058, "grad_norm": 0.35090444424170447, "learning_rate": 5.2293301899045363e-05, "loss": 0.8945, "step": 5084 }, { "epoch": 2.3029891304347827, "grad_norm": 0.3869544533442879, "learning_rate": 5.228126607375831e-05, "loss": 0.8864, "step": 5085 }, { "epoch": 2.3034420289855073, "grad_norm": 0.378000253412213, "learning_rate": 5.2269229020693635e-05, "loss": 0.8708, "step": 5086 }, { "epoch": 2.303894927536232, "grad_norm": 0.34518551488325216, "learning_rate": 5.2257190741054697e-05, "loss": 0.9011, "step": 5087 }, { "epoch": 2.3043478260869565, "grad_norm": 0.3345985254809163, "learning_rate": 5.2245151236044974e-05, "loss": 0.8662, "step": 5088 }, { "epoch": 2.304800724637681, "grad_norm": 0.2727385963005962, "learning_rate": 5.223311050686807e-05, "loss": 0.871, "step": 5089 }, { "epoch": 2.3052536231884058, "grad_norm": 0.25795443198853496, "learning_rate": 5.222106855472774e-05, "loss": 0.8815, "step": 5090 }, { "epoch": 2.3057065217391304, "grad_norm": 0.28825982756025564, "learning_rate": 5.220902538082781e-05, "loss": 0.8708, "step": 5091 }, { "epoch": 2.306159420289855, "grad_norm": 0.341857305908484, "learning_rate": 5.219698098637226e-05, "loss": 0.8767, "step": 5092 }, { "epoch": 2.3066123188405796, "grad_norm": 0.25264724510646636, "learning_rate": 5.2184935372565204e-05, "loss": 0.8423, "step": 5093 }, { "epoch": 2.307065217391304, "grad_norm": 0.239639314660974, "learning_rate": 5.217288854061085e-05, "loss": 0.884, "step": 5094 }, { "epoch": 2.307518115942029, "grad_norm": 0.33411975197662125, "learning_rate": 5.216084049171353e-05, "loss": 0.8757, "step": 5095 }, { "epoch": 2.3079710144927534, "grad_norm": 0.3534251924338687, "learning_rate": 5.2148791227077715e-05, "loss": 0.8441, "step": 5096 }, { "epoch": 2.3084239130434785, "grad_norm": 0.25688158051766635, "learning_rate": 5.2136740747907996e-05, "loss": 0.8601, "step": 5097 }, { "epoch": 2.308876811594203, "grad_norm": 0.2204796709075867, "learning_rate": 5.212468905540907e-05, "loss": 0.8684, "step": 5098 }, { "epoch": 2.3093297101449277, "grad_norm": 0.24773872011078202, "learning_rate": 5.2112636150785774e-05, "loss": 0.8685, "step": 5099 }, { "epoch": 2.3097826086956523, "grad_norm": 0.25749295327702115, "learning_rate": 5.210058203524304e-05, "loss": 0.875, "step": 5100 }, { "epoch": 2.310235507246377, "grad_norm": 0.22366448579684156, "learning_rate": 5.208852670998594e-05, "loss": 0.8579, "step": 5101 }, { "epoch": 2.3106884057971016, "grad_norm": 0.24294146464380573, "learning_rate": 5.207647017621967e-05, "loss": 0.8512, "step": 5102 }, { "epoch": 2.311141304347826, "grad_norm": 0.2590446604264153, "learning_rate": 5.206441243514955e-05, "loss": 0.8818, "step": 5103 }, { "epoch": 2.3115942028985508, "grad_norm": 0.3560226654398005, "learning_rate": 5.2052353487981e-05, "loss": 0.8865, "step": 5104 }, { "epoch": 2.3120471014492754, "grad_norm": 0.261950528929026, "learning_rate": 5.204029333591956e-05, "loss": 0.8762, "step": 5105 }, { "epoch": 2.3125, "grad_norm": 0.21240038393894078, "learning_rate": 5.202823198017092e-05, "loss": 0.8661, "step": 5106 }, { "epoch": 2.3129528985507246, "grad_norm": 0.3063565376126379, "learning_rate": 5.201616942194088e-05, "loss": 0.8675, "step": 5107 }, { "epoch": 2.3134057971014492, "grad_norm": 0.32593038515567413, "learning_rate": 5.200410566243534e-05, "loss": 0.8728, "step": 5108 }, { "epoch": 2.313858695652174, "grad_norm": 0.2823614592764224, "learning_rate": 5.1992040702860325e-05, "loss": 0.8554, "step": 5109 }, { "epoch": 2.3143115942028984, "grad_norm": 0.32205734964106353, "learning_rate": 5.1979974544422e-05, "loss": 0.887, "step": 5110 }, { "epoch": 2.314764492753623, "grad_norm": 0.3789202320508188, "learning_rate": 5.196790718832665e-05, "loss": 0.8691, "step": 5111 }, { "epoch": 2.3152173913043477, "grad_norm": 0.4268654205601833, "learning_rate": 5.195583863578064e-05, "loss": 0.8851, "step": 5112 }, { "epoch": 2.3156702898550723, "grad_norm": 0.5044313321178171, "learning_rate": 5.194376888799049e-05, "loss": 0.8537, "step": 5113 }, { "epoch": 2.316123188405797, "grad_norm": 0.5524816336099153, "learning_rate": 5.193169794616285e-05, "loss": 0.861, "step": 5114 }, { "epoch": 2.3165760869565215, "grad_norm": 0.5283359041054899, "learning_rate": 5.191962581150446e-05, "loss": 0.8819, "step": 5115 }, { "epoch": 2.3170289855072466, "grad_norm": 0.5403413838218109, "learning_rate": 5.190755248522218e-05, "loss": 0.8887, "step": 5116 }, { "epoch": 2.317481884057971, "grad_norm": 0.5406836348996045, "learning_rate": 5.1895477968523013e-05, "loss": 0.8583, "step": 5117 }, { "epoch": 2.317934782608696, "grad_norm": 0.5269626270167319, "learning_rate": 5.188340226261406e-05, "loss": 0.8648, "step": 5118 }, { "epoch": 2.3183876811594204, "grad_norm": 0.45118984520369626, "learning_rate": 5.1871325368702553e-05, "loss": 0.8641, "step": 5119 }, { "epoch": 2.318840579710145, "grad_norm": 0.36086060449375235, "learning_rate": 5.1859247287995846e-05, "loss": 0.8851, "step": 5120 }, { "epoch": 2.3192934782608696, "grad_norm": 0.24225336710372777, "learning_rate": 5.1847168021701384e-05, "loss": 0.8784, "step": 5121 }, { "epoch": 2.3197463768115942, "grad_norm": 0.3310325640026359, "learning_rate": 5.183508757102676e-05, "loss": 0.8904, "step": 5122 }, { "epoch": 2.320199275362319, "grad_norm": 0.37774266588379857, "learning_rate": 5.1823005937179684e-05, "loss": 0.885, "step": 5123 }, { "epoch": 2.3206521739130435, "grad_norm": 0.35927671188921884, "learning_rate": 5.1810923121367975e-05, "loss": 0.8603, "step": 5124 }, { "epoch": 2.321105072463768, "grad_norm": 0.4191522662383335, "learning_rate": 5.1798839124799554e-05, "loss": 0.8749, "step": 5125 }, { "epoch": 2.3215579710144927, "grad_norm": 0.42924426575977587, "learning_rate": 5.1786753948682486e-05, "loss": 0.8925, "step": 5126 }, { "epoch": 2.3220108695652173, "grad_norm": 0.3491277782498874, "learning_rate": 5.177466759422496e-05, "loss": 0.8635, "step": 5127 }, { "epoch": 2.322463768115942, "grad_norm": 0.2823259842369387, "learning_rate": 5.176258006263526e-05, "loss": 0.8653, "step": 5128 }, { "epoch": 2.3229166666666665, "grad_norm": 0.26992590196502775, "learning_rate": 5.1750491355121776e-05, "loss": 0.8762, "step": 5129 }, { "epoch": 2.323369565217391, "grad_norm": 0.2911180493981485, "learning_rate": 5.1738401472893055e-05, "loss": 0.8752, "step": 5130 }, { "epoch": 2.323822463768116, "grad_norm": 0.3000246698012474, "learning_rate": 5.1726310417157744e-05, "loss": 0.8669, "step": 5131 }, { "epoch": 2.324275362318841, "grad_norm": 0.34519199716323656, "learning_rate": 5.1714218189124605e-05, "loss": 0.8818, "step": 5132 }, { "epoch": 2.3247282608695654, "grad_norm": 0.364297993873116, "learning_rate": 5.1702124790002497e-05, "loss": 0.883, "step": 5133 }, { "epoch": 2.32518115942029, "grad_norm": 0.3512181581570169, "learning_rate": 5.1690030221000434e-05, "loss": 0.8903, "step": 5134 }, { "epoch": 2.3256340579710146, "grad_norm": 0.33763464118970604, "learning_rate": 5.1677934483327525e-05, "loss": 0.8776, "step": 5135 }, { "epoch": 2.3260869565217392, "grad_norm": 0.3708882818529672, "learning_rate": 5.1665837578193e-05, "loss": 0.8823, "step": 5136 }, { "epoch": 2.326539855072464, "grad_norm": 0.39738625500894026, "learning_rate": 5.1653739506806216e-05, "loss": 0.8633, "step": 5137 }, { "epoch": 2.3269927536231885, "grad_norm": 0.364461191623312, "learning_rate": 5.164164027037661e-05, "loss": 0.8828, "step": 5138 }, { "epoch": 2.327445652173913, "grad_norm": 0.3998108430870921, "learning_rate": 5.1629539870113786e-05, "loss": 0.8813, "step": 5139 }, { "epoch": 2.3278985507246377, "grad_norm": 0.2922816821961714, "learning_rate": 5.161743830722743e-05, "loss": 0.864, "step": 5140 }, { "epoch": 2.3283514492753623, "grad_norm": 0.22393989761798022, "learning_rate": 5.160533558292738e-05, "loss": 0.8817, "step": 5141 }, { "epoch": 2.328804347826087, "grad_norm": 0.3555841686207376, "learning_rate": 5.159323169842352e-05, "loss": 0.8664, "step": 5142 }, { "epoch": 2.3292572463768115, "grad_norm": 0.4603285049964745, "learning_rate": 5.158112665492591e-05, "loss": 0.8794, "step": 5143 }, { "epoch": 2.329710144927536, "grad_norm": 0.509702047070327, "learning_rate": 5.1569020453644726e-05, "loss": 0.8942, "step": 5144 }, { "epoch": 2.3301630434782608, "grad_norm": 0.5385601055755465, "learning_rate": 5.1556913095790234e-05, "loss": 0.8745, "step": 5145 }, { "epoch": 2.3306159420289854, "grad_norm": 0.48901173800485714, "learning_rate": 5.154480458257283e-05, "loss": 0.8844, "step": 5146 }, { "epoch": 2.33106884057971, "grad_norm": 0.42915979355248374, "learning_rate": 5.1532694915203005e-05, "loss": 0.8784, "step": 5147 }, { "epoch": 2.3315217391304346, "grad_norm": 0.441812068898866, "learning_rate": 5.15205840948914e-05, "loss": 0.8605, "step": 5148 }, { "epoch": 2.331974637681159, "grad_norm": 0.4380692037480383, "learning_rate": 5.150847212284875e-05, "loss": 0.8749, "step": 5149 }, { "epoch": 2.332427536231884, "grad_norm": 0.3486907335053125, "learning_rate": 5.149635900028591e-05, "loss": 0.8852, "step": 5150 }, { "epoch": 2.332880434782609, "grad_norm": 0.31936686927667957, "learning_rate": 5.1484244728413834e-05, "loss": 0.862, "step": 5151 }, { "epoch": 2.3333333333333335, "grad_norm": 0.27640317436816825, "learning_rate": 5.1472129308443616e-05, "loss": 0.8917, "step": 5152 }, { "epoch": 2.333786231884058, "grad_norm": 0.27141256672777686, "learning_rate": 5.146001274158646e-05, "loss": 0.8727, "step": 5153 }, { "epoch": 2.3342391304347827, "grad_norm": 0.18458179019563697, "learning_rate": 5.144789502905366e-05, "loss": 0.8648, "step": 5154 }, { "epoch": 2.3346920289855073, "grad_norm": 0.22191111848846995, "learning_rate": 5.1435776172056665e-05, "loss": 0.8744, "step": 5155 }, { "epoch": 2.335144927536232, "grad_norm": 0.22793525158171885, "learning_rate": 5.1423656171806986e-05, "loss": 0.8731, "step": 5156 }, { "epoch": 2.3355978260869565, "grad_norm": 0.28096449444742283, "learning_rate": 5.141153502951631e-05, "loss": 0.8549, "step": 5157 }, { "epoch": 2.336050724637681, "grad_norm": 0.292032428525056, "learning_rate": 5.1399412746396394e-05, "loss": 0.8779, "step": 5158 }, { "epoch": 2.3365036231884058, "grad_norm": 0.3017790765112371, "learning_rate": 5.1387289323659114e-05, "loss": 0.8586, "step": 5159 }, { "epoch": 2.3369565217391304, "grad_norm": 0.32856355233863843, "learning_rate": 5.137516476251648e-05, "loss": 0.8754, "step": 5160 }, { "epoch": 2.337409420289855, "grad_norm": 0.35379910042075413, "learning_rate": 5.13630390641806e-05, "loss": 0.8676, "step": 5161 }, { "epoch": 2.3378623188405796, "grad_norm": 0.4109782387967013, "learning_rate": 5.1350912229863704e-05, "loss": 0.866, "step": 5162 }, { "epoch": 2.338315217391304, "grad_norm": 0.36981405273867024, "learning_rate": 5.133878426077812e-05, "loss": 0.8881, "step": 5163 }, { "epoch": 2.338768115942029, "grad_norm": 0.3385837668525731, "learning_rate": 5.132665515813631e-05, "loss": 0.8836, "step": 5164 }, { "epoch": 2.3392210144927534, "grad_norm": 0.35888428030972996, "learning_rate": 5.131452492315083e-05, "loss": 0.856, "step": 5165 }, { "epoch": 2.3396739130434785, "grad_norm": 0.33661714034488255, "learning_rate": 5.130239355703436e-05, "loss": 0.8751, "step": 5166 }, { "epoch": 2.340126811594203, "grad_norm": 0.27875248318510915, "learning_rate": 5.129026106099971e-05, "loss": 0.8815, "step": 5167 }, { "epoch": 2.3405797101449277, "grad_norm": 0.24737887957856744, "learning_rate": 5.127812743625976e-05, "loss": 0.8522, "step": 5168 }, { "epoch": 2.3410326086956523, "grad_norm": 0.2327779629435453, "learning_rate": 5.126599268402753e-05, "loss": 0.8744, "step": 5169 }, { "epoch": 2.341485507246377, "grad_norm": 0.29653018041896484, "learning_rate": 5.125385680551617e-05, "loss": 0.8965, "step": 5170 }, { "epoch": 2.3419384057971016, "grad_norm": 0.34847290524702623, "learning_rate": 5.124171980193891e-05, "loss": 0.8688, "step": 5171 }, { "epoch": 2.342391304347826, "grad_norm": 0.347594620538964, "learning_rate": 5.122958167450911e-05, "loss": 0.8942, "step": 5172 }, { "epoch": 2.3428442028985508, "grad_norm": 0.2675266352389, "learning_rate": 5.121744242444023e-05, "loss": 0.8714, "step": 5173 }, { "epoch": 2.3432971014492754, "grad_norm": 0.21417580953877555, "learning_rate": 5.1205302052945856e-05, "loss": 0.8836, "step": 5174 }, { "epoch": 2.34375, "grad_norm": 0.3289004525021296, "learning_rate": 5.1193160561239694e-05, "loss": 0.8628, "step": 5175 }, { "epoch": 2.3442028985507246, "grad_norm": 0.43433225678147425, "learning_rate": 5.118101795053552e-05, "loss": 0.8815, "step": 5176 }, { "epoch": 2.3446557971014492, "grad_norm": 0.4569030172058593, "learning_rate": 5.116887422204726e-05, "loss": 0.8653, "step": 5177 }, { "epoch": 2.345108695652174, "grad_norm": 0.43529491704272333, "learning_rate": 5.115672937698895e-05, "loss": 0.8735, "step": 5178 }, { "epoch": 2.3455615942028984, "grad_norm": 0.5123667998395673, "learning_rate": 5.114458341657473e-05, "loss": 0.8596, "step": 5179 }, { "epoch": 2.346014492753623, "grad_norm": 0.5553371889664812, "learning_rate": 5.113243634201884e-05, "loss": 0.855, "step": 5180 }, { "epoch": 2.3464673913043477, "grad_norm": 0.49662711924759817, "learning_rate": 5.112028815453565e-05, "loss": 0.8823, "step": 5181 }, { "epoch": 2.3469202898550723, "grad_norm": 0.45420888543369103, "learning_rate": 5.1108138855339633e-05, "loss": 0.8733, "step": 5182 }, { "epoch": 2.347373188405797, "grad_norm": 0.46451006926241106, "learning_rate": 5.109598844564537e-05, "loss": 0.8846, "step": 5183 }, { "epoch": 2.3478260869565215, "grad_norm": 0.35341285864062566, "learning_rate": 5.108383692666756e-05, "loss": 0.8542, "step": 5184 }, { "epoch": 2.3482789855072466, "grad_norm": 0.2918409248297297, "learning_rate": 5.107168429962101e-05, "loss": 0.8873, "step": 5185 }, { "epoch": 2.348731884057971, "grad_norm": 0.294945029882274, "learning_rate": 5.1059530565720624e-05, "loss": 0.8667, "step": 5186 }, { "epoch": 2.349184782608696, "grad_norm": 0.231700305389497, "learning_rate": 5.104737572618145e-05, "loss": 0.8919, "step": 5187 }, { "epoch": 2.3496376811594204, "grad_norm": 0.3331838468135691, "learning_rate": 5.103521978221863e-05, "loss": 0.871, "step": 5188 }, { "epoch": 2.350090579710145, "grad_norm": 0.39674971075811866, "learning_rate": 5.102306273504739e-05, "loss": 0.8543, "step": 5189 }, { "epoch": 2.3505434782608696, "grad_norm": 0.44922634078518037, "learning_rate": 5.10109045858831e-05, "loss": 0.8711, "step": 5190 }, { "epoch": 2.3509963768115942, "grad_norm": 0.5183250327562345, "learning_rate": 5.099874533594123e-05, "loss": 0.8852, "step": 5191 }, { "epoch": 2.351449275362319, "grad_norm": 0.5681464111642783, "learning_rate": 5.0986584986437366e-05, "loss": 0.8961, "step": 5192 }, { "epoch": 2.3519021739130435, "grad_norm": 0.5217167426054109, "learning_rate": 5.0974423538587194e-05, "loss": 0.8673, "step": 5193 }, { "epoch": 2.352355072463768, "grad_norm": 0.4507091766259378, "learning_rate": 5.09622609936065e-05, "loss": 0.859, "step": 5194 }, { "epoch": 2.3528079710144927, "grad_norm": 0.33566868077055334, "learning_rate": 5.09500973527112e-05, "loss": 0.879, "step": 5195 }, { "epoch": 2.3532608695652173, "grad_norm": 0.2534266137961187, "learning_rate": 5.093793261711733e-05, "loss": 0.8632, "step": 5196 }, { "epoch": 2.353713768115942, "grad_norm": 0.29941329648783443, "learning_rate": 5.0925766788041e-05, "loss": 0.8784, "step": 5197 }, { "epoch": 2.3541666666666665, "grad_norm": 0.34903695939258605, "learning_rate": 5.091359986669845e-05, "loss": 0.8784, "step": 5198 }, { "epoch": 2.354619565217391, "grad_norm": 0.4484428269655672, "learning_rate": 5.0901431854306024e-05, "loss": 0.8761, "step": 5199 }, { "epoch": 2.355072463768116, "grad_norm": 0.4522610322663896, "learning_rate": 5.0889262752080185e-05, "loss": 0.8808, "step": 5200 }, { "epoch": 2.355525362318841, "grad_norm": 0.49400053848974235, "learning_rate": 5.0877092561237484e-05, "loss": 0.8843, "step": 5201 }, { "epoch": 2.3559782608695654, "grad_norm": 0.47026787102743256, "learning_rate": 5.086492128299461e-05, "loss": 0.8715, "step": 5202 }, { "epoch": 2.35643115942029, "grad_norm": 0.3588149180747251, "learning_rate": 5.0852748918568335e-05, "loss": 0.8638, "step": 5203 }, { "epoch": 2.3568840579710146, "grad_norm": 0.24645511161093978, "learning_rate": 5.084057546917557e-05, "loss": 0.874, "step": 5204 }, { "epoch": 2.3573369565217392, "grad_norm": 0.2833604772786992, "learning_rate": 5.082840093603327e-05, "loss": 0.8586, "step": 5205 }, { "epoch": 2.357789855072464, "grad_norm": 0.27690018486443146, "learning_rate": 5.0816225320358586e-05, "loss": 0.8779, "step": 5206 }, { "epoch": 2.3582427536231885, "grad_norm": 0.34036352701043776, "learning_rate": 5.08040486233687e-05, "loss": 0.8841, "step": 5207 }, { "epoch": 2.358695652173913, "grad_norm": 0.4443186052980098, "learning_rate": 5.0791870846280974e-05, "loss": 0.8785, "step": 5208 }, { "epoch": 2.3591485507246377, "grad_norm": 0.460076997687842, "learning_rate": 5.077969199031282e-05, "loss": 0.8821, "step": 5209 }, { "epoch": 2.3596014492753623, "grad_norm": 0.4670647080844096, "learning_rate": 5.076751205668176e-05, "loss": 0.8818, "step": 5210 }, { "epoch": 2.360054347826087, "grad_norm": 0.48987728789419277, "learning_rate": 5.075533104660546e-05, "loss": 0.8729, "step": 5211 }, { "epoch": 2.3605072463768115, "grad_norm": 0.4123393487125801, "learning_rate": 5.074314896130167e-05, "loss": 0.8776, "step": 5212 }, { "epoch": 2.360960144927536, "grad_norm": 0.3635757134150173, "learning_rate": 5.073096580198827e-05, "loss": 0.8531, "step": 5213 }, { "epoch": 2.3614130434782608, "grad_norm": 0.3616028160621931, "learning_rate": 5.071878156988319e-05, "loss": 0.8787, "step": 5214 }, { "epoch": 2.3618659420289854, "grad_norm": 0.31684369402611845, "learning_rate": 5.0706596266204555e-05, "loss": 0.878, "step": 5215 }, { "epoch": 2.36231884057971, "grad_norm": 0.22353056359872686, "learning_rate": 5.069440989217051e-05, "loss": 0.9161, "step": 5216 }, { "epoch": 2.3627717391304346, "grad_norm": 0.2587360251469153, "learning_rate": 5.068222244899938e-05, "loss": 0.8749, "step": 5217 }, { "epoch": 2.363224637681159, "grad_norm": 0.31718776777980034, "learning_rate": 5.067003393790952e-05, "loss": 0.8557, "step": 5218 }, { "epoch": 2.363677536231884, "grad_norm": 0.3333825839615583, "learning_rate": 5.065784436011947e-05, "loss": 0.869, "step": 5219 }, { "epoch": 2.364130434782609, "grad_norm": 0.4086667196313561, "learning_rate": 5.064565371684783e-05, "loss": 0.8787, "step": 5220 }, { "epoch": 2.3645833333333335, "grad_norm": 0.4613308683787748, "learning_rate": 5.0633462009313315e-05, "loss": 0.885, "step": 5221 }, { "epoch": 2.365036231884058, "grad_norm": 0.47342249234648415, "learning_rate": 5.0621269238734764e-05, "loss": 0.8448, "step": 5222 }, { "epoch": 2.3654891304347827, "grad_norm": 0.43520730452482675, "learning_rate": 5.0609075406331084e-05, "loss": 0.8955, "step": 5223 }, { "epoch": 2.3659420289855073, "grad_norm": 0.4212942217439035, "learning_rate": 5.059688051332132e-05, "loss": 0.8784, "step": 5224 }, { "epoch": 2.366394927536232, "grad_norm": 0.4088139522351096, "learning_rate": 5.058468456092464e-05, "loss": 0.8823, "step": 5225 }, { "epoch": 2.3668478260869565, "grad_norm": 0.4190713632283784, "learning_rate": 5.0572487550360256e-05, "loss": 0.8742, "step": 5226 }, { "epoch": 2.367300724637681, "grad_norm": 0.41254305754695303, "learning_rate": 5.0560289482847534e-05, "loss": 0.8653, "step": 5227 }, { "epoch": 2.3677536231884058, "grad_norm": 0.38047536619437305, "learning_rate": 5.054809035960595e-05, "loss": 0.8733, "step": 5228 }, { "epoch": 2.3682065217391304, "grad_norm": 0.36206290011378905, "learning_rate": 5.0535890181855044e-05, "loss": 0.8782, "step": 5229 }, { "epoch": 2.368659420289855, "grad_norm": 0.24709903814983, "learning_rate": 5.052368895081453e-05, "loss": 0.8794, "step": 5230 }, { "epoch": 2.3691123188405796, "grad_norm": 0.2942355879528333, "learning_rate": 5.051148666770413e-05, "loss": 0.8933, "step": 5231 }, { "epoch": 2.369565217391304, "grad_norm": 0.3061427801476798, "learning_rate": 5.0499283333743754e-05, "loss": 0.873, "step": 5232 }, { "epoch": 2.370018115942029, "grad_norm": 0.2100795393323515, "learning_rate": 5.048707895015338e-05, "loss": 0.8583, "step": 5233 }, { "epoch": 2.3704710144927534, "grad_norm": 0.2865102872221596, "learning_rate": 5.047487351815312e-05, "loss": 0.8734, "step": 5234 }, { "epoch": 2.3709239130434785, "grad_norm": 0.3020067859835802, "learning_rate": 5.046266703896315e-05, "loss": 0.8753, "step": 5235 }, { "epoch": 2.371376811594203, "grad_norm": 0.27675080181305534, "learning_rate": 5.045045951380378e-05, "loss": 0.8784, "step": 5236 }, { "epoch": 2.3718297101449277, "grad_norm": 0.269641684391541, "learning_rate": 5.0438250943895404e-05, "loss": 0.8884, "step": 5237 }, { "epoch": 2.3722826086956523, "grad_norm": 0.25788695268733935, "learning_rate": 5.042604133045855e-05, "loss": 0.8997, "step": 5238 }, { "epoch": 2.372735507246377, "grad_norm": 0.28476879334436866, "learning_rate": 5.0413830674713815e-05, "loss": 0.9022, "step": 5239 }, { "epoch": 2.3731884057971016, "grad_norm": 0.29098112118953273, "learning_rate": 5.0401618977881924e-05, "loss": 0.8584, "step": 5240 }, { "epoch": 2.373641304347826, "grad_norm": 0.30847594206893525, "learning_rate": 5.03894062411837e-05, "loss": 0.8607, "step": 5241 }, { "epoch": 2.3740942028985508, "grad_norm": 0.41404033271361423, "learning_rate": 5.037719246584008e-05, "loss": 0.8837, "step": 5242 }, { "epoch": 2.3745471014492754, "grad_norm": 0.4705605529969835, "learning_rate": 5.036497765307207e-05, "loss": 0.8467, "step": 5243 }, { "epoch": 2.375, "grad_norm": 0.49000583208697224, "learning_rate": 5.0352761804100835e-05, "loss": 0.8892, "step": 5244 }, { "epoch": 2.3754528985507246, "grad_norm": 0.4598541822772822, "learning_rate": 5.034054492014758e-05, "loss": 0.8705, "step": 5245 }, { "epoch": 2.3759057971014492, "grad_norm": 0.30760059385622424, "learning_rate": 5.0328327002433686e-05, "loss": 0.867, "step": 5246 }, { "epoch": 2.376358695652174, "grad_norm": 0.25918231271842174, "learning_rate": 5.031610805218057e-05, "loss": 0.875, "step": 5247 }, { "epoch": 2.3768115942028984, "grad_norm": 0.3555014183171479, "learning_rate": 5.030388807060977e-05, "loss": 0.8905, "step": 5248 }, { "epoch": 2.377264492753623, "grad_norm": 0.3855345110972102, "learning_rate": 5.0291667058942964e-05, "loss": 0.8793, "step": 5249 }, { "epoch": 2.3777173913043477, "grad_norm": 0.4122865718757217, "learning_rate": 5.0279445018401894e-05, "loss": 0.8824, "step": 5250 }, { "epoch": 2.3781702898550723, "grad_norm": 0.4381914246693972, "learning_rate": 5.02672219502084e-05, "loss": 0.8849, "step": 5251 }, { "epoch": 2.378623188405797, "grad_norm": 0.4302512554333977, "learning_rate": 5.0254997855584474e-05, "loss": 0.8634, "step": 5252 }, { "epoch": 2.3790760869565215, "grad_norm": 0.35794913906626175, "learning_rate": 5.024277273575216e-05, "loss": 0.8701, "step": 5253 }, { "epoch": 2.3795289855072466, "grad_norm": 0.3079147552009781, "learning_rate": 5.023054659193362e-05, "loss": 0.8751, "step": 5254 }, { "epoch": 2.379981884057971, "grad_norm": 0.24471482577256942, "learning_rate": 5.021831942535113e-05, "loss": 0.8749, "step": 5255 }, { "epoch": 2.380434782608696, "grad_norm": 0.20754264953628462, "learning_rate": 5.020609123722705e-05, "loss": 0.8717, "step": 5256 }, { "epoch": 2.3808876811594204, "grad_norm": 0.22195998497440247, "learning_rate": 5.019386202878386e-05, "loss": 0.8616, "step": 5257 }, { "epoch": 2.381340579710145, "grad_norm": 0.20652270273670797, "learning_rate": 5.018163180124412e-05, "loss": 0.8771, "step": 5258 }, { "epoch": 2.3817934782608696, "grad_norm": 0.23611406398005608, "learning_rate": 5.016940055583053e-05, "loss": 0.8925, "step": 5259 }, { "epoch": 2.3822463768115942, "grad_norm": 0.21873516944339716, "learning_rate": 5.015716829376584e-05, "loss": 0.8628, "step": 5260 }, { "epoch": 2.382699275362319, "grad_norm": 0.19225719309388145, "learning_rate": 5.0144935016272944e-05, "loss": 0.8726, "step": 5261 }, { "epoch": 2.3831521739130435, "grad_norm": 0.23049364501804423, "learning_rate": 5.0132700724574814e-05, "loss": 0.8707, "step": 5262 }, { "epoch": 2.383605072463768, "grad_norm": 0.29347632901170995, "learning_rate": 5.0120465419894554e-05, "loss": 0.8752, "step": 5263 }, { "epoch": 2.3840579710144927, "grad_norm": 0.31580923629577473, "learning_rate": 5.010822910345531e-05, "loss": 0.8744, "step": 5264 }, { "epoch": 2.3845108695652173, "grad_norm": 0.3072523069934345, "learning_rate": 5.009599177648039e-05, "loss": 0.87, "step": 5265 }, { "epoch": 2.384963768115942, "grad_norm": 0.32044630500591004, "learning_rate": 5.008375344019318e-05, "loss": 0.8991, "step": 5266 }, { "epoch": 2.3854166666666665, "grad_norm": 0.2975880408077756, "learning_rate": 5.007151409581715e-05, "loss": 0.8862, "step": 5267 }, { "epoch": 2.385869565217391, "grad_norm": 0.3261381158063891, "learning_rate": 5.0059273744575906e-05, "loss": 0.8795, "step": 5268 }, { "epoch": 2.386322463768116, "grad_norm": 0.37514808816813167, "learning_rate": 5.004703238769312e-05, "loss": 0.8543, "step": 5269 }, { "epoch": 2.386775362318841, "grad_norm": 0.40748706215557984, "learning_rate": 5.0034790026392574e-05, "loss": 0.8736, "step": 5270 }, { "epoch": 2.3872282608695654, "grad_norm": 0.39967841363576845, "learning_rate": 5.002254666189818e-05, "loss": 0.8919, "step": 5271 }, { "epoch": 2.38768115942029, "grad_norm": 0.3905705280974013, "learning_rate": 5.001030229543392e-05, "loss": 0.86, "step": 5272 }, { "epoch": 2.3881340579710146, "grad_norm": 0.3605267539213544, "learning_rate": 4.9998056928223865e-05, "loss": 0.8818, "step": 5273 }, { "epoch": 2.3885869565217392, "grad_norm": 0.3691439216861918, "learning_rate": 4.998581056149222e-05, "loss": 0.8876, "step": 5274 }, { "epoch": 2.389039855072464, "grad_norm": 0.33430742324307067, "learning_rate": 4.9973563196463266e-05, "loss": 0.8615, "step": 5275 }, { "epoch": 2.3894927536231885, "grad_norm": 0.32270017967328357, "learning_rate": 4.99613148343614e-05, "loss": 0.8556, "step": 5276 }, { "epoch": 2.389945652173913, "grad_norm": 0.2976136854509647, "learning_rate": 4.99490654764111e-05, "loss": 0.8898, "step": 5277 }, { "epoch": 2.3903985507246377, "grad_norm": 0.2580085808187967, "learning_rate": 4.9936815123836956e-05, "loss": 0.8813, "step": 5278 }, { "epoch": 2.3908514492753623, "grad_norm": 0.21954704736939828, "learning_rate": 4.9924563777863666e-05, "loss": 0.8659, "step": 5279 }, { "epoch": 2.391304347826087, "grad_norm": 0.22134332506575602, "learning_rate": 4.9912311439716e-05, "loss": 0.8624, "step": 5280 }, { "epoch": 2.3917572463768115, "grad_norm": 0.22570525382823572, "learning_rate": 4.990005811061886e-05, "loss": 0.8854, "step": 5281 }, { "epoch": 2.392210144927536, "grad_norm": 0.232065095631546, "learning_rate": 4.988780379179722e-05, "loss": 0.9016, "step": 5282 }, { "epoch": 2.3926630434782608, "grad_norm": 0.20308253599838286, "learning_rate": 4.9875548484476166e-05, "loss": 0.8795, "step": 5283 }, { "epoch": 2.3931159420289854, "grad_norm": 0.3131624793818251, "learning_rate": 4.986329218988088e-05, "loss": 0.8688, "step": 5284 }, { "epoch": 2.39356884057971, "grad_norm": 0.3022785837719026, "learning_rate": 4.9851034909236655e-05, "loss": 0.8663, "step": 5285 }, { "epoch": 2.3940217391304346, "grad_norm": 0.24817945081593154, "learning_rate": 4.983877664376884e-05, "loss": 0.8686, "step": 5286 }, { "epoch": 2.394474637681159, "grad_norm": 0.30673482922209055, "learning_rate": 4.982651739470294e-05, "loss": 0.8795, "step": 5287 }, { "epoch": 2.394927536231884, "grad_norm": 0.2733648516736479, "learning_rate": 4.981425716326453e-05, "loss": 0.8653, "step": 5288 }, { "epoch": 2.395380434782609, "grad_norm": 0.3795904723131756, "learning_rate": 4.980199595067928e-05, "loss": 0.8876, "step": 5289 }, { "epoch": 2.3958333333333335, "grad_norm": 0.31930132022045676, "learning_rate": 4.9789733758172956e-05, "loss": 0.8668, "step": 5290 }, { "epoch": 2.396286231884058, "grad_norm": 0.2553884655829558, "learning_rate": 4.9777470586971444e-05, "loss": 0.8855, "step": 5291 }, { "epoch": 2.3967391304347827, "grad_norm": 0.29099931028054915, "learning_rate": 4.9765206438300694e-05, "loss": 0.8796, "step": 5292 }, { "epoch": 2.3971920289855073, "grad_norm": 0.21976252241629002, "learning_rate": 4.975294131338679e-05, "loss": 0.8724, "step": 5293 }, { "epoch": 2.397644927536232, "grad_norm": 0.24358405597076022, "learning_rate": 4.974067521345588e-05, "loss": 0.8488, "step": 5294 }, { "epoch": 2.3980978260869565, "grad_norm": 0.21791094320486673, "learning_rate": 4.972840813973424e-05, "loss": 0.8582, "step": 5295 }, { "epoch": 2.398550724637681, "grad_norm": 0.27610572495654506, "learning_rate": 4.971614009344822e-05, "loss": 0.8729, "step": 5296 }, { "epoch": 2.3990036231884058, "grad_norm": 0.3358631893169515, "learning_rate": 4.970387107582427e-05, "loss": 0.866, "step": 5297 }, { "epoch": 2.3994565217391304, "grad_norm": 0.42357130967763157, "learning_rate": 4.9691601088088967e-05, "loss": 0.8918, "step": 5298 }, { "epoch": 2.399909420289855, "grad_norm": 0.4240274719961398, "learning_rate": 4.967933013146894e-05, "loss": 0.8729, "step": 5299 }, { "epoch": 2.4003623188405796, "grad_norm": 0.3991055565781149, "learning_rate": 4.966705820719093e-05, "loss": 0.8994, "step": 5300 }, { "epoch": 2.400815217391304, "grad_norm": 0.40619837602542586, "learning_rate": 4.96547853164818e-05, "loss": 0.867, "step": 5301 }, { "epoch": 2.401268115942029, "grad_norm": 0.3751159808129083, "learning_rate": 4.9642511460568495e-05, "loss": 0.8781, "step": 5302 }, { "epoch": 2.4017210144927534, "grad_norm": 0.2963352859528538, "learning_rate": 4.9630236640678025e-05, "loss": 0.8727, "step": 5303 }, { "epoch": 2.4021739130434785, "grad_norm": 0.17683152906615213, "learning_rate": 4.961796085803754e-05, "loss": 0.8684, "step": 5304 }, { "epoch": 2.402626811594203, "grad_norm": 0.20732831526895237, "learning_rate": 4.960568411387427e-05, "loss": 0.8664, "step": 5305 }, { "epoch": 2.4030797101449277, "grad_norm": 0.383960266142177, "learning_rate": 4.959340640941554e-05, "loss": 0.8945, "step": 5306 }, { "epoch": 2.4035326086956523, "grad_norm": 0.5426974605883466, "learning_rate": 4.958112774588876e-05, "loss": 0.8612, "step": 5307 }, { "epoch": 2.403985507246377, "grad_norm": 0.5347241831910593, "learning_rate": 4.956884812452146e-05, "loss": 0.877, "step": 5308 }, { "epoch": 2.4044384057971016, "grad_norm": 0.5003253393556747, "learning_rate": 4.9556567546541256e-05, "loss": 0.8633, "step": 5309 }, { "epoch": 2.404891304347826, "grad_norm": 0.433325149209193, "learning_rate": 4.954428601317585e-05, "loss": 0.8644, "step": 5310 }, { "epoch": 2.4053442028985508, "grad_norm": 0.37280480113995973, "learning_rate": 4.953200352565304e-05, "loss": 0.8877, "step": 5311 }, { "epoch": 2.4057971014492754, "grad_norm": 0.3081571973237557, "learning_rate": 4.9519720085200734e-05, "loss": 0.8859, "step": 5312 }, { "epoch": 2.40625, "grad_norm": 0.29421566530980425, "learning_rate": 4.950743569304693e-05, "loss": 0.8594, "step": 5313 }, { "epoch": 2.4067028985507246, "grad_norm": 0.31331149707375333, "learning_rate": 4.9495150350419716e-05, "loss": 0.8815, "step": 5314 }, { "epoch": 2.4071557971014492, "grad_norm": 0.35772851514584647, "learning_rate": 4.9482864058547265e-05, "loss": 0.893, "step": 5315 }, { "epoch": 2.407608695652174, "grad_norm": 0.4234000024955328, "learning_rate": 4.9470576818657873e-05, "loss": 0.8684, "step": 5316 }, { "epoch": 2.4080615942028984, "grad_norm": 0.4736408723435168, "learning_rate": 4.9458288631979916e-05, "loss": 0.8529, "step": 5317 }, { "epoch": 2.408514492753623, "grad_norm": 0.5750130024007448, "learning_rate": 4.9445999499741854e-05, "loss": 0.8743, "step": 5318 }, { "epoch": 2.4089673913043477, "grad_norm": 0.6475511629370548, "learning_rate": 4.9433709423172245e-05, "loss": 0.8869, "step": 5319 }, { "epoch": 2.4094202898550723, "grad_norm": 0.5795924736612001, "learning_rate": 4.942141840349976e-05, "loss": 0.8938, "step": 5320 }, { "epoch": 2.409873188405797, "grad_norm": 0.36400286865609166, "learning_rate": 4.940912644195314e-05, "loss": 0.8642, "step": 5321 }, { "epoch": 2.4103260869565215, "grad_norm": 0.277946139607492, "learning_rate": 4.939683353976125e-05, "loss": 0.8766, "step": 5322 }, { "epoch": 2.4107789855072466, "grad_norm": 0.30060235971548854, "learning_rate": 4.9384539698153023e-05, "loss": 0.845, "step": 5323 }, { "epoch": 2.411231884057971, "grad_norm": 0.2743611436849062, "learning_rate": 4.937224491835748e-05, "loss": 0.8511, "step": 5324 }, { "epoch": 2.411684782608696, "grad_norm": 0.3935961802146981, "learning_rate": 4.9359949201603764e-05, "loss": 0.8575, "step": 5325 }, { "epoch": 2.4121376811594204, "grad_norm": 0.4192928037156446, "learning_rate": 4.93476525491211e-05, "loss": 0.8591, "step": 5326 }, { "epoch": 2.412590579710145, "grad_norm": 0.5107858449394309, "learning_rate": 4.93353549621388e-05, "loss": 0.8832, "step": 5327 }, { "epoch": 2.4130434782608696, "grad_norm": 0.6269499222198804, "learning_rate": 4.932305644188627e-05, "loss": 0.8794, "step": 5328 }, { "epoch": 2.4134963768115942, "grad_norm": 0.5998705573604248, "learning_rate": 4.9310756989593006e-05, "loss": 0.8843, "step": 5329 }, { "epoch": 2.413949275362319, "grad_norm": 0.5545394179962835, "learning_rate": 4.9298456606488616e-05, "loss": 0.8706, "step": 5330 }, { "epoch": 2.4144021739130435, "grad_norm": 0.5565764177178146, "learning_rate": 4.9286155293802796e-05, "loss": 0.8672, "step": 5331 }, { "epoch": 2.414855072463768, "grad_norm": 0.5010437687887355, "learning_rate": 4.92738530527653e-05, "loss": 0.8648, "step": 5332 }, { "epoch": 2.4153079710144927, "grad_norm": 0.40769591320981774, "learning_rate": 4.9261549884606024e-05, "loss": 0.8809, "step": 5333 }, { "epoch": 2.4157608695652173, "grad_norm": 0.2950212938678329, "learning_rate": 4.9249245790554924e-05, "loss": 0.8957, "step": 5334 }, { "epoch": 2.416213768115942, "grad_norm": 0.24125603518569425, "learning_rate": 4.9236940771842075e-05, "loss": 0.8751, "step": 5335 }, { "epoch": 2.4166666666666665, "grad_norm": 0.33480130476776826, "learning_rate": 4.922463482969761e-05, "loss": 0.8491, "step": 5336 }, { "epoch": 2.417119565217391, "grad_norm": 0.2924126617956189, "learning_rate": 4.921232796535179e-05, "loss": 0.8891, "step": 5337 }, { "epoch": 2.417572463768116, "grad_norm": 0.22351167763022364, "learning_rate": 4.920002018003495e-05, "loss": 0.8707, "step": 5338 }, { "epoch": 2.418025362318841, "grad_norm": 0.3144786805162463, "learning_rate": 4.9187711474977506e-05, "loss": 0.8861, "step": 5339 }, { "epoch": 2.4184782608695654, "grad_norm": 0.31617049039264244, "learning_rate": 4.9175401851409986e-05, "loss": 0.8898, "step": 5340 }, { "epoch": 2.41893115942029, "grad_norm": 0.34931554608163856, "learning_rate": 4.9163091310563e-05, "loss": 0.8901, "step": 5341 }, { "epoch": 2.4193840579710146, "grad_norm": 0.39982123111611273, "learning_rate": 4.915077985366727e-05, "loss": 0.8668, "step": 5342 }, { "epoch": 2.4198369565217392, "grad_norm": 0.4119145819313707, "learning_rate": 4.913846748195356e-05, "loss": 0.8615, "step": 5343 }, { "epoch": 2.420289855072464, "grad_norm": 0.37794279765970046, "learning_rate": 4.91261541966528e-05, "loss": 0.8769, "step": 5344 }, { "epoch": 2.4207427536231885, "grad_norm": 0.24570307831228186, "learning_rate": 4.911383999899592e-05, "loss": 0.8622, "step": 5345 }, { "epoch": 2.421195652173913, "grad_norm": 0.2980383506398808, "learning_rate": 4.910152489021402e-05, "loss": 0.8744, "step": 5346 }, { "epoch": 2.4216485507246377, "grad_norm": 0.3759007478357938, "learning_rate": 4.908920887153825e-05, "loss": 0.8661, "step": 5347 }, { "epoch": 2.4221014492753623, "grad_norm": 0.3219943874074149, "learning_rate": 4.907689194419988e-05, "loss": 0.8767, "step": 5348 }, { "epoch": 2.422554347826087, "grad_norm": 0.33530660796470196, "learning_rate": 4.906457410943022e-05, "loss": 0.8803, "step": 5349 }, { "epoch": 2.4230072463768115, "grad_norm": 0.3790116963451799, "learning_rate": 4.9052255368460736e-05, "loss": 0.8698, "step": 5350 }, { "epoch": 2.423460144927536, "grad_norm": 0.3229956002912773, "learning_rate": 4.9039935722522926e-05, "loss": 0.8791, "step": 5351 }, { "epoch": 2.4239130434782608, "grad_norm": 0.29706768928981947, "learning_rate": 4.902761517284843e-05, "loss": 0.8744, "step": 5352 }, { "epoch": 2.4243659420289854, "grad_norm": 0.3770184873865813, "learning_rate": 4.901529372066893e-05, "loss": 0.8651, "step": 5353 }, { "epoch": 2.42481884057971, "grad_norm": 0.3044027453036504, "learning_rate": 4.900297136721624e-05, "loss": 0.8742, "step": 5354 }, { "epoch": 2.4252717391304346, "grad_norm": 0.2416919207997444, "learning_rate": 4.8990648113722226e-05, "loss": 0.8568, "step": 5355 }, { "epoch": 2.425724637681159, "grad_norm": 0.25773638019317635, "learning_rate": 4.8978323961418887e-05, "loss": 0.8603, "step": 5356 }, { "epoch": 2.426177536231884, "grad_norm": 0.24283603139367116, "learning_rate": 4.896599891153826e-05, "loss": 0.8684, "step": 5357 }, { "epoch": 2.426630434782609, "grad_norm": 0.24383188346537185, "learning_rate": 4.8953672965312515e-05, "loss": 0.8817, "step": 5358 }, { "epoch": 2.4270833333333335, "grad_norm": 0.25775482587121434, "learning_rate": 4.89413461239739e-05, "loss": 0.8907, "step": 5359 }, { "epoch": 2.427536231884058, "grad_norm": 0.3104354124555436, "learning_rate": 4.892901838875476e-05, "loss": 0.869, "step": 5360 }, { "epoch": 2.4279891304347827, "grad_norm": 0.2790659746070461, "learning_rate": 4.891668976088747e-05, "loss": 0.8665, "step": 5361 }, { "epoch": 2.4284420289855073, "grad_norm": 0.20718852121564685, "learning_rate": 4.890436024160459e-05, "loss": 0.8864, "step": 5362 }, { "epoch": 2.428894927536232, "grad_norm": 0.25409509528129304, "learning_rate": 4.8892029832138704e-05, "loss": 0.879, "step": 5363 }, { "epoch": 2.4293478260869565, "grad_norm": 0.29134898919672186, "learning_rate": 4.8879698533722514e-05, "loss": 0.8625, "step": 5364 }, { "epoch": 2.429800724637681, "grad_norm": 0.22460883926366837, "learning_rate": 4.88673663475888e-05, "loss": 0.8918, "step": 5365 }, { "epoch": 2.4302536231884058, "grad_norm": 0.2664472892039953, "learning_rate": 4.88550332749704e-05, "loss": 0.8765, "step": 5366 }, { "epoch": 2.4307065217391304, "grad_norm": 0.32724928204236803, "learning_rate": 4.884269931710029e-05, "loss": 0.8518, "step": 5367 }, { "epoch": 2.431159420289855, "grad_norm": 0.27731809815276004, "learning_rate": 4.883036447521152e-05, "loss": 0.87, "step": 5368 }, { "epoch": 2.4316123188405796, "grad_norm": 0.26630908483779603, "learning_rate": 4.8818028750537235e-05, "loss": 0.8812, "step": 5369 }, { "epoch": 2.432065217391304, "grad_norm": 0.279657378809937, "learning_rate": 4.880569214431063e-05, "loss": 0.8618, "step": 5370 }, { "epoch": 2.432518115942029, "grad_norm": 0.28592107752715396, "learning_rate": 4.879335465776502e-05, "loss": 0.8714, "step": 5371 }, { "epoch": 2.4329710144927534, "grad_norm": 0.2476639159713202, "learning_rate": 4.878101629213382e-05, "loss": 0.8713, "step": 5372 }, { "epoch": 2.4334239130434785, "grad_norm": 0.2770326181497434, "learning_rate": 4.876867704865052e-05, "loss": 0.8695, "step": 5373 }, { "epoch": 2.433876811594203, "grad_norm": 0.2851926117343674, "learning_rate": 4.875633692854866e-05, "loss": 0.8888, "step": 5374 }, { "epoch": 2.4343297101449277, "grad_norm": 0.3766150119706782, "learning_rate": 4.874399593306192e-05, "loss": 0.8877, "step": 5375 }, { "epoch": 2.4347826086956523, "grad_norm": 0.35072430432719315, "learning_rate": 4.8731654063424066e-05, "loss": 0.857, "step": 5376 }, { "epoch": 2.435235507246377, "grad_norm": 0.380575093545192, "learning_rate": 4.871931132086892e-05, "loss": 0.8912, "step": 5377 }, { "epoch": 2.4356884057971016, "grad_norm": 0.6282788993372753, "learning_rate": 4.8706967706630395e-05, "loss": 0.8777, "step": 5378 }, { "epoch": 2.436141304347826, "grad_norm": 0.3547253761870583, "learning_rate": 4.869462322194251e-05, "loss": 0.8886, "step": 5379 }, { "epoch": 2.4365942028985508, "grad_norm": 0.2823576447369973, "learning_rate": 4.868227786803936e-05, "loss": 0.8779, "step": 5380 }, { "epoch": 2.4370471014492754, "grad_norm": 0.28891857749545913, "learning_rate": 4.866993164615515e-05, "loss": 0.8623, "step": 5381 }, { "epoch": 2.4375, "grad_norm": 0.27439240060126485, "learning_rate": 4.8657584557524116e-05, "loss": 0.8649, "step": 5382 }, { "epoch": 2.4379528985507246, "grad_norm": 0.2160893458559295, "learning_rate": 4.8645236603380645e-05, "loss": 0.8578, "step": 5383 }, { "epoch": 2.4384057971014492, "grad_norm": 0.21787593912503386, "learning_rate": 4.8632887784959165e-05, "loss": 0.888, "step": 5384 }, { "epoch": 2.438858695652174, "grad_norm": 0.23740416768435793, "learning_rate": 4.862053810349421e-05, "loss": 0.8664, "step": 5385 }, { "epoch": 2.4393115942028984, "grad_norm": 0.24963636669756684, "learning_rate": 4.860818756022042e-05, "loss": 0.8501, "step": 5386 }, { "epoch": 2.439764492753623, "grad_norm": 0.2990634851659448, "learning_rate": 4.8595836156372465e-05, "loss": 0.8798, "step": 5387 }, { "epoch": 2.4402173913043477, "grad_norm": 0.40448668084327855, "learning_rate": 4.858348389318515e-05, "loss": 0.884, "step": 5388 }, { "epoch": 2.4406702898550723, "grad_norm": 0.5016187450712583, "learning_rate": 4.8571130771893354e-05, "loss": 0.886, "step": 5389 }, { "epoch": 2.441123188405797, "grad_norm": 0.47541930198775906, "learning_rate": 4.855877679373202e-05, "loss": 0.871, "step": 5390 }, { "epoch": 2.4415760869565215, "grad_norm": 0.40894554127828114, "learning_rate": 4.8546421959936216e-05, "loss": 0.8606, "step": 5391 }, { "epoch": 2.4420289855072466, "grad_norm": 0.3864297583334922, "learning_rate": 4.8534066271741065e-05, "loss": 0.8633, "step": 5392 }, { "epoch": 2.442481884057971, "grad_norm": 0.3493663672240203, "learning_rate": 4.8521709730381783e-05, "loss": 0.8639, "step": 5393 }, { "epoch": 2.442934782608696, "grad_norm": 0.2588239237475014, "learning_rate": 4.850935233709368e-05, "loss": 0.8558, "step": 5394 }, { "epoch": 2.4433876811594204, "grad_norm": 0.22952191025046892, "learning_rate": 4.849699409311214e-05, "loss": 0.87, "step": 5395 }, { "epoch": 2.443840579710145, "grad_norm": 0.26010269734255237, "learning_rate": 4.848463499967262e-05, "loss": 0.8478, "step": 5396 }, { "epoch": 2.4442934782608696, "grad_norm": 0.17335364309389553, "learning_rate": 4.847227505801071e-05, "loss": 0.8769, "step": 5397 }, { "epoch": 2.4447463768115942, "grad_norm": 0.2479071862194796, "learning_rate": 4.845991426936203e-05, "loss": 0.8776, "step": 5398 }, { "epoch": 2.445199275362319, "grad_norm": 0.2733237815692924, "learning_rate": 4.844755263496233e-05, "loss": 0.8825, "step": 5399 }, { "epoch": 2.4456521739130435, "grad_norm": 0.18784882470978254, "learning_rate": 4.843519015604739e-05, "loss": 0.8844, "step": 5400 }, { "epoch": 2.446105072463768, "grad_norm": 0.21242161172695007, "learning_rate": 4.8422826833853126e-05, "loss": 0.883, "step": 5401 }, { "epoch": 2.4465579710144927, "grad_norm": 0.26671852715993166, "learning_rate": 4.8410462669615516e-05, "loss": 0.8982, "step": 5402 }, { "epoch": 2.4470108695652173, "grad_norm": 0.24363575143598937, "learning_rate": 4.839809766457063e-05, "loss": 0.8884, "step": 5403 }, { "epoch": 2.447463768115942, "grad_norm": 0.23571684057896689, "learning_rate": 4.838573181995461e-05, "loss": 0.8666, "step": 5404 }, { "epoch": 2.4479166666666665, "grad_norm": 0.24404198535812588, "learning_rate": 4.837336513700369e-05, "loss": 0.8599, "step": 5405 }, { "epoch": 2.448369565217391, "grad_norm": 0.3144859839505691, "learning_rate": 4.836099761695419e-05, "loss": 0.8911, "step": 5406 }, { "epoch": 2.448822463768116, "grad_norm": 0.3753292388009527, "learning_rate": 4.8348629261042516e-05, "loss": 0.8598, "step": 5407 }, { "epoch": 2.449275362318841, "grad_norm": 0.28628379161963313, "learning_rate": 4.833626007050513e-05, "loss": 0.8793, "step": 5408 }, { "epoch": 2.4497282608695654, "grad_norm": 0.2294850164946791, "learning_rate": 4.832389004657862e-05, "loss": 0.8612, "step": 5409 }, { "epoch": 2.45018115942029, "grad_norm": 0.3490161460082433, "learning_rate": 4.831151919049964e-05, "loss": 0.8739, "step": 5410 }, { "epoch": 2.4506340579710146, "grad_norm": 0.27706178712721147, "learning_rate": 4.82991475035049e-05, "loss": 0.8496, "step": 5411 }, { "epoch": 2.4510869565217392, "grad_norm": 0.20394418835281025, "learning_rate": 4.828677498683125e-05, "loss": 0.8694, "step": 5412 }, { "epoch": 2.451539855072464, "grad_norm": 0.28326432162763165, "learning_rate": 4.827440164171557e-05, "loss": 0.8757, "step": 5413 }, { "epoch": 2.4519927536231885, "grad_norm": 0.2985378928384454, "learning_rate": 4.826202746939484e-05, "loss": 0.8724, "step": 5414 }, { "epoch": 2.452445652173913, "grad_norm": 0.26850313375424156, "learning_rate": 4.824965247110614e-05, "loss": 0.8797, "step": 5415 }, { "epoch": 2.4528985507246377, "grad_norm": 0.32513510563493403, "learning_rate": 4.82372766480866e-05, "loss": 0.8907, "step": 5416 }, { "epoch": 2.4533514492753623, "grad_norm": 0.2953912303027937, "learning_rate": 4.822490000157347e-05, "loss": 0.8759, "step": 5417 }, { "epoch": 2.453804347826087, "grad_norm": 0.23632654179747536, "learning_rate": 4.8212522532804046e-05, "loss": 0.866, "step": 5418 }, { "epoch": 2.4542572463768115, "grad_norm": 0.18912427969648468, "learning_rate": 4.8200144243015724e-05, "loss": 0.8706, "step": 5419 }, { "epoch": 2.454710144927536, "grad_norm": 0.19165167386858334, "learning_rate": 4.8187765133446006e-05, "loss": 0.8933, "step": 5420 }, { "epoch": 2.4551630434782608, "grad_norm": 0.19431740538831224, "learning_rate": 4.817538520533242e-05, "loss": 0.8615, "step": 5421 }, { "epoch": 2.4556159420289854, "grad_norm": 0.22615987009678304, "learning_rate": 4.816300445991262e-05, "loss": 0.8763, "step": 5422 }, { "epoch": 2.45606884057971, "grad_norm": 0.19297833226426284, "learning_rate": 4.815062289842433e-05, "loss": 0.8513, "step": 5423 }, { "epoch": 2.4565217391304346, "grad_norm": 0.23501019445666496, "learning_rate": 4.8138240522105365e-05, "loss": 0.8763, "step": 5424 }, { "epoch": 2.456974637681159, "grad_norm": 0.2629312973220744, "learning_rate": 4.812585733219359e-05, "loss": 0.883, "step": 5425 }, { "epoch": 2.457427536231884, "grad_norm": 0.2412863577610878, "learning_rate": 4.8113473329926975e-05, "loss": 0.8734, "step": 5426 }, { "epoch": 2.457880434782609, "grad_norm": 0.35718618045683936, "learning_rate": 4.8101088516543574e-05, "loss": 0.8676, "step": 5427 }, { "epoch": 2.4583333333333335, "grad_norm": 0.37184623047105414, "learning_rate": 4.808870289328153e-05, "loss": 0.8932, "step": 5428 }, { "epoch": 2.458786231884058, "grad_norm": 0.32707156381852837, "learning_rate": 4.807631646137903e-05, "loss": 0.8511, "step": 5429 }, { "epoch": 2.4592391304347827, "grad_norm": 0.33327793309161524, "learning_rate": 4.806392922207438e-05, "loss": 0.8758, "step": 5430 }, { "epoch": 2.4596920289855073, "grad_norm": 0.29220096079713104, "learning_rate": 4.8051541176605934e-05, "loss": 0.8839, "step": 5431 }, { "epoch": 2.460144927536232, "grad_norm": 0.24561522139422226, "learning_rate": 4.803915232621217e-05, "loss": 0.8694, "step": 5432 }, { "epoch": 2.4605978260869565, "grad_norm": 0.23524104938588083, "learning_rate": 4.802676267213161e-05, "loss": 0.8779, "step": 5433 }, { "epoch": 2.461050724637681, "grad_norm": 0.263415077742197, "learning_rate": 4.801437221560286e-05, "loss": 0.9079, "step": 5434 }, { "epoch": 2.4615036231884058, "grad_norm": 0.3100190074206922, "learning_rate": 4.8001980957864624e-05, "loss": 0.875, "step": 5435 }, { "epoch": 2.4619565217391304, "grad_norm": 0.33655182212795365, "learning_rate": 4.7989588900155675e-05, "loss": 0.8574, "step": 5436 }, { "epoch": 2.462409420289855, "grad_norm": 0.43769131695258995, "learning_rate": 4.7977196043714856e-05, "loss": 0.8729, "step": 5437 }, { "epoch": 2.4628623188405796, "grad_norm": 0.48144993460063756, "learning_rate": 4.7964802389781114e-05, "loss": 0.861, "step": 5438 }, { "epoch": 2.463315217391304, "grad_norm": 0.5044156726079991, "learning_rate": 4.7952407939593446e-05, "loss": 0.8715, "step": 5439 }, { "epoch": 2.463768115942029, "grad_norm": 0.45000643265870777, "learning_rate": 4.794001269439096e-05, "loss": 0.8771, "step": 5440 }, { "epoch": 2.4642210144927534, "grad_norm": 0.38843996041330375, "learning_rate": 4.792761665541283e-05, "loss": 0.88, "step": 5441 }, { "epoch": 2.4646739130434785, "grad_norm": 0.2580746586081194, "learning_rate": 4.791521982389829e-05, "loss": 0.8718, "step": 5442 }, { "epoch": 2.465126811594203, "grad_norm": 0.24974981032765836, "learning_rate": 4.790282220108669e-05, "loss": 0.8956, "step": 5443 }, { "epoch": 2.4655797101449277, "grad_norm": 0.3951245907440686, "learning_rate": 4.789042378821742e-05, "loss": 0.8618, "step": 5444 }, { "epoch": 2.4660326086956523, "grad_norm": 0.5344614990880141, "learning_rate": 4.7878024586529994e-05, "loss": 0.8536, "step": 5445 }, { "epoch": 2.466485507246377, "grad_norm": 0.5441235456787211, "learning_rate": 4.786562459726396e-05, "loss": 0.8764, "step": 5446 }, { "epoch": 2.4669384057971016, "grad_norm": 0.46726593669374794, "learning_rate": 4.7853223821658965e-05, "loss": 0.8713, "step": 5447 }, { "epoch": 2.467391304347826, "grad_norm": 0.4481151150824303, "learning_rate": 4.784082226095474e-05, "loss": 0.876, "step": 5448 }, { "epoch": 2.4678442028985508, "grad_norm": 0.2986710067324889, "learning_rate": 4.7828419916391105e-05, "loss": 0.8789, "step": 5449 }, { "epoch": 2.4682971014492754, "grad_norm": 0.234969118470105, "learning_rate": 4.7816016789207913e-05, "loss": 0.8585, "step": 5450 }, { "epoch": 2.46875, "grad_norm": 0.23243783302297463, "learning_rate": 4.780361288064514e-05, "loss": 0.8968, "step": 5451 }, { "epoch": 2.4692028985507246, "grad_norm": 0.31042668622321673, "learning_rate": 4.7791208191942815e-05, "loss": 0.8753, "step": 5452 }, { "epoch": 2.4696557971014492, "grad_norm": 0.271971292041256, "learning_rate": 4.777880272434107e-05, "loss": 0.8692, "step": 5453 }, { "epoch": 2.470108695652174, "grad_norm": 0.3006524974036897, "learning_rate": 4.7766396479080086e-05, "loss": 0.8758, "step": 5454 }, { "epoch": 2.4705615942028984, "grad_norm": 0.3675069648835495, "learning_rate": 4.7753989457400137e-05, "loss": 0.8819, "step": 5455 }, { "epoch": 2.471014492753623, "grad_norm": 0.3591129843079386, "learning_rate": 4.774158166054157e-05, "loss": 0.8903, "step": 5456 }, { "epoch": 2.4714673913043477, "grad_norm": 0.22933403723395523, "learning_rate": 4.772917308974483e-05, "loss": 0.872, "step": 5457 }, { "epoch": 2.4719202898550723, "grad_norm": 0.22973032853874978, "learning_rate": 4.771676374625041e-05, "loss": 0.8727, "step": 5458 }, { "epoch": 2.472373188405797, "grad_norm": 0.2737456147332083, "learning_rate": 4.7704353631298875e-05, "loss": 0.8811, "step": 5459 }, { "epoch": 2.4728260869565215, "grad_norm": 0.2519507647462811, "learning_rate": 4.7691942746130904e-05, "loss": 0.8661, "step": 5460 }, { "epoch": 2.4732789855072466, "grad_norm": 0.2414860017472934, "learning_rate": 4.767953109198723e-05, "loss": 0.8758, "step": 5461 }, { "epoch": 2.473731884057971, "grad_norm": 0.3048027290534194, "learning_rate": 4.766711867010867e-05, "loss": 0.8873, "step": 5462 }, { "epoch": 2.474184782608696, "grad_norm": 0.3056661499580036, "learning_rate": 4.7654705481736094e-05, "loss": 0.855, "step": 5463 }, { "epoch": 2.4746376811594204, "grad_norm": 0.2616153603622335, "learning_rate": 4.764229152811049e-05, "loss": 0.884, "step": 5464 }, { "epoch": 2.475090579710145, "grad_norm": 0.272944864891121, "learning_rate": 4.762987681047289e-05, "loss": 0.8613, "step": 5465 }, { "epoch": 2.4755434782608696, "grad_norm": 0.2770794233969573, "learning_rate": 4.761746133006442e-05, "loss": 0.8786, "step": 5466 }, { "epoch": 2.4759963768115942, "grad_norm": 0.23695458413140083, "learning_rate": 4.760504508812626e-05, "loss": 0.8584, "step": 5467 }, { "epoch": 2.476449275362319, "grad_norm": 0.1992010955994558, "learning_rate": 4.759262808589969e-05, "loss": 0.8772, "step": 5468 }, { "epoch": 2.4769021739130435, "grad_norm": 0.3232989621956689, "learning_rate": 4.7580210324626064e-05, "loss": 0.8816, "step": 5469 }, { "epoch": 2.477355072463768, "grad_norm": 0.30999205364589666, "learning_rate": 4.756779180554681e-05, "loss": 0.8795, "step": 5470 }, { "epoch": 2.4778079710144927, "grad_norm": 0.20218761939927837, "learning_rate": 4.7555372529903404e-05, "loss": 0.8698, "step": 5471 }, { "epoch": 2.4782608695652173, "grad_norm": 0.25483242992056976, "learning_rate": 4.7542952498937437e-05, "loss": 0.8641, "step": 5472 }, { "epoch": 2.478713768115942, "grad_norm": 0.2961904675136123, "learning_rate": 4.753053171389055e-05, "loss": 0.8573, "step": 5473 }, { "epoch": 2.4791666666666665, "grad_norm": 0.23961416159261065, "learning_rate": 4.751811017600448e-05, "loss": 0.8715, "step": 5474 }, { "epoch": 2.479619565217391, "grad_norm": 0.2539364109483992, "learning_rate": 4.7505687886521026e-05, "loss": 0.8703, "step": 5475 }, { "epoch": 2.480072463768116, "grad_norm": 0.34796005036679284, "learning_rate": 4.7493264846682055e-05, "loss": 0.8829, "step": 5476 }, { "epoch": 2.480525362318841, "grad_norm": 0.2292789799260602, "learning_rate": 4.7480841057729524e-05, "loss": 0.8955, "step": 5477 }, { "epoch": 2.4809782608695654, "grad_norm": 0.2557666259572596, "learning_rate": 4.7468416520905454e-05, "loss": 0.8929, "step": 5478 }, { "epoch": 2.48143115942029, "grad_norm": 0.35709026592162557, "learning_rate": 4.745599123745196e-05, "loss": 0.866, "step": 5479 }, { "epoch": 2.4818840579710146, "grad_norm": 0.26467008020227445, "learning_rate": 4.7443565208611196e-05, "loss": 0.8753, "step": 5480 }, { "epoch": 2.4823369565217392, "grad_norm": 0.20959995461455216, "learning_rate": 4.743113843562543e-05, "loss": 0.8826, "step": 5481 }, { "epoch": 2.482789855072464, "grad_norm": 0.31040471722666296, "learning_rate": 4.741871091973697e-05, "loss": 0.8743, "step": 5482 }, { "epoch": 2.4832427536231885, "grad_norm": 0.2219299517309192, "learning_rate": 4.740628266218823e-05, "loss": 0.8524, "step": 5483 }, { "epoch": 2.483695652173913, "grad_norm": 0.19904878573924983, "learning_rate": 4.739385366422168e-05, "loss": 0.8738, "step": 5484 }, { "epoch": 2.4841485507246377, "grad_norm": 0.2564755013362236, "learning_rate": 4.7381423927079854e-05, "loss": 0.8731, "step": 5485 }, { "epoch": 2.4846014492753623, "grad_norm": 0.2488255044395118, "learning_rate": 4.736899345200539e-05, "loss": 0.8801, "step": 5486 }, { "epoch": 2.485054347826087, "grad_norm": 0.23798121708384143, "learning_rate": 4.735656224024097e-05, "loss": 0.887, "step": 5487 }, { "epoch": 2.4855072463768115, "grad_norm": 0.19872978390203977, "learning_rate": 4.734413029302936e-05, "loss": 0.8716, "step": 5488 }, { "epoch": 2.485960144927536, "grad_norm": 0.23713607391077554, "learning_rate": 4.733169761161342e-05, "loss": 0.8676, "step": 5489 }, { "epoch": 2.4864130434782608, "grad_norm": 0.22589760696270764, "learning_rate": 4.7319264197236034e-05, "loss": 0.8931, "step": 5490 }, { "epoch": 2.4868659420289854, "grad_norm": 0.17620448513290587, "learning_rate": 4.7306830051140225e-05, "loss": 0.8744, "step": 5491 }, { "epoch": 2.48731884057971, "grad_norm": 0.18575307321860043, "learning_rate": 4.729439517456903e-05, "loss": 0.8762, "step": 5492 }, { "epoch": 2.4877717391304346, "grad_norm": 0.19748929801750154, "learning_rate": 4.728195956876558e-05, "loss": 0.8712, "step": 5493 }, { "epoch": 2.488224637681159, "grad_norm": 0.20592212425477396, "learning_rate": 4.726952323497311e-05, "loss": 0.8686, "step": 5494 }, { "epoch": 2.488677536231884, "grad_norm": 0.2134870813602346, "learning_rate": 4.725708617443487e-05, "loss": 0.8834, "step": 5495 }, { "epoch": 2.489130434782609, "grad_norm": 0.2730222954775631, "learning_rate": 4.7244648388394235e-05, "loss": 0.8806, "step": 5496 }, { "epoch": 2.4895833333333335, "grad_norm": 0.3449701273148723, "learning_rate": 4.723220987809462e-05, "loss": 0.8774, "step": 5497 }, { "epoch": 2.490036231884058, "grad_norm": 0.3711569040777031, "learning_rate": 4.7219770644779514e-05, "loss": 0.8852, "step": 5498 }, { "epoch": 2.4904891304347827, "grad_norm": 0.35723759652775705, "learning_rate": 4.720733068969249e-05, "loss": 0.8722, "step": 5499 }, { "epoch": 2.4909420289855073, "grad_norm": 0.26352463648268604, "learning_rate": 4.719489001407722e-05, "loss": 0.871, "step": 5500 }, { "epoch": 2.491394927536232, "grad_norm": 0.235010906071634, "learning_rate": 4.718244861917738e-05, "loss": 0.8796, "step": 5501 }, { "epoch": 2.4918478260869565, "grad_norm": 0.23266129858304152, "learning_rate": 4.717000650623677e-05, "loss": 0.8456, "step": 5502 }, { "epoch": 2.492300724637681, "grad_norm": 0.2814044639829595, "learning_rate": 4.715756367649924e-05, "loss": 0.8914, "step": 5503 }, { "epoch": 2.4927536231884058, "grad_norm": 0.4009970508967857, "learning_rate": 4.714512013120874e-05, "loss": 0.8893, "step": 5504 }, { "epoch": 2.4932065217391304, "grad_norm": 0.4262908885537154, "learning_rate": 4.713267587160926e-05, "loss": 0.8629, "step": 5505 }, { "epoch": 2.493659420289855, "grad_norm": 0.43552829275141836, "learning_rate": 4.712023089894486e-05, "loss": 0.8759, "step": 5506 }, { "epoch": 2.4941123188405796, "grad_norm": 0.43919822414458615, "learning_rate": 4.710778521445969e-05, "loss": 0.8791, "step": 5507 }, { "epoch": 2.494565217391304, "grad_norm": 0.32592167445258924, "learning_rate": 4.709533881939798e-05, "loss": 0.8764, "step": 5508 }, { "epoch": 2.495018115942029, "grad_norm": 0.3324476587120806, "learning_rate": 4.708289171500401e-05, "loss": 0.8651, "step": 5509 }, { "epoch": 2.4954710144927534, "grad_norm": 0.36586448477554373, "learning_rate": 4.707044390252212e-05, "loss": 0.8882, "step": 5510 }, { "epoch": 2.4959239130434785, "grad_norm": 0.2260019059682862, "learning_rate": 4.7057995383196745e-05, "loss": 0.8656, "step": 5511 }, { "epoch": 2.496376811594203, "grad_norm": 0.3292782561438116, "learning_rate": 4.7045546158272404e-05, "loss": 0.8893, "step": 5512 }, { "epoch": 2.4968297101449277, "grad_norm": 0.47908987053173213, "learning_rate": 4.7033096228993634e-05, "loss": 0.8822, "step": 5513 }, { "epoch": 2.4972826086956523, "grad_norm": 0.477869791500093, "learning_rate": 4.702064559660511e-05, "loss": 0.8791, "step": 5514 }, { "epoch": 2.497735507246377, "grad_norm": 0.4630462575940673, "learning_rate": 4.70081942623515e-05, "loss": 0.8683, "step": 5515 }, { "epoch": 2.4981884057971016, "grad_norm": 0.4444197031933878, "learning_rate": 4.6995742227477606e-05, "loss": 0.8755, "step": 5516 }, { "epoch": 2.498641304347826, "grad_norm": 0.35422681046058435, "learning_rate": 4.698328949322828e-05, "loss": 0.8765, "step": 5517 }, { "epoch": 2.4990942028985508, "grad_norm": 0.2888450418241275, "learning_rate": 4.6970836060848434e-05, "loss": 0.8762, "step": 5518 }, { "epoch": 2.4995471014492754, "grad_norm": 0.28256323822111146, "learning_rate": 4.695838193158306e-05, "loss": 0.8686, "step": 5519 }, { "epoch": 2.5, "grad_norm": 0.3131575008464696, "learning_rate": 4.694592710667723e-05, "loss": 0.872, "step": 5520 }, { "epoch": 2.5004528985507246, "grad_norm": 0.31616328251950815, "learning_rate": 4.693347158737604e-05, "loss": 0.8785, "step": 5521 }, { "epoch": 2.5009057971014492, "grad_norm": 0.3375020827521136, "learning_rate": 4.692101537492472e-05, "loss": 0.8624, "step": 5522 }, { "epoch": 2.501358695652174, "grad_norm": 0.3958787288653256, "learning_rate": 4.690855847056853e-05, "loss": 0.8827, "step": 5523 }, { "epoch": 2.5018115942028984, "grad_norm": 0.4588925635347199, "learning_rate": 4.689610087555279e-05, "loss": 0.891, "step": 5524 }, { "epoch": 2.502264492753623, "grad_norm": 0.4533032894687713, "learning_rate": 4.688364259112293e-05, "loss": 0.8862, "step": 5525 }, { "epoch": 2.5027173913043477, "grad_norm": 0.3993646677251451, "learning_rate": 4.687118361852441e-05, "loss": 0.8983, "step": 5526 }, { "epoch": 2.5031702898550723, "grad_norm": 0.2944348172239361, "learning_rate": 4.685872395900277e-05, "loss": 0.8756, "step": 5527 }, { "epoch": 2.503623188405797, "grad_norm": 0.26026818732265067, "learning_rate": 4.684626361380363e-05, "loss": 0.8672, "step": 5528 }, { "epoch": 2.5040760869565215, "grad_norm": 0.23998186275124658, "learning_rate": 4.683380258417266e-05, "loss": 0.863, "step": 5529 }, { "epoch": 2.504528985507246, "grad_norm": 0.2451018303045621, "learning_rate": 4.6821340871355635e-05, "loss": 0.8716, "step": 5530 }, { "epoch": 2.504981884057971, "grad_norm": 0.198723783978387, "learning_rate": 4.680887847659834e-05, "loss": 0.8415, "step": 5531 }, { "epoch": 2.505434782608696, "grad_norm": 0.23886553876096434, "learning_rate": 4.679641540114667e-05, "loss": 0.8843, "step": 5532 }, { "epoch": 2.5058876811594204, "grad_norm": 0.24167443136007108, "learning_rate": 4.678395164624659e-05, "loss": 0.8705, "step": 5533 }, { "epoch": 2.506340579710145, "grad_norm": 0.1887386578061915, "learning_rate": 4.677148721314412e-05, "loss": 0.8787, "step": 5534 }, { "epoch": 2.5067934782608696, "grad_norm": 0.2011222706853691, "learning_rate": 4.675902210308534e-05, "loss": 0.8695, "step": 5535 }, { "epoch": 2.5072463768115942, "grad_norm": 0.2433382085997767, "learning_rate": 4.6746556317316394e-05, "loss": 0.8739, "step": 5536 }, { "epoch": 2.507699275362319, "grad_norm": 0.2598712139563439, "learning_rate": 4.6734089857083536e-05, "loss": 0.8861, "step": 5537 }, { "epoch": 2.5081521739130435, "grad_norm": 0.31682667100483697, "learning_rate": 4.672162272363305e-05, "loss": 0.8951, "step": 5538 }, { "epoch": 2.508605072463768, "grad_norm": 0.38950590771565075, "learning_rate": 4.670915491821127e-05, "loss": 0.8768, "step": 5539 }, { "epoch": 2.5090579710144927, "grad_norm": 0.36520712977316233, "learning_rate": 4.6696686442064664e-05, "loss": 0.8885, "step": 5540 }, { "epoch": 2.5095108695652173, "grad_norm": 0.4152654989230476, "learning_rate": 4.668421729643968e-05, "loss": 0.8786, "step": 5541 }, { "epoch": 2.509963768115942, "grad_norm": 0.46033922553538914, "learning_rate": 4.667174748258291e-05, "loss": 0.8747, "step": 5542 }, { "epoch": 2.5104166666666665, "grad_norm": 0.3941862566104195, "learning_rate": 4.6659277001740984e-05, "loss": 0.874, "step": 5543 }, { "epoch": 2.5108695652173916, "grad_norm": 0.2614061480714026, "learning_rate": 4.664680585516057e-05, "loss": 0.8558, "step": 5544 }, { "epoch": 2.511322463768116, "grad_norm": 0.27677941168839804, "learning_rate": 4.663433404408844e-05, "loss": 0.8774, "step": 5545 }, { "epoch": 2.511775362318841, "grad_norm": 0.27246017097840985, "learning_rate": 4.662186156977143e-05, "loss": 0.8914, "step": 5546 }, { "epoch": 2.5122282608695654, "grad_norm": 0.31814497658376195, "learning_rate": 4.660938843345642e-05, "loss": 0.8979, "step": 5547 }, { "epoch": 2.51268115942029, "grad_norm": 0.45757949313487134, "learning_rate": 4.659691463639038e-05, "loss": 0.888, "step": 5548 }, { "epoch": 2.5131340579710146, "grad_norm": 0.586860470816259, "learning_rate": 4.658444017982033e-05, "loss": 0.896, "step": 5549 }, { "epoch": 2.5135869565217392, "grad_norm": 0.6635195975845838, "learning_rate": 4.657196506499335e-05, "loss": 0.8575, "step": 5550 }, { "epoch": 2.514039855072464, "grad_norm": 0.5782008702334468, "learning_rate": 4.655948929315662e-05, "loss": 0.8706, "step": 5551 }, { "epoch": 2.5144927536231885, "grad_norm": 0.44934245786550653, "learning_rate": 4.6547012865557354e-05, "loss": 0.8795, "step": 5552 }, { "epoch": 2.514945652173913, "grad_norm": 0.36076318687949577, "learning_rate": 4.653453578344283e-05, "loss": 0.8505, "step": 5553 }, { "epoch": 2.5153985507246377, "grad_norm": 0.2359886397212859, "learning_rate": 4.652205804806041e-05, "loss": 0.8689, "step": 5554 }, { "epoch": 2.5158514492753623, "grad_norm": 0.26424959858714303, "learning_rate": 4.6509579660657525e-05, "loss": 0.87, "step": 5555 }, { "epoch": 2.516304347826087, "grad_norm": 0.46923322670714546, "learning_rate": 4.6497100622481625e-05, "loss": 0.8668, "step": 5556 }, { "epoch": 2.5167572463768115, "grad_norm": 0.5979604316173759, "learning_rate": 4.648462093478029e-05, "loss": 0.8473, "step": 5557 }, { "epoch": 2.517210144927536, "grad_norm": 0.6263167697690032, "learning_rate": 4.647214059880112e-05, "loss": 0.8875, "step": 5558 }, { "epoch": 2.5176630434782608, "grad_norm": 0.5767964787703721, "learning_rate": 4.645965961579181e-05, "loss": 0.8794, "step": 5559 }, { "epoch": 2.5181159420289854, "grad_norm": 0.42511035258170665, "learning_rate": 4.6447177987000075e-05, "loss": 0.8802, "step": 5560 }, { "epoch": 2.51856884057971, "grad_norm": 0.2591165298868241, "learning_rate": 4.6434695713673746e-05, "loss": 0.8767, "step": 5561 }, { "epoch": 2.5190217391304346, "grad_norm": 0.20799273712672764, "learning_rate": 4.64222127970607e-05, "loss": 0.8731, "step": 5562 }, { "epoch": 2.519474637681159, "grad_norm": 0.38144663351574654, "learning_rate": 4.6409729238408844e-05, "loss": 0.8735, "step": 5563 }, { "epoch": 2.519927536231884, "grad_norm": 0.5516850046166141, "learning_rate": 4.639724503896622e-05, "loss": 0.8882, "step": 5564 }, { "epoch": 2.5203804347826084, "grad_norm": 0.697618361624477, "learning_rate": 4.638476019998086e-05, "loss": 0.8723, "step": 5565 }, { "epoch": 2.5208333333333335, "grad_norm": 0.8310101992195531, "learning_rate": 4.637227472270091e-05, "loss": 0.8855, "step": 5566 }, { "epoch": 2.521286231884058, "grad_norm": 0.759209069825428, "learning_rate": 4.6359788608374553e-05, "loss": 0.8875, "step": 5567 }, { "epoch": 2.5217391304347827, "grad_norm": 0.6117051257597683, "learning_rate": 4.634730185825006e-05, "loss": 0.8868, "step": 5568 }, { "epoch": 2.5221920289855073, "grad_norm": 0.4799547217874617, "learning_rate": 4.6334814473575735e-05, "loss": 0.8673, "step": 5569 }, { "epoch": 2.522644927536232, "grad_norm": 0.3614936063621735, "learning_rate": 4.6322326455599974e-05, "loss": 0.8765, "step": 5570 }, { "epoch": 2.5230978260869565, "grad_norm": 0.2250214554865788, "learning_rate": 4.630983780557121e-05, "loss": 0.8605, "step": 5571 }, { "epoch": 2.523550724637681, "grad_norm": 0.21742062141764393, "learning_rate": 4.629734852473797e-05, "loss": 0.8581, "step": 5572 }, { "epoch": 2.5240036231884058, "grad_norm": 0.3362641832490791, "learning_rate": 4.6284858614348815e-05, "loss": 0.8564, "step": 5573 }, { "epoch": 2.5244565217391304, "grad_norm": 0.36847153042108743, "learning_rate": 4.627236807565238e-05, "loss": 0.8668, "step": 5574 }, { "epoch": 2.524909420289855, "grad_norm": 0.38338096795016813, "learning_rate": 4.625987690989738e-05, "loss": 0.8777, "step": 5575 }, { "epoch": 2.5253623188405796, "grad_norm": 0.32418417436077707, "learning_rate": 4.624738511833257e-05, "loss": 0.866, "step": 5576 }, { "epoch": 2.525815217391304, "grad_norm": 0.260728531839039, "learning_rate": 4.623489270220676e-05, "loss": 0.8639, "step": 5577 }, { "epoch": 2.526268115942029, "grad_norm": 0.2553330361596334, "learning_rate": 4.622239966276884e-05, "loss": 0.8673, "step": 5578 }, { "epoch": 2.526721014492754, "grad_norm": 0.23309190324153012, "learning_rate": 4.620990600126777e-05, "loss": 0.8683, "step": 5579 }, { "epoch": 2.5271739130434785, "grad_norm": 0.2404072738326107, "learning_rate": 4.619741171895257e-05, "loss": 0.8737, "step": 5580 }, { "epoch": 2.527626811594203, "grad_norm": 0.3172537009037581, "learning_rate": 4.618491681707228e-05, "loss": 0.8841, "step": 5581 }, { "epoch": 2.5280797101449277, "grad_norm": 0.3005578026341218, "learning_rate": 4.617242129687606e-05, "loss": 0.8725, "step": 5582 }, { "epoch": 2.5285326086956523, "grad_norm": 0.28942623483162344, "learning_rate": 4.6159925159613106e-05, "loss": 0.8686, "step": 5583 }, { "epoch": 2.528985507246377, "grad_norm": 0.31121962901013106, "learning_rate": 4.614742840653267e-05, "loss": 0.872, "step": 5584 }, { "epoch": 2.5294384057971016, "grad_norm": 0.2481127398076993, "learning_rate": 4.613493103888409e-05, "loss": 0.873, "step": 5585 }, { "epoch": 2.529891304347826, "grad_norm": 0.20942523579093675, "learning_rate": 4.612243305791672e-05, "loss": 0.877, "step": 5586 }, { "epoch": 2.5303442028985508, "grad_norm": 0.1798650538109418, "learning_rate": 4.610993446488e-05, "loss": 0.8643, "step": 5587 }, { "epoch": 2.5307971014492754, "grad_norm": 0.24735985114396988, "learning_rate": 4.609743526102347e-05, "loss": 0.8718, "step": 5588 }, { "epoch": 2.53125, "grad_norm": 0.24974931797252434, "learning_rate": 4.608493544759667e-05, "loss": 0.8775, "step": 5589 }, { "epoch": 2.5317028985507246, "grad_norm": 0.16722529016400678, "learning_rate": 4.6072435025849235e-05, "loss": 0.8552, "step": 5590 }, { "epoch": 2.5321557971014492, "grad_norm": 0.2641495110242741, "learning_rate": 4.605993399703084e-05, "loss": 0.8709, "step": 5591 }, { "epoch": 2.532608695652174, "grad_norm": 0.29008734759831684, "learning_rate": 4.604743236239124e-05, "loss": 0.8734, "step": 5592 }, { "epoch": 2.5330615942028984, "grad_norm": 0.28698018128758657, "learning_rate": 4.603493012318025e-05, "loss": 0.8858, "step": 5593 }, { "epoch": 2.533514492753623, "grad_norm": 0.2560463351352233, "learning_rate": 4.6022427280647734e-05, "loss": 0.8807, "step": 5594 }, { "epoch": 2.5339673913043477, "grad_norm": 0.2140258349889512, "learning_rate": 4.600992383604361e-05, "loss": 0.8804, "step": 5595 }, { "epoch": 2.5344202898550723, "grad_norm": 0.22062576669237888, "learning_rate": 4.599741979061788e-05, "loss": 0.8631, "step": 5596 }, { "epoch": 2.534873188405797, "grad_norm": 0.20979662650923778, "learning_rate": 4.59849151456206e-05, "loss": 0.8693, "step": 5597 }, { "epoch": 2.5353260869565215, "grad_norm": 0.1851733426770897, "learning_rate": 4.597240990230185e-05, "loss": 0.8882, "step": 5598 }, { "epoch": 2.535778985507246, "grad_norm": 0.19668895515996637, "learning_rate": 4.595990406191184e-05, "loss": 0.8894, "step": 5599 }, { "epoch": 2.536231884057971, "grad_norm": 0.2135140773305466, "learning_rate": 4.5947397625700756e-05, "loss": 0.8799, "step": 5600 }, { "epoch": 2.536684782608696, "grad_norm": 0.2439767599083818, "learning_rate": 4.593489059491893e-05, "loss": 0.8634, "step": 5601 }, { "epoch": 2.5371376811594204, "grad_norm": 0.2001608642332677, "learning_rate": 4.5922382970816665e-05, "loss": 0.8489, "step": 5602 }, { "epoch": 2.537590579710145, "grad_norm": 0.2498160245541176, "learning_rate": 4.59098747546444e-05, "loss": 0.8857, "step": 5603 }, { "epoch": 2.5380434782608696, "grad_norm": 0.2657837763619205, "learning_rate": 4.5897365947652584e-05, "loss": 0.8614, "step": 5604 }, { "epoch": 2.5384963768115942, "grad_norm": 0.2194883201503141, "learning_rate": 4.5884856551091756e-05, "loss": 0.872, "step": 5605 }, { "epoch": 2.538949275362319, "grad_norm": 0.24675412975084984, "learning_rate": 4.5872346566212486e-05, "loss": 0.8601, "step": 5606 }, { "epoch": 2.5394021739130435, "grad_norm": 0.27548417607406395, "learning_rate": 4.5859835994265434e-05, "loss": 0.8637, "step": 5607 }, { "epoch": 2.539855072463768, "grad_norm": 0.2076809999455994, "learning_rate": 4.584732483650128e-05, "loss": 0.8674, "step": 5608 }, { "epoch": 2.5403079710144927, "grad_norm": 0.20985779093836934, "learning_rate": 4.58348130941708e-05, "loss": 0.87, "step": 5609 }, { "epoch": 2.5407608695652173, "grad_norm": 0.25347475732857727, "learning_rate": 4.582230076852482e-05, "loss": 0.8858, "step": 5610 }, { "epoch": 2.541213768115942, "grad_norm": 0.23082938245247397, "learning_rate": 4.58097878608142e-05, "loss": 0.8743, "step": 5611 }, { "epoch": 2.5416666666666665, "grad_norm": 0.21782692973573425, "learning_rate": 4.579727437228987e-05, "loss": 0.8621, "step": 5612 }, { "epoch": 2.5421195652173916, "grad_norm": 0.2673448309997498, "learning_rate": 4.578476030420284e-05, "loss": 0.8518, "step": 5613 }, { "epoch": 2.542572463768116, "grad_norm": 0.25117152412932314, "learning_rate": 4.5772245657804166e-05, "loss": 0.8734, "step": 5614 }, { "epoch": 2.543025362318841, "grad_norm": 0.27951887871508835, "learning_rate": 4.575973043434494e-05, "loss": 0.8752, "step": 5615 }, { "epoch": 2.5434782608695654, "grad_norm": 0.5141080278943254, "learning_rate": 4.574721463507633e-05, "loss": 0.8757, "step": 5616 }, { "epoch": 2.54393115942029, "grad_norm": 0.5631962009039634, "learning_rate": 4.5734698261249575e-05, "loss": 0.8705, "step": 5617 }, { "epoch": 2.5443840579710146, "grad_norm": 0.585539124914504, "learning_rate": 4.572218131411594e-05, "loss": 0.8594, "step": 5618 }, { "epoch": 2.5448369565217392, "grad_norm": 0.6622463961234542, "learning_rate": 4.570966379492678e-05, "loss": 0.8864, "step": 5619 }, { "epoch": 2.545289855072464, "grad_norm": 0.7307316434535645, "learning_rate": 4.5697145704933477e-05, "loss": 0.8928, "step": 5620 }, { "epoch": 2.5457427536231885, "grad_norm": 0.6249113520039213, "learning_rate": 4.568462704538748e-05, "loss": 0.8592, "step": 5621 }, { "epoch": 2.546195652173913, "grad_norm": 0.4658276808180646, "learning_rate": 4.567210781754033e-05, "loss": 0.8704, "step": 5622 }, { "epoch": 2.5466485507246377, "grad_norm": 0.3735109358158944, "learning_rate": 4.565958802264356e-05, "loss": 0.8765, "step": 5623 }, { "epoch": 2.5471014492753623, "grad_norm": 0.28181626864445675, "learning_rate": 4.56470676619488e-05, "loss": 0.8781, "step": 5624 }, { "epoch": 2.547554347826087, "grad_norm": 0.390779222517657, "learning_rate": 4.563454673670776e-05, "loss": 0.8773, "step": 5625 }, { "epoch": 2.5480072463768115, "grad_norm": 0.585242696234287, "learning_rate": 4.562202524817214e-05, "loss": 0.8988, "step": 5626 }, { "epoch": 2.548460144927536, "grad_norm": 0.6344211154602712, "learning_rate": 4.560950319759375e-05, "loss": 0.85, "step": 5627 }, { "epoch": 2.5489130434782608, "grad_norm": 0.625615835361286, "learning_rate": 4.559698058622443e-05, "loss": 0.8765, "step": 5628 }, { "epoch": 2.5493659420289854, "grad_norm": 0.5904880409836325, "learning_rate": 4.558445741531609e-05, "loss": 0.8711, "step": 5629 }, { "epoch": 2.54981884057971, "grad_norm": 0.3583753193893843, "learning_rate": 4.55719336861207e-05, "loss": 0.8677, "step": 5630 }, { "epoch": 2.5502717391304346, "grad_norm": 0.2032206562923508, "learning_rate": 4.555940939989027e-05, "loss": 0.8471, "step": 5631 }, { "epoch": 2.550724637681159, "grad_norm": 0.26513907280494753, "learning_rate": 4.554688455787688e-05, "loss": 0.88, "step": 5632 }, { "epoch": 2.551177536231884, "grad_norm": 0.3521444573520082, "learning_rate": 4.5534359161332634e-05, "loss": 0.8521, "step": 5633 }, { "epoch": 2.5516304347826084, "grad_norm": 0.3316023179335421, "learning_rate": 4.552183321150974e-05, "loss": 0.869, "step": 5634 }, { "epoch": 2.5520833333333335, "grad_norm": 0.3903784561900844, "learning_rate": 4.550930670966043e-05, "loss": 0.8786, "step": 5635 }, { "epoch": 2.552536231884058, "grad_norm": 0.42563531586204734, "learning_rate": 4.5496779657037e-05, "loss": 0.8872, "step": 5636 }, { "epoch": 2.5529891304347827, "grad_norm": 0.39825681152061576, "learning_rate": 4.54842520548918e-05, "loss": 0.8798, "step": 5637 }, { "epoch": 2.5534420289855073, "grad_norm": 0.4282985357680943, "learning_rate": 4.5471723904477234e-05, "loss": 0.8668, "step": 5638 }, { "epoch": 2.553894927536232, "grad_norm": 0.5151149252807758, "learning_rate": 4.5459195207045745e-05, "loss": 0.8586, "step": 5639 }, { "epoch": 2.5543478260869565, "grad_norm": 0.5762999538273054, "learning_rate": 4.5446665963849874e-05, "loss": 0.873, "step": 5640 }, { "epoch": 2.554800724637681, "grad_norm": 0.5553814351518197, "learning_rate": 4.543413617614217e-05, "loss": 0.8722, "step": 5641 }, { "epoch": 2.5552536231884058, "grad_norm": 0.5357583958967774, "learning_rate": 4.5421605845175254e-05, "loss": 0.8613, "step": 5642 }, { "epoch": 2.5557065217391304, "grad_norm": 0.46702340022619115, "learning_rate": 4.540907497220182e-05, "loss": 0.8602, "step": 5643 }, { "epoch": 2.556159420289855, "grad_norm": 0.28785859892595334, "learning_rate": 4.539654355847459e-05, "loss": 0.8656, "step": 5644 }, { "epoch": 2.5566123188405796, "grad_norm": 0.2313400735519284, "learning_rate": 4.5384011605246343e-05, "loss": 0.8537, "step": 5645 }, { "epoch": 2.557065217391304, "grad_norm": 0.3269717064043279, "learning_rate": 4.537147911376993e-05, "loss": 0.8738, "step": 5646 }, { "epoch": 2.557518115942029, "grad_norm": 0.40043409247667006, "learning_rate": 4.535894608529823e-05, "loss": 0.8697, "step": 5647 }, { "epoch": 2.557971014492754, "grad_norm": 0.427547349486506, "learning_rate": 4.53464125210842e-05, "loss": 0.8674, "step": 5648 }, { "epoch": 2.5584239130434785, "grad_norm": 0.4982620011558666, "learning_rate": 4.533387842238084e-05, "loss": 0.8623, "step": 5649 }, { "epoch": 2.558876811594203, "grad_norm": 0.5709344047721214, "learning_rate": 4.53213437904412e-05, "loss": 0.8543, "step": 5650 }, { "epoch": 2.5593297101449277, "grad_norm": 0.4694655277683832, "learning_rate": 4.53088086265184e-05, "loss": 0.8816, "step": 5651 }, { "epoch": 2.5597826086956523, "grad_norm": 0.320388047319885, "learning_rate": 4.529627293186558e-05, "loss": 0.8833, "step": 5652 }, { "epoch": 2.560235507246377, "grad_norm": 0.21526415680818484, "learning_rate": 4.528373670773597e-05, "loss": 0.8697, "step": 5653 }, { "epoch": 2.5606884057971016, "grad_norm": 0.38565538780919756, "learning_rate": 4.527119995538281e-05, "loss": 0.861, "step": 5654 }, { "epoch": 2.561141304347826, "grad_norm": 0.5050647710565229, "learning_rate": 4.5258662676059445e-05, "loss": 0.8807, "step": 5655 }, { "epoch": 2.5615942028985508, "grad_norm": 0.5128668426539185, "learning_rate": 4.5246124871019246e-05, "loss": 0.8809, "step": 5656 }, { "epoch": 2.5620471014492754, "grad_norm": 0.5600381304637688, "learning_rate": 4.523358654151563e-05, "loss": 0.8682, "step": 5657 }, { "epoch": 2.5625, "grad_norm": 0.5221135979479026, "learning_rate": 4.522104768880208e-05, "loss": 0.8588, "step": 5658 }, { "epoch": 2.5629528985507246, "grad_norm": 0.3404137327811273, "learning_rate": 4.52085083141321e-05, "loss": 0.8565, "step": 5659 }, { "epoch": 2.5634057971014492, "grad_norm": 0.1818972190793959, "learning_rate": 4.5195968418759304e-05, "loss": 0.8621, "step": 5660 }, { "epoch": 2.563858695652174, "grad_norm": 0.3411805988381861, "learning_rate": 4.518342800393731e-05, "loss": 0.8549, "step": 5661 }, { "epoch": 2.5643115942028984, "grad_norm": 0.39625014744240583, "learning_rate": 4.51708870709198e-05, "loss": 0.8724, "step": 5662 }, { "epoch": 2.564764492753623, "grad_norm": 0.3724276758407814, "learning_rate": 4.515834562096052e-05, "loss": 0.8444, "step": 5663 }, { "epoch": 2.5652173913043477, "grad_norm": 0.34387887419397467, "learning_rate": 4.514580365531325e-05, "loss": 0.8592, "step": 5664 }, { "epoch": 2.5656702898550723, "grad_norm": 0.25370164272911144, "learning_rate": 4.513326117523185e-05, "loss": 0.8764, "step": 5665 }, { "epoch": 2.566123188405797, "grad_norm": 0.21343444999291306, "learning_rate": 4.512071818197017e-05, "loss": 0.8666, "step": 5666 }, { "epoch": 2.5665760869565215, "grad_norm": 0.22406181359776187, "learning_rate": 4.5108174676782196e-05, "loss": 0.8572, "step": 5667 }, { "epoch": 2.567028985507246, "grad_norm": 0.20674330451936895, "learning_rate": 4.50956306609219e-05, "loss": 0.8677, "step": 5668 }, { "epoch": 2.567481884057971, "grad_norm": 0.25138336236992215, "learning_rate": 4.508308613564333e-05, "loss": 0.8824, "step": 5669 }, { "epoch": 2.567934782608696, "grad_norm": 0.2170806750174942, "learning_rate": 4.507054110220059e-05, "loss": 0.8622, "step": 5670 }, { "epoch": 2.5683876811594204, "grad_norm": 0.22288950926546428, "learning_rate": 4.505799556184782e-05, "loss": 0.8656, "step": 5671 }, { "epoch": 2.568840579710145, "grad_norm": 0.24859026092744393, "learning_rate": 4.504544951583922e-05, "loss": 0.8727, "step": 5672 }, { "epoch": 2.5692934782608696, "grad_norm": 0.21450568201503248, "learning_rate": 4.503290296542903e-05, "loss": 0.8767, "step": 5673 }, { "epoch": 2.5697463768115942, "grad_norm": 0.19922888769460853, "learning_rate": 4.502035591187156e-05, "loss": 0.8581, "step": 5674 }, { "epoch": 2.570199275362319, "grad_norm": 0.24187864967606687, "learning_rate": 4.5007808356421145e-05, "loss": 0.8754, "step": 5675 }, { "epoch": 2.5706521739130435, "grad_norm": 0.2165593603652902, "learning_rate": 4.4995260300332204e-05, "loss": 0.8837, "step": 5676 }, { "epoch": 2.571105072463768, "grad_norm": 0.2247405049322665, "learning_rate": 4.4982711744859167e-05, "loss": 0.8746, "step": 5677 }, { "epoch": 2.5715579710144927, "grad_norm": 0.21302394994146184, "learning_rate": 4.4970162691256556e-05, "loss": 0.8638, "step": 5678 }, { "epoch": 2.5720108695652173, "grad_norm": 0.2153538973126454, "learning_rate": 4.4957613140778895e-05, "loss": 0.9035, "step": 5679 }, { "epoch": 2.572463768115942, "grad_norm": 0.18536532156240473, "learning_rate": 4.4945063094680784e-05, "loss": 0.872, "step": 5680 }, { "epoch": 2.5729166666666665, "grad_norm": 0.270569872364493, "learning_rate": 4.4932512554216886e-05, "loss": 0.8777, "step": 5681 }, { "epoch": 2.5733695652173916, "grad_norm": 0.28792661731671887, "learning_rate": 4.49199615206419e-05, "loss": 0.8813, "step": 5682 }, { "epoch": 2.573822463768116, "grad_norm": 0.28795432776830915, "learning_rate": 4.4907409995210554e-05, "loss": 0.8584, "step": 5683 }, { "epoch": 2.574275362318841, "grad_norm": 0.30349201657847424, "learning_rate": 4.489485797917766e-05, "loss": 0.8812, "step": 5684 }, { "epoch": 2.5747282608695654, "grad_norm": 0.304358494498049, "learning_rate": 4.488230547379805e-05, "loss": 0.8657, "step": 5685 }, { "epoch": 2.57518115942029, "grad_norm": 0.22748754885421862, "learning_rate": 4.4869752480326645e-05, "loss": 0.8685, "step": 5686 }, { "epoch": 2.5756340579710146, "grad_norm": 0.2397996836696356, "learning_rate": 4.485719900001835e-05, "loss": 0.8772, "step": 5687 }, { "epoch": 2.5760869565217392, "grad_norm": 0.2414927045207118, "learning_rate": 4.484464503412817e-05, "loss": 0.8733, "step": 5688 }, { "epoch": 2.576539855072464, "grad_norm": 0.18661993963737308, "learning_rate": 4.483209058391115e-05, "loss": 0.8778, "step": 5689 }, { "epoch": 2.5769927536231885, "grad_norm": 0.2278445426033646, "learning_rate": 4.481953565062239e-05, "loss": 0.8699, "step": 5690 }, { "epoch": 2.577445652173913, "grad_norm": 0.23456618630181533, "learning_rate": 4.480698023551701e-05, "loss": 0.8799, "step": 5691 }, { "epoch": 2.5778985507246377, "grad_norm": 0.2840353820350236, "learning_rate": 4.4794424339850196e-05, "loss": 0.8803, "step": 5692 }, { "epoch": 2.5783514492753623, "grad_norm": 0.262395447167844, "learning_rate": 4.478186796487718e-05, "loss": 0.883, "step": 5693 }, { "epoch": 2.578804347826087, "grad_norm": 0.2575964056223032, "learning_rate": 4.476931111185325e-05, "loss": 0.8647, "step": 5694 }, { "epoch": 2.5792572463768115, "grad_norm": 0.37843129419885624, "learning_rate": 4.475675378203373e-05, "loss": 0.8554, "step": 5695 }, { "epoch": 2.579710144927536, "grad_norm": 0.34351532478707647, "learning_rate": 4.474419597667399e-05, "loss": 0.8799, "step": 5696 }, { "epoch": 2.5801630434782608, "grad_norm": 0.24127103636923408, "learning_rate": 4.473163769702947e-05, "loss": 0.8753, "step": 5697 }, { "epoch": 2.5806159420289854, "grad_norm": 0.2381585130079766, "learning_rate": 4.471907894435562e-05, "loss": 0.865, "step": 5698 }, { "epoch": 2.58106884057971, "grad_norm": 0.21056573237568738, "learning_rate": 4.470651971990799e-05, "loss": 0.8732, "step": 5699 }, { "epoch": 2.5815217391304346, "grad_norm": 0.1912258665987948, "learning_rate": 4.469396002494211e-05, "loss": 0.8753, "step": 5700 }, { "epoch": 2.581974637681159, "grad_norm": 0.3214304414814498, "learning_rate": 4.4681399860713606e-05, "loss": 0.886, "step": 5701 }, { "epoch": 2.582427536231884, "grad_norm": 0.3488927817673525, "learning_rate": 4.466883922847815e-05, "loss": 0.8943, "step": 5702 }, { "epoch": 2.5828804347826084, "grad_norm": 0.3043603153054017, "learning_rate": 4.465627812949144e-05, "loss": 0.8717, "step": 5703 }, { "epoch": 2.5833333333333335, "grad_norm": 0.3174864208575511, "learning_rate": 4.464371656500921e-05, "loss": 0.8646, "step": 5704 }, { "epoch": 2.583786231884058, "grad_norm": 0.23684371580995245, "learning_rate": 4.463115453628729e-05, "loss": 0.8544, "step": 5705 }, { "epoch": 2.5842391304347827, "grad_norm": 0.17522226086776105, "learning_rate": 4.461859204458151e-05, "loss": 0.8574, "step": 5706 }, { "epoch": 2.5846920289855073, "grad_norm": 0.18262315256453174, "learning_rate": 4.4606029091147764e-05, "loss": 0.8812, "step": 5707 }, { "epoch": 2.585144927536232, "grad_norm": 0.21763243016393843, "learning_rate": 4.4593465677241985e-05, "loss": 0.859, "step": 5708 }, { "epoch": 2.5855978260869565, "grad_norm": 0.1651560181972132, "learning_rate": 4.458090180412017e-05, "loss": 0.8737, "step": 5709 }, { "epoch": 2.586050724637681, "grad_norm": 0.20802661156254554, "learning_rate": 4.4568337473038336e-05, "loss": 0.8546, "step": 5710 }, { "epoch": 2.5865036231884058, "grad_norm": 0.2757926440121578, "learning_rate": 4.455577268525258e-05, "loss": 0.8713, "step": 5711 }, { "epoch": 2.5869565217391304, "grad_norm": 0.2467379945905209, "learning_rate": 4.4543207442019e-05, "loss": 0.8897, "step": 5712 }, { "epoch": 2.587409420289855, "grad_norm": 0.2081487064761063, "learning_rate": 4.4530641744593774e-05, "loss": 0.8576, "step": 5713 }, { "epoch": 2.5878623188405796, "grad_norm": 0.1787191406315958, "learning_rate": 4.4518075594233106e-05, "loss": 0.843, "step": 5714 }, { "epoch": 2.588315217391304, "grad_norm": 0.2246655385767487, "learning_rate": 4.450550899219328e-05, "loss": 0.8542, "step": 5715 }, { "epoch": 2.588768115942029, "grad_norm": 0.19304159593447084, "learning_rate": 4.449294193973056e-05, "loss": 0.8664, "step": 5716 }, { "epoch": 2.589221014492754, "grad_norm": 0.2173032245114637, "learning_rate": 4.448037443810132e-05, "loss": 0.8776, "step": 5717 }, { "epoch": 2.5896739130434785, "grad_norm": 0.25388123527872264, "learning_rate": 4.4467806488561954e-05, "loss": 0.8558, "step": 5718 }, { "epoch": 2.590126811594203, "grad_norm": 0.2491200616450722, "learning_rate": 4.445523809236888e-05, "loss": 0.8856, "step": 5719 }, { "epoch": 2.5905797101449277, "grad_norm": 0.2001054302688928, "learning_rate": 4.4442669250778614e-05, "loss": 0.8827, "step": 5720 }, { "epoch": 2.5910326086956523, "grad_norm": 0.28987939467132634, "learning_rate": 4.443009996504765e-05, "loss": 0.8799, "step": 5721 }, { "epoch": 2.591485507246377, "grad_norm": 0.2783884588051299, "learning_rate": 4.4417530236432566e-05, "loss": 0.8741, "step": 5722 }, { "epoch": 2.5919384057971016, "grad_norm": 0.25048974023452475, "learning_rate": 4.440496006619e-05, "loss": 0.8725, "step": 5723 }, { "epoch": 2.592391304347826, "grad_norm": 0.31956806820601075, "learning_rate": 4.43923894555766e-05, "loss": 0.846, "step": 5724 }, { "epoch": 2.5928442028985508, "grad_norm": 0.34813173320870966, "learning_rate": 4.437981840584905e-05, "loss": 0.8798, "step": 5725 }, { "epoch": 2.5932971014492754, "grad_norm": 0.31753505382565994, "learning_rate": 4.436724691826411e-05, "loss": 0.8731, "step": 5726 }, { "epoch": 2.59375, "grad_norm": 0.27679611604527593, "learning_rate": 4.4354674994078585e-05, "loss": 0.879, "step": 5727 }, { "epoch": 2.5942028985507246, "grad_norm": 0.27814750889034234, "learning_rate": 4.434210263454931e-05, "loss": 0.864, "step": 5728 }, { "epoch": 2.5946557971014492, "grad_norm": 0.31017715025389564, "learning_rate": 4.432952984093313e-05, "loss": 0.8712, "step": 5729 }, { "epoch": 2.595108695652174, "grad_norm": 0.3172700501094903, "learning_rate": 4.431695661448701e-05, "loss": 0.885, "step": 5730 }, { "epoch": 2.5955615942028984, "grad_norm": 0.33471858470206955, "learning_rate": 4.430438295646789e-05, "loss": 0.8826, "step": 5731 }, { "epoch": 2.596014492753623, "grad_norm": 0.31861333495526967, "learning_rate": 4.4291808868132786e-05, "loss": 0.8731, "step": 5732 }, { "epoch": 2.5964673913043477, "grad_norm": 0.23540072040950077, "learning_rate": 4.4279234350738746e-05, "loss": 0.8625, "step": 5733 }, { "epoch": 2.5969202898550723, "grad_norm": 0.25044685069283623, "learning_rate": 4.426665940554287e-05, "loss": 0.8651, "step": 5734 }, { "epoch": 2.597373188405797, "grad_norm": 0.2495848808444457, "learning_rate": 4.425408403380228e-05, "loss": 0.8666, "step": 5735 }, { "epoch": 2.5978260869565215, "grad_norm": 0.19742200029273735, "learning_rate": 4.424150823677419e-05, "loss": 0.8489, "step": 5736 }, { "epoch": 2.598278985507246, "grad_norm": 0.24644221876732178, "learning_rate": 4.422893201571579e-05, "loss": 0.8613, "step": 5737 }, { "epoch": 2.598731884057971, "grad_norm": 0.2835348337020532, "learning_rate": 4.421635537188435e-05, "loss": 0.8783, "step": 5738 }, { "epoch": 2.599184782608696, "grad_norm": 0.30320131642916476, "learning_rate": 4.420377830653719e-05, "loss": 0.8848, "step": 5739 }, { "epoch": 2.5996376811594204, "grad_norm": 0.3102840773830216, "learning_rate": 4.419120082093164e-05, "loss": 0.8836, "step": 5740 }, { "epoch": 2.600090579710145, "grad_norm": 0.3110002882317414, "learning_rate": 4.417862291632512e-05, "loss": 0.8742, "step": 5741 }, { "epoch": 2.6005434782608696, "grad_norm": 0.316701801271367, "learning_rate": 4.416604459397503e-05, "loss": 0.8641, "step": 5742 }, { "epoch": 2.6009963768115942, "grad_norm": 0.2941767233422764, "learning_rate": 4.4153465855138864e-05, "loss": 0.8647, "step": 5743 }, { "epoch": 2.601449275362319, "grad_norm": 0.19431885028934134, "learning_rate": 4.414088670107413e-05, "loss": 0.8909, "step": 5744 }, { "epoch": 2.6019021739130435, "grad_norm": 0.19326971487016156, "learning_rate": 4.412830713303841e-05, "loss": 0.8677, "step": 5745 }, { "epoch": 2.602355072463768, "grad_norm": 0.2301845495735688, "learning_rate": 4.411572715228927e-05, "loss": 0.8649, "step": 5746 }, { "epoch": 2.6028079710144927, "grad_norm": 0.24205178727409205, "learning_rate": 4.410314676008436e-05, "loss": 0.8834, "step": 5747 }, { "epoch": 2.6032608695652173, "grad_norm": 0.2953754723728503, "learning_rate": 4.409056595768137e-05, "loss": 0.8651, "step": 5748 }, { "epoch": 2.603713768115942, "grad_norm": 0.3070749877504125, "learning_rate": 4.4077984746338026e-05, "loss": 0.8799, "step": 5749 }, { "epoch": 2.6041666666666665, "grad_norm": 0.31051308731079685, "learning_rate": 4.406540312731208e-05, "loss": 0.8818, "step": 5750 }, { "epoch": 2.6046195652173916, "grad_norm": 0.3156529703583027, "learning_rate": 4.405282110186134e-05, "loss": 0.8776, "step": 5751 }, { "epoch": 2.605072463768116, "grad_norm": 0.3356240554338801, "learning_rate": 4.404023867124365e-05, "loss": 0.8849, "step": 5752 }, { "epoch": 2.605525362318841, "grad_norm": 0.2412210293585717, "learning_rate": 4.402765583671691e-05, "loss": 0.8765, "step": 5753 }, { "epoch": 2.6059782608695654, "grad_norm": 0.20960054089232477, "learning_rate": 4.401507259953902e-05, "loss": 0.881, "step": 5754 }, { "epoch": 2.60643115942029, "grad_norm": 0.18060146276964406, "learning_rate": 4.4002488960967964e-05, "loss": 0.8729, "step": 5755 }, { "epoch": 2.6068840579710146, "grad_norm": 0.18387861274809345, "learning_rate": 4.3989904922261746e-05, "loss": 0.861, "step": 5756 }, { "epoch": 2.6073369565217392, "grad_norm": 0.19971610733523223, "learning_rate": 4.397732048467841e-05, "loss": 0.8754, "step": 5757 }, { "epoch": 2.607789855072464, "grad_norm": 0.20927313122920463, "learning_rate": 4.396473564947604e-05, "loss": 0.8947, "step": 5758 }, { "epoch": 2.6082427536231885, "grad_norm": 0.19486490010766494, "learning_rate": 4.395215041791277e-05, "loss": 0.8632, "step": 5759 }, { "epoch": 2.608695652173913, "grad_norm": 0.20386230354758927, "learning_rate": 4.3939564791246744e-05, "loss": 0.8713, "step": 5760 }, { "epoch": 2.6091485507246377, "grad_norm": 0.25423421118285217, "learning_rate": 4.39269787707362e-05, "loss": 0.8778, "step": 5761 }, { "epoch": 2.6096014492753623, "grad_norm": 0.3183833901978331, "learning_rate": 4.391439235763938e-05, "loss": 0.8619, "step": 5762 }, { "epoch": 2.610054347826087, "grad_norm": 0.3419028103499635, "learning_rate": 4.3901805553214525e-05, "loss": 0.8844, "step": 5763 }, { "epoch": 2.6105072463768115, "grad_norm": 0.3499895631186666, "learning_rate": 4.3889218358719994e-05, "loss": 0.8686, "step": 5764 }, { "epoch": 2.610960144927536, "grad_norm": 0.3465664493139378, "learning_rate": 4.387663077541414e-05, "loss": 0.888, "step": 5765 }, { "epoch": 2.6114130434782608, "grad_norm": 0.27158851227044095, "learning_rate": 4.3864042804555363e-05, "loss": 0.8649, "step": 5766 }, { "epoch": 2.6118659420289854, "grad_norm": 0.18544944658899332, "learning_rate": 4.385145444740211e-05, "loss": 0.884, "step": 5767 }, { "epoch": 2.61231884057971, "grad_norm": 0.2093753505677767, "learning_rate": 4.3838865705212854e-05, "loss": 0.9024, "step": 5768 }, { "epoch": 2.6127717391304346, "grad_norm": 0.2898373084471052, "learning_rate": 4.3826276579246095e-05, "loss": 0.86, "step": 5769 }, { "epoch": 2.613224637681159, "grad_norm": 0.3590315827124731, "learning_rate": 4.3813687070760416e-05, "loss": 0.9093, "step": 5770 }, { "epoch": 2.613677536231884, "grad_norm": 0.30034022909646135, "learning_rate": 4.38010971810144e-05, "loss": 0.8883, "step": 5771 }, { "epoch": 2.6141304347826084, "grad_norm": 0.2259189631100891, "learning_rate": 4.3788506911266667e-05, "loss": 0.8708, "step": 5772 }, { "epoch": 2.6145833333333335, "grad_norm": 0.2416665722387569, "learning_rate": 4.37759162627759e-05, "loss": 0.8758, "step": 5773 }, { "epoch": 2.615036231884058, "grad_norm": 0.16621614628483, "learning_rate": 4.376332523680079e-05, "loss": 0.8646, "step": 5774 }, { "epoch": 2.6154891304347827, "grad_norm": 0.22005074522167697, "learning_rate": 4.3750733834600114e-05, "loss": 0.868, "step": 5775 }, { "epoch": 2.6159420289855073, "grad_norm": 0.2736435956395162, "learning_rate": 4.373814205743261e-05, "loss": 0.8753, "step": 5776 }, { "epoch": 2.616394927536232, "grad_norm": 0.2703876074321549, "learning_rate": 4.372554990655713e-05, "loss": 0.8815, "step": 5777 }, { "epoch": 2.6168478260869565, "grad_norm": 0.26454111412031217, "learning_rate": 4.3712957383232524e-05, "loss": 0.8789, "step": 5778 }, { "epoch": 2.617300724637681, "grad_norm": 0.2694101789250738, "learning_rate": 4.370036448871769e-05, "loss": 0.858, "step": 5779 }, { "epoch": 2.6177536231884058, "grad_norm": 0.20247612814940744, "learning_rate": 4.368777122427155e-05, "loss": 0.8723, "step": 5780 }, { "epoch": 2.6182065217391304, "grad_norm": 0.2623845988779585, "learning_rate": 4.367517759115307e-05, "loss": 0.8726, "step": 5781 }, { "epoch": 2.618659420289855, "grad_norm": 0.3407710219796899, "learning_rate": 4.3662583590621266e-05, "loss": 0.8865, "step": 5782 }, { "epoch": 2.6191123188405796, "grad_norm": 0.32967829422496464, "learning_rate": 4.364998922393518e-05, "loss": 0.8801, "step": 5783 }, { "epoch": 2.619565217391304, "grad_norm": 0.4521514110283182, "learning_rate": 4.3637394492353886e-05, "loss": 0.8816, "step": 5784 }, { "epoch": 2.620018115942029, "grad_norm": 0.35049799508675505, "learning_rate": 4.36247993971365e-05, "loss": 0.8791, "step": 5785 }, { "epoch": 2.620471014492754, "grad_norm": 0.23848626073897594, "learning_rate": 4.3612203939542176e-05, "loss": 0.8671, "step": 5786 }, { "epoch": 2.6209239130434785, "grad_norm": 0.23713842558842216, "learning_rate": 4.3599608120830094e-05, "loss": 0.8716, "step": 5787 }, { "epoch": 2.621376811594203, "grad_norm": 0.195214888153204, "learning_rate": 4.358701194225949e-05, "loss": 0.8431, "step": 5788 }, { "epoch": 2.6218297101449277, "grad_norm": 0.22796941014758068, "learning_rate": 4.357441540508962e-05, "loss": 0.8801, "step": 5789 }, { "epoch": 2.6222826086956523, "grad_norm": 0.31925578842170177, "learning_rate": 4.3561818510579774e-05, "loss": 0.8468, "step": 5790 }, { "epoch": 2.622735507246377, "grad_norm": 0.2968668284696698, "learning_rate": 4.3549221259989286e-05, "loss": 0.8885, "step": 5791 }, { "epoch": 2.6231884057971016, "grad_norm": 0.2976259284816325, "learning_rate": 4.353662365457753e-05, "loss": 0.8975, "step": 5792 }, { "epoch": 2.623641304347826, "grad_norm": 0.26932839357036564, "learning_rate": 4.35240256956039e-05, "loss": 0.871, "step": 5793 }, { "epoch": 2.6240942028985508, "grad_norm": 0.2135409198080569, "learning_rate": 4.351142738432783e-05, "loss": 0.8836, "step": 5794 }, { "epoch": 2.6245471014492754, "grad_norm": 0.3184503312089119, "learning_rate": 4.3498828722008805e-05, "loss": 0.8666, "step": 5795 }, { "epoch": 2.625, "grad_norm": 0.300899570731643, "learning_rate": 4.348622970990634e-05, "loss": 0.8693, "step": 5796 }, { "epoch": 2.6254528985507246, "grad_norm": 0.2971725660790538, "learning_rate": 4.347363034927995e-05, "loss": 0.8807, "step": 5797 }, { "epoch": 2.6259057971014492, "grad_norm": 0.2503744086954991, "learning_rate": 4.346103064138923e-05, "loss": 0.8479, "step": 5798 }, { "epoch": 2.626358695652174, "grad_norm": 0.19351642607675815, "learning_rate": 4.3448430587493796e-05, "loss": 0.87, "step": 5799 }, { "epoch": 2.6268115942028984, "grad_norm": 0.2537797800611189, "learning_rate": 4.3435830188853304e-05, "loss": 0.8703, "step": 5800 }, { "epoch": 2.627264492753623, "grad_norm": 0.2568779361599437, "learning_rate": 4.342322944672741e-05, "loss": 0.8722, "step": 5801 }, { "epoch": 2.6277173913043477, "grad_norm": 0.23188759068769926, "learning_rate": 4.3410628362375844e-05, "loss": 0.8783, "step": 5802 }, { "epoch": 2.6281702898550723, "grad_norm": 0.2123318165271853, "learning_rate": 4.339802693705836e-05, "loss": 0.8724, "step": 5803 }, { "epoch": 2.628623188405797, "grad_norm": 0.21817751546110645, "learning_rate": 4.3385425172034744e-05, "loss": 0.8775, "step": 5804 }, { "epoch": 2.6290760869565215, "grad_norm": 0.16393202520987143, "learning_rate": 4.3372823068564805e-05, "loss": 0.8703, "step": 5805 }, { "epoch": 2.629528985507246, "grad_norm": 0.1959204861730338, "learning_rate": 4.3360220627908383e-05, "loss": 0.8665, "step": 5806 }, { "epoch": 2.629981884057971, "grad_norm": 0.20018197184077455, "learning_rate": 4.334761785132539e-05, "loss": 0.8959, "step": 5807 }, { "epoch": 2.630434782608696, "grad_norm": 0.23738295162263093, "learning_rate": 4.333501474007574e-05, "loss": 0.8773, "step": 5808 }, { "epoch": 2.6308876811594204, "grad_norm": 0.21578794095879456, "learning_rate": 4.332241129541939e-05, "loss": 0.8814, "step": 5809 }, { "epoch": 2.631340579710145, "grad_norm": 0.2147574403904211, "learning_rate": 4.330980751861631e-05, "loss": 0.8779, "step": 5810 }, { "epoch": 2.6317934782608696, "grad_norm": 0.21053404603883266, "learning_rate": 4.3297203410926526e-05, "loss": 0.8765, "step": 5811 }, { "epoch": 2.6322463768115942, "grad_norm": 0.22447742452439837, "learning_rate": 4.3284598973610085e-05, "loss": 0.8855, "step": 5812 }, { "epoch": 2.632699275362319, "grad_norm": 0.196436920652347, "learning_rate": 4.32719942079271e-05, "loss": 0.8624, "step": 5813 }, { "epoch": 2.6331521739130435, "grad_norm": 0.18983303309640115, "learning_rate": 4.3259389115137654e-05, "loss": 0.8785, "step": 5814 }, { "epoch": 2.633605072463768, "grad_norm": 0.22685687171081673, "learning_rate": 4.3246783696501905e-05, "loss": 0.872, "step": 5815 }, { "epoch": 2.6340579710144927, "grad_norm": 0.3097694382222484, "learning_rate": 4.3234177953280055e-05, "loss": 0.854, "step": 5816 }, { "epoch": 2.6345108695652173, "grad_norm": 0.30084653492614494, "learning_rate": 4.322157188673232e-05, "loss": 0.894, "step": 5817 }, { "epoch": 2.634963768115942, "grad_norm": 0.25320300079595687, "learning_rate": 4.3208965498118924e-05, "loss": 0.8728, "step": 5818 }, { "epoch": 2.6354166666666665, "grad_norm": 0.24654010420711453, "learning_rate": 4.3196358788700164e-05, "loss": 0.8845, "step": 5819 }, { "epoch": 2.6358695652173916, "grad_norm": 0.2024661973861191, "learning_rate": 4.3183751759736354e-05, "loss": 0.8705, "step": 5820 }, { "epoch": 2.636322463768116, "grad_norm": 0.19991971080603535, "learning_rate": 4.317114441248784e-05, "loss": 0.8689, "step": 5821 }, { "epoch": 2.636775362318841, "grad_norm": 0.22107084574180846, "learning_rate": 4.315853674821498e-05, "loss": 0.8606, "step": 5822 }, { "epoch": 2.6372282608695654, "grad_norm": 0.2232881549101415, "learning_rate": 4.3145928768178205e-05, "loss": 0.8731, "step": 5823 }, { "epoch": 2.63768115942029, "grad_norm": 0.2356309249553189, "learning_rate": 4.313332047363795e-05, "loss": 0.8723, "step": 5824 }, { "epoch": 2.6381340579710146, "grad_norm": 0.23166236088240624, "learning_rate": 4.312071186585469e-05, "loss": 0.8843, "step": 5825 }, { "epoch": 2.6385869565217392, "grad_norm": 0.19663005993469743, "learning_rate": 4.31081029460889e-05, "loss": 0.8651, "step": 5826 }, { "epoch": 2.639039855072464, "grad_norm": 0.24361275838918214, "learning_rate": 4.3095493715601144e-05, "loss": 0.8588, "step": 5827 }, { "epoch": 2.6394927536231885, "grad_norm": 0.19504243437478802, "learning_rate": 4.308288417565196e-05, "loss": 0.8806, "step": 5828 }, { "epoch": 2.639945652173913, "grad_norm": 0.2264160832323669, "learning_rate": 4.307027432750198e-05, "loss": 0.8904, "step": 5829 }, { "epoch": 2.6403985507246377, "grad_norm": 0.22624984315966248, "learning_rate": 4.305766417241181e-05, "loss": 0.8808, "step": 5830 }, { "epoch": 2.6408514492753623, "grad_norm": 0.1782704133833528, "learning_rate": 4.30450537116421e-05, "loss": 0.8742, "step": 5831 }, { "epoch": 2.641304347826087, "grad_norm": 0.25039255221052104, "learning_rate": 4.303244294645354e-05, "loss": 0.8827, "step": 5832 }, { "epoch": 2.6417572463768115, "grad_norm": 0.23792785909485023, "learning_rate": 4.3019831878106855e-05, "loss": 0.8695, "step": 5833 }, { "epoch": 2.642210144927536, "grad_norm": 0.24311870071278482, "learning_rate": 4.30072205078628e-05, "loss": 0.8948, "step": 5834 }, { "epoch": 2.6426630434782608, "grad_norm": 0.21735051637532465, "learning_rate": 4.299460883698214e-05, "loss": 0.8865, "step": 5835 }, { "epoch": 2.6431159420289854, "grad_norm": 0.21475465287191353, "learning_rate": 4.298199686672569e-05, "loss": 0.8805, "step": 5836 }, { "epoch": 2.64356884057971, "grad_norm": 0.20592873980453205, "learning_rate": 4.2969384598354284e-05, "loss": 0.8582, "step": 5837 }, { "epoch": 2.6440217391304346, "grad_norm": 0.2062150100853468, "learning_rate": 4.295677203312881e-05, "loss": 0.8623, "step": 5838 }, { "epoch": 2.644474637681159, "grad_norm": 0.29964052675830755, "learning_rate": 4.294415917231013e-05, "loss": 0.8709, "step": 5839 }, { "epoch": 2.644927536231884, "grad_norm": 0.3825964188083746, "learning_rate": 4.293154601715921e-05, "loss": 0.8711, "step": 5840 }, { "epoch": 2.6453804347826084, "grad_norm": 0.32490654845122846, "learning_rate": 4.291893256893697e-05, "loss": 0.868, "step": 5841 }, { "epoch": 2.6458333333333335, "grad_norm": 0.2983792135921932, "learning_rate": 4.290631882890443e-05, "loss": 0.8677, "step": 5842 }, { "epoch": 2.646286231884058, "grad_norm": 0.23190614269253884, "learning_rate": 4.2893704798322586e-05, "loss": 0.8646, "step": 5843 }, { "epoch": 2.6467391304347827, "grad_norm": 0.18325227850125628, "learning_rate": 4.288109047845249e-05, "loss": 0.8734, "step": 5844 }, { "epoch": 2.6471920289855073, "grad_norm": 0.2077899980241696, "learning_rate": 4.286847587055521e-05, "loss": 0.8847, "step": 5845 }, { "epoch": 2.647644927536232, "grad_norm": 0.2386302463692566, "learning_rate": 4.285586097589186e-05, "loss": 0.8748, "step": 5846 }, { "epoch": 2.6480978260869565, "grad_norm": 0.3078210026636928, "learning_rate": 4.2843245795723554e-05, "loss": 0.8796, "step": 5847 }, { "epoch": 2.648550724637681, "grad_norm": 0.35279343667318613, "learning_rate": 4.283063033131147e-05, "loss": 0.8748, "step": 5848 }, { "epoch": 2.6490036231884058, "grad_norm": 0.3655581788131125, "learning_rate": 4.2818014583916766e-05, "loss": 0.8539, "step": 5849 }, { "epoch": 2.6494565217391304, "grad_norm": 0.24176145652360853, "learning_rate": 4.2805398554800695e-05, "loss": 0.8773, "step": 5850 }, { "epoch": 2.649909420289855, "grad_norm": 0.2310769474060846, "learning_rate": 4.279278224522449e-05, "loss": 0.8801, "step": 5851 }, { "epoch": 2.6503623188405796, "grad_norm": 0.34613626540273507, "learning_rate": 4.27801656564494e-05, "loss": 0.8624, "step": 5852 }, { "epoch": 2.650815217391304, "grad_norm": 0.36280756742851195, "learning_rate": 4.276754878973676e-05, "loss": 0.8683, "step": 5853 }, { "epoch": 2.651268115942029, "grad_norm": 0.37359686781537077, "learning_rate": 4.275493164634787e-05, "loss": 0.8859, "step": 5854 }, { "epoch": 2.651721014492754, "grad_norm": 0.3800786178055471, "learning_rate": 4.274231422754411e-05, "loss": 0.8698, "step": 5855 }, { "epoch": 2.6521739130434785, "grad_norm": 0.37523085815794555, "learning_rate": 4.272969653458685e-05, "loss": 0.8676, "step": 5856 }, { "epoch": 2.652626811594203, "grad_norm": 0.3204285118470007, "learning_rate": 4.2717078568737485e-05, "loss": 0.8672, "step": 5857 }, { "epoch": 2.6530797101449277, "grad_norm": 0.3075789071149555, "learning_rate": 4.2704460331257484e-05, "loss": 0.8568, "step": 5858 }, { "epoch": 2.6535326086956523, "grad_norm": 0.24622001572615546, "learning_rate": 4.26918418234083e-05, "loss": 0.9042, "step": 5859 }, { "epoch": 2.653985507246377, "grad_norm": 0.20145363113719886, "learning_rate": 4.267922304645142e-05, "loss": 0.8674, "step": 5860 }, { "epoch": 2.6544384057971016, "grad_norm": 0.19562545012713978, "learning_rate": 4.266660400164837e-05, "loss": 0.8856, "step": 5861 }, { "epoch": 2.654891304347826, "grad_norm": 0.23530599577372538, "learning_rate": 4.265398469026067e-05, "loss": 0.8831, "step": 5862 }, { "epoch": 2.6553442028985508, "grad_norm": 0.21319934056346904, "learning_rate": 4.264136511354995e-05, "loss": 0.8582, "step": 5863 }, { "epoch": 2.6557971014492754, "grad_norm": 0.20688613600391678, "learning_rate": 4.262874527277775e-05, "loss": 0.8751, "step": 5864 }, { "epoch": 2.65625, "grad_norm": 0.2378256950031439, "learning_rate": 4.261612516920573e-05, "loss": 0.8719, "step": 5865 }, { "epoch": 2.6567028985507246, "grad_norm": 0.24641031796718119, "learning_rate": 4.260350480409552e-05, "loss": 0.855, "step": 5866 }, { "epoch": 2.6571557971014492, "grad_norm": 0.20699216815340357, "learning_rate": 4.259088417870882e-05, "loss": 0.8593, "step": 5867 }, { "epoch": 2.657608695652174, "grad_norm": 0.2490679407305972, "learning_rate": 4.257826329430733e-05, "loss": 0.8709, "step": 5868 }, { "epoch": 2.6580615942028984, "grad_norm": 0.2884147898546209, "learning_rate": 4.256564215215275e-05, "loss": 0.8652, "step": 5869 }, { "epoch": 2.658514492753623, "grad_norm": 0.2587107405967696, "learning_rate": 4.2553020753506875e-05, "loss": 0.8906, "step": 5870 }, { "epoch": 2.6589673913043477, "grad_norm": 0.26889719078537416, "learning_rate": 4.2540399099631465e-05, "loss": 0.8723, "step": 5871 }, { "epoch": 2.6594202898550723, "grad_norm": 0.2441242445549249, "learning_rate": 4.252777719178833e-05, "loss": 0.8684, "step": 5872 }, { "epoch": 2.659873188405797, "grad_norm": 0.3072924006590988, "learning_rate": 4.2515155031239306e-05, "loss": 0.8731, "step": 5873 }, { "epoch": 2.6603260869565215, "grad_norm": 0.2415187533659859, "learning_rate": 4.250253261924624e-05, "loss": 0.8725, "step": 5874 }, { "epoch": 2.660778985507246, "grad_norm": 0.17241973853296136, "learning_rate": 4.248990995707104e-05, "loss": 0.8796, "step": 5875 }, { "epoch": 2.661231884057971, "grad_norm": 0.26636966089918807, "learning_rate": 4.2477287045975585e-05, "loss": 0.8683, "step": 5876 }, { "epoch": 2.661684782608696, "grad_norm": 0.2594305808668593, "learning_rate": 4.246466388722182e-05, "loss": 0.8666, "step": 5877 }, { "epoch": 2.6621376811594204, "grad_norm": 0.16034451513413361, "learning_rate": 4.2452040482071694e-05, "loss": 0.8759, "step": 5878 }, { "epoch": 2.662590579710145, "grad_norm": 0.27614317215491674, "learning_rate": 4.2439416831787204e-05, "loss": 0.8578, "step": 5879 }, { "epoch": 2.6630434782608696, "grad_norm": 0.2911930272261758, "learning_rate": 4.242679293763035e-05, "loss": 0.8639, "step": 5880 }, { "epoch": 2.6634963768115942, "grad_norm": 0.20873714224520884, "learning_rate": 4.241416880086316e-05, "loss": 0.866, "step": 5881 }, { "epoch": 2.663949275362319, "grad_norm": 0.21163612340230878, "learning_rate": 4.240154442274768e-05, "loss": 0.8741, "step": 5882 }, { "epoch": 2.6644021739130435, "grad_norm": 0.23528106389290726, "learning_rate": 4.2388919804545993e-05, "loss": 0.8758, "step": 5883 }, { "epoch": 2.664855072463768, "grad_norm": 0.2277152337237947, "learning_rate": 4.2376294947520226e-05, "loss": 0.877, "step": 5884 }, { "epoch": 2.6653079710144927, "grad_norm": 0.21035953126778104, "learning_rate": 4.2363669852932475e-05, "loss": 0.8859, "step": 5885 }, { "epoch": 2.6657608695652173, "grad_norm": 0.22741243236342173, "learning_rate": 4.235104452204492e-05, "loss": 0.8793, "step": 5886 }, { "epoch": 2.666213768115942, "grad_norm": 0.19983706689341127, "learning_rate": 4.233841895611971e-05, "loss": 0.8822, "step": 5887 }, { "epoch": 2.6666666666666665, "grad_norm": 0.20035846814222402, "learning_rate": 4.2325793156419035e-05, "loss": 0.8802, "step": 5888 }, { "epoch": 2.6671195652173916, "grad_norm": 0.3031587352339886, "learning_rate": 4.2313167124205154e-05, "loss": 0.8797, "step": 5889 }, { "epoch": 2.667572463768116, "grad_norm": 0.3807981916768925, "learning_rate": 4.230054086074028e-05, "loss": 0.856, "step": 5890 }, { "epoch": 2.668025362318841, "grad_norm": 0.33920860830359745, "learning_rate": 4.228791436728669e-05, "loss": 0.873, "step": 5891 }, { "epoch": 2.6684782608695654, "grad_norm": 0.2723825049916238, "learning_rate": 4.227528764510668e-05, "loss": 0.8627, "step": 5892 }, { "epoch": 2.66893115942029, "grad_norm": 0.24187065285801404, "learning_rate": 4.226266069546256e-05, "loss": 0.8833, "step": 5893 }, { "epoch": 2.6693840579710146, "grad_norm": 0.20774690148549552, "learning_rate": 4.225003351961666e-05, "loss": 0.8702, "step": 5894 }, { "epoch": 2.6698369565217392, "grad_norm": 0.20437720535718865, "learning_rate": 4.223740611883134e-05, "loss": 0.8788, "step": 5895 }, { "epoch": 2.670289855072464, "grad_norm": 0.2209532845454224, "learning_rate": 4.2224778494368986e-05, "loss": 0.8847, "step": 5896 }, { "epoch": 2.6707427536231885, "grad_norm": 0.24972258603026262, "learning_rate": 4.2212150647492e-05, "loss": 0.8786, "step": 5897 }, { "epoch": 2.671195652173913, "grad_norm": 0.2567019779247641, "learning_rate": 4.219952257946282e-05, "loss": 0.8649, "step": 5898 }, { "epoch": 2.6716485507246377, "grad_norm": 0.26701204912292537, "learning_rate": 4.218689429154387e-05, "loss": 0.859, "step": 5899 }, { "epoch": 2.6721014492753623, "grad_norm": 0.1923515868567132, "learning_rate": 4.2174265784997635e-05, "loss": 0.8719, "step": 5900 }, { "epoch": 2.672554347826087, "grad_norm": 0.2118625505801057, "learning_rate": 4.216163706108661e-05, "loss": 0.8768, "step": 5901 }, { "epoch": 2.6730072463768115, "grad_norm": 0.2105192887669108, "learning_rate": 4.2149008121073305e-05, "loss": 0.8772, "step": 5902 }, { "epoch": 2.673460144927536, "grad_norm": 0.20403612405594931, "learning_rate": 4.213637896622023e-05, "loss": 0.8859, "step": 5903 }, { "epoch": 2.6739130434782608, "grad_norm": 0.23312448526614046, "learning_rate": 4.2123749597789984e-05, "loss": 0.8576, "step": 5904 }, { "epoch": 2.6743659420289854, "grad_norm": 0.24957262096632302, "learning_rate": 4.211112001704513e-05, "loss": 0.8662, "step": 5905 }, { "epoch": 2.67481884057971, "grad_norm": 0.28056327385301616, "learning_rate": 4.209849022524825e-05, "loss": 0.8697, "step": 5906 }, { "epoch": 2.6752717391304346, "grad_norm": 0.1992044908409931, "learning_rate": 4.2085860223661984e-05, "loss": 0.8675, "step": 5907 }, { "epoch": 2.675724637681159, "grad_norm": 0.28018107250181956, "learning_rate": 4.207323001354896e-05, "loss": 0.8735, "step": 5908 }, { "epoch": 2.676177536231884, "grad_norm": 0.2837123027783233, "learning_rate": 4.206059959617185e-05, "loss": 0.8649, "step": 5909 }, { "epoch": 2.6766304347826084, "grad_norm": 0.21749972014045307, "learning_rate": 4.204796897279334e-05, "loss": 0.8626, "step": 5910 }, { "epoch": 2.6770833333333335, "grad_norm": 0.33092330664445974, "learning_rate": 4.203533814467611e-05, "loss": 0.8906, "step": 5911 }, { "epoch": 2.677536231884058, "grad_norm": 0.34153205134994113, "learning_rate": 4.202270711308291e-05, "loss": 0.8677, "step": 5912 }, { "epoch": 2.6779891304347827, "grad_norm": 0.2940078822817186, "learning_rate": 4.2010075879276474e-05, "loss": 0.8689, "step": 5913 }, { "epoch": 2.6784420289855073, "grad_norm": 0.24325629765574386, "learning_rate": 4.199744444451957e-05, "loss": 0.8845, "step": 5914 }, { "epoch": 2.678894927536232, "grad_norm": 0.23984914178056, "learning_rate": 4.198481281007498e-05, "loss": 0.8828, "step": 5915 }, { "epoch": 2.6793478260869565, "grad_norm": 0.25819722370997295, "learning_rate": 4.1972180977205495e-05, "loss": 0.8757, "step": 5916 }, { "epoch": 2.679800724637681, "grad_norm": 0.3331945457043144, "learning_rate": 4.195954894717396e-05, "loss": 0.8561, "step": 5917 }, { "epoch": 2.6802536231884058, "grad_norm": 0.4058596206234269, "learning_rate": 4.1946916721243215e-05, "loss": 0.8722, "step": 5918 }, { "epoch": 2.6807065217391304, "grad_norm": 0.4467794689784759, "learning_rate": 4.193428430067612e-05, "loss": 0.8835, "step": 5919 }, { "epoch": 2.681159420289855, "grad_norm": 0.4838178604113971, "learning_rate": 4.1921651686735555e-05, "loss": 0.9104, "step": 5920 }, { "epoch": 2.6816123188405796, "grad_norm": 0.42281147642755884, "learning_rate": 4.190901888068443e-05, "loss": 0.8704, "step": 5921 }, { "epoch": 2.682065217391304, "grad_norm": 0.33007050343089706, "learning_rate": 4.189638588378566e-05, "loss": 0.867, "step": 5922 }, { "epoch": 2.682518115942029, "grad_norm": 0.30403801110511913, "learning_rate": 4.18837526973022e-05, "loss": 0.8488, "step": 5923 }, { "epoch": 2.682971014492754, "grad_norm": 0.2318797325844851, "learning_rate": 4.187111932249698e-05, "loss": 0.8739, "step": 5924 }, { "epoch": 2.6834239130434785, "grad_norm": 0.3608317860779662, "learning_rate": 4.1858485760632996e-05, "loss": 0.8741, "step": 5925 }, { "epoch": 2.683876811594203, "grad_norm": 0.3972889034802442, "learning_rate": 4.184585201297325e-05, "loss": 0.8555, "step": 5926 }, { "epoch": 2.6843297101449277, "grad_norm": 0.26584240700773554, "learning_rate": 4.183321808078076e-05, "loss": 0.8685, "step": 5927 }, { "epoch": 2.6847826086956523, "grad_norm": 0.196274942515407, "learning_rate": 4.1820583965318555e-05, "loss": 0.8681, "step": 5928 }, { "epoch": 2.685235507246377, "grad_norm": 0.2871945522091861, "learning_rate": 4.180794966784968e-05, "loss": 0.8759, "step": 5929 }, { "epoch": 2.6856884057971016, "grad_norm": 0.4461366657328242, "learning_rate": 4.17953151896372e-05, "loss": 0.8655, "step": 5930 }, { "epoch": 2.686141304347826, "grad_norm": 0.5010831194354894, "learning_rate": 4.1782680531944234e-05, "loss": 0.8643, "step": 5931 }, { "epoch": 2.6865942028985508, "grad_norm": 0.4398913151575453, "learning_rate": 4.1770045696033866e-05, "loss": 0.8697, "step": 5932 }, { "epoch": 2.6870471014492754, "grad_norm": 0.27016707011631524, "learning_rate": 4.1757410683169214e-05, "loss": 0.8617, "step": 5933 }, { "epoch": 2.6875, "grad_norm": 0.18926186111964885, "learning_rate": 4.174477549461345e-05, "loss": 0.8749, "step": 5934 }, { "epoch": 2.6879528985507246, "grad_norm": 0.22864322231408096, "learning_rate": 4.17321401316297e-05, "loss": 0.8602, "step": 5935 }, { "epoch": 2.6884057971014492, "grad_norm": 0.3033658497642938, "learning_rate": 4.171950459548116e-05, "loss": 0.8403, "step": 5936 }, { "epoch": 2.688858695652174, "grad_norm": 0.3106644012105579, "learning_rate": 4.170686888743102e-05, "loss": 0.8461, "step": 5937 }, { "epoch": 2.6893115942028984, "grad_norm": 0.2706515276704555, "learning_rate": 4.169423300874249e-05, "loss": 0.8546, "step": 5938 }, { "epoch": 2.689764492753623, "grad_norm": 0.25302282410894106, "learning_rate": 4.168159696067882e-05, "loss": 0.8544, "step": 5939 }, { "epoch": 2.6902173913043477, "grad_norm": 0.18437270642038492, "learning_rate": 4.166896074450322e-05, "loss": 0.8457, "step": 5940 }, { "epoch": 2.6906702898550723, "grad_norm": 0.18331310515252175, "learning_rate": 4.165632436147897e-05, "loss": 0.8745, "step": 5941 }, { "epoch": 2.691123188405797, "grad_norm": 0.2599209783073537, "learning_rate": 4.1643687812869344e-05, "loss": 0.876, "step": 5942 }, { "epoch": 2.6915760869565215, "grad_norm": 0.36088478857396167, "learning_rate": 4.163105109993765e-05, "loss": 0.8596, "step": 5943 }, { "epoch": 2.692028985507246, "grad_norm": 0.39374682189363475, "learning_rate": 4.16184142239472e-05, "loss": 0.8874, "step": 5944 }, { "epoch": 2.692481884057971, "grad_norm": 0.33356981787052226, "learning_rate": 4.1605777186161305e-05, "loss": 0.8709, "step": 5945 }, { "epoch": 2.692934782608696, "grad_norm": 0.29108506908917015, "learning_rate": 4.159313998784332e-05, "loss": 0.8733, "step": 5946 }, { "epoch": 2.6933876811594204, "grad_norm": 0.28985786625498916, "learning_rate": 4.1580502630256614e-05, "loss": 0.8588, "step": 5947 }, { "epoch": 2.693840579710145, "grad_norm": 0.2650503908975153, "learning_rate": 4.1567865114664554e-05, "loss": 0.8759, "step": 5948 }, { "epoch": 2.6942934782608696, "grad_norm": 0.236116424723334, "learning_rate": 4.155522744233053e-05, "loss": 0.8743, "step": 5949 }, { "epoch": 2.6947463768115942, "grad_norm": 0.25824202042720307, "learning_rate": 4.154258961451795e-05, "loss": 0.8634, "step": 5950 }, { "epoch": 2.695199275362319, "grad_norm": 0.2769105293316691, "learning_rate": 4.152995163249025e-05, "loss": 0.8724, "step": 5951 }, { "epoch": 2.6956521739130435, "grad_norm": 0.17535525553897408, "learning_rate": 4.151731349751086e-05, "loss": 0.8448, "step": 5952 }, { "epoch": 2.696105072463768, "grad_norm": 0.2635086304365243, "learning_rate": 4.150467521084323e-05, "loss": 0.8727, "step": 5953 }, { "epoch": 2.6965579710144927, "grad_norm": 0.28538383612566615, "learning_rate": 4.1492036773750835e-05, "loss": 0.8914, "step": 5954 }, { "epoch": 2.6970108695652173, "grad_norm": 0.2699177916564289, "learning_rate": 4.147939818749716e-05, "loss": 0.8803, "step": 5955 }, { "epoch": 2.697463768115942, "grad_norm": 0.23415795265223924, "learning_rate": 4.146675945334572e-05, "loss": 0.8893, "step": 5956 }, { "epoch": 2.6979166666666665, "grad_norm": 0.24860553191127785, "learning_rate": 4.145412057256e-05, "loss": 0.8679, "step": 5957 }, { "epoch": 2.6983695652173916, "grad_norm": 0.29284994022264105, "learning_rate": 4.144148154640354e-05, "loss": 0.8592, "step": 5958 }, { "epoch": 2.698822463768116, "grad_norm": 0.34470278029611456, "learning_rate": 4.142884237613988e-05, "loss": 0.8745, "step": 5959 }, { "epoch": 2.699275362318841, "grad_norm": 0.3334941580190158, "learning_rate": 4.1416203063032594e-05, "loss": 0.8691, "step": 5960 }, { "epoch": 2.6997282608695654, "grad_norm": 0.35662063081008194, "learning_rate": 4.140356360834524e-05, "loss": 0.8735, "step": 5961 }, { "epoch": 2.70018115942029, "grad_norm": 0.3994784672511301, "learning_rate": 4.13909240133414e-05, "loss": 0.8801, "step": 5962 }, { "epoch": 2.7006340579710146, "grad_norm": 0.423245610448463, "learning_rate": 4.1378284279284685e-05, "loss": 0.8576, "step": 5963 }, { "epoch": 2.7010869565217392, "grad_norm": 0.34431691058306335, "learning_rate": 4.136564440743872e-05, "loss": 0.8807, "step": 5964 }, { "epoch": 2.701539855072464, "grad_norm": 0.28275499942374305, "learning_rate": 4.135300439906711e-05, "loss": 0.8658, "step": 5965 }, { "epoch": 2.7019927536231885, "grad_norm": 0.1964880935249965, "learning_rate": 4.134036425543351e-05, "loss": 0.8552, "step": 5966 }, { "epoch": 2.702445652173913, "grad_norm": 0.23448006970311608, "learning_rate": 4.132772397780157e-05, "loss": 0.866, "step": 5967 }, { "epoch": 2.7028985507246377, "grad_norm": 0.3569839262436756, "learning_rate": 4.1315083567434954e-05, "loss": 0.8647, "step": 5968 }, { "epoch": 2.7033514492753623, "grad_norm": 0.4670552008739952, "learning_rate": 4.130244302559736e-05, "loss": 0.8527, "step": 5969 }, { "epoch": 2.703804347826087, "grad_norm": 0.4710862262474254, "learning_rate": 4.128980235355247e-05, "loss": 0.8852, "step": 5970 }, { "epoch": 2.7042572463768115, "grad_norm": 0.3999903035891253, "learning_rate": 4.127716155256399e-05, "loss": 0.8816, "step": 5971 }, { "epoch": 2.704710144927536, "grad_norm": 0.3581910893266623, "learning_rate": 4.126452062389565e-05, "loss": 0.8606, "step": 5972 }, { "epoch": 2.7051630434782608, "grad_norm": 0.2702075733121791, "learning_rate": 4.125187956881117e-05, "loss": 0.8857, "step": 5973 }, { "epoch": 2.7056159420289854, "grad_norm": 0.195829407105793, "learning_rate": 4.123923838857431e-05, "loss": 0.8678, "step": 5974 }, { "epoch": 2.70606884057971, "grad_norm": 0.1973596246750702, "learning_rate": 4.1226597084448834e-05, "loss": 0.8764, "step": 5975 }, { "epoch": 2.7065217391304346, "grad_norm": 0.24430861932566897, "learning_rate": 4.12139556576985e-05, "loss": 0.8714, "step": 5976 }, { "epoch": 2.706974637681159, "grad_norm": 0.2678098849601567, "learning_rate": 4.1201314109587096e-05, "loss": 0.8448, "step": 5977 }, { "epoch": 2.707427536231884, "grad_norm": 0.8729361079780317, "learning_rate": 4.118867244137841e-05, "loss": 0.9024, "step": 5978 }, { "epoch": 2.7078804347826084, "grad_norm": 0.2318049596557114, "learning_rate": 4.117603065433626e-05, "loss": 0.8778, "step": 5979 }, { "epoch": 2.7083333333333335, "grad_norm": 0.27155825409213313, "learning_rate": 4.116338874972446e-05, "loss": 0.8644, "step": 5980 }, { "epoch": 2.708786231884058, "grad_norm": 0.2538571838078644, "learning_rate": 4.1150746728806844e-05, "loss": 0.8876, "step": 5981 }, { "epoch": 2.7092391304347827, "grad_norm": 0.20746786031285108, "learning_rate": 4.113810459284726e-05, "loss": 0.8607, "step": 5982 }, { "epoch": 2.7096920289855073, "grad_norm": 0.2692335764362856, "learning_rate": 4.112546234310955e-05, "loss": 0.8787, "step": 5983 }, { "epoch": 2.710144927536232, "grad_norm": 0.397795881034228, "learning_rate": 4.111281998085758e-05, "loss": 0.8672, "step": 5984 }, { "epoch": 2.7105978260869565, "grad_norm": 0.44515377204845197, "learning_rate": 4.110017750735524e-05, "loss": 0.8954, "step": 5985 }, { "epoch": 2.711050724637681, "grad_norm": 0.36922320159812766, "learning_rate": 4.108753492386642e-05, "loss": 0.8679, "step": 5986 }, { "epoch": 2.7115036231884058, "grad_norm": 0.29236370039163284, "learning_rate": 4.1074892231655e-05, "loss": 0.8825, "step": 5987 }, { "epoch": 2.7119565217391304, "grad_norm": 0.24350744159710336, "learning_rate": 4.10622494319849e-05, "loss": 0.8908, "step": 5988 }, { "epoch": 2.712409420289855, "grad_norm": 0.2533712404054044, "learning_rate": 4.104960652612004e-05, "loss": 0.8636, "step": 5989 }, { "epoch": 2.7128623188405796, "grad_norm": 0.23895089483179305, "learning_rate": 4.103696351532437e-05, "loss": 0.8637, "step": 5990 }, { "epoch": 2.713315217391304, "grad_norm": 0.18815092085143173, "learning_rate": 4.102432040086179e-05, "loss": 0.8608, "step": 5991 }, { "epoch": 2.713768115942029, "grad_norm": 0.25448961946086246, "learning_rate": 4.101167718399628e-05, "loss": 0.8636, "step": 5992 }, { "epoch": 2.714221014492754, "grad_norm": 0.20862363553944094, "learning_rate": 4.099903386599181e-05, "loss": 0.8759, "step": 5993 }, { "epoch": 2.7146739130434785, "grad_norm": 0.20308545633518996, "learning_rate": 4.098639044811234e-05, "loss": 0.8737, "step": 5994 }, { "epoch": 2.715126811594203, "grad_norm": 0.2493025490347, "learning_rate": 4.0973746931621855e-05, "loss": 0.8951, "step": 5995 }, { "epoch": 2.7155797101449277, "grad_norm": 0.2400100927906403, "learning_rate": 4.096110331778434e-05, "loss": 0.8996, "step": 5996 }, { "epoch": 2.7160326086956523, "grad_norm": 0.2689439512849955, "learning_rate": 4.09484596078638e-05, "loss": 0.8659, "step": 5997 }, { "epoch": 2.716485507246377, "grad_norm": 0.20165349462593982, "learning_rate": 4.093581580312427e-05, "loss": 0.8663, "step": 5998 }, { "epoch": 2.7169384057971016, "grad_norm": 0.240130144591981, "learning_rate": 4.092317190482972e-05, "loss": 0.8797, "step": 5999 }, { "epoch": 2.717391304347826, "grad_norm": 0.21322558388577964, "learning_rate": 4.091052791424423e-05, "loss": 0.8808, "step": 6000 }, { "epoch": 2.7178442028985508, "grad_norm": 0.17867752957001018, "learning_rate": 4.089788383263182e-05, "loss": 0.8778, "step": 6001 }, { "epoch": 2.7182971014492754, "grad_norm": 0.24023783182385364, "learning_rate": 4.088523966125654e-05, "loss": 0.8865, "step": 6002 }, { "epoch": 2.71875, "grad_norm": 0.2586605886466044, "learning_rate": 4.087259540138245e-05, "loss": 0.8645, "step": 6003 }, { "epoch": 2.7192028985507246, "grad_norm": 0.2169115390155628, "learning_rate": 4.085995105427361e-05, "loss": 0.8638, "step": 6004 }, { "epoch": 2.7196557971014492, "grad_norm": 0.21110888896636187, "learning_rate": 4.0847306621194096e-05, "loss": 0.8739, "step": 6005 }, { "epoch": 2.720108695652174, "grad_norm": 0.1930200720432791, "learning_rate": 4.0834662103408005e-05, "loss": 0.8968, "step": 6006 }, { "epoch": 2.7205615942028984, "grad_norm": 0.18344044104090448, "learning_rate": 4.082201750217943e-05, "loss": 0.8796, "step": 6007 }, { "epoch": 2.721014492753623, "grad_norm": 0.22897019829265622, "learning_rate": 4.080937281877245e-05, "loss": 0.8901, "step": 6008 }, { "epoch": 2.7214673913043477, "grad_norm": 0.214084999706329, "learning_rate": 4.0796728054451186e-05, "loss": 0.8752, "step": 6009 }, { "epoch": 2.7219202898550723, "grad_norm": 0.1894886471932834, "learning_rate": 4.078408321047977e-05, "loss": 0.8772, "step": 6010 }, { "epoch": 2.722373188405797, "grad_norm": 0.22333909539186983, "learning_rate": 4.0771438288122294e-05, "loss": 0.8723, "step": 6011 }, { "epoch": 2.7228260869565215, "grad_norm": 0.27169422987003267, "learning_rate": 4.075879328864293e-05, "loss": 0.8811, "step": 6012 }, { "epoch": 2.723278985507246, "grad_norm": 0.27577798660449604, "learning_rate": 4.0746148213305784e-05, "loss": 0.8851, "step": 6013 }, { "epoch": 2.723731884057971, "grad_norm": 0.2353394596307959, "learning_rate": 4.073350306337502e-05, "loss": 0.8696, "step": 6014 }, { "epoch": 2.724184782608696, "grad_norm": 0.339933066989283, "learning_rate": 4.072085784011479e-05, "loss": 0.879, "step": 6015 }, { "epoch": 2.7246376811594204, "grad_norm": 0.22339622182506236, "learning_rate": 4.070821254478927e-05, "loss": 0.8731, "step": 6016 }, { "epoch": 2.725090579710145, "grad_norm": 0.24879580032956322, "learning_rate": 4.0695567178662607e-05, "loss": 0.8715, "step": 6017 }, { "epoch": 2.7255434782608696, "grad_norm": 0.30036124076824217, "learning_rate": 4.0682921742999e-05, "loss": 0.8935, "step": 6018 }, { "epoch": 2.7259963768115942, "grad_norm": 0.24936909508692456, "learning_rate": 4.067027623906262e-05, "loss": 0.8543, "step": 6019 }, { "epoch": 2.726449275362319, "grad_norm": 0.2818013791069272, "learning_rate": 4.0657630668117665e-05, "loss": 0.861, "step": 6020 }, { "epoch": 2.7269021739130435, "grad_norm": 0.3039689756980323, "learning_rate": 4.064498503142833e-05, "loss": 0.8603, "step": 6021 }, { "epoch": 2.727355072463768, "grad_norm": 0.2034557187443631, "learning_rate": 4.063233933025881e-05, "loss": 0.8606, "step": 6022 }, { "epoch": 2.7278079710144927, "grad_norm": 0.2504216093937474, "learning_rate": 4.061969356587333e-05, "loss": 0.8748, "step": 6023 }, { "epoch": 2.7282608695652173, "grad_norm": 0.2587645219983987, "learning_rate": 4.060704773953611e-05, "loss": 0.865, "step": 6024 }, { "epoch": 2.728713768115942, "grad_norm": 0.22167511633881296, "learning_rate": 4.059440185251136e-05, "loss": 0.8615, "step": 6025 }, { "epoch": 2.7291666666666665, "grad_norm": 0.2672084613097601, "learning_rate": 4.058175590606332e-05, "loss": 0.8736, "step": 6026 }, { "epoch": 2.7296195652173916, "grad_norm": 0.2654093526139069, "learning_rate": 4.056910990145621e-05, "loss": 0.877, "step": 6027 }, { "epoch": 2.730072463768116, "grad_norm": 0.2467082201537665, "learning_rate": 4.055646383995429e-05, "loss": 0.8973, "step": 6028 }, { "epoch": 2.730525362318841, "grad_norm": 0.23079851837537596, "learning_rate": 4.054381772282179e-05, "loss": 0.8753, "step": 6029 }, { "epoch": 2.7309782608695654, "grad_norm": 0.1657010664745083, "learning_rate": 4.053117155132298e-05, "loss": 0.8714, "step": 6030 }, { "epoch": 2.73143115942029, "grad_norm": 0.23354226716535098, "learning_rate": 4.05185253267221e-05, "loss": 0.8735, "step": 6031 }, { "epoch": 2.7318840579710146, "grad_norm": 0.17753091543830501, "learning_rate": 4.0505879050283424e-05, "loss": 0.8738, "step": 6032 }, { "epoch": 2.7323369565217392, "grad_norm": 0.25549307900688567, "learning_rate": 4.049323272327122e-05, "loss": 0.8813, "step": 6033 }, { "epoch": 2.732789855072464, "grad_norm": 0.2941308769946493, "learning_rate": 4.048058634694976e-05, "loss": 0.8842, "step": 6034 }, { "epoch": 2.7332427536231885, "grad_norm": 0.325784824515614, "learning_rate": 4.046793992258332e-05, "loss": 0.871, "step": 6035 }, { "epoch": 2.733695652173913, "grad_norm": 0.3281297894616176, "learning_rate": 4.045529345143617e-05, "loss": 0.8794, "step": 6036 }, { "epoch": 2.7341485507246377, "grad_norm": 0.25366349666723853, "learning_rate": 4.0442646934772635e-05, "loss": 0.877, "step": 6037 }, { "epoch": 2.7346014492753623, "grad_norm": 0.18840714413897652, "learning_rate": 4.0430000373856976e-05, "loss": 0.8701, "step": 6038 }, { "epoch": 2.735054347826087, "grad_norm": 0.17053459649662736, "learning_rate": 4.04173537699535e-05, "loss": 0.8534, "step": 6039 }, { "epoch": 2.7355072463768115, "grad_norm": 0.1907204850772322, "learning_rate": 4.04047071243265e-05, "loss": 0.8721, "step": 6040 }, { "epoch": 2.735960144927536, "grad_norm": 0.1857413390871107, "learning_rate": 4.039206043824029e-05, "loss": 0.8699, "step": 6041 }, { "epoch": 2.7364130434782608, "grad_norm": 0.19388134141449848, "learning_rate": 4.0379413712959176e-05, "loss": 0.8796, "step": 6042 }, { "epoch": 2.7368659420289854, "grad_norm": 0.22086580919972676, "learning_rate": 4.0366766949747484e-05, "loss": 0.8703, "step": 6043 }, { "epoch": 2.73731884057971, "grad_norm": 0.22366557942918133, "learning_rate": 4.03541201498695e-05, "loss": 0.8851, "step": 6044 }, { "epoch": 2.7377717391304346, "grad_norm": 0.20817378861356084, "learning_rate": 4.034147331458958e-05, "loss": 0.8666, "step": 6045 }, { "epoch": 2.738224637681159, "grad_norm": 0.201927408764328, "learning_rate": 4.0328826445172027e-05, "loss": 0.867, "step": 6046 }, { "epoch": 2.738677536231884, "grad_norm": 0.2715155407801207, "learning_rate": 4.0316179542881164e-05, "loss": 0.8822, "step": 6047 }, { "epoch": 2.7391304347826084, "grad_norm": 0.28605248240929493, "learning_rate": 4.0303532608981336e-05, "loss": 0.8748, "step": 6048 }, { "epoch": 2.7395833333333335, "grad_norm": 0.29800730368246986, "learning_rate": 4.029088564473688e-05, "loss": 0.877, "step": 6049 }, { "epoch": 2.740036231884058, "grad_norm": 0.25433730084225786, "learning_rate": 4.027823865141211e-05, "loss": 0.8875, "step": 6050 }, { "epoch": 2.7404891304347827, "grad_norm": 0.18500383799813241, "learning_rate": 4.0265591630271393e-05, "loss": 0.843, "step": 6051 }, { "epoch": 2.7409420289855073, "grad_norm": 0.27461866296511517, "learning_rate": 4.025294458257905e-05, "loss": 0.8661, "step": 6052 }, { "epoch": 2.741394927536232, "grad_norm": 0.25003468967649195, "learning_rate": 4.024029750959944e-05, "loss": 0.8672, "step": 6053 }, { "epoch": 2.7418478260869565, "grad_norm": 0.20610126016026314, "learning_rate": 4.0227650412596916e-05, "loss": 0.8657, "step": 6054 }, { "epoch": 2.742300724637681, "grad_norm": 0.2867349502446445, "learning_rate": 4.021500329283582e-05, "loss": 0.8747, "step": 6055 }, { "epoch": 2.7427536231884058, "grad_norm": 0.3737821222223535, "learning_rate": 4.020235615158049e-05, "loss": 0.8845, "step": 6056 }, { "epoch": 2.7432065217391304, "grad_norm": 0.3231567279220437, "learning_rate": 4.0189708990095295e-05, "loss": 0.8869, "step": 6057 }, { "epoch": 2.743659420289855, "grad_norm": 0.27627153140839283, "learning_rate": 4.017706180964461e-05, "loss": 0.8768, "step": 6058 }, { "epoch": 2.7441123188405796, "grad_norm": 0.24802546623465738, "learning_rate": 4.016441461149276e-05, "loss": 0.8734, "step": 6059 }, { "epoch": 2.744565217391304, "grad_norm": 0.1824054155825354, "learning_rate": 4.015176739690413e-05, "loss": 0.8602, "step": 6060 }, { "epoch": 2.745018115942029, "grad_norm": 0.2533631606239129, "learning_rate": 4.0139120167143065e-05, "loss": 0.8773, "step": 6061 }, { "epoch": 2.745471014492754, "grad_norm": 0.35590780023377566, "learning_rate": 4.012647292347394e-05, "loss": 0.8776, "step": 6062 }, { "epoch": 2.7459239130434785, "grad_norm": 0.4999226339436973, "learning_rate": 4.0113825667161114e-05, "loss": 0.8691, "step": 6063 }, { "epoch": 2.746376811594203, "grad_norm": 0.5361324896026001, "learning_rate": 4.010117839946896e-05, "loss": 0.8693, "step": 6064 }, { "epoch": 2.7468297101449277, "grad_norm": 0.45179279515314674, "learning_rate": 4.008853112166183e-05, "loss": 0.8933, "step": 6065 }, { "epoch": 2.7472826086956523, "grad_norm": 0.3710155009968634, "learning_rate": 4.0075883835004124e-05, "loss": 0.8723, "step": 6066 }, { "epoch": 2.747735507246377, "grad_norm": 0.2863478423933489, "learning_rate": 4.006323654076017e-05, "loss": 0.8666, "step": 6067 }, { "epoch": 2.7481884057971016, "grad_norm": 0.1842582760721918, "learning_rate": 4.005058924019436e-05, "loss": 0.8731, "step": 6068 }, { "epoch": 2.748641304347826, "grad_norm": 0.2725596915598498, "learning_rate": 4.0037941934571075e-05, "loss": 0.8574, "step": 6069 }, { "epoch": 2.7490942028985508, "grad_norm": 0.3627662782016933, "learning_rate": 4.002529462515468e-05, "loss": 0.8872, "step": 6070 }, { "epoch": 2.7495471014492754, "grad_norm": 0.4272468644903024, "learning_rate": 4.001264731320953e-05, "loss": 0.8458, "step": 6071 }, { "epoch": 2.75, "grad_norm": 0.44109764222086684, "learning_rate": 4e-05, "loss": 0.8832, "step": 6072 }, { "epoch": 2.7504528985507246, "grad_norm": 0.417239250053351, "learning_rate": 3.998735268679049e-05, "loss": 0.8754, "step": 6073 }, { "epoch": 2.7509057971014492, "grad_norm": 0.26888249309928963, "learning_rate": 3.997470537484533e-05, "loss": 0.8666, "step": 6074 }, { "epoch": 2.751358695652174, "grad_norm": 0.20572436038396846, "learning_rate": 3.9962058065428945e-05, "loss": 0.8617, "step": 6075 }, { "epoch": 2.7518115942028984, "grad_norm": 0.3200953644933957, "learning_rate": 3.9949410759805645e-05, "loss": 0.8772, "step": 6076 }, { "epoch": 2.752264492753623, "grad_norm": 0.3878390174280516, "learning_rate": 3.993676345923985e-05, "loss": 0.8843, "step": 6077 }, { "epoch": 2.7527173913043477, "grad_norm": 0.47391376435453103, "learning_rate": 3.992411616499589e-05, "loss": 0.8835, "step": 6078 }, { "epoch": 2.7531702898550723, "grad_norm": 0.5005310285474973, "learning_rate": 3.991146887833818e-05, "loss": 0.842, "step": 6079 }, { "epoch": 2.753623188405797, "grad_norm": 0.3947351688568495, "learning_rate": 3.9898821600531054e-05, "loss": 0.892, "step": 6080 }, { "epoch": 2.7540760869565215, "grad_norm": 0.2971704373223171, "learning_rate": 3.9886174332838906e-05, "loss": 0.8653, "step": 6081 }, { "epoch": 2.754528985507246, "grad_norm": 0.25483740971644686, "learning_rate": 3.987352707652607e-05, "loss": 0.8727, "step": 6082 }, { "epoch": 2.754981884057971, "grad_norm": 0.25514121619277613, "learning_rate": 3.9860879832856955e-05, "loss": 0.8788, "step": 6083 }, { "epoch": 2.755434782608696, "grad_norm": 0.20731366314382674, "learning_rate": 3.984823260309589e-05, "loss": 0.8829, "step": 6084 }, { "epoch": 2.7558876811594204, "grad_norm": 0.3466735163985649, "learning_rate": 3.983558538850726e-05, "loss": 0.888, "step": 6085 }, { "epoch": 2.756340579710145, "grad_norm": 0.45681434174174546, "learning_rate": 3.98229381903554e-05, "loss": 0.8628, "step": 6086 }, { "epoch": 2.7567934782608696, "grad_norm": 0.33854324683146153, "learning_rate": 3.981029100990471e-05, "loss": 0.8853, "step": 6087 }, { "epoch": 2.7572463768115942, "grad_norm": 0.30870577179822334, "learning_rate": 3.9797643848419524e-05, "loss": 0.8845, "step": 6088 }, { "epoch": 2.757699275362319, "grad_norm": 0.27275063579931097, "learning_rate": 3.978499670716421e-05, "loss": 0.8912, "step": 6089 }, { "epoch": 2.7581521739130435, "grad_norm": 0.27662412967035616, "learning_rate": 3.977234958740309e-05, "loss": 0.8706, "step": 6090 }, { "epoch": 2.758605072463768, "grad_norm": 0.37374426215965534, "learning_rate": 3.975970249040057e-05, "loss": 0.8927, "step": 6091 }, { "epoch": 2.7590579710144927, "grad_norm": 0.2341792892296097, "learning_rate": 3.974705541742096e-05, "loss": 0.8731, "step": 6092 }, { "epoch": 2.7595108695652173, "grad_norm": 0.3220695258124783, "learning_rate": 3.973440836972863e-05, "loss": 0.8727, "step": 6093 }, { "epoch": 2.759963768115942, "grad_norm": 0.23588185677204138, "learning_rate": 3.97217613485879e-05, "loss": 0.8569, "step": 6094 }, { "epoch": 2.7604166666666665, "grad_norm": 0.2649404457041557, "learning_rate": 3.970911435526314e-05, "loss": 0.8861, "step": 6095 }, { "epoch": 2.7608695652173916, "grad_norm": 0.266751660905497, "learning_rate": 3.969646739101868e-05, "loss": 0.8786, "step": 6096 }, { "epoch": 2.761322463768116, "grad_norm": 0.24097632621231693, "learning_rate": 3.9683820457118856e-05, "loss": 0.8717, "step": 6097 }, { "epoch": 2.761775362318841, "grad_norm": 0.2527395386173191, "learning_rate": 3.9671173554827994e-05, "loss": 0.8472, "step": 6098 }, { "epoch": 2.7622282608695654, "grad_norm": 0.24269494668558741, "learning_rate": 3.965852668541044e-05, "loss": 0.87, "step": 6099 }, { "epoch": 2.76268115942029, "grad_norm": 0.2669232658006028, "learning_rate": 3.964587985013051e-05, "loss": 0.8675, "step": 6100 }, { "epoch": 2.7631340579710146, "grad_norm": 0.23035736750388186, "learning_rate": 3.963323305025254e-05, "loss": 0.8638, "step": 6101 }, { "epoch": 2.7635869565217392, "grad_norm": 0.25981514671422984, "learning_rate": 3.962058628704083e-05, "loss": 0.9038, "step": 6102 }, { "epoch": 2.764039855072464, "grad_norm": 0.33511749137316044, "learning_rate": 3.960793956175972e-05, "loss": 0.8815, "step": 6103 }, { "epoch": 2.7644927536231885, "grad_norm": 0.301360455900483, "learning_rate": 3.959529287567351e-05, "loss": 0.8765, "step": 6104 }, { "epoch": 2.764945652173913, "grad_norm": 0.24109266780142186, "learning_rate": 3.958264623004653e-05, "loss": 0.8714, "step": 6105 }, { "epoch": 2.7653985507246377, "grad_norm": 0.22569014472926718, "learning_rate": 3.956999962614304e-05, "loss": 0.8828, "step": 6106 }, { "epoch": 2.7658514492753623, "grad_norm": 0.24365805753550554, "learning_rate": 3.955735306522738e-05, "loss": 0.841, "step": 6107 }, { "epoch": 2.766304347826087, "grad_norm": 0.2176313173655609, "learning_rate": 3.9544706548563837e-05, "loss": 0.8718, "step": 6108 }, { "epoch": 2.7667572463768115, "grad_norm": 0.26209304938477235, "learning_rate": 3.953206007741671e-05, "loss": 0.8743, "step": 6109 }, { "epoch": 2.767210144927536, "grad_norm": 0.28739284935070847, "learning_rate": 3.9519413653050255e-05, "loss": 0.8653, "step": 6110 }, { "epoch": 2.7676630434782608, "grad_norm": 0.23740118553513512, "learning_rate": 3.95067672767288e-05, "loss": 0.8822, "step": 6111 }, { "epoch": 2.7681159420289854, "grad_norm": 0.2677661033298586, "learning_rate": 3.949412094971659e-05, "loss": 0.878, "step": 6112 }, { "epoch": 2.76856884057971, "grad_norm": 0.23885879398018445, "learning_rate": 3.948147467327792e-05, "loss": 0.8734, "step": 6113 }, { "epoch": 2.7690217391304346, "grad_norm": 0.21882323058940287, "learning_rate": 3.946882844867703e-05, "loss": 0.8488, "step": 6114 }, { "epoch": 2.769474637681159, "grad_norm": 0.15942318630495803, "learning_rate": 3.945618227717823e-05, "loss": 0.8702, "step": 6115 }, { "epoch": 2.769927536231884, "grad_norm": 0.2483036223516976, "learning_rate": 3.944353616004572e-05, "loss": 0.8653, "step": 6116 }, { "epoch": 2.7703804347826084, "grad_norm": 0.2218405759889766, "learning_rate": 3.9430890098543815e-05, "loss": 0.8498, "step": 6117 }, { "epoch": 2.7708333333333335, "grad_norm": 0.20569138874414505, "learning_rate": 3.9418244093936694e-05, "loss": 0.8923, "step": 6118 }, { "epoch": 2.771286231884058, "grad_norm": 0.26761817856949743, "learning_rate": 3.940559814748866e-05, "loss": 0.8434, "step": 6119 }, { "epoch": 2.7717391304347827, "grad_norm": 0.22893531884876608, "learning_rate": 3.93929522604639e-05, "loss": 0.8484, "step": 6120 }, { "epoch": 2.7721920289855073, "grad_norm": 0.1855631360326568, "learning_rate": 3.9380306434126675e-05, "loss": 0.8638, "step": 6121 }, { "epoch": 2.772644927536232, "grad_norm": 0.2779363103052644, "learning_rate": 3.9367660669741196e-05, "loss": 0.8679, "step": 6122 }, { "epoch": 2.7730978260869565, "grad_norm": 0.2705760993396201, "learning_rate": 3.9355014968571696e-05, "loss": 0.8652, "step": 6123 }, { "epoch": 2.773550724637681, "grad_norm": 0.28507523689026704, "learning_rate": 3.934236933188235e-05, "loss": 0.8492, "step": 6124 }, { "epoch": 2.7740036231884058, "grad_norm": 0.29654697664479684, "learning_rate": 3.9329723760937395e-05, "loss": 0.865, "step": 6125 }, { "epoch": 2.7744565217391304, "grad_norm": 0.3262360958465348, "learning_rate": 3.9317078257001014e-05, "loss": 0.8641, "step": 6126 }, { "epoch": 2.774909420289855, "grad_norm": 0.266733408890506, "learning_rate": 3.9304432821337414e-05, "loss": 0.8637, "step": 6127 }, { "epoch": 2.7753623188405796, "grad_norm": 0.1838661169977412, "learning_rate": 3.9291787455210746e-05, "loss": 0.8571, "step": 6128 }, { "epoch": 2.775815217391304, "grad_norm": 0.17652439086024158, "learning_rate": 3.9279142159885224e-05, "loss": 0.8664, "step": 6129 }, { "epoch": 2.776268115942029, "grad_norm": 0.20396003226695736, "learning_rate": 3.9266496936624994e-05, "loss": 0.8984, "step": 6130 }, { "epoch": 2.776721014492754, "grad_norm": 0.20232933914546528, "learning_rate": 3.9253851786694236e-05, "loss": 0.8763, "step": 6131 }, { "epoch": 2.7771739130434785, "grad_norm": 0.21723917903303197, "learning_rate": 3.924120671135708e-05, "loss": 0.8783, "step": 6132 }, { "epoch": 2.777626811594203, "grad_norm": 0.22901117953541253, "learning_rate": 3.922856171187771e-05, "loss": 0.867, "step": 6133 }, { "epoch": 2.7780797101449277, "grad_norm": 0.2104646237569141, "learning_rate": 3.9215916789520246e-05, "loss": 0.8557, "step": 6134 }, { "epoch": 2.7785326086956523, "grad_norm": 0.1830361200918725, "learning_rate": 3.920327194554883e-05, "loss": 0.879, "step": 6135 }, { "epoch": 2.778985507246377, "grad_norm": 0.18788516859111876, "learning_rate": 3.919062718122755e-05, "loss": 0.8916, "step": 6136 }, { "epoch": 2.7794384057971016, "grad_norm": 0.17264428248578315, "learning_rate": 3.917798249782058e-05, "loss": 0.8489, "step": 6137 }, { "epoch": 2.779891304347826, "grad_norm": 0.20270952082283616, "learning_rate": 3.9165337896591995e-05, "loss": 0.8717, "step": 6138 }, { "epoch": 2.7803442028985508, "grad_norm": 0.20685920541411718, "learning_rate": 3.915269337880591e-05, "loss": 0.8762, "step": 6139 }, { "epoch": 2.7807971014492754, "grad_norm": 0.1956070427951778, "learning_rate": 3.914004894572639e-05, "loss": 0.8663, "step": 6140 }, { "epoch": 2.78125, "grad_norm": 0.2436828300081023, "learning_rate": 3.912740459861756e-05, "loss": 0.8721, "step": 6141 }, { "epoch": 2.7817028985507246, "grad_norm": 0.30033605747531755, "learning_rate": 3.911476033874347e-05, "loss": 0.8919, "step": 6142 }, { "epoch": 2.7821557971014492, "grad_norm": 0.22755123494055876, "learning_rate": 3.91021161673682e-05, "loss": 0.8518, "step": 6143 }, { "epoch": 2.782608695652174, "grad_norm": 0.18569996592280477, "learning_rate": 3.908947208575577e-05, "loss": 0.8684, "step": 6144 }, { "epoch": 2.7830615942028984, "grad_norm": 0.24077264959654843, "learning_rate": 3.9076828095170286e-05, "loss": 0.8627, "step": 6145 }, { "epoch": 2.783514492753623, "grad_norm": 0.21385670111223995, "learning_rate": 3.9064184196875753e-05, "loss": 0.8804, "step": 6146 }, { "epoch": 2.7839673913043477, "grad_norm": 0.21783920291074702, "learning_rate": 3.9051540392136213e-05, "loss": 0.8813, "step": 6147 }, { "epoch": 2.7844202898550723, "grad_norm": 0.2112925023273617, "learning_rate": 3.903889668221567e-05, "loss": 0.894, "step": 6148 }, { "epoch": 2.784873188405797, "grad_norm": 0.18600211603709374, "learning_rate": 3.902625306837816e-05, "loss": 0.8578, "step": 6149 }, { "epoch": 2.7853260869565215, "grad_norm": 0.22228823360351604, "learning_rate": 3.901360955188767e-05, "loss": 0.8904, "step": 6150 }, { "epoch": 2.785778985507246, "grad_norm": 0.1895794683608238, "learning_rate": 3.9000966134008203e-05, "loss": 0.8886, "step": 6151 }, { "epoch": 2.786231884057971, "grad_norm": 0.21695077901639173, "learning_rate": 3.898832281600371e-05, "loss": 0.8598, "step": 6152 }, { "epoch": 2.786684782608696, "grad_norm": 0.2826352591176207, "learning_rate": 3.897567959913822e-05, "loss": 0.8712, "step": 6153 }, { "epoch": 2.7871376811594204, "grad_norm": 0.2800008668941422, "learning_rate": 3.8963036484675645e-05, "loss": 0.8764, "step": 6154 }, { "epoch": 2.787590579710145, "grad_norm": 0.25189819302717636, "learning_rate": 3.895039347387997e-05, "loss": 0.8758, "step": 6155 }, { "epoch": 2.7880434782608696, "grad_norm": 0.2550714748613504, "learning_rate": 3.8937750568015106e-05, "loss": 0.8784, "step": 6156 }, { "epoch": 2.7884963768115942, "grad_norm": 0.23195712625532705, "learning_rate": 3.892510776834501e-05, "loss": 0.8714, "step": 6157 }, { "epoch": 2.788949275362319, "grad_norm": 0.18605293064146658, "learning_rate": 3.891246507613359e-05, "loss": 0.8699, "step": 6158 }, { "epoch": 2.7894021739130435, "grad_norm": 0.19750060548000212, "learning_rate": 3.889982249264477e-05, "loss": 0.877, "step": 6159 }, { "epoch": 2.789855072463768, "grad_norm": 0.21963084400760452, "learning_rate": 3.888718001914242e-05, "loss": 0.873, "step": 6160 }, { "epoch": 2.7903079710144927, "grad_norm": 0.19801094599644153, "learning_rate": 3.8874537656890466e-05, "loss": 0.8722, "step": 6161 }, { "epoch": 2.7907608695652173, "grad_norm": 0.2788412623623372, "learning_rate": 3.886189540715275e-05, "loss": 0.8638, "step": 6162 }, { "epoch": 2.791213768115942, "grad_norm": 0.2926352471711787, "learning_rate": 3.884925327119316e-05, "loss": 0.8736, "step": 6163 }, { "epoch": 2.7916666666666665, "grad_norm": 0.28905104101555384, "learning_rate": 3.8836611250275546e-05, "loss": 0.8475, "step": 6164 }, { "epoch": 2.7921195652173916, "grad_norm": 0.22969344241806006, "learning_rate": 3.882396934566375e-05, "loss": 0.8786, "step": 6165 }, { "epoch": 2.792572463768116, "grad_norm": 0.17351304738068316, "learning_rate": 3.8811327558621596e-05, "loss": 0.868, "step": 6166 }, { "epoch": 2.793025362318841, "grad_norm": 0.2295220604483657, "learning_rate": 3.879868589041292e-05, "loss": 0.8869, "step": 6167 }, { "epoch": 2.7934782608695654, "grad_norm": 0.23733405398146817, "learning_rate": 3.878604434230151e-05, "loss": 0.8662, "step": 6168 }, { "epoch": 2.79393115942029, "grad_norm": 0.22445929534447104, "learning_rate": 3.877340291555118e-05, "loss": 0.8554, "step": 6169 }, { "epoch": 2.7943840579710146, "grad_norm": 0.2081444944495309, "learning_rate": 3.876076161142568e-05, "loss": 0.8751, "step": 6170 }, { "epoch": 2.7948369565217392, "grad_norm": 0.15957983653908175, "learning_rate": 3.8748120431188834e-05, "loss": 0.8691, "step": 6171 }, { "epoch": 2.795289855072464, "grad_norm": 0.19670882590352018, "learning_rate": 3.873547937610436e-05, "loss": 0.8818, "step": 6172 }, { "epoch": 2.7957427536231885, "grad_norm": 0.20879057205840346, "learning_rate": 3.8722838447436025e-05, "loss": 0.849, "step": 6173 }, { "epoch": 2.796195652173913, "grad_norm": 0.2705991220483276, "learning_rate": 3.8710197646447535e-05, "loss": 0.8478, "step": 6174 }, { "epoch": 2.7966485507246377, "grad_norm": 0.28959143189001973, "learning_rate": 3.8697556974402653e-05, "loss": 0.8774, "step": 6175 }, { "epoch": 2.7971014492753623, "grad_norm": 0.22679591584107991, "learning_rate": 3.868491643256505e-05, "loss": 0.8847, "step": 6176 }, { "epoch": 2.797554347826087, "grad_norm": 0.21174958084342707, "learning_rate": 3.867227602219844e-05, "loss": 0.8715, "step": 6177 }, { "epoch": 2.7980072463768115, "grad_norm": 0.1668892840862616, "learning_rate": 3.86596357445665e-05, "loss": 0.8684, "step": 6178 }, { "epoch": 2.798460144927536, "grad_norm": 0.22127550399868975, "learning_rate": 3.8646995600932895e-05, "loss": 0.872, "step": 6179 }, { "epoch": 2.7989130434782608, "grad_norm": 0.27335567754160833, "learning_rate": 3.8634355592561286e-05, "loss": 0.8626, "step": 6180 }, { "epoch": 2.7993659420289854, "grad_norm": 0.2904285389678149, "learning_rate": 3.862171572071532e-05, "loss": 0.8787, "step": 6181 }, { "epoch": 2.79981884057971, "grad_norm": 0.28244018580848124, "learning_rate": 3.86090759866586e-05, "loss": 0.8625, "step": 6182 }, { "epoch": 2.8002717391304346, "grad_norm": 0.23418329596485202, "learning_rate": 3.8596436391654774e-05, "loss": 0.8684, "step": 6183 }, { "epoch": 2.800724637681159, "grad_norm": 0.243979670842854, "learning_rate": 3.858379693696742e-05, "loss": 0.8896, "step": 6184 }, { "epoch": 2.801177536231884, "grad_norm": 0.22967296544784035, "learning_rate": 3.857115762386014e-05, "loss": 0.8587, "step": 6185 }, { "epoch": 2.8016304347826084, "grad_norm": 0.18857686877328259, "learning_rate": 3.8558518453596466e-05, "loss": 0.8607, "step": 6186 }, { "epoch": 2.8020833333333335, "grad_norm": 0.1610279098560916, "learning_rate": 3.854587942744002e-05, "loss": 0.8547, "step": 6187 }, { "epoch": 2.802536231884058, "grad_norm": 0.16337578749956536, "learning_rate": 3.85332405466543e-05, "loss": 0.8814, "step": 6188 }, { "epoch": 2.8029891304347827, "grad_norm": 0.15771290175328492, "learning_rate": 3.8520601812502844e-05, "loss": 0.8591, "step": 6189 }, { "epoch": 2.8034420289855073, "grad_norm": 0.18463106744773675, "learning_rate": 3.8507963226249165e-05, "loss": 0.8883, "step": 6190 }, { "epoch": 2.803894927536232, "grad_norm": 0.27652661427774017, "learning_rate": 3.849532478915678e-05, "loss": 0.8697, "step": 6191 }, { "epoch": 2.8043478260869565, "grad_norm": 0.28852251837238596, "learning_rate": 3.848268650248915e-05, "loss": 0.8765, "step": 6192 }, { "epoch": 2.804800724637681, "grad_norm": 0.22566652410878965, "learning_rate": 3.847004836750976e-05, "loss": 0.8801, "step": 6193 }, { "epoch": 2.8052536231884058, "grad_norm": 0.16767125978608755, "learning_rate": 3.845741038548205e-05, "loss": 0.8628, "step": 6194 }, { "epoch": 2.8057065217391304, "grad_norm": 0.18569537592984892, "learning_rate": 3.844477255766948e-05, "loss": 0.8762, "step": 6195 }, { "epoch": 2.806159420289855, "grad_norm": 0.2652692744839217, "learning_rate": 3.843213488533545e-05, "loss": 0.8794, "step": 6196 }, { "epoch": 2.8066123188405796, "grad_norm": 0.33624616804901036, "learning_rate": 3.84194973697434e-05, "loss": 0.898, "step": 6197 }, { "epoch": 2.807065217391304, "grad_norm": 0.33191376594429206, "learning_rate": 3.8406860012156695e-05, "loss": 0.8585, "step": 6198 }, { "epoch": 2.807518115942029, "grad_norm": 0.32589708306927245, "learning_rate": 3.83942228138387e-05, "loss": 0.8578, "step": 6199 }, { "epoch": 2.807971014492754, "grad_norm": 0.2465254579964102, "learning_rate": 3.8381585776052824e-05, "loss": 0.8632, "step": 6200 }, { "epoch": 2.8084239130434785, "grad_norm": 0.16948523299176266, "learning_rate": 3.836894890006236e-05, "loss": 0.8816, "step": 6201 }, { "epoch": 2.808876811594203, "grad_norm": 0.2431822597731178, "learning_rate": 3.8356312187130676e-05, "loss": 0.855, "step": 6202 }, { "epoch": 2.8093297101449277, "grad_norm": 0.30315072377842683, "learning_rate": 3.8343675638521044e-05, "loss": 0.8875, "step": 6203 }, { "epoch": 2.8097826086956523, "grad_norm": 0.3282253183541544, "learning_rate": 3.8331039255496805e-05, "loss": 0.8473, "step": 6204 }, { "epoch": 2.810235507246377, "grad_norm": 0.3355753519856146, "learning_rate": 3.83184030393212e-05, "loss": 0.8728, "step": 6205 }, { "epoch": 2.8106884057971016, "grad_norm": 0.3374500856705175, "learning_rate": 3.830576699125753e-05, "loss": 0.8663, "step": 6206 }, { "epoch": 2.811141304347826, "grad_norm": 0.2816452910588511, "learning_rate": 3.8293131112568984e-05, "loss": 0.8794, "step": 6207 }, { "epoch": 2.8115942028985508, "grad_norm": 0.24826345944437994, "learning_rate": 3.828049540451886e-05, "loss": 0.8994, "step": 6208 }, { "epoch": 2.8120471014492754, "grad_norm": 0.2791075362643782, "learning_rate": 3.8267859868370305e-05, "loss": 0.875, "step": 6209 }, { "epoch": 2.8125, "grad_norm": 0.2957669580609672, "learning_rate": 3.825522450538657e-05, "loss": 0.8713, "step": 6210 }, { "epoch": 2.8129528985507246, "grad_norm": 0.23285226022161468, "learning_rate": 3.824258931683079e-05, "loss": 0.8607, "step": 6211 }, { "epoch": 2.8134057971014492, "grad_norm": 0.16699397954458559, "learning_rate": 3.8229954303966154e-05, "loss": 0.8765, "step": 6212 }, { "epoch": 2.813858695652174, "grad_norm": 0.2333344029086485, "learning_rate": 3.821731946805577e-05, "loss": 0.8732, "step": 6213 }, { "epoch": 2.8143115942028984, "grad_norm": 0.31633741293688594, "learning_rate": 3.820468481036281e-05, "loss": 0.8555, "step": 6214 }, { "epoch": 2.814764492753623, "grad_norm": 0.3912263806582276, "learning_rate": 3.8192050332150336e-05, "loss": 0.8726, "step": 6215 }, { "epoch": 2.8152173913043477, "grad_norm": 0.47475207535804653, "learning_rate": 3.8179416034681465e-05, "loss": 0.882, "step": 6216 }, { "epoch": 2.8156702898550723, "grad_norm": 0.5077006850794294, "learning_rate": 3.8166781919219246e-05, "loss": 0.8828, "step": 6217 }, { "epoch": 2.816123188405797, "grad_norm": 0.3985700149783064, "learning_rate": 3.815414798702676e-05, "loss": 0.8832, "step": 6218 }, { "epoch": 2.8165760869565215, "grad_norm": 0.23664575513805997, "learning_rate": 3.814151423936702e-05, "loss": 0.8691, "step": 6219 }, { "epoch": 2.817028985507246, "grad_norm": 0.20532716148962366, "learning_rate": 3.812888067750304e-05, "loss": 0.8764, "step": 6220 }, { "epoch": 2.817481884057971, "grad_norm": 0.2606355789431962, "learning_rate": 3.8116247302697816e-05, "loss": 0.8527, "step": 6221 }, { "epoch": 2.817934782608696, "grad_norm": 0.3689460127008861, "learning_rate": 3.8103614116214354e-05, "loss": 0.873, "step": 6222 }, { "epoch": 2.8183876811594204, "grad_norm": 0.4947916151222166, "learning_rate": 3.809098111931558e-05, "loss": 0.8551, "step": 6223 }, { "epoch": 2.818840579710145, "grad_norm": 0.5211386989233109, "learning_rate": 3.807834831326446e-05, "loss": 0.8572, "step": 6224 }, { "epoch": 2.8192934782608696, "grad_norm": 0.39119330077648495, "learning_rate": 3.8065715699323884e-05, "loss": 0.8759, "step": 6225 }, { "epoch": 2.8197463768115942, "grad_norm": 0.22689536656480627, "learning_rate": 3.80530832787568e-05, "loss": 0.8771, "step": 6226 }, { "epoch": 2.820199275362319, "grad_norm": 0.1779525621957458, "learning_rate": 3.804045105282605e-05, "loss": 0.8926, "step": 6227 }, { "epoch": 2.8206521739130435, "grad_norm": 0.33930793464565506, "learning_rate": 3.8027819022794525e-05, "loss": 0.8627, "step": 6228 }, { "epoch": 2.821105072463768, "grad_norm": 0.49140767998969953, "learning_rate": 3.801518718992504e-05, "loss": 0.8744, "step": 6229 }, { "epoch": 2.8215579710144927, "grad_norm": 0.5426478370393489, "learning_rate": 3.800255555548045e-05, "loss": 0.8747, "step": 6230 }, { "epoch": 2.8220108695652173, "grad_norm": 0.4168825270640996, "learning_rate": 3.798992412072354e-05, "loss": 0.8843, "step": 6231 }, { "epoch": 2.822463768115942, "grad_norm": 0.2534512001486609, "learning_rate": 3.797729288691711e-05, "loss": 0.8604, "step": 6232 }, { "epoch": 2.8229166666666665, "grad_norm": 0.20898101923203447, "learning_rate": 3.79646618553239e-05, "loss": 0.8829, "step": 6233 }, { "epoch": 2.8233695652173916, "grad_norm": 0.2820786821947322, "learning_rate": 3.795203102720669e-05, "loss": 0.8714, "step": 6234 }, { "epoch": 2.823822463768116, "grad_norm": 0.3688074283722897, "learning_rate": 3.7939400403828167e-05, "loss": 0.8722, "step": 6235 }, { "epoch": 2.824275362318841, "grad_norm": 0.44917126467356916, "learning_rate": 3.792676998645106e-05, "loss": 0.8909, "step": 6236 }, { "epoch": 2.8247282608695654, "grad_norm": 0.4065781743692406, "learning_rate": 3.791413977633803e-05, "loss": 0.8515, "step": 6237 }, { "epoch": 2.82518115942029, "grad_norm": 0.215894287847927, "learning_rate": 3.790150977475177e-05, "loss": 0.8672, "step": 6238 }, { "epoch": 2.8256340579710146, "grad_norm": 0.20368892281369638, "learning_rate": 3.788887998295489e-05, "loss": 0.8512, "step": 6239 }, { "epoch": 2.8260869565217392, "grad_norm": 0.26534398326121134, "learning_rate": 3.7876250402210036e-05, "loss": 0.861, "step": 6240 }, { "epoch": 2.826539855072464, "grad_norm": 0.23882426353457467, "learning_rate": 3.7863621033779774e-05, "loss": 0.8739, "step": 6241 }, { "epoch": 2.8269927536231885, "grad_norm": 0.27072798329034375, "learning_rate": 3.785099187892672e-05, "loss": 0.8628, "step": 6242 }, { "epoch": 2.827445652173913, "grad_norm": 0.1759053679874751, "learning_rate": 3.7838362938913406e-05, "loss": 0.8618, "step": 6243 }, { "epoch": 2.8278985507246377, "grad_norm": 0.24730101495957185, "learning_rate": 3.7825734215002386e-05, "loss": 0.8726, "step": 6244 }, { "epoch": 2.8283514492753623, "grad_norm": 0.34425188078051655, "learning_rate": 3.7813105708456134e-05, "loss": 0.8704, "step": 6245 }, { "epoch": 2.828804347826087, "grad_norm": 0.31860996404481895, "learning_rate": 3.78004774205372e-05, "loss": 0.861, "step": 6246 }, { "epoch": 2.8292572463768115, "grad_norm": 0.26362501000713545, "learning_rate": 3.778784935250801e-05, "loss": 0.8719, "step": 6247 }, { "epoch": 2.829710144927536, "grad_norm": 0.19910190412247222, "learning_rate": 3.7775221505631034e-05, "loss": 0.8771, "step": 6248 }, { "epoch": 2.8301630434782608, "grad_norm": 0.19672941872710925, "learning_rate": 3.776259388116867e-05, "loss": 0.8634, "step": 6249 }, { "epoch": 2.8306159420289854, "grad_norm": 0.2065275721420361, "learning_rate": 3.774996648038336e-05, "loss": 0.8571, "step": 6250 }, { "epoch": 2.83106884057971, "grad_norm": 0.22289050540455935, "learning_rate": 3.773733930453745e-05, "loss": 0.8443, "step": 6251 }, { "epoch": 2.8315217391304346, "grad_norm": 0.18839837060289857, "learning_rate": 3.7724712354893335e-05, "loss": 0.8725, "step": 6252 }, { "epoch": 2.831974637681159, "grad_norm": 0.23322232953795355, "learning_rate": 3.771208563271332e-05, "loss": 0.8945, "step": 6253 }, { "epoch": 2.832427536231884, "grad_norm": 0.14469210094748552, "learning_rate": 3.769945913925974e-05, "loss": 0.8606, "step": 6254 }, { "epoch": 2.8328804347826084, "grad_norm": 0.21724271117989946, "learning_rate": 3.768683287579486e-05, "loss": 0.8468, "step": 6255 }, { "epoch": 2.8333333333333335, "grad_norm": 0.18969665016417428, "learning_rate": 3.767420684358097e-05, "loss": 0.8694, "step": 6256 }, { "epoch": 2.833786231884058, "grad_norm": 0.15979092541143333, "learning_rate": 3.766158104388031e-05, "loss": 0.8648, "step": 6257 }, { "epoch": 2.8342391304347827, "grad_norm": 0.20486687432883866, "learning_rate": 3.76489554779551e-05, "loss": 0.8733, "step": 6258 }, { "epoch": 2.8346920289855073, "grad_norm": 0.1811711406824404, "learning_rate": 3.763633014706752e-05, "loss": 0.8889, "step": 6259 }, { "epoch": 2.835144927536232, "grad_norm": 0.2542719196303944, "learning_rate": 3.762370505247978e-05, "loss": 0.8747, "step": 6260 }, { "epoch": 2.8355978260869565, "grad_norm": 0.2861311950628597, "learning_rate": 3.7611080195454006e-05, "loss": 0.882, "step": 6261 }, { "epoch": 2.836050724637681, "grad_norm": 0.3160185826322148, "learning_rate": 3.7598455577252336e-05, "loss": 0.8787, "step": 6262 }, { "epoch": 2.8365036231884058, "grad_norm": 0.32590299147433577, "learning_rate": 3.758583119913685e-05, "loss": 0.8833, "step": 6263 }, { "epoch": 2.8369565217391304, "grad_norm": 0.18295386012513076, "learning_rate": 3.757320706236966e-05, "loss": 0.8866, "step": 6264 }, { "epoch": 2.837409420289855, "grad_norm": 0.2586891973847998, "learning_rate": 3.75605831682128e-05, "loss": 0.8544, "step": 6265 }, { "epoch": 2.8378623188405796, "grad_norm": 0.2147917579310473, "learning_rate": 3.754795951792832e-05, "loss": 0.8909, "step": 6266 }, { "epoch": 2.838315217391304, "grad_norm": 0.30724941081486673, "learning_rate": 3.7535336112778186e-05, "loss": 0.8531, "step": 6267 }, { "epoch": 2.838768115942029, "grad_norm": 0.3655113262735624, "learning_rate": 3.752271295402443e-05, "loss": 0.9041, "step": 6268 }, { "epoch": 2.839221014492754, "grad_norm": 0.41325450350056164, "learning_rate": 3.751009004292897e-05, "loss": 0.8873, "step": 6269 }, { "epoch": 2.8396739130434785, "grad_norm": 0.33592728669434024, "learning_rate": 3.7497467380753766e-05, "loss": 0.8726, "step": 6270 }, { "epoch": 2.840126811594203, "grad_norm": 0.2134536487872839, "learning_rate": 3.7484844968760693e-05, "loss": 0.875, "step": 6271 }, { "epoch": 2.8405797101449277, "grad_norm": 0.19627718589732301, "learning_rate": 3.7472222808211676e-05, "loss": 0.8847, "step": 6272 }, { "epoch": 2.8410326086956523, "grad_norm": 0.1796535897137677, "learning_rate": 3.745960090036854e-05, "loss": 0.8806, "step": 6273 }, { "epoch": 2.841485507246377, "grad_norm": 0.2358273251360706, "learning_rate": 3.7446979246493145e-05, "loss": 0.9047, "step": 6274 }, { "epoch": 2.8419384057971016, "grad_norm": 0.2597153999170818, "learning_rate": 3.743435784784725e-05, "loss": 0.8757, "step": 6275 }, { "epoch": 2.842391304347826, "grad_norm": 0.2298166596493873, "learning_rate": 3.742173670569269e-05, "loss": 0.8827, "step": 6276 }, { "epoch": 2.8428442028985508, "grad_norm": 0.19494435362178367, "learning_rate": 3.7409115821291184e-05, "loss": 0.8681, "step": 6277 }, { "epoch": 2.8432971014492754, "grad_norm": 0.21088759359700956, "learning_rate": 3.739649519590449e-05, "loss": 0.8954, "step": 6278 }, { "epoch": 2.84375, "grad_norm": 0.17994929987498015, "learning_rate": 3.738387483079428e-05, "loss": 0.8802, "step": 6279 }, { "epoch": 2.8442028985507246, "grad_norm": 0.2069091324635064, "learning_rate": 3.737125472722226e-05, "loss": 0.8787, "step": 6280 }, { "epoch": 2.8446557971014492, "grad_norm": 0.23255043879998252, "learning_rate": 3.7358634886450064e-05, "loss": 0.8604, "step": 6281 }, { "epoch": 2.845108695652174, "grad_norm": 0.202606912990747, "learning_rate": 3.7346015309739334e-05, "loss": 0.8657, "step": 6282 }, { "epoch": 2.8455615942028984, "grad_norm": 0.2769371981424688, "learning_rate": 3.733339599835164e-05, "loss": 0.8868, "step": 6283 }, { "epoch": 2.846014492753623, "grad_norm": 0.2565584848152109, "learning_rate": 3.732077695354859e-05, "loss": 0.8864, "step": 6284 }, { "epoch": 2.8464673913043477, "grad_norm": 0.17684187898549555, "learning_rate": 3.7308158176591714e-05, "loss": 0.8838, "step": 6285 }, { "epoch": 2.8469202898550723, "grad_norm": 0.2095668893567001, "learning_rate": 3.729553966874253e-05, "loss": 0.878, "step": 6286 }, { "epoch": 2.847373188405797, "grad_norm": 0.19939563156438778, "learning_rate": 3.7282921431262515e-05, "loss": 0.883, "step": 6287 }, { "epoch": 2.8478260869565215, "grad_norm": 0.22641330126096718, "learning_rate": 3.727030346541317e-05, "loss": 0.8819, "step": 6288 }, { "epoch": 2.848278985507246, "grad_norm": 0.22681186315179977, "learning_rate": 3.72576857724559e-05, "loss": 0.8785, "step": 6289 }, { "epoch": 2.848731884057971, "grad_norm": 0.2691185859508474, "learning_rate": 3.724506835365214e-05, "loss": 0.9009, "step": 6290 }, { "epoch": 2.849184782608696, "grad_norm": 0.2805808024053036, "learning_rate": 3.723245121026325e-05, "loss": 0.8903, "step": 6291 }, { "epoch": 2.8496376811594204, "grad_norm": 0.25808446119444883, "learning_rate": 3.7219834343550606e-05, "loss": 0.877, "step": 6292 }, { "epoch": 2.850090579710145, "grad_norm": 0.17878734626899953, "learning_rate": 3.720721775477551e-05, "loss": 0.8803, "step": 6293 }, { "epoch": 2.8505434782608696, "grad_norm": 0.18240265016471904, "learning_rate": 3.719460144519932e-05, "loss": 0.8702, "step": 6294 }, { "epoch": 2.8509963768115942, "grad_norm": 0.20156087402477363, "learning_rate": 3.7181985416083234e-05, "loss": 0.8826, "step": 6295 }, { "epoch": 2.851449275362319, "grad_norm": 0.24647148057942797, "learning_rate": 3.716936966868855e-05, "loss": 0.8665, "step": 6296 }, { "epoch": 2.8519021739130435, "grad_norm": 0.23239112865129405, "learning_rate": 3.7156754204276446e-05, "loss": 0.8722, "step": 6297 }, { "epoch": 2.852355072463768, "grad_norm": 0.21796627457294285, "learning_rate": 3.714413902410815e-05, "loss": 0.8629, "step": 6298 }, { "epoch": 2.8528079710144927, "grad_norm": 0.20673064969894428, "learning_rate": 3.7131524129444796e-05, "loss": 0.885, "step": 6299 }, { "epoch": 2.8532608695652173, "grad_norm": 0.20359920488694142, "learning_rate": 3.711890952154752e-05, "loss": 0.8851, "step": 6300 }, { "epoch": 2.853713768115942, "grad_norm": 0.2043846259094407, "learning_rate": 3.7106295201677414e-05, "loss": 0.8683, "step": 6301 }, { "epoch": 2.8541666666666665, "grad_norm": 0.20625966922095754, "learning_rate": 3.709368117109558e-05, "loss": 0.8659, "step": 6302 }, { "epoch": 2.8546195652173916, "grad_norm": 0.22509867443377082, "learning_rate": 3.708106743106303e-05, "loss": 0.8902, "step": 6303 }, { "epoch": 2.855072463768116, "grad_norm": 0.21981361254305662, "learning_rate": 3.7068453982840813e-05, "loss": 0.8707, "step": 6304 }, { "epoch": 2.855525362318841, "grad_norm": 0.22950503591859078, "learning_rate": 3.7055840827689874e-05, "loss": 0.8647, "step": 6305 }, { "epoch": 2.8559782608695654, "grad_norm": 0.23677933811028684, "learning_rate": 3.704322796687121e-05, "loss": 0.8883, "step": 6306 }, { "epoch": 2.85643115942029, "grad_norm": 0.24452840451423305, "learning_rate": 3.703061540164572e-05, "loss": 0.8676, "step": 6307 }, { "epoch": 2.8568840579710146, "grad_norm": 0.28764046510150576, "learning_rate": 3.7018003133274324e-05, "loss": 0.8523, "step": 6308 }, { "epoch": 2.8573369565217392, "grad_norm": 0.217187376890612, "learning_rate": 3.7005391163017866e-05, "loss": 0.8692, "step": 6309 }, { "epoch": 2.857789855072464, "grad_norm": 0.17810177435908564, "learning_rate": 3.6992779492137206e-05, "loss": 0.8667, "step": 6310 }, { "epoch": 2.8582427536231885, "grad_norm": 0.2445978437074232, "learning_rate": 3.6980168121893145e-05, "loss": 0.8774, "step": 6311 }, { "epoch": 2.858695652173913, "grad_norm": 0.24185820206895944, "learning_rate": 3.696755705354647e-05, "loss": 0.8725, "step": 6312 }, { "epoch": 2.8591485507246377, "grad_norm": 0.21439659958783647, "learning_rate": 3.6954946288357906e-05, "loss": 0.8923, "step": 6313 }, { "epoch": 2.8596014492753623, "grad_norm": 0.19298298120483964, "learning_rate": 3.6942335827588204e-05, "loss": 0.8698, "step": 6314 }, { "epoch": 2.860054347826087, "grad_norm": 0.1518171332565994, "learning_rate": 3.692972567249803e-05, "loss": 0.8749, "step": 6315 }, { "epoch": 2.8605072463768115, "grad_norm": 0.16903704614851212, "learning_rate": 3.691711582434804e-05, "loss": 0.8855, "step": 6316 }, { "epoch": 2.860960144927536, "grad_norm": 0.24746332603170995, "learning_rate": 3.690450628439886e-05, "loss": 0.872, "step": 6317 }, { "epoch": 2.8614130434782608, "grad_norm": 0.25064082702601737, "learning_rate": 3.689189705391111e-05, "loss": 0.8795, "step": 6318 }, { "epoch": 2.8618659420289854, "grad_norm": 0.21221576440619394, "learning_rate": 3.687928813414533e-05, "loss": 0.878, "step": 6319 }, { "epoch": 2.86231884057971, "grad_norm": 0.1716481245149205, "learning_rate": 3.686667952636206e-05, "loss": 0.8477, "step": 6320 }, { "epoch": 2.8627717391304346, "grad_norm": 0.168801369267579, "learning_rate": 3.685407123182181e-05, "loss": 0.8754, "step": 6321 }, { "epoch": 2.863224637681159, "grad_norm": 0.23212879147910062, "learning_rate": 3.6841463251785024e-05, "loss": 0.8909, "step": 6322 }, { "epoch": 2.863677536231884, "grad_norm": 0.3594145535251441, "learning_rate": 3.682885558751218e-05, "loss": 0.8606, "step": 6323 }, { "epoch": 2.8641304347826084, "grad_norm": 0.43186999010146754, "learning_rate": 3.681624824026366e-05, "loss": 0.8588, "step": 6324 }, { "epoch": 2.8645833333333335, "grad_norm": 0.3703746378410668, "learning_rate": 3.6803641211299856e-05, "loss": 0.8552, "step": 6325 }, { "epoch": 2.865036231884058, "grad_norm": 0.22563740997465898, "learning_rate": 3.679103450188109e-05, "loss": 0.8742, "step": 6326 }, { "epoch": 2.8654891304347827, "grad_norm": 0.15684653391449221, "learning_rate": 3.67784281132677e-05, "loss": 0.8741, "step": 6327 }, { "epoch": 2.8659420289855073, "grad_norm": 0.2948281521277112, "learning_rate": 3.676582204671995e-05, "loss": 0.8679, "step": 6328 }, { "epoch": 2.866394927536232, "grad_norm": 0.35193678589015204, "learning_rate": 3.675321630349811e-05, "loss": 0.8648, "step": 6329 }, { "epoch": 2.8668478260869565, "grad_norm": 0.25723789949749104, "learning_rate": 3.6740610884862366e-05, "loss": 0.8857, "step": 6330 }, { "epoch": 2.867300724637681, "grad_norm": 0.1697539750002498, "learning_rate": 3.672800579207293e-05, "loss": 0.8493, "step": 6331 }, { "epoch": 2.8677536231884058, "grad_norm": 0.2044442355555697, "learning_rate": 3.671540102638993e-05, "loss": 0.8674, "step": 6332 }, { "epoch": 2.8682065217391304, "grad_norm": 0.19249342793081733, "learning_rate": 3.6702796589073494e-05, "loss": 0.8716, "step": 6333 }, { "epoch": 2.868659420289855, "grad_norm": 0.1645649970615499, "learning_rate": 3.66901924813837e-05, "loss": 0.8729, "step": 6334 }, { "epoch": 2.8691123188405796, "grad_norm": 0.2142682565951439, "learning_rate": 3.667758870458063e-05, "loss": 0.8876, "step": 6335 }, { "epoch": 2.869565217391304, "grad_norm": 0.22125554350985865, "learning_rate": 3.6664985259924265e-05, "loss": 0.9075, "step": 6336 }, { "epoch": 2.870018115942029, "grad_norm": 0.2032828272430876, "learning_rate": 3.6652382148674624e-05, "loss": 0.878, "step": 6337 }, { "epoch": 2.870471014492754, "grad_norm": 0.22950038541576628, "learning_rate": 3.663977937209162e-05, "loss": 0.8795, "step": 6338 }, { "epoch": 2.8709239130434785, "grad_norm": 0.1886878610598715, "learning_rate": 3.662717693143522e-05, "loss": 0.8934, "step": 6339 }, { "epoch": 2.871376811594203, "grad_norm": 0.1857256887331153, "learning_rate": 3.661457482796527e-05, "loss": 0.8896, "step": 6340 }, { "epoch": 2.8718297101449277, "grad_norm": 0.18480276717137228, "learning_rate": 3.660197306294165e-05, "loss": 0.8751, "step": 6341 }, { "epoch": 2.8722826086956523, "grad_norm": 0.2209507925586453, "learning_rate": 3.658937163762416e-05, "loss": 0.8542, "step": 6342 }, { "epoch": 2.872735507246377, "grad_norm": 0.1973165588979511, "learning_rate": 3.657677055327261e-05, "loss": 0.8746, "step": 6343 }, { "epoch": 2.8731884057971016, "grad_norm": 0.21004624131572214, "learning_rate": 3.656416981114671e-05, "loss": 0.8895, "step": 6344 }, { "epoch": 2.873641304347826, "grad_norm": 0.18956999716488582, "learning_rate": 3.655156941250621e-05, "loss": 0.8662, "step": 6345 }, { "epoch": 2.8740942028985508, "grad_norm": 0.22343165277578067, "learning_rate": 3.653896935861078e-05, "loss": 0.8746, "step": 6346 }, { "epoch": 2.8745471014492754, "grad_norm": 0.25065421592674403, "learning_rate": 3.652636965072007e-05, "loss": 0.8517, "step": 6347 }, { "epoch": 2.875, "grad_norm": 0.2149154716670697, "learning_rate": 3.6513770290093674e-05, "loss": 0.8664, "step": 6348 }, { "epoch": 2.8754528985507246, "grad_norm": 0.1722795236449415, "learning_rate": 3.65011712779912e-05, "loss": 0.874, "step": 6349 }, { "epoch": 2.8759057971014492, "grad_norm": 0.17780345049072505, "learning_rate": 3.648857261567218e-05, "loss": 0.8822, "step": 6350 }, { "epoch": 2.876358695652174, "grad_norm": 0.21117432816394444, "learning_rate": 3.6475974304396126e-05, "loss": 0.8921, "step": 6351 }, { "epoch": 2.8768115942028984, "grad_norm": 0.19979187595465403, "learning_rate": 3.646337634542248e-05, "loss": 0.8907, "step": 6352 }, { "epoch": 2.877264492753623, "grad_norm": 0.2293094211833571, "learning_rate": 3.645077874001073e-05, "loss": 0.8803, "step": 6353 }, { "epoch": 2.8777173913043477, "grad_norm": 0.2306852666362131, "learning_rate": 3.643818148942024e-05, "loss": 0.8676, "step": 6354 }, { "epoch": 2.8781702898550723, "grad_norm": 0.1511564918442321, "learning_rate": 3.64255845949104e-05, "loss": 0.8626, "step": 6355 }, { "epoch": 2.878623188405797, "grad_norm": 0.18526236056321266, "learning_rate": 3.6412988057740514e-05, "loss": 0.8885, "step": 6356 }, { "epoch": 2.8790760869565215, "grad_norm": 0.23397554134855406, "learning_rate": 3.640039187916992e-05, "loss": 0.8787, "step": 6357 }, { "epoch": 2.879528985507246, "grad_norm": 0.22072923112349022, "learning_rate": 3.6387796060457844e-05, "loss": 0.8872, "step": 6358 }, { "epoch": 2.879981884057971, "grad_norm": 0.2768005939962878, "learning_rate": 3.637520060286352e-05, "loss": 0.8606, "step": 6359 }, { "epoch": 2.880434782608696, "grad_norm": 0.25070572704718663, "learning_rate": 3.636260550764613e-05, "loss": 0.8638, "step": 6360 }, { "epoch": 2.8808876811594204, "grad_norm": 0.17567432229780158, "learning_rate": 3.635001077606484e-05, "loss": 0.8576, "step": 6361 }, { "epoch": 2.881340579710145, "grad_norm": 0.21811534759712958, "learning_rate": 3.633741640937875e-05, "loss": 0.8765, "step": 6362 }, { "epoch": 2.8817934782608696, "grad_norm": 0.26362903882736444, "learning_rate": 3.632482240884695e-05, "loss": 0.8792, "step": 6363 }, { "epoch": 2.8822463768115942, "grad_norm": 0.2533014724686922, "learning_rate": 3.6312228775728465e-05, "loss": 0.8682, "step": 6364 }, { "epoch": 2.882699275362319, "grad_norm": 0.26312845834747023, "learning_rate": 3.629963551128233e-05, "loss": 0.8565, "step": 6365 }, { "epoch": 2.8831521739130435, "grad_norm": 0.2808436778563259, "learning_rate": 3.628704261676749e-05, "loss": 0.8678, "step": 6366 }, { "epoch": 2.883605072463768, "grad_norm": 0.21617085658196616, "learning_rate": 3.627445009344289e-05, "loss": 0.8681, "step": 6367 }, { "epoch": 2.8840579710144927, "grad_norm": 0.2006403480989453, "learning_rate": 3.6261857942567396e-05, "loss": 0.8739, "step": 6368 }, { "epoch": 2.8845108695652173, "grad_norm": 0.28054011934812373, "learning_rate": 3.624926616539991e-05, "loss": 0.8557, "step": 6369 }, { "epoch": 2.884963768115942, "grad_norm": 0.3459042295703538, "learning_rate": 3.623667476319922e-05, "loss": 0.8644, "step": 6370 }, { "epoch": 2.8854166666666665, "grad_norm": 0.40028941040735017, "learning_rate": 3.622408373722412e-05, "loss": 0.8851, "step": 6371 }, { "epoch": 2.8858695652173916, "grad_norm": 0.33158491716528926, "learning_rate": 3.621149308873334e-05, "loss": 0.8838, "step": 6372 }, { "epoch": 2.886322463768116, "grad_norm": 0.246283232675594, "learning_rate": 3.619890281898562e-05, "loss": 0.8736, "step": 6373 }, { "epoch": 2.886775362318841, "grad_norm": 0.18317600074345186, "learning_rate": 3.61863129292396e-05, "loss": 0.8683, "step": 6374 }, { "epoch": 2.8872282608695654, "grad_norm": 0.24369372263207514, "learning_rate": 3.6173723420753925e-05, "loss": 0.8734, "step": 6375 }, { "epoch": 2.88768115942029, "grad_norm": 0.30283316978111113, "learning_rate": 3.6161134294787166e-05, "loss": 0.8875, "step": 6376 }, { "epoch": 2.8881340579710146, "grad_norm": 0.25403175503563835, "learning_rate": 3.614854555259791e-05, "loss": 0.8737, "step": 6377 }, { "epoch": 2.8885869565217392, "grad_norm": 0.24788021790825707, "learning_rate": 3.613595719544465e-05, "loss": 0.875, "step": 6378 }, { "epoch": 2.889039855072464, "grad_norm": 0.213271796115494, "learning_rate": 3.612336922458588e-05, "loss": 0.8741, "step": 6379 }, { "epoch": 2.8894927536231885, "grad_norm": 0.22649344341741717, "learning_rate": 3.611078164128002e-05, "loss": 0.877, "step": 6380 }, { "epoch": 2.889945652173913, "grad_norm": 0.18775878106162552, "learning_rate": 3.6098194446785495e-05, "loss": 0.8654, "step": 6381 }, { "epoch": 2.8903985507246377, "grad_norm": 0.19803039208057874, "learning_rate": 3.608560764236064e-05, "loss": 0.8826, "step": 6382 }, { "epoch": 2.8908514492753623, "grad_norm": 0.20681885704357267, "learning_rate": 3.607302122926381e-05, "loss": 0.8689, "step": 6383 }, { "epoch": 2.891304347826087, "grad_norm": 0.15946415261910168, "learning_rate": 3.6060435208753256e-05, "loss": 0.8761, "step": 6384 }, { "epoch": 2.8917572463768115, "grad_norm": 0.2000941825945142, "learning_rate": 3.6047849582087246e-05, "loss": 0.8673, "step": 6385 }, { "epoch": 2.892210144927536, "grad_norm": 0.19133524323853576, "learning_rate": 3.6035264350523964e-05, "loss": 0.9005, "step": 6386 }, { "epoch": 2.8926630434782608, "grad_norm": 0.21389206730860919, "learning_rate": 3.6022679515321597e-05, "loss": 0.8574, "step": 6387 }, { "epoch": 2.8931159420289854, "grad_norm": 0.17966053226314954, "learning_rate": 3.601009507773826e-05, "loss": 0.8686, "step": 6388 }, { "epoch": 2.89356884057971, "grad_norm": 0.189832339557191, "learning_rate": 3.599751103903205e-05, "loss": 0.8651, "step": 6389 }, { "epoch": 2.8940217391304346, "grad_norm": 0.20417782851298344, "learning_rate": 3.5984927400460984e-05, "loss": 0.8582, "step": 6390 }, { "epoch": 2.894474637681159, "grad_norm": 0.1627145644870973, "learning_rate": 3.59723441632831e-05, "loss": 0.8662, "step": 6391 }, { "epoch": 2.894927536231884, "grad_norm": 0.2185519087755998, "learning_rate": 3.595976132875636e-05, "loss": 0.9028, "step": 6392 }, { "epoch": 2.8953804347826084, "grad_norm": 0.2875444372680144, "learning_rate": 3.5947178898138674e-05, "loss": 0.8625, "step": 6393 }, { "epoch": 2.8958333333333335, "grad_norm": 0.21867433058858318, "learning_rate": 3.5934596872687924e-05, "loss": 0.8665, "step": 6394 }, { "epoch": 2.896286231884058, "grad_norm": 0.1810755721790658, "learning_rate": 3.592201525366199e-05, "loss": 0.8616, "step": 6395 }, { "epoch": 2.8967391304347827, "grad_norm": 0.2941752110302448, "learning_rate": 3.590943404231863e-05, "loss": 0.8815, "step": 6396 }, { "epoch": 2.8971920289855073, "grad_norm": 0.2955609215905663, "learning_rate": 3.5896853239915655e-05, "loss": 0.8676, "step": 6397 }, { "epoch": 2.897644927536232, "grad_norm": 0.21266278793560373, "learning_rate": 3.588427284771074e-05, "loss": 0.8595, "step": 6398 }, { "epoch": 2.8980978260869565, "grad_norm": 0.2961112502404921, "learning_rate": 3.587169286696161e-05, "loss": 0.8634, "step": 6399 }, { "epoch": 2.898550724637681, "grad_norm": 0.3089679705476195, "learning_rate": 3.585911329892587e-05, "loss": 0.8588, "step": 6400 }, { "epoch": 2.8990036231884058, "grad_norm": 0.36749529152947846, "learning_rate": 3.584653414486115e-05, "loss": 0.8564, "step": 6401 }, { "epoch": 2.8994565217391304, "grad_norm": 0.38877726744140384, "learning_rate": 3.583395540602497e-05, "loss": 0.8768, "step": 6402 }, { "epoch": 2.899909420289855, "grad_norm": 0.29447054906473646, "learning_rate": 3.5821377083674894e-05, "loss": 0.8741, "step": 6403 }, { "epoch": 2.9003623188405796, "grad_norm": 0.1943851817080598, "learning_rate": 3.580879917906837e-05, "loss": 0.8747, "step": 6404 }, { "epoch": 2.900815217391304, "grad_norm": 0.2012260362916952, "learning_rate": 3.579622169346283e-05, "loss": 0.8568, "step": 6405 }, { "epoch": 2.901268115942029, "grad_norm": 0.23305404536095808, "learning_rate": 3.578364462811565e-05, "loss": 0.8792, "step": 6406 }, { "epoch": 2.901721014492754, "grad_norm": 0.2854821937213063, "learning_rate": 3.5771067984284225e-05, "loss": 0.8763, "step": 6407 }, { "epoch": 2.9021739130434785, "grad_norm": 0.2363760418533482, "learning_rate": 3.575849176322582e-05, "loss": 0.8723, "step": 6408 }, { "epoch": 2.902626811594203, "grad_norm": 0.28821578867428194, "learning_rate": 3.5745915966197724e-05, "loss": 0.8764, "step": 6409 }, { "epoch": 2.9030797101449277, "grad_norm": 0.2884725811884699, "learning_rate": 3.5733340594457136e-05, "loss": 0.8697, "step": 6410 }, { "epoch": 2.9035326086956523, "grad_norm": 0.19073883121273766, "learning_rate": 3.572076564926127e-05, "loss": 0.8607, "step": 6411 }, { "epoch": 2.903985507246377, "grad_norm": 0.26542616842702454, "learning_rate": 3.570819113186723e-05, "loss": 0.8641, "step": 6412 }, { "epoch": 2.9044384057971016, "grad_norm": 0.28850745620948937, "learning_rate": 3.5695617043532124e-05, "loss": 0.8739, "step": 6413 }, { "epoch": 2.904891304347826, "grad_norm": 0.2030355173938551, "learning_rate": 3.568304338551299e-05, "loss": 0.855, "step": 6414 }, { "epoch": 2.9053442028985508, "grad_norm": 0.24672031298385724, "learning_rate": 3.567047015906687e-05, "loss": 0.864, "step": 6415 }, { "epoch": 2.9057971014492754, "grad_norm": 0.2838355646469118, "learning_rate": 3.5657897365450706e-05, "loss": 0.8779, "step": 6416 }, { "epoch": 2.90625, "grad_norm": 0.23946225032724172, "learning_rate": 3.564532500592143e-05, "loss": 0.8724, "step": 6417 }, { "epoch": 2.9067028985507246, "grad_norm": 0.22196607877666844, "learning_rate": 3.5632753081735895e-05, "loss": 0.8711, "step": 6418 }, { "epoch": 2.9071557971014492, "grad_norm": 0.32959784305368356, "learning_rate": 3.5620181594150966e-05, "loss": 0.8758, "step": 6419 }, { "epoch": 2.907608695652174, "grad_norm": 0.3583111388140093, "learning_rate": 3.560761054442342e-05, "loss": 0.8887, "step": 6420 }, { "epoch": 2.9080615942028984, "grad_norm": 0.28275108364375623, "learning_rate": 3.559503993381001e-05, "loss": 0.8721, "step": 6421 }, { "epoch": 2.908514492753623, "grad_norm": 0.2006975487091996, "learning_rate": 3.558246976356743e-05, "loss": 0.8894, "step": 6422 }, { "epoch": 2.9089673913043477, "grad_norm": 0.3386444825218075, "learning_rate": 3.556990003495236e-05, "loss": 0.873, "step": 6423 }, { "epoch": 2.9094202898550723, "grad_norm": 0.22398006927791445, "learning_rate": 3.55573307492214e-05, "loss": 0.8569, "step": 6424 }, { "epoch": 2.909873188405797, "grad_norm": 0.2525638889932092, "learning_rate": 3.5544761907631127e-05, "loss": 0.8845, "step": 6425 }, { "epoch": 2.9103260869565215, "grad_norm": 0.3229548130278622, "learning_rate": 3.553219351143805e-05, "loss": 0.8594, "step": 6426 }, { "epoch": 2.910778985507246, "grad_norm": 0.24260050752273266, "learning_rate": 3.551962556189869e-05, "loss": 0.844, "step": 6427 }, { "epoch": 2.911231884057971, "grad_norm": 0.23108038079529747, "learning_rate": 3.5507058060269446e-05, "loss": 0.8559, "step": 6428 }, { "epoch": 2.911684782608696, "grad_norm": 0.24394964666341273, "learning_rate": 3.5494491007806736e-05, "loss": 0.8484, "step": 6429 }, { "epoch": 2.9121376811594204, "grad_norm": 0.19388822171550704, "learning_rate": 3.5481924405766894e-05, "loss": 0.871, "step": 6430 }, { "epoch": 2.912590579710145, "grad_norm": 0.23136903267367975, "learning_rate": 3.546935825540624e-05, "loss": 0.8814, "step": 6431 }, { "epoch": 2.9130434782608696, "grad_norm": 0.2771570159564028, "learning_rate": 3.5456792557981e-05, "loss": 0.8841, "step": 6432 }, { "epoch": 2.9134963768115942, "grad_norm": 0.2787148997743089, "learning_rate": 3.544422731474743e-05, "loss": 0.8642, "step": 6433 }, { "epoch": 2.913949275362319, "grad_norm": 0.23077449557491245, "learning_rate": 3.5431662526961664e-05, "loss": 0.8682, "step": 6434 }, { "epoch": 2.9144021739130435, "grad_norm": 0.1907325966593529, "learning_rate": 3.541909819587984e-05, "loss": 0.8827, "step": 6435 }, { "epoch": 2.914855072463768, "grad_norm": 0.18118634737455294, "learning_rate": 3.540653432275801e-05, "loss": 0.8721, "step": 6436 }, { "epoch": 2.9153079710144927, "grad_norm": 0.18633096912478822, "learning_rate": 3.539397090885225e-05, "loss": 0.8562, "step": 6437 }, { "epoch": 2.9157608695652173, "grad_norm": 0.21113403238024214, "learning_rate": 3.5381407955418496e-05, "loss": 0.8639, "step": 6438 }, { "epoch": 2.916213768115942, "grad_norm": 0.2150515476541742, "learning_rate": 3.536884546371272e-05, "loss": 0.89, "step": 6439 }, { "epoch": 2.9166666666666665, "grad_norm": 0.22422342878767113, "learning_rate": 3.535628343499079e-05, "loss": 0.8722, "step": 6440 }, { "epoch": 2.9171195652173916, "grad_norm": 0.2546035008068183, "learning_rate": 3.534372187050857e-05, "loss": 0.8727, "step": 6441 }, { "epoch": 2.917572463768116, "grad_norm": 0.1942571754047168, "learning_rate": 3.533116077152185e-05, "loss": 0.888, "step": 6442 }, { "epoch": 2.918025362318841, "grad_norm": 0.17622055277541582, "learning_rate": 3.53186001392864e-05, "loss": 0.8838, "step": 6443 }, { "epoch": 2.9184782608695654, "grad_norm": 0.17064828667909576, "learning_rate": 3.530603997505789e-05, "loss": 0.8719, "step": 6444 }, { "epoch": 2.91893115942029, "grad_norm": 0.21488744528128903, "learning_rate": 3.529348028009202e-05, "loss": 0.8732, "step": 6445 }, { "epoch": 2.9193840579710146, "grad_norm": 0.22360093291551908, "learning_rate": 3.5280921055644387e-05, "loss": 0.868, "step": 6446 }, { "epoch": 2.9198369565217392, "grad_norm": 0.29305240689711815, "learning_rate": 3.526836230297054e-05, "loss": 0.8826, "step": 6447 }, { "epoch": 2.920289855072464, "grad_norm": 0.35758619767417227, "learning_rate": 3.525580402332603e-05, "loss": 0.8709, "step": 6448 }, { "epoch": 2.9207427536231885, "grad_norm": 0.37227242075951683, "learning_rate": 3.5243246217966284e-05, "loss": 0.8734, "step": 6449 }, { "epoch": 2.921195652173913, "grad_norm": 0.3486235299726957, "learning_rate": 3.523068888814677e-05, "loss": 0.8764, "step": 6450 }, { "epoch": 2.9216485507246377, "grad_norm": 0.36367880386407797, "learning_rate": 3.521813203512284e-05, "loss": 0.8731, "step": 6451 }, { "epoch": 2.9221014492753623, "grad_norm": 0.279555516428967, "learning_rate": 3.520557566014983e-05, "loss": 0.872, "step": 6452 }, { "epoch": 2.922554347826087, "grad_norm": 0.18469217246172506, "learning_rate": 3.5193019764483005e-05, "loss": 0.9034, "step": 6453 }, { "epoch": 2.9230072463768115, "grad_norm": 0.29457016483443654, "learning_rate": 3.5180464349377626e-05, "loss": 0.8664, "step": 6454 }, { "epoch": 2.923460144927536, "grad_norm": 0.29401741993820907, "learning_rate": 3.516790941608886e-05, "loss": 0.8811, "step": 6455 }, { "epoch": 2.9239130434782608, "grad_norm": 0.2905899363535621, "learning_rate": 3.5155354965871846e-05, "loss": 0.8783, "step": 6456 }, { "epoch": 2.9243659420289854, "grad_norm": 0.21994429326715073, "learning_rate": 3.5142800999981665e-05, "loss": 0.8861, "step": 6457 }, { "epoch": 2.92481884057971, "grad_norm": 0.18172424268878842, "learning_rate": 3.513024751967338e-05, "loss": 0.8887, "step": 6458 }, { "epoch": 2.9252717391304346, "grad_norm": 0.23398368789056975, "learning_rate": 3.511769452620196e-05, "loss": 0.8689, "step": 6459 }, { "epoch": 2.925724637681159, "grad_norm": 0.25906353979603824, "learning_rate": 3.5105142020822366e-05, "loss": 0.859, "step": 6460 }, { "epoch": 2.926177536231884, "grad_norm": 0.2320585242403237, "learning_rate": 3.509259000478945e-05, "loss": 0.8716, "step": 6461 }, { "epoch": 2.9266304347826084, "grad_norm": 0.231410859134153, "learning_rate": 3.5080038479358124e-05, "loss": 0.8805, "step": 6462 }, { "epoch": 2.9270833333333335, "grad_norm": 0.159542400836153, "learning_rate": 3.506748744578312e-05, "loss": 0.855, "step": 6463 }, { "epoch": 2.927536231884058, "grad_norm": 0.26097732055328476, "learning_rate": 3.5054936905319236e-05, "loss": 0.8842, "step": 6464 }, { "epoch": 2.9279891304347827, "grad_norm": 0.2994526117950906, "learning_rate": 3.5042386859221125e-05, "loss": 0.8445, "step": 6465 }, { "epoch": 2.9284420289855073, "grad_norm": 0.27663240986299176, "learning_rate": 3.5029837308743464e-05, "loss": 0.8627, "step": 6466 }, { "epoch": 2.928894927536232, "grad_norm": 0.25080173891482, "learning_rate": 3.501728825514085e-05, "loss": 0.8709, "step": 6467 }, { "epoch": 2.9293478260869565, "grad_norm": 0.2152484017259538, "learning_rate": 3.5004739699667816e-05, "loss": 0.8815, "step": 6468 }, { "epoch": 2.929800724637681, "grad_norm": 0.24382185077925572, "learning_rate": 3.499219164357886e-05, "loss": 0.878, "step": 6469 }, { "epoch": 2.9302536231884058, "grad_norm": 0.1933252292819737, "learning_rate": 3.497964408812846e-05, "loss": 0.8627, "step": 6470 }, { "epoch": 2.9307065217391304, "grad_norm": 0.3761531750611396, "learning_rate": 3.496709703457098e-05, "loss": 0.8504, "step": 6471 }, { "epoch": 2.931159420289855, "grad_norm": 0.3637792581790523, "learning_rate": 3.495455048416081e-05, "loss": 0.8609, "step": 6472 }, { "epoch": 2.9316123188405796, "grad_norm": 0.278323418190629, "learning_rate": 3.4942004438152196e-05, "loss": 0.8782, "step": 6473 }, { "epoch": 2.932065217391304, "grad_norm": 0.2403158369792561, "learning_rate": 3.492945889779943e-05, "loss": 0.8671, "step": 6474 }, { "epoch": 2.932518115942029, "grad_norm": 0.2025892536806978, "learning_rate": 3.4916913864356675e-05, "loss": 0.8709, "step": 6475 }, { "epoch": 2.932971014492754, "grad_norm": 0.193180174183392, "learning_rate": 3.4904369339078117e-05, "loss": 0.895, "step": 6476 }, { "epoch": 2.9334239130434785, "grad_norm": 0.20918612883686824, "learning_rate": 3.4891825323217824e-05, "loss": 0.8894, "step": 6477 }, { "epoch": 2.933876811594203, "grad_norm": 0.20883245284340204, "learning_rate": 3.4879281818029846e-05, "loss": 0.8744, "step": 6478 }, { "epoch": 2.9343297101449277, "grad_norm": 0.16683877772854766, "learning_rate": 3.486673882476817e-05, "loss": 0.8754, "step": 6479 }, { "epoch": 2.9347826086956523, "grad_norm": 0.1454372466072306, "learning_rate": 3.485419634468676e-05, "loss": 0.8785, "step": 6480 }, { "epoch": 2.935235507246377, "grad_norm": 0.19095084306831495, "learning_rate": 3.4841654379039495e-05, "loss": 0.8543, "step": 6481 }, { "epoch": 2.9356884057971016, "grad_norm": 0.16595218970138617, "learning_rate": 3.482911292908022e-05, "loss": 0.8733, "step": 6482 }, { "epoch": 2.936141304347826, "grad_norm": 0.17402000987012253, "learning_rate": 3.48165719960627e-05, "loss": 0.8702, "step": 6483 }, { "epoch": 2.9365942028985508, "grad_norm": 0.1828676466792323, "learning_rate": 3.4804031581240716e-05, "loss": 0.8628, "step": 6484 }, { "epoch": 2.9370471014492754, "grad_norm": 0.1949161562460648, "learning_rate": 3.479149168586791e-05, "loss": 0.8737, "step": 6485 }, { "epoch": 2.9375, "grad_norm": 0.1883919143490623, "learning_rate": 3.477895231119795e-05, "loss": 0.8663, "step": 6486 }, { "epoch": 2.9379528985507246, "grad_norm": 0.21001273997403105, "learning_rate": 3.476641345848438e-05, "loss": 0.8845, "step": 6487 }, { "epoch": 2.9384057971014492, "grad_norm": 0.2512307624839612, "learning_rate": 3.475387512898077e-05, "loss": 0.8724, "step": 6488 }, { "epoch": 2.938858695652174, "grad_norm": 0.21715697638291157, "learning_rate": 3.474133732394056e-05, "loss": 0.8686, "step": 6489 }, { "epoch": 2.9393115942028984, "grad_norm": 0.2252925957077141, "learning_rate": 3.472880004461721e-05, "loss": 0.8814, "step": 6490 }, { "epoch": 2.939764492753623, "grad_norm": 0.19251143659211167, "learning_rate": 3.471626329226405e-05, "loss": 0.8788, "step": 6491 }, { "epoch": 2.9402173913043477, "grad_norm": 0.19615438807348168, "learning_rate": 3.470372706813444e-05, "loss": 0.8556, "step": 6492 }, { "epoch": 2.9406702898550723, "grad_norm": 0.25595286140291323, "learning_rate": 3.4691191373481614e-05, "loss": 0.8454, "step": 6493 }, { "epoch": 2.941123188405797, "grad_norm": 0.2008353873234775, "learning_rate": 3.4678656209558814e-05, "loss": 0.8705, "step": 6494 }, { "epoch": 2.9415760869565215, "grad_norm": 0.1540090112966068, "learning_rate": 3.4666121577619164e-05, "loss": 0.8786, "step": 6495 }, { "epoch": 2.942028985507246, "grad_norm": 0.21570104407947, "learning_rate": 3.465358747891581e-05, "loss": 0.8675, "step": 6496 }, { "epoch": 2.942481884057971, "grad_norm": 0.2554021549674277, "learning_rate": 3.464105391470178e-05, "loss": 0.8734, "step": 6497 }, { "epoch": 2.942934782608696, "grad_norm": 0.24822411105136594, "learning_rate": 3.462852088623009e-05, "loss": 0.8731, "step": 6498 }, { "epoch": 2.9433876811594204, "grad_norm": 0.220067512967915, "learning_rate": 3.461598839475367e-05, "loss": 0.8665, "step": 6499 }, { "epoch": 2.943840579710145, "grad_norm": 0.2282248152892477, "learning_rate": 3.460345644152543e-05, "loss": 0.8965, "step": 6500 }, { "epoch": 2.9442934782608696, "grad_norm": 0.15981339477889286, "learning_rate": 3.4590925027798194e-05, "loss": 0.8605, "step": 6501 }, { "epoch": 2.9447463768115942, "grad_norm": 0.2356775580939227, "learning_rate": 3.457839415482476e-05, "loss": 0.8611, "step": 6502 }, { "epoch": 2.945199275362319, "grad_norm": 0.2266638241335822, "learning_rate": 3.456586382385784e-05, "loss": 0.8587, "step": 6503 }, { "epoch": 2.9456521739130435, "grad_norm": 0.20357269642633644, "learning_rate": 3.4553334036150146e-05, "loss": 0.8654, "step": 6504 }, { "epoch": 2.946105072463768, "grad_norm": 0.21310561167655978, "learning_rate": 3.454080479295427e-05, "loss": 0.8761, "step": 6505 }, { "epoch": 2.9465579710144927, "grad_norm": 0.23320080384677325, "learning_rate": 3.4528276095522786e-05, "loss": 0.8708, "step": 6506 }, { "epoch": 2.9470108695652173, "grad_norm": 0.2593861990412982, "learning_rate": 3.45157479451082e-05, "loss": 0.862, "step": 6507 }, { "epoch": 2.947463768115942, "grad_norm": 0.322270052285742, "learning_rate": 3.450322034296301e-05, "loss": 0.8704, "step": 6508 }, { "epoch": 2.9479166666666665, "grad_norm": 0.26188362142669735, "learning_rate": 3.4490693290339576e-05, "loss": 0.8758, "step": 6509 }, { "epoch": 2.9483695652173916, "grad_norm": 0.19953809057421476, "learning_rate": 3.447816678849028e-05, "loss": 0.8795, "step": 6510 }, { "epoch": 2.948822463768116, "grad_norm": 0.21114826581125087, "learning_rate": 3.446564083866737e-05, "loss": 0.8771, "step": 6511 }, { "epoch": 2.949275362318841, "grad_norm": 0.20042131964641574, "learning_rate": 3.4453115442123136e-05, "loss": 0.8748, "step": 6512 }, { "epoch": 2.9497282608695654, "grad_norm": 0.212585079030076, "learning_rate": 3.4440590600109734e-05, "loss": 0.8499, "step": 6513 }, { "epoch": 2.95018115942029, "grad_norm": 0.2622211165341937, "learning_rate": 3.442806631387931e-05, "loss": 0.8693, "step": 6514 }, { "epoch": 2.9506340579710146, "grad_norm": 0.31843165144011315, "learning_rate": 3.441554258468391e-05, "loss": 0.8638, "step": 6515 }, { "epoch": 2.9510869565217392, "grad_norm": 0.2826909779633618, "learning_rate": 3.440301941377558e-05, "loss": 0.8654, "step": 6516 }, { "epoch": 2.951539855072464, "grad_norm": 0.24557399052601334, "learning_rate": 3.4390496802406255e-05, "loss": 0.8748, "step": 6517 }, { "epoch": 2.9519927536231885, "grad_norm": 0.16889293522887344, "learning_rate": 3.437797475182787e-05, "loss": 0.8676, "step": 6518 }, { "epoch": 2.952445652173913, "grad_norm": 0.18484452212503555, "learning_rate": 3.436545326329225e-05, "loss": 0.8912, "step": 6519 }, { "epoch": 2.9528985507246377, "grad_norm": 0.22863116019878957, "learning_rate": 3.4352932338051204e-05, "loss": 0.8754, "step": 6520 }, { "epoch": 2.9533514492753623, "grad_norm": 0.19768423953841924, "learning_rate": 3.4340411977356445e-05, "loss": 0.8769, "step": 6521 }, { "epoch": 2.953804347826087, "grad_norm": 0.18889790641465637, "learning_rate": 3.432789218245968e-05, "loss": 0.8519, "step": 6522 }, { "epoch": 2.9542572463768115, "grad_norm": 0.14707627369676152, "learning_rate": 3.431537295461252e-05, "loss": 0.8428, "step": 6523 }, { "epoch": 2.954710144927536, "grad_norm": 0.1955147012329621, "learning_rate": 3.430285429506654e-05, "loss": 0.8623, "step": 6524 }, { "epoch": 2.9551630434782608, "grad_norm": 0.248139372491634, "learning_rate": 3.429033620507323e-05, "loss": 0.9027, "step": 6525 }, { "epoch": 2.9556159420289854, "grad_norm": 0.22194936408993962, "learning_rate": 3.427781868588406e-05, "loss": 0.8648, "step": 6526 }, { "epoch": 2.95606884057971, "grad_norm": 0.2147195086154827, "learning_rate": 3.426530173875043e-05, "loss": 0.8796, "step": 6527 }, { "epoch": 2.9565217391304346, "grad_norm": 0.18193701941263418, "learning_rate": 3.425278536492368e-05, "loss": 0.875, "step": 6528 }, { "epoch": 2.956974637681159, "grad_norm": 0.17922608942649226, "learning_rate": 3.424026956565506e-05, "loss": 0.8606, "step": 6529 }, { "epoch": 2.957427536231884, "grad_norm": 0.14614087007120877, "learning_rate": 3.422775434219584e-05, "loss": 0.8488, "step": 6530 }, { "epoch": 2.9578804347826084, "grad_norm": 0.18871034787497237, "learning_rate": 3.421523969579716e-05, "loss": 0.8959, "step": 6531 }, { "epoch": 2.9583333333333335, "grad_norm": 0.21531597057659535, "learning_rate": 3.4202725627710136e-05, "loss": 0.8587, "step": 6532 }, { "epoch": 2.958786231884058, "grad_norm": 0.20395980644420575, "learning_rate": 3.4190212139185805e-05, "loss": 0.8587, "step": 6533 }, { "epoch": 2.9592391304347827, "grad_norm": 0.17796279691286576, "learning_rate": 3.417769923147519e-05, "loss": 0.8919, "step": 6534 }, { "epoch": 2.9596920289855073, "grad_norm": 0.2102243192673048, "learning_rate": 3.41651869058292e-05, "loss": 0.8607, "step": 6535 }, { "epoch": 2.960144927536232, "grad_norm": 0.18941453702352343, "learning_rate": 3.415267516349873e-05, "loss": 0.8672, "step": 6536 }, { "epoch": 2.9605978260869565, "grad_norm": 0.1844070740564069, "learning_rate": 3.414016400573457e-05, "loss": 0.8682, "step": 6537 }, { "epoch": 2.961050724637681, "grad_norm": 0.22208241946240162, "learning_rate": 3.412765343378752e-05, "loss": 0.879, "step": 6538 }, { "epoch": 2.9615036231884058, "grad_norm": 0.19617751971397246, "learning_rate": 3.411514344890826e-05, "loss": 0.8485, "step": 6539 }, { "epoch": 2.9619565217391304, "grad_norm": 0.1898908689127067, "learning_rate": 3.410263405234743e-05, "loss": 0.873, "step": 6540 }, { "epoch": 2.962409420289855, "grad_norm": 0.21648349518279833, "learning_rate": 3.4090125245355605e-05, "loss": 0.8592, "step": 6541 }, { "epoch": 2.9628623188405796, "grad_norm": 0.2530598674100872, "learning_rate": 3.407761702918334e-05, "loss": 0.8791, "step": 6542 }, { "epoch": 2.963315217391304, "grad_norm": 0.18912447396562596, "learning_rate": 3.4065109405081084e-05, "loss": 0.8583, "step": 6543 }, { "epoch": 2.963768115942029, "grad_norm": 0.2617293007833807, "learning_rate": 3.405260237429925e-05, "loss": 0.8797, "step": 6544 }, { "epoch": 2.964221014492754, "grad_norm": 0.1998673795306355, "learning_rate": 3.404009593808817e-05, "loss": 0.8779, "step": 6545 }, { "epoch": 2.9646739130434785, "grad_norm": 0.2606646155469806, "learning_rate": 3.4027590097698154e-05, "loss": 0.8688, "step": 6546 }, { "epoch": 2.965126811594203, "grad_norm": 0.2264357702046376, "learning_rate": 3.401508485437941e-05, "loss": 0.8609, "step": 6547 }, { "epoch": 2.9655797101449277, "grad_norm": 0.2040001697797138, "learning_rate": 3.400258020938213e-05, "loss": 0.8796, "step": 6548 }, { "epoch": 2.9660326086956523, "grad_norm": 0.2064788172428049, "learning_rate": 3.3990076163956386e-05, "loss": 0.8606, "step": 6549 }, { "epoch": 2.966485507246377, "grad_norm": 0.21402889509438097, "learning_rate": 3.397757271935228e-05, "loss": 0.8696, "step": 6550 }, { "epoch": 2.9669384057971016, "grad_norm": 0.26644382402487415, "learning_rate": 3.3965069876819755e-05, "loss": 0.8854, "step": 6551 }, { "epoch": 2.967391304347826, "grad_norm": 0.3007752799111617, "learning_rate": 3.395256763760877e-05, "loss": 0.8937, "step": 6552 }, { "epoch": 2.9678442028985508, "grad_norm": 0.3048114556066479, "learning_rate": 3.394006600296916e-05, "loss": 0.8865, "step": 6553 }, { "epoch": 2.9682971014492754, "grad_norm": 0.23230000178442214, "learning_rate": 3.392756497415078e-05, "loss": 0.8636, "step": 6554 }, { "epoch": 2.96875, "grad_norm": 0.15821294769232921, "learning_rate": 3.3915064552403336e-05, "loss": 0.8713, "step": 6555 }, { "epoch": 2.9692028985507246, "grad_norm": 0.16208792750623122, "learning_rate": 3.3902564738976545e-05, "loss": 0.8639, "step": 6556 }, { "epoch": 2.9696557971014492, "grad_norm": 0.19460003667010475, "learning_rate": 3.389006553512e-05, "loss": 0.8915, "step": 6557 }, { "epoch": 2.970108695652174, "grad_norm": 0.26625677727228986, "learning_rate": 3.3877566942083294e-05, "loss": 0.8819, "step": 6558 }, { "epoch": 2.9705615942028984, "grad_norm": 0.3026664999231167, "learning_rate": 3.386506896111592e-05, "loss": 0.8678, "step": 6559 }, { "epoch": 2.971014492753623, "grad_norm": 0.22352600533787675, "learning_rate": 3.3852571593467335e-05, "loss": 0.882, "step": 6560 }, { "epoch": 2.9714673913043477, "grad_norm": 0.17068713164201987, "learning_rate": 3.384007484038689e-05, "loss": 0.8642, "step": 6561 }, { "epoch": 2.9719202898550723, "grad_norm": 0.211110770786446, "learning_rate": 3.382757870312395e-05, "loss": 0.8575, "step": 6562 }, { "epoch": 2.972373188405797, "grad_norm": 0.2633843794413417, "learning_rate": 3.381508318292772e-05, "loss": 0.866, "step": 6563 }, { "epoch": 2.9728260869565215, "grad_norm": 0.30610288553721593, "learning_rate": 3.3802588281047444e-05, "loss": 0.8548, "step": 6564 }, { "epoch": 2.973278985507246, "grad_norm": 0.2884687222035888, "learning_rate": 3.379009399873223e-05, "loss": 0.8603, "step": 6565 }, { "epoch": 2.973731884057971, "grad_norm": 0.24784694898075518, "learning_rate": 3.377760033723117e-05, "loss": 0.8772, "step": 6566 }, { "epoch": 2.974184782608696, "grad_norm": 0.1899776126406438, "learning_rate": 3.376510729779325e-05, "loss": 0.8922, "step": 6567 }, { "epoch": 2.9746376811594204, "grad_norm": 0.17125214894084256, "learning_rate": 3.3752614881667444e-05, "loss": 0.8831, "step": 6568 }, { "epoch": 2.975090579710145, "grad_norm": 0.22445044651185, "learning_rate": 3.374012309010264e-05, "loss": 0.87, "step": 6569 }, { "epoch": 2.9755434782608696, "grad_norm": 0.28434155994449783, "learning_rate": 3.3727631924347625e-05, "loss": 0.8908, "step": 6570 }, { "epoch": 2.9759963768115942, "grad_norm": 0.36333815016673937, "learning_rate": 3.3715141385651205e-05, "loss": 0.8773, "step": 6571 }, { "epoch": 2.976449275362319, "grad_norm": 0.3507998107535178, "learning_rate": 3.370265147526204e-05, "loss": 0.8672, "step": 6572 }, { "epoch": 2.9769021739130435, "grad_norm": 0.23760718608836895, "learning_rate": 3.3690162194428804e-05, "loss": 0.8824, "step": 6573 }, { "epoch": 2.977355072463768, "grad_norm": 0.24232966925477495, "learning_rate": 3.3677673544400046e-05, "loss": 0.8655, "step": 6574 }, { "epoch": 2.9778079710144927, "grad_norm": 0.28151724400762096, "learning_rate": 3.3665185526424285e-05, "loss": 0.8896, "step": 6575 }, { "epoch": 2.9782608695652173, "grad_norm": 0.33260168646949667, "learning_rate": 3.3652698141749946e-05, "loss": 0.8776, "step": 6576 }, { "epoch": 2.978713768115942, "grad_norm": 0.26825505508038033, "learning_rate": 3.364021139162546e-05, "loss": 0.8751, "step": 6577 }, { "epoch": 2.9791666666666665, "grad_norm": 0.21176020619795502, "learning_rate": 3.3627725277299103e-05, "loss": 0.8803, "step": 6578 }, { "epoch": 2.9796195652173916, "grad_norm": 0.196133455477677, "learning_rate": 3.361523980001916e-05, "loss": 0.8556, "step": 6579 }, { "epoch": 2.980072463768116, "grad_norm": 0.20857999143828898, "learning_rate": 3.3602754961033794e-05, "loss": 0.8629, "step": 6580 }, { "epoch": 2.980525362318841, "grad_norm": 0.19345491014410277, "learning_rate": 3.359027076159116e-05, "loss": 0.8763, "step": 6581 }, { "epoch": 2.9809782608695654, "grad_norm": 0.22933402405428727, "learning_rate": 3.357778720293932e-05, "loss": 0.8675, "step": 6582 }, { "epoch": 2.98143115942029, "grad_norm": 0.1620638457614074, "learning_rate": 3.3565304286326274e-05, "loss": 0.8667, "step": 6583 }, { "epoch": 2.9818840579710146, "grad_norm": 0.20029679309603463, "learning_rate": 3.355282201299994e-05, "loss": 0.8544, "step": 6584 }, { "epoch": 2.9823369565217392, "grad_norm": 0.23734083461928535, "learning_rate": 3.354034038420821e-05, "loss": 0.8582, "step": 6585 }, { "epoch": 2.982789855072464, "grad_norm": 0.20758338527398956, "learning_rate": 3.35278594011989e-05, "loss": 0.8682, "step": 6586 }, { "epoch": 2.9832427536231885, "grad_norm": 0.2021462612428492, "learning_rate": 3.351537906521973e-05, "loss": 0.8668, "step": 6587 }, { "epoch": 2.983695652173913, "grad_norm": 0.17846181082275422, "learning_rate": 3.350289937751839e-05, "loss": 0.8553, "step": 6588 }, { "epoch": 2.9841485507246377, "grad_norm": 0.1799117859287552, "learning_rate": 3.34904203393425e-05, "loss": 0.8795, "step": 6589 }, { "epoch": 2.9846014492753623, "grad_norm": 0.24715563056229747, "learning_rate": 3.34779419519396e-05, "loss": 0.8593, "step": 6590 }, { "epoch": 2.985054347826087, "grad_norm": 0.27810553492723056, "learning_rate": 3.346546421655719e-05, "loss": 0.851, "step": 6591 }, { "epoch": 2.9855072463768115, "grad_norm": 0.31234889868835014, "learning_rate": 3.345298713444265e-05, "loss": 0.8603, "step": 6592 }, { "epoch": 2.985960144927536, "grad_norm": 0.21094844810713895, "learning_rate": 3.344051070684339e-05, "loss": 0.8665, "step": 6593 }, { "epoch": 2.9864130434782608, "grad_norm": 0.20424687235804131, "learning_rate": 3.3428034935006655e-05, "loss": 0.8744, "step": 6594 }, { "epoch": 2.9868659420289854, "grad_norm": 0.20667876658587456, "learning_rate": 3.341555982017969e-05, "loss": 0.8754, "step": 6595 }, { "epoch": 2.98731884057971, "grad_norm": 0.22800117718767843, "learning_rate": 3.3403085363609626e-05, "loss": 0.8887, "step": 6596 }, { "epoch": 2.9877717391304346, "grad_norm": 0.2542229969060977, "learning_rate": 3.339061156654359e-05, "loss": 0.895, "step": 6597 }, { "epoch": 2.988224637681159, "grad_norm": 0.22951673194867311, "learning_rate": 3.337813843022859e-05, "loss": 0.8643, "step": 6598 }, { "epoch": 2.988677536231884, "grad_norm": 0.19357325606972325, "learning_rate": 3.336566595591158e-05, "loss": 0.8844, "step": 6599 }, { "epoch": 2.9891304347826084, "grad_norm": 0.17428045638466863, "learning_rate": 3.3353194144839444e-05, "loss": 0.8624, "step": 6600 }, { "epoch": 2.9895833333333335, "grad_norm": 0.2282249453570923, "learning_rate": 3.3340722998259036e-05, "loss": 0.8634, "step": 6601 }, { "epoch": 2.990036231884058, "grad_norm": 0.22396058227833177, "learning_rate": 3.33282525174171e-05, "loss": 0.8799, "step": 6602 }, { "epoch": 2.9904891304347827, "grad_norm": 0.19034430881193512, "learning_rate": 3.331578270356034e-05, "loss": 0.8559, "step": 6603 }, { "epoch": 2.9909420289855073, "grad_norm": 0.16431045674729736, "learning_rate": 3.3303313557935357e-05, "loss": 0.8777, "step": 6604 }, { "epoch": 2.991394927536232, "grad_norm": 0.23054453072195275, "learning_rate": 3.329084508178874e-05, "loss": 0.8854, "step": 6605 }, { "epoch": 2.9918478260869565, "grad_norm": 0.2169841331479425, "learning_rate": 3.327837727636696e-05, "loss": 0.8852, "step": 6606 }, { "epoch": 2.992300724637681, "grad_norm": 0.254918309990888, "learning_rate": 3.326591014291648e-05, "loss": 0.8744, "step": 6607 }, { "epoch": 2.9927536231884058, "grad_norm": 0.23466530255908105, "learning_rate": 3.325344368268361e-05, "loss": 0.8553, "step": 6608 }, { "epoch": 2.9932065217391304, "grad_norm": 0.20416656995397633, "learning_rate": 3.324097789691468e-05, "loss": 0.8754, "step": 6609 }, { "epoch": 2.993659420289855, "grad_norm": 0.21314877290401896, "learning_rate": 3.322851278685589e-05, "loss": 0.8857, "step": 6610 }, { "epoch": 2.9941123188405796, "grad_norm": 0.2662493094171669, "learning_rate": 3.3216048353753424e-05, "loss": 0.889, "step": 6611 }, { "epoch": 2.994565217391304, "grad_norm": 0.2819298239814187, "learning_rate": 3.3203584598853335e-05, "loss": 0.8651, "step": 6612 }, { "epoch": 2.995018115942029, "grad_norm": 0.2551125724003299, "learning_rate": 3.319112152340168e-05, "loss": 0.8804, "step": 6613 }, { "epoch": 2.995471014492754, "grad_norm": 0.19527827264167755, "learning_rate": 3.317865912864438e-05, "loss": 0.8694, "step": 6614 }, { "epoch": 2.9959239130434785, "grad_norm": 0.23316243294342096, "learning_rate": 3.3166197415827345e-05, "loss": 0.8851, "step": 6615 }, { "epoch": 2.996376811594203, "grad_norm": 0.21674820596314376, "learning_rate": 3.3153736386196384e-05, "loss": 0.8731, "step": 6616 }, { "epoch": 2.9968297101449277, "grad_norm": 0.2806665471692988, "learning_rate": 3.314127604099725e-05, "loss": 0.8937, "step": 6617 }, { "epoch": 2.9972826086956523, "grad_norm": 0.2674828834168081, "learning_rate": 3.31288163814756e-05, "loss": 0.8786, "step": 6618 }, { "epoch": 2.997735507246377, "grad_norm": 0.20702701159593093, "learning_rate": 3.3116357408877084e-05, "loss": 0.8831, "step": 6619 }, { "epoch": 2.9981884057971016, "grad_norm": 0.18744846527639591, "learning_rate": 3.3103899124447216e-05, "loss": 0.8544, "step": 6620 }, { "epoch": 2.998641304347826, "grad_norm": 0.20525326968868982, "learning_rate": 3.309144152943149e-05, "loss": 0.8698, "step": 6621 }, { "epoch": 2.9990942028985508, "grad_norm": 0.42104557253203145, "learning_rate": 3.3078984625075284e-05, "loss": 0.8891, "step": 6622 }, { "epoch": 2.9995471014492754, "grad_norm": 0.2388668308156623, "learning_rate": 3.306652841262397e-05, "loss": 0.8653, "step": 6623 }, { "epoch": 3.0, "grad_norm": 0.19414042462294942, "learning_rate": 3.305407289332279e-05, "loss": 0.8818, "step": 6624 }, { "epoch": 3.0004528985507246, "grad_norm": 0.171799700034463, "learning_rate": 3.304161806841695e-05, "loss": 0.8574, "step": 6625 }, { "epoch": 3.0009057971014492, "grad_norm": 0.21095602502157595, "learning_rate": 3.302916393915157e-05, "loss": 0.8608, "step": 6626 }, { "epoch": 3.001358695652174, "grad_norm": 0.17038824636916652, "learning_rate": 3.3016710506771734e-05, "loss": 0.8566, "step": 6627 }, { "epoch": 3.0018115942028984, "grad_norm": 0.1514150359189018, "learning_rate": 3.300425777252241e-05, "loss": 0.8601, "step": 6628 }, { "epoch": 3.002264492753623, "grad_norm": 0.17220707639832952, "learning_rate": 3.299180573764853e-05, "loss": 0.888, "step": 6629 }, { "epoch": 3.0027173913043477, "grad_norm": 0.1861520645306728, "learning_rate": 3.297935440339491e-05, "loss": 0.8713, "step": 6630 }, { "epoch": 3.0031702898550723, "grad_norm": 0.19260788624234035, "learning_rate": 3.296690377100637e-05, "loss": 0.8598, "step": 6631 }, { "epoch": 3.003623188405797, "grad_norm": 0.17590779189313793, "learning_rate": 3.295445384172761e-05, "loss": 0.8505, "step": 6632 }, { "epoch": 3.004076086956522, "grad_norm": 0.18017055919723546, "learning_rate": 3.294200461680326e-05, "loss": 0.8361, "step": 6633 }, { "epoch": 3.0045289855072466, "grad_norm": 0.17718722797451394, "learning_rate": 3.292955609747788e-05, "loss": 0.8688, "step": 6634 }, { "epoch": 3.004981884057971, "grad_norm": 0.17933424212700327, "learning_rate": 3.2917108284996004e-05, "loss": 0.878, "step": 6635 }, { "epoch": 3.005434782608696, "grad_norm": 0.17016001707657097, "learning_rate": 3.2904661180602024e-05, "loss": 0.8551, "step": 6636 }, { "epoch": 3.0058876811594204, "grad_norm": 0.20158056870515342, "learning_rate": 3.2892214785540315e-05, "loss": 0.86, "step": 6637 }, { "epoch": 3.006340579710145, "grad_norm": 0.17769828165209414, "learning_rate": 3.287976910105514e-05, "loss": 0.8715, "step": 6638 }, { "epoch": 3.0067934782608696, "grad_norm": 0.2248035196815619, "learning_rate": 3.2867324128390756e-05, "loss": 0.8667, "step": 6639 }, { "epoch": 3.0072463768115942, "grad_norm": 0.2452634591806196, "learning_rate": 3.285487986879127e-05, "loss": 0.8438, "step": 6640 }, { "epoch": 3.007699275362319, "grad_norm": 0.15896349142864777, "learning_rate": 3.284243632350077e-05, "loss": 0.8534, "step": 6641 }, { "epoch": 3.0081521739130435, "grad_norm": 0.21410294102486868, "learning_rate": 3.282999349376324e-05, "loss": 0.8727, "step": 6642 }, { "epoch": 3.008605072463768, "grad_norm": 0.22725094593019754, "learning_rate": 3.2817551380822633e-05, "loss": 0.8549, "step": 6643 }, { "epoch": 3.0090579710144927, "grad_norm": 0.18484462856035488, "learning_rate": 3.2805109985922796e-05, "loss": 0.8605, "step": 6644 }, { "epoch": 3.0095108695652173, "grad_norm": 0.1949902858753216, "learning_rate": 3.2792669310307515e-05, "loss": 0.8615, "step": 6645 }, { "epoch": 3.009963768115942, "grad_norm": 0.18479618151214966, "learning_rate": 3.278022935522049e-05, "loss": 0.8736, "step": 6646 }, { "epoch": 3.0104166666666665, "grad_norm": 0.16832474348539703, "learning_rate": 3.2767790121905396e-05, "loss": 0.8776, "step": 6647 }, { "epoch": 3.010869565217391, "grad_norm": 0.19006868142699443, "learning_rate": 3.2755351611605765e-05, "loss": 0.8582, "step": 6648 }, { "epoch": 3.0113224637681157, "grad_norm": 0.1864583186832977, "learning_rate": 3.2742913825565145e-05, "loss": 0.8552, "step": 6649 }, { "epoch": 3.011775362318841, "grad_norm": 0.16983648991736036, "learning_rate": 3.27304767650269e-05, "loss": 0.8628, "step": 6650 }, { "epoch": 3.0122282608695654, "grad_norm": 0.20070099530845223, "learning_rate": 3.2718040431234425e-05, "loss": 0.8496, "step": 6651 }, { "epoch": 3.01268115942029, "grad_norm": 0.2006371613205729, "learning_rate": 3.2705604825430976e-05, "loss": 0.8416, "step": 6652 }, { "epoch": 3.0131340579710146, "grad_norm": 0.18508550106564997, "learning_rate": 3.269316994885978e-05, "loss": 0.8701, "step": 6653 }, { "epoch": 3.0135869565217392, "grad_norm": 0.16006168649174426, "learning_rate": 3.2680735802763966e-05, "loss": 0.8441, "step": 6654 }, { "epoch": 3.014039855072464, "grad_norm": 0.15267726847359672, "learning_rate": 3.2668302388386594e-05, "loss": 0.8523, "step": 6655 }, { "epoch": 3.0144927536231885, "grad_norm": 0.2226291620285511, "learning_rate": 3.265586970697064e-05, "loss": 0.863, "step": 6656 }, { "epoch": 3.014945652173913, "grad_norm": 0.29145940461912845, "learning_rate": 3.264343775975904e-05, "loss": 0.8558, "step": 6657 }, { "epoch": 3.0153985507246377, "grad_norm": 0.2710077272175235, "learning_rate": 3.2631006547994616e-05, "loss": 0.8578, "step": 6658 }, { "epoch": 3.0158514492753623, "grad_norm": 0.18490699376719996, "learning_rate": 3.261857607292015e-05, "loss": 0.8719, "step": 6659 }, { "epoch": 3.016304347826087, "grad_norm": 0.24073309143292326, "learning_rate": 3.260614633577832e-05, "loss": 0.862, "step": 6660 }, { "epoch": 3.0167572463768115, "grad_norm": 0.1841034280038283, "learning_rate": 3.2593717337811775e-05, "loss": 0.8529, "step": 6661 }, { "epoch": 3.017210144927536, "grad_norm": 0.21968086214220395, "learning_rate": 3.258128908026303e-05, "loss": 0.8545, "step": 6662 }, { "epoch": 3.0176630434782608, "grad_norm": 0.24513139504300732, "learning_rate": 3.256886156437458e-05, "loss": 0.8486, "step": 6663 }, { "epoch": 3.0181159420289854, "grad_norm": 0.17662224641824592, "learning_rate": 3.2556434791388804e-05, "loss": 0.8587, "step": 6664 }, { "epoch": 3.01856884057971, "grad_norm": 0.21366132173751123, "learning_rate": 3.254400876254805e-05, "loss": 0.8567, "step": 6665 }, { "epoch": 3.0190217391304346, "grad_norm": 0.20265310127437136, "learning_rate": 3.253158347909455e-05, "loss": 0.8553, "step": 6666 }, { "epoch": 3.019474637681159, "grad_norm": 0.22533771677791337, "learning_rate": 3.2519158942270496e-05, "loss": 0.8517, "step": 6667 }, { "epoch": 3.0199275362318843, "grad_norm": 0.18027786361296233, "learning_rate": 3.250673515331795e-05, "loss": 0.882, "step": 6668 }, { "epoch": 3.020380434782609, "grad_norm": 0.19184769704455093, "learning_rate": 3.249431211347899e-05, "loss": 0.8483, "step": 6669 }, { "epoch": 3.0208333333333335, "grad_norm": 0.1617800843600765, "learning_rate": 3.248188982399553e-05, "loss": 0.8755, "step": 6670 }, { "epoch": 3.021286231884058, "grad_norm": 0.17075135616432072, "learning_rate": 3.2469468286109465e-05, "loss": 0.8476, "step": 6671 }, { "epoch": 3.0217391304347827, "grad_norm": 0.229228038246179, "learning_rate": 3.245704750106257e-05, "loss": 0.8466, "step": 6672 }, { "epoch": 3.0221920289855073, "grad_norm": 0.20142279789346157, "learning_rate": 3.244462747009661e-05, "loss": 0.8456, "step": 6673 }, { "epoch": 3.022644927536232, "grad_norm": 0.16692428348442379, "learning_rate": 3.2432208194453204e-05, "loss": 0.8507, "step": 6674 }, { "epoch": 3.0230978260869565, "grad_norm": 0.22700502723391436, "learning_rate": 3.241978967537395e-05, "loss": 0.8386, "step": 6675 }, { "epoch": 3.023550724637681, "grad_norm": 0.19015359716381727, "learning_rate": 3.240737191410031e-05, "loss": 0.8545, "step": 6676 }, { "epoch": 3.0240036231884058, "grad_norm": 0.20496097159586224, "learning_rate": 3.239495491187375e-05, "loss": 0.8593, "step": 6677 }, { "epoch": 3.0244565217391304, "grad_norm": 0.2360327994507851, "learning_rate": 3.238253866993559e-05, "loss": 0.8763, "step": 6678 }, { "epoch": 3.024909420289855, "grad_norm": 0.2200176937375178, "learning_rate": 3.2370123189527125e-05, "loss": 0.8552, "step": 6679 }, { "epoch": 3.0253623188405796, "grad_norm": 0.24403477614292898, "learning_rate": 3.2357708471889514e-05, "loss": 0.8475, "step": 6680 }, { "epoch": 3.025815217391304, "grad_norm": 0.2366586604144798, "learning_rate": 3.234529451826391e-05, "loss": 0.8601, "step": 6681 }, { "epoch": 3.026268115942029, "grad_norm": 0.16531471696301678, "learning_rate": 3.233288132989134e-05, "loss": 0.8544, "step": 6682 }, { "epoch": 3.0267210144927534, "grad_norm": 0.22531102510275144, "learning_rate": 3.2320468908012776e-05, "loss": 0.8503, "step": 6683 }, { "epoch": 3.027173913043478, "grad_norm": 0.23495830803452936, "learning_rate": 3.2308057253869095e-05, "loss": 0.852, "step": 6684 }, { "epoch": 3.027626811594203, "grad_norm": 0.17339180745470287, "learning_rate": 3.229564636870113e-05, "loss": 0.8594, "step": 6685 }, { "epoch": 3.0280797101449277, "grad_norm": 0.24999190369561305, "learning_rate": 3.2283236253749605e-05, "loss": 0.8592, "step": 6686 }, { "epoch": 3.0285326086956523, "grad_norm": 0.2462907918348973, "learning_rate": 3.227082691025518e-05, "loss": 0.8521, "step": 6687 }, { "epoch": 3.028985507246377, "grad_norm": 0.19153298619500542, "learning_rate": 3.225841833945843e-05, "loss": 0.8621, "step": 6688 }, { "epoch": 3.0294384057971016, "grad_norm": 0.2584759493629631, "learning_rate": 3.224601054259987e-05, "loss": 0.8431, "step": 6689 }, { "epoch": 3.029891304347826, "grad_norm": 0.21620959868578746, "learning_rate": 3.2233603520919913e-05, "loss": 0.8536, "step": 6690 }, { "epoch": 3.0303442028985508, "grad_norm": 0.17930079667937482, "learning_rate": 3.2221197275658945e-05, "loss": 0.8613, "step": 6691 }, { "epoch": 3.0307971014492754, "grad_norm": 0.20744971525281783, "learning_rate": 3.2208791808057185e-05, "loss": 0.8465, "step": 6692 }, { "epoch": 3.03125, "grad_norm": 0.15644242083570223, "learning_rate": 3.219638711935488e-05, "loss": 0.8614, "step": 6693 }, { "epoch": 3.0317028985507246, "grad_norm": 0.1799504677840616, "learning_rate": 3.218398321079211e-05, "loss": 0.8582, "step": 6694 }, { "epoch": 3.0321557971014492, "grad_norm": 0.1658022557393418, "learning_rate": 3.21715800836089e-05, "loss": 0.8515, "step": 6695 }, { "epoch": 3.032608695652174, "grad_norm": 0.15280175277487693, "learning_rate": 3.215917773904528e-05, "loss": 0.8611, "step": 6696 }, { "epoch": 3.0330615942028984, "grad_norm": 0.18188028238898823, "learning_rate": 3.214677617834105e-05, "loss": 0.8533, "step": 6697 }, { "epoch": 3.033514492753623, "grad_norm": 0.14702065164039654, "learning_rate": 3.2134375402736065e-05, "loss": 0.8644, "step": 6698 }, { "epoch": 3.0339673913043477, "grad_norm": 0.19032043559245693, "learning_rate": 3.212197541347002e-05, "loss": 0.8525, "step": 6699 }, { "epoch": 3.0344202898550723, "grad_norm": 0.19217670202342668, "learning_rate": 3.210957621178259e-05, "loss": 0.8686, "step": 6700 }, { "epoch": 3.034873188405797, "grad_norm": 0.1870083383265315, "learning_rate": 3.209717779891332e-05, "loss": 0.8701, "step": 6701 }, { "epoch": 3.035326086956522, "grad_norm": 0.19653789317084117, "learning_rate": 3.208478017610173e-05, "loss": 0.8606, "step": 6702 }, { "epoch": 3.0357789855072466, "grad_norm": 0.1884872080761389, "learning_rate": 3.207238334458718e-05, "loss": 0.8661, "step": 6703 }, { "epoch": 3.036231884057971, "grad_norm": 0.15678837075334456, "learning_rate": 3.205998730560905e-05, "loss": 0.8725, "step": 6704 }, { "epoch": 3.036684782608696, "grad_norm": 0.16145993013417448, "learning_rate": 3.204759206040657e-05, "loss": 0.847, "step": 6705 }, { "epoch": 3.0371376811594204, "grad_norm": 0.15859157610396346, "learning_rate": 3.203519761021891e-05, "loss": 0.8505, "step": 6706 }, { "epoch": 3.037590579710145, "grad_norm": 0.21259341603258364, "learning_rate": 3.202280395628515e-05, "loss": 0.8534, "step": 6707 }, { "epoch": 3.0380434782608696, "grad_norm": 0.19116098862217787, "learning_rate": 3.2010411099844345e-05, "loss": 0.8593, "step": 6708 }, { "epoch": 3.0384963768115942, "grad_norm": 0.161393616024798, "learning_rate": 3.199801904213539e-05, "loss": 0.8753, "step": 6709 }, { "epoch": 3.038949275362319, "grad_norm": 0.16009262916961473, "learning_rate": 3.198562778439715e-05, "loss": 0.8484, "step": 6710 }, { "epoch": 3.0394021739130435, "grad_norm": 0.16133061805033558, "learning_rate": 3.19732373278684e-05, "loss": 0.8496, "step": 6711 }, { "epoch": 3.039855072463768, "grad_norm": 0.15851866720591737, "learning_rate": 3.1960847673787844e-05, "loss": 0.8572, "step": 6712 }, { "epoch": 3.0403079710144927, "grad_norm": 0.1684442586501275, "learning_rate": 3.194845882339407e-05, "loss": 0.8565, "step": 6713 }, { "epoch": 3.0407608695652173, "grad_norm": 0.1569097964080471, "learning_rate": 3.193607077792565e-05, "loss": 0.8458, "step": 6714 }, { "epoch": 3.041213768115942, "grad_norm": 0.16418717899660698, "learning_rate": 3.192368353862098e-05, "loss": 0.8668, "step": 6715 }, { "epoch": 3.0416666666666665, "grad_norm": 0.1755538422931892, "learning_rate": 3.191129710671849e-05, "loss": 0.8537, "step": 6716 }, { "epoch": 3.042119565217391, "grad_norm": 0.1788933761874822, "learning_rate": 3.189891148345644e-05, "loss": 0.8442, "step": 6717 }, { "epoch": 3.0425724637681157, "grad_norm": 0.16185957406864432, "learning_rate": 3.188652667007305e-05, "loss": 0.8555, "step": 6718 }, { "epoch": 3.043025362318841, "grad_norm": 0.19395640969294348, "learning_rate": 3.187414266780643e-05, "loss": 0.8583, "step": 6719 }, { "epoch": 3.0434782608695654, "grad_norm": 0.18960365066938084, "learning_rate": 3.1861759477894656e-05, "loss": 0.8831, "step": 6720 }, { "epoch": 3.04393115942029, "grad_norm": 0.16946210499456618, "learning_rate": 3.184937710157568e-05, "loss": 0.8483, "step": 6721 }, { "epoch": 3.0443840579710146, "grad_norm": 0.21349155875968123, "learning_rate": 3.18369955400874e-05, "loss": 0.8402, "step": 6722 }, { "epoch": 3.0448369565217392, "grad_norm": 0.15960499313200419, "learning_rate": 3.182461479466759e-05, "loss": 0.8666, "step": 6723 }, { "epoch": 3.045289855072464, "grad_norm": 0.2622477358928996, "learning_rate": 3.1812234866554014e-05, "loss": 0.8565, "step": 6724 }, { "epoch": 3.0457427536231885, "grad_norm": 0.2634457535151387, "learning_rate": 3.179985575698429e-05, "loss": 0.8664, "step": 6725 }, { "epoch": 3.046195652173913, "grad_norm": 0.3007938244154644, "learning_rate": 3.178747746719598e-05, "loss": 0.876, "step": 6726 }, { "epoch": 3.0466485507246377, "grad_norm": 0.2669530891200618, "learning_rate": 3.1775099998426545e-05, "loss": 0.8378, "step": 6727 }, { "epoch": 3.0471014492753623, "grad_norm": 0.3075326641075669, "learning_rate": 3.1762723351913414e-05, "loss": 0.8781, "step": 6728 }, { "epoch": 3.047554347826087, "grad_norm": 0.15392265332691343, "learning_rate": 3.175034752889388e-05, "loss": 0.8566, "step": 6729 }, { "epoch": 3.0480072463768115, "grad_norm": 0.23675093087944393, "learning_rate": 3.173797253060518e-05, "loss": 0.8585, "step": 6730 }, { "epoch": 3.048460144927536, "grad_norm": 0.25563770020622106, "learning_rate": 3.172559835828444e-05, "loss": 0.8424, "step": 6731 }, { "epoch": 3.0489130434782608, "grad_norm": 0.18554941650982273, "learning_rate": 3.171322501316876e-05, "loss": 0.8545, "step": 6732 }, { "epoch": 3.0493659420289854, "grad_norm": 0.2039898295476159, "learning_rate": 3.1700852496495106e-05, "loss": 0.8596, "step": 6733 }, { "epoch": 3.04981884057971, "grad_norm": 0.17861394573857237, "learning_rate": 3.168848080950038e-05, "loss": 0.8402, "step": 6734 }, { "epoch": 3.0502717391304346, "grad_norm": 0.22388932795582286, "learning_rate": 3.167610995342139e-05, "loss": 0.8679, "step": 6735 }, { "epoch": 3.050724637681159, "grad_norm": 0.27067891731215094, "learning_rate": 3.166373992949489e-05, "loss": 0.8482, "step": 6736 }, { "epoch": 3.0511775362318843, "grad_norm": 0.2554803956777641, "learning_rate": 3.16513707389575e-05, "loss": 0.858, "step": 6737 }, { "epoch": 3.051630434782609, "grad_norm": 0.2219932232510756, "learning_rate": 3.1639002383045834e-05, "loss": 0.8672, "step": 6738 }, { "epoch": 3.0520833333333335, "grad_norm": 0.14286135229437275, "learning_rate": 3.162663486299632e-05, "loss": 0.8581, "step": 6739 }, { "epoch": 3.052536231884058, "grad_norm": 0.18482288658342716, "learning_rate": 3.161426818004541e-05, "loss": 0.8534, "step": 6740 }, { "epoch": 3.0529891304347827, "grad_norm": 0.20663217535164013, "learning_rate": 3.160190233542938e-05, "loss": 0.8469, "step": 6741 }, { "epoch": 3.0534420289855073, "grad_norm": 0.20919875073123437, "learning_rate": 3.15895373303845e-05, "loss": 0.8705, "step": 6742 }, { "epoch": 3.053894927536232, "grad_norm": 0.17672190669173124, "learning_rate": 3.157717316614689e-05, "loss": 0.8609, "step": 6743 }, { "epoch": 3.0543478260869565, "grad_norm": 0.18427415672688258, "learning_rate": 3.156480984395263e-05, "loss": 0.8648, "step": 6744 }, { "epoch": 3.054800724637681, "grad_norm": 0.2026357494832107, "learning_rate": 3.1552447365037686e-05, "loss": 0.8434, "step": 6745 }, { "epoch": 3.0552536231884058, "grad_norm": 0.15931439372934592, "learning_rate": 3.1540085730637975e-05, "loss": 0.856, "step": 6746 }, { "epoch": 3.0557065217391304, "grad_norm": 0.21906560214397727, "learning_rate": 3.15277249419893e-05, "loss": 0.8521, "step": 6747 }, { "epoch": 3.056159420289855, "grad_norm": 0.2391770189228384, "learning_rate": 3.151536500032739e-05, "loss": 0.8626, "step": 6748 }, { "epoch": 3.0566123188405796, "grad_norm": 0.18115958353528722, "learning_rate": 3.150300590688787e-05, "loss": 0.8831, "step": 6749 }, { "epoch": 3.057065217391304, "grad_norm": 0.21574730726520386, "learning_rate": 3.1490647662906336e-05, "loss": 0.8714, "step": 6750 }, { "epoch": 3.057518115942029, "grad_norm": 0.16429038420393535, "learning_rate": 3.147829026961822e-05, "loss": 0.855, "step": 6751 }, { "epoch": 3.0579710144927534, "grad_norm": 0.2054152044874009, "learning_rate": 3.146593372825895e-05, "loss": 0.8636, "step": 6752 }, { "epoch": 3.058423913043478, "grad_norm": 0.2279024610220602, "learning_rate": 3.145357804006379e-05, "loss": 0.8634, "step": 6753 }, { "epoch": 3.058876811594203, "grad_norm": 0.19152982925670256, "learning_rate": 3.144122320626799e-05, "loss": 0.839, "step": 6754 }, { "epoch": 3.0593297101449277, "grad_norm": 0.20931562236644669, "learning_rate": 3.142886922810666e-05, "loss": 0.8459, "step": 6755 }, { "epoch": 3.0597826086956523, "grad_norm": 0.17998101465254754, "learning_rate": 3.141651610681486e-05, "loss": 0.852, "step": 6756 }, { "epoch": 3.060235507246377, "grad_norm": 0.2449873518524221, "learning_rate": 3.140416384362754e-05, "loss": 0.8648, "step": 6757 }, { "epoch": 3.0606884057971016, "grad_norm": 0.2462620712274549, "learning_rate": 3.1391812439779594e-05, "loss": 0.874, "step": 6758 }, { "epoch": 3.061141304347826, "grad_norm": 0.23371001246158754, "learning_rate": 3.1379461896505784e-05, "loss": 0.8702, "step": 6759 }, { "epoch": 3.0615942028985508, "grad_norm": 0.27829179602159837, "learning_rate": 3.136711221504085e-05, "loss": 0.8483, "step": 6760 }, { "epoch": 3.0620471014492754, "grad_norm": 0.2073448170018715, "learning_rate": 3.135476339661936e-05, "loss": 0.8569, "step": 6761 }, { "epoch": 3.0625, "grad_norm": 0.2052470992565739, "learning_rate": 3.134241544247589e-05, "loss": 0.8689, "step": 6762 }, { "epoch": 3.0629528985507246, "grad_norm": 0.25530012968418764, "learning_rate": 3.133006835384486e-05, "loss": 0.8735, "step": 6763 }, { "epoch": 3.0634057971014492, "grad_norm": 0.2652808301193937, "learning_rate": 3.1317722131960645e-05, "loss": 0.8696, "step": 6764 }, { "epoch": 3.063858695652174, "grad_norm": 0.32528810060948676, "learning_rate": 3.13053767780575e-05, "loss": 0.866, "step": 6765 }, { "epoch": 3.0643115942028984, "grad_norm": 0.23658441918620837, "learning_rate": 3.129303229336962e-05, "loss": 0.8811, "step": 6766 }, { "epoch": 3.064764492753623, "grad_norm": 0.21444178110243836, "learning_rate": 3.1280688679131096e-05, "loss": 0.867, "step": 6767 }, { "epoch": 3.0652173913043477, "grad_norm": 0.23665437849703044, "learning_rate": 3.126834593657595e-05, "loss": 0.852, "step": 6768 }, { "epoch": 3.0656702898550723, "grad_norm": 0.2847237500398125, "learning_rate": 3.125600406693807e-05, "loss": 0.8632, "step": 6769 }, { "epoch": 3.066123188405797, "grad_norm": 0.30756524276826264, "learning_rate": 3.124366307145135e-05, "loss": 0.845, "step": 6770 }, { "epoch": 3.066576086956522, "grad_norm": 0.2352020881803264, "learning_rate": 3.1231322951349494e-05, "loss": 0.856, "step": 6771 }, { "epoch": 3.0670289855072466, "grad_norm": 0.18412401060377565, "learning_rate": 3.121898370786619e-05, "loss": 0.8828, "step": 6772 }, { "epoch": 3.067481884057971, "grad_norm": 0.2229559813779249, "learning_rate": 3.120664534223498e-05, "loss": 0.8658, "step": 6773 }, { "epoch": 3.067934782608696, "grad_norm": 0.23221809874919475, "learning_rate": 3.1194307855689386e-05, "loss": 0.8418, "step": 6774 }, { "epoch": 3.0683876811594204, "grad_norm": 0.28437738926475314, "learning_rate": 3.118197124946278e-05, "loss": 0.8503, "step": 6775 }, { "epoch": 3.068840579710145, "grad_norm": 0.262976911784404, "learning_rate": 3.116963552478849e-05, "loss": 0.8864, "step": 6776 }, { "epoch": 3.0692934782608696, "grad_norm": 0.18205074130949472, "learning_rate": 3.1157300682899715e-05, "loss": 0.8456, "step": 6777 }, { "epoch": 3.0697463768115942, "grad_norm": 0.19498319287372576, "learning_rate": 3.1144966725029616e-05, "loss": 0.8523, "step": 6778 }, { "epoch": 3.070199275362319, "grad_norm": 0.2196476055733904, "learning_rate": 3.113263365241122e-05, "loss": 0.8589, "step": 6779 }, { "epoch": 3.0706521739130435, "grad_norm": 0.25179547830132576, "learning_rate": 3.11203014662775e-05, "loss": 0.8797, "step": 6780 }, { "epoch": 3.071105072463768, "grad_norm": 0.28915920384240656, "learning_rate": 3.110797016786129e-05, "loss": 0.8515, "step": 6781 }, { "epoch": 3.0715579710144927, "grad_norm": 0.29897467896799756, "learning_rate": 3.1095639758395416e-05, "loss": 0.8596, "step": 6782 }, { "epoch": 3.0720108695652173, "grad_norm": 0.2899091436034674, "learning_rate": 3.1083310239112524e-05, "loss": 0.8629, "step": 6783 }, { "epoch": 3.072463768115942, "grad_norm": 0.16146311881038955, "learning_rate": 3.107098161124526e-05, "loss": 0.8539, "step": 6784 }, { "epoch": 3.0729166666666665, "grad_norm": 0.17718210077931895, "learning_rate": 3.1058653876026105e-05, "loss": 0.8582, "step": 6785 }, { "epoch": 3.073369565217391, "grad_norm": 0.17480573948024292, "learning_rate": 3.104632703468749e-05, "loss": 0.8641, "step": 6786 }, { "epoch": 3.0738224637681157, "grad_norm": 0.14981118939805171, "learning_rate": 3.103400108846174e-05, "loss": 0.8408, "step": 6787 }, { "epoch": 3.074275362318841, "grad_norm": 0.16259312684247082, "learning_rate": 3.102167603858113e-05, "loss": 0.8645, "step": 6788 }, { "epoch": 3.0747282608695654, "grad_norm": 0.18778102047290196, "learning_rate": 3.1009351886277774e-05, "loss": 0.8454, "step": 6789 }, { "epoch": 3.07518115942029, "grad_norm": 0.16547004475001656, "learning_rate": 3.0997028632783774e-05, "loss": 0.8662, "step": 6790 }, { "epoch": 3.0756340579710146, "grad_norm": 0.17521651247740724, "learning_rate": 3.098470627933107e-05, "loss": 0.8505, "step": 6791 }, { "epoch": 3.0760869565217392, "grad_norm": 0.1474331783732181, "learning_rate": 3.097238482715158e-05, "loss": 0.847, "step": 6792 }, { "epoch": 3.076539855072464, "grad_norm": 0.1771911520969385, "learning_rate": 3.0960064277477074e-05, "loss": 0.8558, "step": 6793 }, { "epoch": 3.0769927536231885, "grad_norm": 0.20610414390080226, "learning_rate": 3.094774463153928e-05, "loss": 0.8655, "step": 6794 }, { "epoch": 3.077445652173913, "grad_norm": 0.15796328575692584, "learning_rate": 3.093542589056978e-05, "loss": 0.8538, "step": 6795 }, { "epoch": 3.0778985507246377, "grad_norm": 0.1888442534169401, "learning_rate": 3.0923108055800136e-05, "loss": 0.87, "step": 6796 }, { "epoch": 3.0783514492753623, "grad_norm": 0.1949571172513104, "learning_rate": 3.091079112846175e-05, "loss": 0.87, "step": 6797 }, { "epoch": 3.078804347826087, "grad_norm": 0.20584827122094665, "learning_rate": 3.0898475109785994e-05, "loss": 0.8625, "step": 6798 }, { "epoch": 3.0792572463768115, "grad_norm": 0.1657239908659467, "learning_rate": 3.0886160001004086e-05, "loss": 0.8502, "step": 6799 }, { "epoch": 3.079710144927536, "grad_norm": 0.198366042517742, "learning_rate": 3.087384580334722e-05, "loss": 0.8657, "step": 6800 }, { "epoch": 3.0801630434782608, "grad_norm": 0.1421671565715627, "learning_rate": 3.086153251804644e-05, "loss": 0.8539, "step": 6801 }, { "epoch": 3.0806159420289854, "grad_norm": 0.19073308097238062, "learning_rate": 3.0849220146332745e-05, "loss": 0.8639, "step": 6802 }, { "epoch": 3.08106884057971, "grad_norm": 0.19630850708429137, "learning_rate": 3.083690868943699e-05, "loss": 0.8546, "step": 6803 }, { "epoch": 3.0815217391304346, "grad_norm": 0.18030680235691943, "learning_rate": 3.082459814859002e-05, "loss": 0.8402, "step": 6804 }, { "epoch": 3.081974637681159, "grad_norm": 0.22131516192992404, "learning_rate": 3.08122885250225e-05, "loss": 0.8579, "step": 6805 }, { "epoch": 3.0824275362318843, "grad_norm": 0.1532207986495874, "learning_rate": 3.0799979819965067e-05, "loss": 0.8566, "step": 6806 }, { "epoch": 3.082880434782609, "grad_norm": 0.18311687788765255, "learning_rate": 3.0787672034648215e-05, "loss": 0.8598, "step": 6807 }, { "epoch": 3.0833333333333335, "grad_norm": 0.1728235471418326, "learning_rate": 3.07753651703024e-05, "loss": 0.8696, "step": 6808 }, { "epoch": 3.083786231884058, "grad_norm": 0.16140484555384568, "learning_rate": 3.076305922815793e-05, "loss": 0.8495, "step": 6809 }, { "epoch": 3.0842391304347827, "grad_norm": 0.1975824415191785, "learning_rate": 3.075075420944509e-05, "loss": 0.8673, "step": 6810 }, { "epoch": 3.0846920289855073, "grad_norm": 0.15556345768141852, "learning_rate": 3.073845011539398e-05, "loss": 0.8422, "step": 6811 }, { "epoch": 3.085144927536232, "grad_norm": 0.19076165986806798, "learning_rate": 3.072614694723472e-05, "loss": 0.8678, "step": 6812 }, { "epoch": 3.0855978260869565, "grad_norm": 0.1999518840155697, "learning_rate": 3.0713844706197224e-05, "loss": 0.8615, "step": 6813 }, { "epoch": 3.086050724637681, "grad_norm": 0.17172478918440565, "learning_rate": 3.07015433935114e-05, "loss": 0.8648, "step": 6814 }, { "epoch": 3.0865036231884058, "grad_norm": 0.21790489862889065, "learning_rate": 3.0689243010406994e-05, "loss": 0.8497, "step": 6815 }, { "epoch": 3.0869565217391304, "grad_norm": 0.19027664118769208, "learning_rate": 3.0676943558113745e-05, "loss": 0.8426, "step": 6816 }, { "epoch": 3.087409420289855, "grad_norm": 0.25728232092972614, "learning_rate": 3.066464503786121e-05, "loss": 0.856, "step": 6817 }, { "epoch": 3.0878623188405796, "grad_norm": 0.22262375930959394, "learning_rate": 3.065234745087891e-05, "loss": 0.8358, "step": 6818 }, { "epoch": 3.088315217391304, "grad_norm": 0.17598527207187747, "learning_rate": 3.064005079839625e-05, "loss": 0.85, "step": 6819 }, { "epoch": 3.088768115942029, "grad_norm": 0.23655064650077917, "learning_rate": 3.0627755081642526e-05, "loss": 0.8565, "step": 6820 }, { "epoch": 3.0892210144927534, "grad_norm": 0.1507100617555858, "learning_rate": 3.0615460301846997e-05, "loss": 0.8768, "step": 6821 }, { "epoch": 3.089673913043478, "grad_norm": 0.25860661465484114, "learning_rate": 3.060316646023876e-05, "loss": 0.8642, "step": 6822 }, { "epoch": 3.090126811594203, "grad_norm": 0.2385983016000661, "learning_rate": 3.059087355804688e-05, "loss": 0.8632, "step": 6823 }, { "epoch": 3.0905797101449277, "grad_norm": 0.21392042248079352, "learning_rate": 3.057858159650025e-05, "loss": 0.8665, "step": 6824 }, { "epoch": 3.0910326086956523, "grad_norm": 0.2876586361567566, "learning_rate": 3.0566290576827775e-05, "loss": 0.851, "step": 6825 }, { "epoch": 3.091485507246377, "grad_norm": 0.22948660400574206, "learning_rate": 3.0554000500258166e-05, "loss": 0.8419, "step": 6826 }, { "epoch": 3.0919384057971016, "grad_norm": 0.2583391390663131, "learning_rate": 3.054171136802011e-05, "loss": 0.8629, "step": 6827 }, { "epoch": 3.092391304347826, "grad_norm": 0.23045729689059638, "learning_rate": 3.052942318134213e-05, "loss": 0.8509, "step": 6828 }, { "epoch": 3.0928442028985508, "grad_norm": 0.19305926838662824, "learning_rate": 3.051713594145275e-05, "loss": 0.8447, "step": 6829 }, { "epoch": 3.0932971014492754, "grad_norm": 0.26946909238621364, "learning_rate": 3.050484964958029e-05, "loss": 0.877, "step": 6830 }, { "epoch": 3.09375, "grad_norm": 0.3285936971230428, "learning_rate": 3.0492564306953083e-05, "loss": 0.8418, "step": 6831 }, { "epoch": 3.0942028985507246, "grad_norm": 0.2911905801452004, "learning_rate": 3.0480279914799276e-05, "loss": 0.8547, "step": 6832 }, { "epoch": 3.0946557971014492, "grad_norm": 0.2309922883111777, "learning_rate": 3.0467996474346978e-05, "loss": 0.8595, "step": 6833 }, { "epoch": 3.095108695652174, "grad_norm": 0.23578716606290662, "learning_rate": 3.045571398682416e-05, "loss": 0.8754, "step": 6834 }, { "epoch": 3.0955615942028984, "grad_norm": 0.3460648493218287, "learning_rate": 3.0443432453458758e-05, "loss": 0.8535, "step": 6835 }, { "epoch": 3.096014492753623, "grad_norm": 0.2803799763209083, "learning_rate": 3.043115187547855e-05, "loss": 0.8315, "step": 6836 }, { "epoch": 3.0964673913043477, "grad_norm": 0.24015611909102566, "learning_rate": 3.0418872254111254e-05, "loss": 0.8411, "step": 6837 }, { "epoch": 3.0969202898550723, "grad_norm": 0.19343313540327844, "learning_rate": 3.040659359058447e-05, "loss": 0.8701, "step": 6838 }, { "epoch": 3.097373188405797, "grad_norm": 0.20312404195432884, "learning_rate": 3.0394315886125746e-05, "loss": 0.8606, "step": 6839 }, { "epoch": 3.097826086956522, "grad_norm": 0.2857903267497234, "learning_rate": 3.0382039141962473e-05, "loss": 0.8692, "step": 6840 }, { "epoch": 3.0982789855072466, "grad_norm": 0.2957803060008469, "learning_rate": 3.0369763359321992e-05, "loss": 0.8531, "step": 6841 }, { "epoch": 3.098731884057971, "grad_norm": 0.20637590129711514, "learning_rate": 3.035748853943152e-05, "loss": 0.8635, "step": 6842 }, { "epoch": 3.099184782608696, "grad_norm": 0.1970855670404413, "learning_rate": 3.0345214683518204e-05, "loss": 0.8439, "step": 6843 }, { "epoch": 3.0996376811594204, "grad_norm": 0.2283834573117366, "learning_rate": 3.0332941792809077e-05, "loss": 0.8712, "step": 6844 }, { "epoch": 3.100090579710145, "grad_norm": 0.18934836661498616, "learning_rate": 3.032066986853108e-05, "loss": 0.8536, "step": 6845 }, { "epoch": 3.1005434782608696, "grad_norm": 0.20183680445455648, "learning_rate": 3.0308398911911047e-05, "loss": 0.8548, "step": 6846 }, { "epoch": 3.1009963768115942, "grad_norm": 0.20501911056756134, "learning_rate": 3.029612892417574e-05, "loss": 0.8603, "step": 6847 }, { "epoch": 3.101449275362319, "grad_norm": 0.25975033382198415, "learning_rate": 3.0283859906551795e-05, "loss": 0.8939, "step": 6848 }, { "epoch": 3.1019021739130435, "grad_norm": 0.23628999978078158, "learning_rate": 3.0271591860265783e-05, "loss": 0.8501, "step": 6849 }, { "epoch": 3.102355072463768, "grad_norm": 0.2338971908045688, "learning_rate": 3.0259324786544128e-05, "loss": 0.851, "step": 6850 }, { "epoch": 3.1028079710144927, "grad_norm": 0.2044693453677495, "learning_rate": 3.0247058686613227e-05, "loss": 0.858, "step": 6851 }, { "epoch": 3.1032608695652173, "grad_norm": 0.4462926856086574, "learning_rate": 3.023479356169932e-05, "loss": 0.8421, "step": 6852 }, { "epoch": 3.103713768115942, "grad_norm": 0.24018879163962759, "learning_rate": 3.022252941302858e-05, "loss": 0.8529, "step": 6853 }, { "epoch": 3.1041666666666665, "grad_norm": 0.20461642429772345, "learning_rate": 3.0210266241827047e-05, "loss": 0.8596, "step": 6854 }, { "epoch": 3.104619565217391, "grad_norm": 0.1720676005917127, "learning_rate": 3.0198004049320735e-05, "loss": 0.8364, "step": 6855 }, { "epoch": 3.1050724637681157, "grad_norm": 0.1855448240611063, "learning_rate": 3.0185742836735483e-05, "loss": 0.8617, "step": 6856 }, { "epoch": 3.105525362318841, "grad_norm": 0.1747440225861715, "learning_rate": 3.0173482605297075e-05, "loss": 0.8596, "step": 6857 }, { "epoch": 3.1059782608695654, "grad_norm": 0.15862991459691136, "learning_rate": 3.016122335623117e-05, "loss": 0.8248, "step": 6858 }, { "epoch": 3.10643115942029, "grad_norm": 0.18597921457438502, "learning_rate": 3.014896509076337e-05, "loss": 0.8561, "step": 6859 }, { "epoch": 3.1068840579710146, "grad_norm": 0.14452484341793526, "learning_rate": 3.013670781011913e-05, "loss": 0.8585, "step": 6860 }, { "epoch": 3.1073369565217392, "grad_norm": 0.2014645329184823, "learning_rate": 3.0124451515523857e-05, "loss": 0.8687, "step": 6861 }, { "epoch": 3.107789855072464, "grad_norm": 0.18530025022224506, "learning_rate": 3.0112196208202787e-05, "loss": 0.8588, "step": 6862 }, { "epoch": 3.1082427536231885, "grad_norm": 0.18074285742470422, "learning_rate": 3.0099941889381154e-05, "loss": 0.8565, "step": 6863 }, { "epoch": 3.108695652173913, "grad_norm": 0.19777111280081422, "learning_rate": 3.008768856028401e-05, "loss": 0.8464, "step": 6864 }, { "epoch": 3.1091485507246377, "grad_norm": 0.16290771804020393, "learning_rate": 3.0075436222136357e-05, "loss": 0.8571, "step": 6865 }, { "epoch": 3.1096014492753623, "grad_norm": 0.1802122417783604, "learning_rate": 3.006318487616305e-05, "loss": 0.8613, "step": 6866 }, { "epoch": 3.110054347826087, "grad_norm": 0.1827753525222863, "learning_rate": 3.005093452358892e-05, "loss": 0.8577, "step": 6867 }, { "epoch": 3.1105072463768115, "grad_norm": 0.2532759535246632, "learning_rate": 3.0038685165638617e-05, "loss": 0.8521, "step": 6868 }, { "epoch": 3.110960144927536, "grad_norm": 0.23945521199405295, "learning_rate": 3.0026436803536758e-05, "loss": 0.8586, "step": 6869 }, { "epoch": 3.1114130434782608, "grad_norm": 0.17381155072956989, "learning_rate": 3.0014189438507793e-05, "loss": 0.848, "step": 6870 }, { "epoch": 3.1118659420289854, "grad_norm": 0.2134541792500267, "learning_rate": 3.0001943071776155e-05, "loss": 0.8638, "step": 6871 }, { "epoch": 3.11231884057971, "grad_norm": 0.18974413243784066, "learning_rate": 2.9989697704566095e-05, "loss": 0.8639, "step": 6872 }, { "epoch": 3.1127717391304346, "grad_norm": 0.19580283279375524, "learning_rate": 2.9977453338101834e-05, "loss": 0.8557, "step": 6873 }, { "epoch": 3.113224637681159, "grad_norm": 0.20368222787619936, "learning_rate": 2.9965209973607436e-05, "loss": 0.8547, "step": 6874 }, { "epoch": 3.1136775362318843, "grad_norm": 0.17542162519730797, "learning_rate": 2.9952967612306906e-05, "loss": 0.8666, "step": 6875 }, { "epoch": 3.114130434782609, "grad_norm": 0.1860764955621095, "learning_rate": 2.9940726255424107e-05, "loss": 0.8644, "step": 6876 }, { "epoch": 3.1145833333333335, "grad_norm": 0.2333878693969874, "learning_rate": 2.9928485904182865e-05, "loss": 0.8499, "step": 6877 }, { "epoch": 3.115036231884058, "grad_norm": 0.28856082055911897, "learning_rate": 2.9916246559806832e-05, "loss": 0.883, "step": 6878 }, { "epoch": 3.1154891304347827, "grad_norm": 0.20575438494544912, "learning_rate": 2.9904008223519624e-05, "loss": 0.8666, "step": 6879 }, { "epoch": 3.1159420289855073, "grad_norm": 0.1812812021010333, "learning_rate": 2.9891770896544695e-05, "loss": 0.8542, "step": 6880 }, { "epoch": 3.116394927536232, "grad_norm": 0.2106092943130064, "learning_rate": 2.9879534580105463e-05, "loss": 0.8621, "step": 6881 }, { "epoch": 3.1168478260869565, "grad_norm": 0.2824132013534304, "learning_rate": 2.9867299275425185e-05, "loss": 0.8683, "step": 6882 }, { "epoch": 3.117300724637681, "grad_norm": 0.2739199887263654, "learning_rate": 2.985506498372707e-05, "loss": 0.8499, "step": 6883 }, { "epoch": 3.1177536231884058, "grad_norm": 0.2255786041862764, "learning_rate": 2.984283170623416e-05, "loss": 0.839, "step": 6884 }, { "epoch": 3.1182065217391304, "grad_norm": 0.16065635966495428, "learning_rate": 2.9830599444169484e-05, "loss": 0.8507, "step": 6885 }, { "epoch": 3.118659420289855, "grad_norm": 0.2429060679946958, "learning_rate": 2.9818368198755886e-05, "loss": 0.8522, "step": 6886 }, { "epoch": 3.1191123188405796, "grad_norm": 0.2400408958388738, "learning_rate": 2.9806137971216157e-05, "loss": 0.8577, "step": 6887 }, { "epoch": 3.119565217391304, "grad_norm": 0.20572915864966382, "learning_rate": 2.9793908762772954e-05, "loss": 0.859, "step": 6888 }, { "epoch": 3.120018115942029, "grad_norm": 0.17167830339016965, "learning_rate": 2.9781680574648885e-05, "loss": 0.8545, "step": 6889 }, { "epoch": 3.1204710144927534, "grad_norm": 0.1590697412018398, "learning_rate": 2.9769453408066388e-05, "loss": 0.8592, "step": 6890 }, { "epoch": 3.120923913043478, "grad_norm": 0.1942543751398582, "learning_rate": 2.9757227264247856e-05, "loss": 0.8743, "step": 6891 }, { "epoch": 3.121376811594203, "grad_norm": 0.14776369965397984, "learning_rate": 2.9745002144415526e-05, "loss": 0.8652, "step": 6892 }, { "epoch": 3.1218297101449277, "grad_norm": 0.17175243444245927, "learning_rate": 2.9732778049791603e-05, "loss": 0.8688, "step": 6893 }, { "epoch": 3.1222826086956523, "grad_norm": 0.16667062164625585, "learning_rate": 2.9720554981598123e-05, "loss": 0.8714, "step": 6894 }, { "epoch": 3.122735507246377, "grad_norm": 0.1499576161915881, "learning_rate": 2.9708332941057053e-05, "loss": 0.8685, "step": 6895 }, { "epoch": 3.1231884057971016, "grad_norm": 0.17785775960706646, "learning_rate": 2.9696111929390234e-05, "loss": 0.8568, "step": 6896 }, { "epoch": 3.123641304347826, "grad_norm": 0.16406660618305435, "learning_rate": 2.9683891947819447e-05, "loss": 0.8708, "step": 6897 }, { "epoch": 3.1240942028985508, "grad_norm": 0.1480019838623719, "learning_rate": 2.967167299756632e-05, "loss": 0.8718, "step": 6898 }, { "epoch": 3.1245471014492754, "grad_norm": 0.17334942809273718, "learning_rate": 2.965945507985242e-05, "loss": 0.8792, "step": 6899 }, { "epoch": 3.125, "grad_norm": 0.13652886508686707, "learning_rate": 2.9647238195899168e-05, "loss": 0.836, "step": 6900 }, { "epoch": 3.1254528985507246, "grad_norm": 0.16837066956517588, "learning_rate": 2.9635022346927934e-05, "loss": 0.862, "step": 6901 }, { "epoch": 3.1259057971014492, "grad_norm": 0.15178764232010628, "learning_rate": 2.962280753415993e-05, "loss": 0.8442, "step": 6902 }, { "epoch": 3.126358695652174, "grad_norm": 0.18347507962762577, "learning_rate": 2.9610593758816314e-05, "loss": 0.8622, "step": 6903 }, { "epoch": 3.1268115942028984, "grad_norm": 0.19375208024457935, "learning_rate": 2.9598381022118086e-05, "loss": 0.8542, "step": 6904 }, { "epoch": 3.127264492753623, "grad_norm": 0.17647089782709813, "learning_rate": 2.95861693252862e-05, "loss": 0.8518, "step": 6905 }, { "epoch": 3.1277173913043477, "grad_norm": 0.19550310368159757, "learning_rate": 2.9573958669541466e-05, "loss": 0.8787, "step": 6906 }, { "epoch": 3.1281702898550723, "grad_norm": 0.15984731151010287, "learning_rate": 2.9561749056104612e-05, "loss": 0.8675, "step": 6907 }, { "epoch": 3.128623188405797, "grad_norm": 0.1715700818202777, "learning_rate": 2.9549540486196232e-05, "loss": 0.8546, "step": 6908 }, { "epoch": 3.1290760869565215, "grad_norm": 0.19906291646645582, "learning_rate": 2.9537332961036863e-05, "loss": 0.8595, "step": 6909 }, { "epoch": 3.1295289855072466, "grad_norm": 0.16628278884802117, "learning_rate": 2.952512648184689e-05, "loss": 0.8693, "step": 6910 }, { "epoch": 3.129981884057971, "grad_norm": 0.17589734720071407, "learning_rate": 2.9512921049846626e-05, "loss": 0.8609, "step": 6911 }, { "epoch": 3.130434782608696, "grad_norm": 0.20440010484517337, "learning_rate": 2.950071666625625e-05, "loss": 0.859, "step": 6912 }, { "epoch": 3.1308876811594204, "grad_norm": 0.170986974758157, "learning_rate": 2.9488513332295883e-05, "loss": 0.8489, "step": 6913 }, { "epoch": 3.131340579710145, "grad_norm": 0.1936060483028358, "learning_rate": 2.9476311049185483e-05, "loss": 0.8264, "step": 6914 }, { "epoch": 3.1317934782608696, "grad_norm": 0.22255627122945343, "learning_rate": 2.946410981814496e-05, "loss": 0.8588, "step": 6915 }, { "epoch": 3.1322463768115942, "grad_norm": 0.23198286160972312, "learning_rate": 2.9451909640394052e-05, "loss": 0.8779, "step": 6916 }, { "epoch": 3.132699275362319, "grad_norm": 0.23647227152871964, "learning_rate": 2.943971051715247e-05, "loss": 0.868, "step": 6917 }, { "epoch": 3.1331521739130435, "grad_norm": 0.254501060523419, "learning_rate": 2.9427512449639744e-05, "loss": 0.8743, "step": 6918 }, { "epoch": 3.133605072463768, "grad_norm": 0.21431479745226642, "learning_rate": 2.9415315439075375e-05, "loss": 0.8433, "step": 6919 }, { "epoch": 3.1340579710144927, "grad_norm": 0.21215879600465995, "learning_rate": 2.9403119486678675e-05, "loss": 0.8688, "step": 6920 }, { "epoch": 3.1345108695652173, "grad_norm": 0.27731409927051504, "learning_rate": 2.939092459366893e-05, "loss": 0.8464, "step": 6921 }, { "epoch": 3.134963768115942, "grad_norm": 0.17463622265286893, "learning_rate": 2.9378730761265243e-05, "loss": 0.8668, "step": 6922 }, { "epoch": 3.1354166666666665, "grad_norm": 0.24476795507159171, "learning_rate": 2.936653799068669e-05, "loss": 0.8769, "step": 6923 }, { "epoch": 3.135869565217391, "grad_norm": 0.18454574591573694, "learning_rate": 2.9354346283152178e-05, "loss": 0.87, "step": 6924 }, { "epoch": 3.1363224637681157, "grad_norm": 0.18520970266053569, "learning_rate": 2.9342155639880542e-05, "loss": 0.8535, "step": 6925 }, { "epoch": 3.136775362318841, "grad_norm": 0.23371266310044495, "learning_rate": 2.9329966062090482e-05, "loss": 0.8501, "step": 6926 }, { "epoch": 3.1372282608695654, "grad_norm": 0.20122784183589495, "learning_rate": 2.931777755100064e-05, "loss": 0.8571, "step": 6927 }, { "epoch": 3.13768115942029, "grad_norm": 0.23744662584708073, "learning_rate": 2.9305590107829495e-05, "loss": 0.8671, "step": 6928 }, { "epoch": 3.1381340579710146, "grad_norm": 0.2276616103513532, "learning_rate": 2.9293403733795462e-05, "loss": 0.8673, "step": 6929 }, { "epoch": 3.1385869565217392, "grad_norm": 0.16173381224303934, "learning_rate": 2.92812184301168e-05, "loss": 0.8302, "step": 6930 }, { "epoch": 3.139039855072464, "grad_norm": 0.164744224735876, "learning_rate": 2.9269034198011743e-05, "loss": 0.8395, "step": 6931 }, { "epoch": 3.1394927536231885, "grad_norm": 0.15875438917035672, "learning_rate": 2.925685103869833e-05, "loss": 0.8511, "step": 6932 }, { "epoch": 3.139945652173913, "grad_norm": 0.17092087502645348, "learning_rate": 2.924466895339455e-05, "loss": 0.8606, "step": 6933 }, { "epoch": 3.1403985507246377, "grad_norm": 0.17018297102991545, "learning_rate": 2.9232487943318245e-05, "loss": 0.8371, "step": 6934 }, { "epoch": 3.1408514492753623, "grad_norm": 0.1820659706598677, "learning_rate": 2.9220308009687195e-05, "loss": 0.8443, "step": 6935 }, { "epoch": 3.141304347826087, "grad_norm": 0.17503555966514808, "learning_rate": 2.9208129153719026e-05, "loss": 0.8518, "step": 6936 }, { "epoch": 3.1417572463768115, "grad_norm": 0.19867989449368864, "learning_rate": 2.91959513766313e-05, "loss": 0.8535, "step": 6937 }, { "epoch": 3.142210144927536, "grad_norm": 0.1475158668069623, "learning_rate": 2.9183774679641418e-05, "loss": 0.8542, "step": 6938 }, { "epoch": 3.1426630434782608, "grad_norm": 0.1877261764907857, "learning_rate": 2.9171599063966737e-05, "loss": 0.8619, "step": 6939 }, { "epoch": 3.1431159420289854, "grad_norm": 0.16709140097174607, "learning_rate": 2.9159424530824448e-05, "loss": 0.8487, "step": 6940 }, { "epoch": 3.14356884057971, "grad_norm": 0.1585592599837944, "learning_rate": 2.9147251081431675e-05, "loss": 0.87, "step": 6941 }, { "epoch": 3.1440217391304346, "grad_norm": 0.17643849409574394, "learning_rate": 2.913507871700541e-05, "loss": 0.8778, "step": 6942 }, { "epoch": 3.144474637681159, "grad_norm": 0.1482482703960249, "learning_rate": 2.9122907438762526e-05, "loss": 0.8461, "step": 6943 }, { "epoch": 3.1449275362318843, "grad_norm": 0.1494970020904982, "learning_rate": 2.9110737247919835e-05, "loss": 0.873, "step": 6944 }, { "epoch": 3.145380434782609, "grad_norm": 0.14662751159578202, "learning_rate": 2.9098568145693992e-05, "loss": 0.8551, "step": 6945 }, { "epoch": 3.1458333333333335, "grad_norm": 0.25836745799539984, "learning_rate": 2.9086400133301573e-05, "loss": 0.8592, "step": 6946 }, { "epoch": 3.146286231884058, "grad_norm": 0.36121866047397955, "learning_rate": 2.9074233211959008e-05, "loss": 0.8582, "step": 6947 }, { "epoch": 3.1467391304347827, "grad_norm": 0.4310512175837931, "learning_rate": 2.906206738288268e-05, "loss": 0.8499, "step": 6948 }, { "epoch": 3.1471920289855073, "grad_norm": 0.413613181059377, "learning_rate": 2.9049902647288805e-05, "loss": 0.8345, "step": 6949 }, { "epoch": 3.147644927536232, "grad_norm": 0.3631991949166147, "learning_rate": 2.9037739006393518e-05, "loss": 0.858, "step": 6950 }, { "epoch": 3.1480978260869565, "grad_norm": 0.37792766327752436, "learning_rate": 2.902557646141282e-05, "loss": 0.858, "step": 6951 }, { "epoch": 3.148550724637681, "grad_norm": 0.30179266001491856, "learning_rate": 2.9013415013562648e-05, "loss": 0.8674, "step": 6952 }, { "epoch": 3.1490036231884058, "grad_norm": 0.1797166912276205, "learning_rate": 2.9001254664058778e-05, "loss": 0.8531, "step": 6953 }, { "epoch": 3.1494565217391304, "grad_norm": 0.4273134518150471, "learning_rate": 2.8989095414116918e-05, "loss": 0.8876, "step": 6954 }, { "epoch": 3.149909420289855, "grad_norm": 0.5040557293214372, "learning_rate": 2.897693726495262e-05, "loss": 0.8651, "step": 6955 }, { "epoch": 3.1503623188405796, "grad_norm": 0.4337990927012841, "learning_rate": 2.8964780217781386e-05, "loss": 0.8683, "step": 6956 }, { "epoch": 3.150815217391304, "grad_norm": 0.36746550259308125, "learning_rate": 2.8952624273818556e-05, "loss": 0.8447, "step": 6957 }, { "epoch": 3.151268115942029, "grad_norm": 0.16805009527747372, "learning_rate": 2.894046943427939e-05, "loss": 0.8582, "step": 6958 }, { "epoch": 3.1517210144927534, "grad_norm": 0.25563705836803674, "learning_rate": 2.8928315700379008e-05, "loss": 0.8502, "step": 6959 }, { "epoch": 3.1521739130434785, "grad_norm": 0.3589303314581988, "learning_rate": 2.8916163073332462e-05, "loss": 0.8693, "step": 6960 }, { "epoch": 3.152626811594203, "grad_norm": 0.4942920977201812, "learning_rate": 2.8904011554354637e-05, "loss": 0.8544, "step": 6961 }, { "epoch": 3.1530797101449277, "grad_norm": 0.48325527063544416, "learning_rate": 2.8891861144660383e-05, "loss": 0.8587, "step": 6962 }, { "epoch": 3.1535326086956523, "grad_norm": 0.23299520223040776, "learning_rate": 2.8879711845464365e-05, "loss": 0.8827, "step": 6963 }, { "epoch": 3.153985507246377, "grad_norm": 0.2607182178918224, "learning_rate": 2.8867563657981173e-05, "loss": 0.8674, "step": 6964 }, { "epoch": 3.1544384057971016, "grad_norm": 0.25049211133827565, "learning_rate": 2.8855416583425278e-05, "loss": 0.8592, "step": 6965 }, { "epoch": 3.154891304347826, "grad_norm": 0.2051393276090291, "learning_rate": 2.8843270623011056e-05, "loss": 0.864, "step": 6966 }, { "epoch": 3.1553442028985508, "grad_norm": 0.24501843557293074, "learning_rate": 2.8831125777952746e-05, "loss": 0.8607, "step": 6967 }, { "epoch": 3.1557971014492754, "grad_norm": 0.1779804406881293, "learning_rate": 2.88189820494645e-05, "loss": 0.8421, "step": 6968 }, { "epoch": 3.15625, "grad_norm": 0.18700012626723325, "learning_rate": 2.8806839438760322e-05, "loss": 0.865, "step": 6969 }, { "epoch": 3.1567028985507246, "grad_norm": 0.26739117307724725, "learning_rate": 2.8794697947054147e-05, "loss": 0.8572, "step": 6970 }, { "epoch": 3.1571557971014492, "grad_norm": 0.2799454614762802, "learning_rate": 2.8782557575559776e-05, "loss": 0.8802, "step": 6971 }, { "epoch": 3.157608695652174, "grad_norm": 0.21595370080170018, "learning_rate": 2.8770418325490907e-05, "loss": 0.8484, "step": 6972 }, { "epoch": 3.1580615942028984, "grad_norm": 0.1552163244346679, "learning_rate": 2.875828019806109e-05, "loss": 0.8394, "step": 6973 }, { "epoch": 3.158514492753623, "grad_norm": 0.24706025455457983, "learning_rate": 2.8746143194483844e-05, "loss": 0.8501, "step": 6974 }, { "epoch": 3.1589673913043477, "grad_norm": 0.24285063021638112, "learning_rate": 2.873400731597248e-05, "loss": 0.8576, "step": 6975 }, { "epoch": 3.1594202898550723, "grad_norm": 0.2034936737894623, "learning_rate": 2.8721872563740264e-05, "loss": 0.8536, "step": 6976 }, { "epoch": 3.159873188405797, "grad_norm": 0.2508868665779995, "learning_rate": 2.870973893900031e-05, "loss": 0.8706, "step": 6977 }, { "epoch": 3.1603260869565215, "grad_norm": 0.19546146037405404, "learning_rate": 2.8697606442965655e-05, "loss": 0.8651, "step": 6978 }, { "epoch": 3.1607789855072466, "grad_norm": 0.21561736379473553, "learning_rate": 2.868547507684919e-05, "loss": 0.8509, "step": 6979 }, { "epoch": 3.161231884057971, "grad_norm": 0.18668900940394687, "learning_rate": 2.8673344841863718e-05, "loss": 0.8784, "step": 6980 }, { "epoch": 3.161684782608696, "grad_norm": 0.17642372265514936, "learning_rate": 2.866121573922189e-05, "loss": 0.8613, "step": 6981 }, { "epoch": 3.1621376811594204, "grad_norm": 0.19158446096944964, "learning_rate": 2.8649087770136312e-05, "loss": 0.8319, "step": 6982 }, { "epoch": 3.162590579710145, "grad_norm": 0.22861962702693922, "learning_rate": 2.8636960935819406e-05, "loss": 0.8742, "step": 6983 }, { "epoch": 3.1630434782608696, "grad_norm": 0.19117318999280145, "learning_rate": 2.8624835237483535e-05, "loss": 0.838, "step": 6984 }, { "epoch": 3.1634963768115942, "grad_norm": 0.18123076627054738, "learning_rate": 2.8612710676340892e-05, "loss": 0.8589, "step": 6985 }, { "epoch": 3.163949275362319, "grad_norm": 0.14395896647675854, "learning_rate": 2.8600587253603626e-05, "loss": 0.8569, "step": 6986 }, { "epoch": 3.1644021739130435, "grad_norm": 0.18093477189718196, "learning_rate": 2.8588464970483704e-05, "loss": 0.8516, "step": 6987 }, { "epoch": 3.164855072463768, "grad_norm": 0.15290525661659987, "learning_rate": 2.857634382819303e-05, "loss": 0.8381, "step": 6988 }, { "epoch": 3.1653079710144927, "grad_norm": 0.18239763790324304, "learning_rate": 2.8564223827943355e-05, "loss": 0.8666, "step": 6989 }, { "epoch": 3.1657608695652173, "grad_norm": 0.1775635080076323, "learning_rate": 2.8552104970946354e-05, "loss": 0.8667, "step": 6990 }, { "epoch": 3.166213768115942, "grad_norm": 0.18518120172178523, "learning_rate": 2.853998725841356e-05, "loss": 0.8545, "step": 6991 }, { "epoch": 3.1666666666666665, "grad_norm": 0.15215424004301145, "learning_rate": 2.8527870691556404e-05, "loss": 0.8483, "step": 6992 }, { "epoch": 3.167119565217391, "grad_norm": 0.2050980728037289, "learning_rate": 2.8515755271586175e-05, "loss": 0.8705, "step": 6993 }, { "epoch": 3.1675724637681157, "grad_norm": 0.16650698329752758, "learning_rate": 2.8503640999714108e-05, "loss": 0.8323, "step": 6994 }, { "epoch": 3.168025362318841, "grad_norm": 0.19117214151182124, "learning_rate": 2.8491527877151264e-05, "loss": 0.8647, "step": 6995 }, { "epoch": 3.1684782608695654, "grad_norm": 0.14911441458576802, "learning_rate": 2.8479415905108614e-05, "loss": 0.8531, "step": 6996 }, { "epoch": 3.16893115942029, "grad_norm": 0.18441442951598042, "learning_rate": 2.8467305084797005e-05, "loss": 0.8661, "step": 6997 }, { "epoch": 3.1693840579710146, "grad_norm": 0.1540494685793611, "learning_rate": 2.8455195417427192e-05, "loss": 0.8637, "step": 6998 }, { "epoch": 3.1698369565217392, "grad_norm": 0.1540319021992441, "learning_rate": 2.8443086904209783e-05, "loss": 0.8571, "step": 6999 }, { "epoch": 3.170289855072464, "grad_norm": 0.12973018260888594, "learning_rate": 2.8430979546355294e-05, "loss": 0.8695, "step": 7000 }, { "epoch": 3.1707427536231885, "grad_norm": 0.1787375792703416, "learning_rate": 2.8418873345074097e-05, "loss": 0.8464, "step": 7001 }, { "epoch": 3.171195652173913, "grad_norm": 0.19015977805034773, "learning_rate": 2.8406768301576506e-05, "loss": 0.8559, "step": 7002 }, { "epoch": 3.1716485507246377, "grad_norm": 0.13403997097233927, "learning_rate": 2.8394664417072636e-05, "loss": 0.8654, "step": 7003 }, { "epoch": 3.1721014492753623, "grad_norm": 0.2078370293860656, "learning_rate": 2.8382561692772576e-05, "loss": 0.8485, "step": 7004 }, { "epoch": 3.172554347826087, "grad_norm": 0.18792404171821378, "learning_rate": 2.837046012988622e-05, "loss": 0.8566, "step": 7005 }, { "epoch": 3.1730072463768115, "grad_norm": 0.13326118604708168, "learning_rate": 2.8358359729623397e-05, "loss": 0.8623, "step": 7006 }, { "epoch": 3.173460144927536, "grad_norm": 0.1830392982187362, "learning_rate": 2.834626049319379e-05, "loss": 0.8556, "step": 7007 }, { "epoch": 3.1739130434782608, "grad_norm": 0.2142318949482808, "learning_rate": 2.8334162421807002e-05, "loss": 0.8796, "step": 7008 }, { "epoch": 3.1743659420289854, "grad_norm": 0.2026293400187406, "learning_rate": 2.8322065516672485e-05, "loss": 0.8616, "step": 7009 }, { "epoch": 3.17481884057971, "grad_norm": 0.17036217658330152, "learning_rate": 2.830996977899958e-05, "loss": 0.864, "step": 7010 }, { "epoch": 3.1752717391304346, "grad_norm": 0.13874148401925723, "learning_rate": 2.8297875209997507e-05, "loss": 0.854, "step": 7011 }, { "epoch": 3.175724637681159, "grad_norm": 0.17243530974980786, "learning_rate": 2.828578181087541e-05, "loss": 0.8599, "step": 7012 }, { "epoch": 3.1761775362318843, "grad_norm": 0.17830921951187292, "learning_rate": 2.827368958284226e-05, "loss": 0.8347, "step": 7013 }, { "epoch": 3.176630434782609, "grad_norm": 0.16980340660028034, "learning_rate": 2.8261598527106948e-05, "loss": 0.8495, "step": 7014 }, { "epoch": 3.1770833333333335, "grad_norm": 0.14567296360262053, "learning_rate": 2.8249508644878224e-05, "loss": 0.8586, "step": 7015 }, { "epoch": 3.177536231884058, "grad_norm": 0.16529978396031536, "learning_rate": 2.8237419937364756e-05, "loss": 0.854, "step": 7016 }, { "epoch": 3.1779891304347827, "grad_norm": 0.20312717578621853, "learning_rate": 2.8225332405775043e-05, "loss": 0.8526, "step": 7017 }, { "epoch": 3.1784420289855073, "grad_norm": 0.24271447514257252, "learning_rate": 2.8213246051317517e-05, "loss": 0.8441, "step": 7018 }, { "epoch": 3.178894927536232, "grad_norm": 0.22987663902244265, "learning_rate": 2.820116087520045e-05, "loss": 0.8631, "step": 7019 }, { "epoch": 3.1793478260869565, "grad_norm": 0.14843042062261053, "learning_rate": 2.818907687863204e-05, "loss": 0.8584, "step": 7020 }, { "epoch": 3.179800724637681, "grad_norm": 0.19612378829625565, "learning_rate": 2.817699406282032e-05, "loss": 0.888, "step": 7021 }, { "epoch": 3.1802536231884058, "grad_norm": 0.23875386809982366, "learning_rate": 2.8164912428973248e-05, "loss": 0.8488, "step": 7022 }, { "epoch": 3.1807065217391304, "grad_norm": 0.2216136395613209, "learning_rate": 2.815283197829862e-05, "loss": 0.8527, "step": 7023 }, { "epoch": 3.181159420289855, "grad_norm": 0.13427115009625834, "learning_rate": 2.8140752712004168e-05, "loss": 0.8596, "step": 7024 }, { "epoch": 3.1816123188405796, "grad_norm": 0.19681827846288033, "learning_rate": 2.8128674631297453e-05, "loss": 0.8625, "step": 7025 }, { "epoch": 3.182065217391304, "grad_norm": 0.23322720423225626, "learning_rate": 2.8116597737385958e-05, "loss": 0.8514, "step": 7026 }, { "epoch": 3.182518115942029, "grad_norm": 0.17750229437312204, "learning_rate": 2.8104522031476996e-05, "loss": 0.8687, "step": 7027 }, { "epoch": 3.1829710144927534, "grad_norm": 0.17593739624482968, "learning_rate": 2.8092447514777834e-05, "loss": 0.8785, "step": 7028 }, { "epoch": 3.1834239130434785, "grad_norm": 0.21010488094484625, "learning_rate": 2.808037418849556e-05, "loss": 0.8611, "step": 7029 }, { "epoch": 3.183876811594203, "grad_norm": 0.2624946122725389, "learning_rate": 2.8068302053837166e-05, "loss": 0.8509, "step": 7030 }, { "epoch": 3.1843297101449277, "grad_norm": 0.19756043568005263, "learning_rate": 2.805623111200951e-05, "loss": 0.8718, "step": 7031 }, { "epoch": 3.1847826086956523, "grad_norm": 0.1332546076447432, "learning_rate": 2.8044161364219378e-05, "loss": 0.8472, "step": 7032 }, { "epoch": 3.185235507246377, "grad_norm": 0.181523333893197, "learning_rate": 2.8032092811673367e-05, "loss": 0.8519, "step": 7033 }, { "epoch": 3.1856884057971016, "grad_norm": 0.19187108208213438, "learning_rate": 2.8020025455578007e-05, "loss": 0.8589, "step": 7034 }, { "epoch": 3.186141304347826, "grad_norm": 0.191082202301021, "learning_rate": 2.8007959297139678e-05, "loss": 0.8519, "step": 7035 }, { "epoch": 3.1865942028985508, "grad_norm": 0.18085631451703918, "learning_rate": 2.7995894337564674e-05, "loss": 0.8552, "step": 7036 }, { "epoch": 3.1870471014492754, "grad_norm": 0.20146597678961545, "learning_rate": 2.7983830578059126e-05, "loss": 0.8687, "step": 7037 }, { "epoch": 3.1875, "grad_norm": 0.23248931525373523, "learning_rate": 2.7971768019829083e-05, "loss": 0.8696, "step": 7038 }, { "epoch": 3.1879528985507246, "grad_norm": 0.17093800417395572, "learning_rate": 2.795970666408044e-05, "loss": 0.8522, "step": 7039 }, { "epoch": 3.1884057971014492, "grad_norm": 0.17544957428514704, "learning_rate": 2.7947646512019012e-05, "loss": 0.8603, "step": 7040 }, { "epoch": 3.188858695652174, "grad_norm": 0.14884774430525088, "learning_rate": 2.7935587564850455e-05, "loss": 0.8535, "step": 7041 }, { "epoch": 3.1893115942028984, "grad_norm": 0.19753374913944508, "learning_rate": 2.7923529823780335e-05, "loss": 0.8523, "step": 7042 }, { "epoch": 3.189764492753623, "grad_norm": 0.18179077623996376, "learning_rate": 2.7911473290014064e-05, "loss": 0.8456, "step": 7043 }, { "epoch": 3.1902173913043477, "grad_norm": 0.1574460648909005, "learning_rate": 2.7899417964756973e-05, "loss": 0.8747, "step": 7044 }, { "epoch": 3.1906702898550723, "grad_norm": 0.16470503082795368, "learning_rate": 2.788736384921424e-05, "loss": 0.8834, "step": 7045 }, { "epoch": 3.191123188405797, "grad_norm": 0.1317554487393134, "learning_rate": 2.7875310944590944e-05, "loss": 0.8514, "step": 7046 }, { "epoch": 3.1915760869565215, "grad_norm": 0.14582354599673888, "learning_rate": 2.7863259252092007e-05, "loss": 0.8452, "step": 7047 }, { "epoch": 3.1920289855072466, "grad_norm": 0.15869966303857624, "learning_rate": 2.7851208772922295e-05, "loss": 0.8677, "step": 7048 }, { "epoch": 3.192481884057971, "grad_norm": 0.14256102543199098, "learning_rate": 2.7839159508286474e-05, "loss": 0.8493, "step": 7049 }, { "epoch": 3.192934782608696, "grad_norm": 0.16891585542678686, "learning_rate": 2.7827111459389163e-05, "loss": 0.8488, "step": 7050 }, { "epoch": 3.1933876811594204, "grad_norm": 0.14934511740333714, "learning_rate": 2.7815064627434806e-05, "loss": 0.8374, "step": 7051 }, { "epoch": 3.193840579710145, "grad_norm": 0.1508643924314046, "learning_rate": 2.780301901362775e-05, "loss": 0.8875, "step": 7052 }, { "epoch": 3.1942934782608696, "grad_norm": 0.17022063913267227, "learning_rate": 2.7790974619172195e-05, "loss": 0.8684, "step": 7053 }, { "epoch": 3.1947463768115942, "grad_norm": 0.17105617179725127, "learning_rate": 2.7778931445272275e-05, "loss": 0.8366, "step": 7054 }, { "epoch": 3.195199275362319, "grad_norm": 0.16018304072484776, "learning_rate": 2.7766889493131933e-05, "loss": 0.8697, "step": 7055 }, { "epoch": 3.1956521739130435, "grad_norm": 0.15741918976926536, "learning_rate": 2.775484876395504e-05, "loss": 0.8463, "step": 7056 }, { "epoch": 3.196105072463768, "grad_norm": 0.16324246926495428, "learning_rate": 2.7742809258945303e-05, "loss": 0.865, "step": 7057 }, { "epoch": 3.1965579710144927, "grad_norm": 0.17617315795542096, "learning_rate": 2.7730770979306368e-05, "loss": 0.8338, "step": 7058 }, { "epoch": 3.1970108695652173, "grad_norm": 0.16518074650257683, "learning_rate": 2.771873392624169e-05, "loss": 0.8761, "step": 7059 }, { "epoch": 3.197463768115942, "grad_norm": 0.22852698004094402, "learning_rate": 2.7706698100954653e-05, "loss": 0.8421, "step": 7060 }, { "epoch": 3.1979166666666665, "grad_norm": 0.14219749398157983, "learning_rate": 2.769466350464847e-05, "loss": 0.8714, "step": 7061 }, { "epoch": 3.198369565217391, "grad_norm": 0.22807349008901295, "learning_rate": 2.7682630138526293e-05, "loss": 0.8549, "step": 7062 }, { "epoch": 3.1988224637681157, "grad_norm": 0.1738975945760526, "learning_rate": 2.7670598003791086e-05, "loss": 0.8554, "step": 7063 }, { "epoch": 3.199275362318841, "grad_norm": 0.17442124306682474, "learning_rate": 2.765856710164575e-05, "loss": 0.8413, "step": 7064 }, { "epoch": 3.1997282608695654, "grad_norm": 0.19099556559318204, "learning_rate": 2.7646537433292997e-05, "loss": 0.8603, "step": 7065 }, { "epoch": 3.20018115942029, "grad_norm": 0.16502643876516573, "learning_rate": 2.763450899993549e-05, "loss": 0.8692, "step": 7066 }, { "epoch": 3.2006340579710146, "grad_norm": 0.163214714943858, "learning_rate": 2.7622481802775713e-05, "loss": 0.8513, "step": 7067 }, { "epoch": 3.2010869565217392, "grad_norm": 0.15670567547228012, "learning_rate": 2.7610455843016035e-05, "loss": 0.8391, "step": 7068 }, { "epoch": 3.201539855072464, "grad_norm": 0.1886347686260183, "learning_rate": 2.759843112185873e-05, "loss": 0.8617, "step": 7069 }, { "epoch": 3.2019927536231885, "grad_norm": 0.15707026088924828, "learning_rate": 2.7586407640505898e-05, "loss": 0.8713, "step": 7070 }, { "epoch": 3.202445652173913, "grad_norm": 0.18307819437621287, "learning_rate": 2.7574385400159585e-05, "loss": 0.8503, "step": 7071 }, { "epoch": 3.2028985507246377, "grad_norm": 0.15263149228430248, "learning_rate": 2.756236440202164e-05, "loss": 0.8866, "step": 7072 }, { "epoch": 3.2033514492753623, "grad_norm": 0.17226556130634335, "learning_rate": 2.755034464729385e-05, "loss": 0.8727, "step": 7073 }, { "epoch": 3.203804347826087, "grad_norm": 0.15712364020343647, "learning_rate": 2.7538326137177814e-05, "loss": 0.8374, "step": 7074 }, { "epoch": 3.2042572463768115, "grad_norm": 0.1585583161752483, "learning_rate": 2.7526308872875075e-05, "loss": 0.8618, "step": 7075 }, { "epoch": 3.204710144927536, "grad_norm": 0.20736670733045381, "learning_rate": 2.7514292855587e-05, "loss": 0.8624, "step": 7076 }, { "epoch": 3.2051630434782608, "grad_norm": 0.20061641029697883, "learning_rate": 2.7502278086514863e-05, "loss": 0.8703, "step": 7077 }, { "epoch": 3.2056159420289854, "grad_norm": 0.16592095301000265, "learning_rate": 2.749026456685977e-05, "loss": 0.837, "step": 7078 }, { "epoch": 3.20606884057971, "grad_norm": 0.189676842923485, "learning_rate": 2.7478252297822767e-05, "loss": 0.8825, "step": 7079 }, { "epoch": 3.2065217391304346, "grad_norm": 0.15617058546327875, "learning_rate": 2.7466241280604716e-05, "loss": 0.8655, "step": 7080 }, { "epoch": 3.206974637681159, "grad_norm": 0.25216668091163263, "learning_rate": 2.7454231516406396e-05, "loss": 0.8484, "step": 7081 }, { "epoch": 3.2074275362318843, "grad_norm": 0.21019265801481016, "learning_rate": 2.7442223006428417e-05, "loss": 0.8596, "step": 7082 }, { "epoch": 3.207880434782609, "grad_norm": 0.1663668624954215, "learning_rate": 2.7430215751871318e-05, "loss": 0.8689, "step": 7083 }, { "epoch": 3.2083333333333335, "grad_norm": 0.21579083428396637, "learning_rate": 2.7418209753935464e-05, "loss": 0.8627, "step": 7084 }, { "epoch": 3.208786231884058, "grad_norm": 0.17459439190734405, "learning_rate": 2.740620501382112e-05, "loss": 0.8555, "step": 7085 }, { "epoch": 3.2092391304347827, "grad_norm": 0.20033615877830718, "learning_rate": 2.7394201532728405e-05, "loss": 0.8624, "step": 7086 }, { "epoch": 3.2096920289855073, "grad_norm": 0.199899320192862, "learning_rate": 2.7382199311857352e-05, "loss": 0.8534, "step": 7087 }, { "epoch": 3.210144927536232, "grad_norm": 0.1789712270308615, "learning_rate": 2.7370198352407826e-05, "loss": 0.8496, "step": 7088 }, { "epoch": 3.2105978260869565, "grad_norm": 0.20743425138683666, "learning_rate": 2.7358198655579587e-05, "loss": 0.866, "step": 7089 }, { "epoch": 3.211050724637681, "grad_norm": 0.15685838688908613, "learning_rate": 2.7346200222572245e-05, "loss": 0.878, "step": 7090 }, { "epoch": 3.2115036231884058, "grad_norm": 0.20252051810296903, "learning_rate": 2.733420305458534e-05, "loss": 0.8712, "step": 7091 }, { "epoch": 3.2119565217391304, "grad_norm": 0.21892358917130406, "learning_rate": 2.7322207152818205e-05, "loss": 0.8593, "step": 7092 }, { "epoch": 3.212409420289855, "grad_norm": 0.1591804637230038, "learning_rate": 2.731021251847014e-05, "loss": 0.86, "step": 7093 }, { "epoch": 3.2128623188405796, "grad_norm": 0.1659900087061573, "learning_rate": 2.7298219152740208e-05, "loss": 0.8585, "step": 7094 }, { "epoch": 3.213315217391304, "grad_norm": 0.14208279420035275, "learning_rate": 2.728622705682745e-05, "loss": 0.8662, "step": 7095 }, { "epoch": 3.213768115942029, "grad_norm": 0.1827058685882327, "learning_rate": 2.7274236231930705e-05, "loss": 0.8446, "step": 7096 }, { "epoch": 3.2142210144927534, "grad_norm": 0.13040025487497167, "learning_rate": 2.7262246679248748e-05, "loss": 0.8663, "step": 7097 }, { "epoch": 3.2146739130434785, "grad_norm": 0.18685697966582623, "learning_rate": 2.7250258399980166e-05, "loss": 0.8423, "step": 7098 }, { "epoch": 3.215126811594203, "grad_norm": 0.17448514813474078, "learning_rate": 2.7238271395323467e-05, "loss": 0.869, "step": 7099 }, { "epoch": 3.2155797101449277, "grad_norm": 0.16425560765220304, "learning_rate": 2.722628566647698e-05, "loss": 0.8657, "step": 7100 }, { "epoch": 3.2160326086956523, "grad_norm": 0.18236699874473913, "learning_rate": 2.7214301214638976e-05, "loss": 0.8477, "step": 7101 }, { "epoch": 3.216485507246377, "grad_norm": 0.17795113733261414, "learning_rate": 2.720231804100753e-05, "loss": 0.8709, "step": 7102 }, { "epoch": 3.2169384057971016, "grad_norm": 0.24760802543190283, "learning_rate": 2.7190336146780637e-05, "loss": 0.8748, "step": 7103 }, { "epoch": 3.217391304347826, "grad_norm": 0.2066799680791498, "learning_rate": 2.717835553315612e-05, "loss": 0.8584, "step": 7104 }, { "epoch": 3.2178442028985508, "grad_norm": 0.14332997493877567, "learning_rate": 2.7166376201331738e-05, "loss": 0.87, "step": 7105 }, { "epoch": 3.2182971014492754, "grad_norm": 0.19450043247613857, "learning_rate": 2.715439815250505e-05, "loss": 0.8535, "step": 7106 }, { "epoch": 3.21875, "grad_norm": 0.21113260182321156, "learning_rate": 2.7142421387873548e-05, "loss": 0.852, "step": 7107 }, { "epoch": 3.2192028985507246, "grad_norm": 0.1865679988778855, "learning_rate": 2.7130445908634537e-05, "loss": 0.8558, "step": 7108 }, { "epoch": 3.2196557971014492, "grad_norm": 0.13223840840892676, "learning_rate": 2.7118471715985256e-05, "loss": 0.8508, "step": 7109 }, { "epoch": 3.220108695652174, "grad_norm": 0.23660450626293666, "learning_rate": 2.7106498811122762e-05, "loss": 0.8596, "step": 7110 }, { "epoch": 3.2205615942028984, "grad_norm": 0.28000177189775377, "learning_rate": 2.709452719524403e-05, "loss": 0.854, "step": 7111 }, { "epoch": 3.221014492753623, "grad_norm": 0.23929703387414475, "learning_rate": 2.7082556869545836e-05, "loss": 0.8593, "step": 7112 }, { "epoch": 3.2214673913043477, "grad_norm": 0.17274976007351753, "learning_rate": 2.7070587835224924e-05, "loss": 0.8645, "step": 7113 }, { "epoch": 3.2219202898550723, "grad_norm": 0.1470117427478846, "learning_rate": 2.705862009347782e-05, "loss": 0.8421, "step": 7114 }, { "epoch": 3.222373188405797, "grad_norm": 0.19603333323159403, "learning_rate": 2.704665364550099e-05, "loss": 0.8504, "step": 7115 }, { "epoch": 3.2228260869565215, "grad_norm": 0.20223593569471435, "learning_rate": 2.70346884924907e-05, "loss": 0.861, "step": 7116 }, { "epoch": 3.2232789855072466, "grad_norm": 0.1616230324415199, "learning_rate": 2.7022724635643167e-05, "loss": 0.867, "step": 7117 }, { "epoch": 3.223731884057971, "grad_norm": 0.14902063577209448, "learning_rate": 2.7010762076154403e-05, "loss": 0.8719, "step": 7118 }, { "epoch": 3.224184782608696, "grad_norm": 0.18741204684984428, "learning_rate": 2.699880081522035e-05, "loss": 0.8515, "step": 7119 }, { "epoch": 3.2246376811594204, "grad_norm": 0.20033232781553095, "learning_rate": 2.698684085403676e-05, "loss": 0.8483, "step": 7120 }, { "epoch": 3.225090579710145, "grad_norm": 0.17006959694925478, "learning_rate": 2.697488219379933e-05, "loss": 0.8471, "step": 7121 }, { "epoch": 3.2255434782608696, "grad_norm": 0.20547469636653273, "learning_rate": 2.696292483570356e-05, "loss": 0.846, "step": 7122 }, { "epoch": 3.2259963768115942, "grad_norm": 0.14632047855399719, "learning_rate": 2.695096878094486e-05, "loss": 0.8697, "step": 7123 }, { "epoch": 3.226449275362319, "grad_norm": 0.21515453211833402, "learning_rate": 2.6939014030718476e-05, "loss": 0.8485, "step": 7124 }, { "epoch": 3.2269021739130435, "grad_norm": 0.23244709637455466, "learning_rate": 2.6927060586219565e-05, "loss": 0.8488, "step": 7125 }, { "epoch": 3.227355072463768, "grad_norm": 0.21744678958886254, "learning_rate": 2.6915108448643124e-05, "loss": 0.8759, "step": 7126 }, { "epoch": 3.2278079710144927, "grad_norm": 0.19671454595353763, "learning_rate": 2.690315761918403e-05, "loss": 0.847, "step": 7127 }, { "epoch": 3.2282608695652173, "grad_norm": 0.14762467945988197, "learning_rate": 2.6891208099037006e-05, "loss": 0.8568, "step": 7128 }, { "epoch": 3.228713768115942, "grad_norm": 0.17576556249197264, "learning_rate": 2.6879259889396697e-05, "loss": 0.8544, "step": 7129 }, { "epoch": 3.2291666666666665, "grad_norm": 0.19398983969183126, "learning_rate": 2.6867312991457563e-05, "loss": 0.894, "step": 7130 }, { "epoch": 3.229619565217391, "grad_norm": 0.1842856504525774, "learning_rate": 2.6855367406413964e-05, "loss": 0.87, "step": 7131 }, { "epoch": 3.2300724637681157, "grad_norm": 0.14405974428766571, "learning_rate": 2.68434231354601e-05, "loss": 0.8578, "step": 7132 }, { "epoch": 3.230525362318841, "grad_norm": 0.1778470612496045, "learning_rate": 2.6831480179790096e-05, "loss": 0.8519, "step": 7133 }, { "epoch": 3.2309782608695654, "grad_norm": 0.12989443446821813, "learning_rate": 2.6819538540597867e-05, "loss": 0.8536, "step": 7134 }, { "epoch": 3.23143115942029, "grad_norm": 0.18042915113831526, "learning_rate": 2.680759821907727e-05, "loss": 0.8776, "step": 7135 }, { "epoch": 3.2318840579710146, "grad_norm": 0.20426847968231868, "learning_rate": 2.679565921642197e-05, "loss": 0.8638, "step": 7136 }, { "epoch": 3.2323369565217392, "grad_norm": 0.19608464825789276, "learning_rate": 2.6783721533825554e-05, "loss": 0.8536, "step": 7137 }, { "epoch": 3.232789855072464, "grad_norm": 0.22323705421714024, "learning_rate": 2.6771785172481426e-05, "loss": 0.8632, "step": 7138 }, { "epoch": 3.2332427536231885, "grad_norm": 0.22909942781966422, "learning_rate": 2.6759850133582914e-05, "loss": 0.868, "step": 7139 }, { "epoch": 3.233695652173913, "grad_norm": 0.20143099457219565, "learning_rate": 2.6747916418323154e-05, "loss": 0.8561, "step": 7140 }, { "epoch": 3.2341485507246377, "grad_norm": 0.17319135713856995, "learning_rate": 2.6735984027895193e-05, "loss": 0.8563, "step": 7141 }, { "epoch": 3.2346014492753623, "grad_norm": 0.18366662352382307, "learning_rate": 2.672405296349192e-05, "loss": 0.8597, "step": 7142 }, { "epoch": 3.235054347826087, "grad_norm": 0.17516404083345055, "learning_rate": 2.671212322630612e-05, "loss": 0.8461, "step": 7143 }, { "epoch": 3.2355072463768115, "grad_norm": 0.18401391037187045, "learning_rate": 2.6700194817530413e-05, "loss": 0.8445, "step": 7144 }, { "epoch": 3.235960144927536, "grad_norm": 0.1595665456065287, "learning_rate": 2.668826773835731e-05, "loss": 0.8611, "step": 7145 }, { "epoch": 3.2364130434782608, "grad_norm": 0.21310525712371223, "learning_rate": 2.6676341989979162e-05, "loss": 0.8788, "step": 7146 }, { "epoch": 3.2368659420289854, "grad_norm": 0.21307364310501498, "learning_rate": 2.6664417573588233e-05, "loss": 0.8608, "step": 7147 }, { "epoch": 3.23731884057971, "grad_norm": 0.16847108116866324, "learning_rate": 2.6652494490376602e-05, "loss": 0.8506, "step": 7148 }, { "epoch": 3.2377717391304346, "grad_norm": 0.26296237752477397, "learning_rate": 2.6640572741536258e-05, "loss": 0.8739, "step": 7149 }, { "epoch": 3.238224637681159, "grad_norm": 0.16892918870928525, "learning_rate": 2.662865232825901e-05, "loss": 0.8694, "step": 7150 }, { "epoch": 3.2386775362318843, "grad_norm": 0.18523174236608556, "learning_rate": 2.6616733251736592e-05, "loss": 0.8521, "step": 7151 }, { "epoch": 3.239130434782609, "grad_norm": 0.17586188347564072, "learning_rate": 2.6604815513160556e-05, "loss": 0.8733, "step": 7152 }, { "epoch": 3.2395833333333335, "grad_norm": 0.165232883747427, "learning_rate": 2.659289911372234e-05, "loss": 0.8493, "step": 7153 }, { "epoch": 3.240036231884058, "grad_norm": 0.24644150559762032, "learning_rate": 2.6580984054613237e-05, "loss": 0.8568, "step": 7154 }, { "epoch": 3.2404891304347827, "grad_norm": 0.296108914625879, "learning_rate": 2.6569070337024432e-05, "loss": 0.8843, "step": 7155 }, { "epoch": 3.2409420289855073, "grad_norm": 0.22353589539532637, "learning_rate": 2.6557157962146943e-05, "loss": 0.8498, "step": 7156 }, { "epoch": 3.241394927536232, "grad_norm": 0.16386708920042808, "learning_rate": 2.6545246931171688e-05, "loss": 0.8513, "step": 7157 }, { "epoch": 3.2418478260869565, "grad_norm": 0.2046395281517599, "learning_rate": 2.653333724528939e-05, "loss": 0.8433, "step": 7158 }, { "epoch": 3.242300724637681, "grad_norm": 0.28632664786489476, "learning_rate": 2.6521428905690734e-05, "loss": 0.8626, "step": 7159 }, { "epoch": 3.2427536231884058, "grad_norm": 0.3183197509304264, "learning_rate": 2.650952191356617e-05, "loss": 0.864, "step": 7160 }, { "epoch": 3.2432065217391304, "grad_norm": 0.24331094095000932, "learning_rate": 2.6497616270106097e-05, "loss": 0.8476, "step": 7161 }, { "epoch": 3.243659420289855, "grad_norm": 0.21496018712911272, "learning_rate": 2.6485711976500703e-05, "loss": 0.8715, "step": 7162 }, { "epoch": 3.2441123188405796, "grad_norm": 0.32450430336354913, "learning_rate": 2.6473809033940114e-05, "loss": 0.8581, "step": 7163 }, { "epoch": 3.244565217391304, "grad_norm": 0.3428964094307738, "learning_rate": 2.646190744361426e-05, "loss": 0.8755, "step": 7164 }, { "epoch": 3.245018115942029, "grad_norm": 0.2884439799712011, "learning_rate": 2.6450007206712975e-05, "loss": 0.8563, "step": 7165 }, { "epoch": 3.2454710144927534, "grad_norm": 0.25582801432999114, "learning_rate": 2.6438108324425937e-05, "loss": 0.8617, "step": 7166 }, { "epoch": 3.2459239130434785, "grad_norm": 0.1908641211197287, "learning_rate": 2.6426210797942707e-05, "loss": 0.8705, "step": 7167 }, { "epoch": 3.246376811594203, "grad_norm": 0.1906058881709472, "learning_rate": 2.6414314628452687e-05, "loss": 0.8639, "step": 7168 }, { "epoch": 3.2468297101449277, "grad_norm": 0.19992708732067352, "learning_rate": 2.640241981714517e-05, "loss": 0.8542, "step": 7169 }, { "epoch": 3.2472826086956523, "grad_norm": 0.22139362906520055, "learning_rate": 2.6390526365209275e-05, "loss": 0.8807, "step": 7170 }, { "epoch": 3.247735507246377, "grad_norm": 0.24046864507084273, "learning_rate": 2.6378634273834044e-05, "loss": 0.8748, "step": 7171 }, { "epoch": 3.2481884057971016, "grad_norm": 0.21191047716533265, "learning_rate": 2.636674354420832e-05, "loss": 0.8595, "step": 7172 }, { "epoch": 3.248641304347826, "grad_norm": 0.21402394039268274, "learning_rate": 2.6354854177520855e-05, "loss": 0.8589, "step": 7173 }, { "epoch": 3.2490942028985508, "grad_norm": 0.23080946785880213, "learning_rate": 2.6342966174960224e-05, "loss": 0.8542, "step": 7174 }, { "epoch": 3.2495471014492754, "grad_norm": 0.25018661682208737, "learning_rate": 2.633107953771492e-05, "loss": 0.8629, "step": 7175 }, { "epoch": 3.25, "grad_norm": 0.2494103684957555, "learning_rate": 2.6319194266973256e-05, "loss": 0.8623, "step": 7176 }, { "epoch": 3.2504528985507246, "grad_norm": 0.23467893554766822, "learning_rate": 2.630731036392342e-05, "loss": 0.8815, "step": 7177 }, { "epoch": 3.2509057971014492, "grad_norm": 0.1415067405417449, "learning_rate": 2.629542782975345e-05, "loss": 0.8437, "step": 7178 }, { "epoch": 3.251358695652174, "grad_norm": 0.2741767480313349, "learning_rate": 2.6283546665651293e-05, "loss": 0.8483, "step": 7179 }, { "epoch": 3.2518115942028984, "grad_norm": 0.2133888170189984, "learning_rate": 2.62716668728047e-05, "loss": 0.8631, "step": 7180 }, { "epoch": 3.252264492753623, "grad_norm": 0.2252786513201821, "learning_rate": 2.625978845240134e-05, "loss": 0.8483, "step": 7181 }, { "epoch": 3.2527173913043477, "grad_norm": 0.15334715004222557, "learning_rate": 2.6247911405628683e-05, "loss": 0.8672, "step": 7182 }, { "epoch": 3.2531702898550723, "grad_norm": 0.16932414813035784, "learning_rate": 2.623603573367413e-05, "loss": 0.8578, "step": 7183 }, { "epoch": 3.253623188405797, "grad_norm": 0.2394044043086193, "learning_rate": 2.6224161437724877e-05, "loss": 0.8498, "step": 7184 }, { "epoch": 3.2540760869565215, "grad_norm": 0.2030385827876385, "learning_rate": 2.621228851896805e-05, "loss": 0.8838, "step": 7185 }, { "epoch": 3.2545289855072466, "grad_norm": 0.13575639189549593, "learning_rate": 2.6200416978590582e-05, "loss": 0.8554, "step": 7186 }, { "epoch": 3.254981884057971, "grad_norm": 0.19806714241452703, "learning_rate": 2.618854681777931e-05, "loss": 0.8556, "step": 7187 }, { "epoch": 3.255434782608696, "grad_norm": 0.1702691987926436, "learning_rate": 2.6176678037720877e-05, "loss": 0.8525, "step": 7188 }, { "epoch": 3.2558876811594204, "grad_norm": 0.1392974707047959, "learning_rate": 2.616481063960186e-05, "loss": 0.8569, "step": 7189 }, { "epoch": 3.256340579710145, "grad_norm": 0.14598847238545426, "learning_rate": 2.615294462460865e-05, "loss": 0.8519, "step": 7190 }, { "epoch": 3.2567934782608696, "grad_norm": 0.15501921337249425, "learning_rate": 2.6141079993927504e-05, "loss": 0.8487, "step": 7191 }, { "epoch": 3.2572463768115942, "grad_norm": 0.16978287125706562, "learning_rate": 2.6129216748744553e-05, "loss": 0.8527, "step": 7192 }, { "epoch": 3.257699275362319, "grad_norm": 0.1488489349699366, "learning_rate": 2.611735489024577e-05, "loss": 0.8535, "step": 7193 }, { "epoch": 3.2581521739130435, "grad_norm": 0.1398762301843464, "learning_rate": 2.610549441961704e-05, "loss": 0.8582, "step": 7194 }, { "epoch": 3.258605072463768, "grad_norm": 0.15274874840943456, "learning_rate": 2.6093635338044035e-05, "loss": 0.8384, "step": 7195 }, { "epoch": 3.2590579710144927, "grad_norm": 0.14672905653483903, "learning_rate": 2.6081777646712354e-05, "loss": 0.8525, "step": 7196 }, { "epoch": 3.2595108695652173, "grad_norm": 0.1520474104324199, "learning_rate": 2.6069921346807398e-05, "loss": 0.8491, "step": 7197 }, { "epoch": 3.259963768115942, "grad_norm": 0.14786630355404465, "learning_rate": 2.605806643951449e-05, "loss": 0.8687, "step": 7198 }, { "epoch": 3.2604166666666665, "grad_norm": 0.17299815231643118, "learning_rate": 2.6046212926018774e-05, "loss": 0.8528, "step": 7199 }, { "epoch": 3.260869565217391, "grad_norm": 0.1701384639909524, "learning_rate": 2.6034360807505267e-05, "loss": 0.8729, "step": 7200 }, { "epoch": 3.261322463768116, "grad_norm": 0.16037891049912478, "learning_rate": 2.6022510085158827e-05, "loss": 0.8751, "step": 7201 }, { "epoch": 3.261775362318841, "grad_norm": 0.1814648988985154, "learning_rate": 2.6010660760164222e-05, "loss": 0.8582, "step": 7202 }, { "epoch": 3.2622282608695654, "grad_norm": 0.17128257015494414, "learning_rate": 2.5998812833706014e-05, "loss": 0.848, "step": 7203 }, { "epoch": 3.26268115942029, "grad_norm": 0.16163226276574202, "learning_rate": 2.5986966306968685e-05, "loss": 0.8786, "step": 7204 }, { "epoch": 3.2631340579710146, "grad_norm": 0.14062843557277271, "learning_rate": 2.5975121181136524e-05, "loss": 0.8593, "step": 7205 }, { "epoch": 3.2635869565217392, "grad_norm": 0.16688364392669253, "learning_rate": 2.5963277457393736e-05, "loss": 0.8621, "step": 7206 }, { "epoch": 3.264039855072464, "grad_norm": 0.1543327764866489, "learning_rate": 2.595143513692434e-05, "loss": 0.8731, "step": 7207 }, { "epoch": 3.2644927536231885, "grad_norm": 0.19589650004395065, "learning_rate": 2.593959422091224e-05, "loss": 0.8747, "step": 7208 }, { "epoch": 3.264945652173913, "grad_norm": 0.12022681859990009, "learning_rate": 2.5927754710541167e-05, "loss": 0.8577, "step": 7209 }, { "epoch": 3.2653985507246377, "grad_norm": 0.20799881633167003, "learning_rate": 2.5915916606994778e-05, "loss": 0.8716, "step": 7210 }, { "epoch": 3.2658514492753623, "grad_norm": 0.16999392771077199, "learning_rate": 2.5904079911456516e-05, "loss": 0.8546, "step": 7211 }, { "epoch": 3.266304347826087, "grad_norm": 0.19567778459077304, "learning_rate": 2.5892244625109725e-05, "loss": 0.8623, "step": 7212 }, { "epoch": 3.2667572463768115, "grad_norm": 0.19201619483758092, "learning_rate": 2.588041074913758e-05, "loss": 0.8636, "step": 7213 }, { "epoch": 3.267210144927536, "grad_norm": 0.16932840205699662, "learning_rate": 2.5868578284723164e-05, "loss": 0.8569, "step": 7214 }, { "epoch": 3.2676630434782608, "grad_norm": 0.21134308134520785, "learning_rate": 2.5856747233049363e-05, "loss": 0.8702, "step": 7215 }, { "epoch": 3.2681159420289854, "grad_norm": 0.2756375900402194, "learning_rate": 2.584491759529896e-05, "loss": 0.8556, "step": 7216 }, { "epoch": 3.26856884057971, "grad_norm": 0.18568332953981107, "learning_rate": 2.583308937265456e-05, "loss": 0.8567, "step": 7217 }, { "epoch": 3.2690217391304346, "grad_norm": 0.14732133004312759, "learning_rate": 2.582126256629868e-05, "loss": 0.8739, "step": 7218 }, { "epoch": 3.269474637681159, "grad_norm": 0.21410458072379107, "learning_rate": 2.5809437177413638e-05, "loss": 0.8535, "step": 7219 }, { "epoch": 3.269927536231884, "grad_norm": 0.15974660807682153, "learning_rate": 2.5797613207181664e-05, "loss": 0.8395, "step": 7220 }, { "epoch": 3.270380434782609, "grad_norm": 0.14605371873612627, "learning_rate": 2.5785790656784782e-05, "loss": 0.8661, "step": 7221 }, { "epoch": 3.2708333333333335, "grad_norm": 0.1681070372089587, "learning_rate": 2.577396952740495e-05, "loss": 0.8474, "step": 7222 }, { "epoch": 3.271286231884058, "grad_norm": 0.18309318123270438, "learning_rate": 2.576214982022392e-05, "loss": 0.8688, "step": 7223 }, { "epoch": 3.2717391304347827, "grad_norm": 0.13865063828361432, "learning_rate": 2.575033153642334e-05, "loss": 0.8673, "step": 7224 }, { "epoch": 3.2721920289855073, "grad_norm": 0.12009150484619435, "learning_rate": 2.573851467718468e-05, "loss": 0.863, "step": 7225 }, { "epoch": 3.272644927536232, "grad_norm": 0.16421699292253375, "learning_rate": 2.5726699243689332e-05, "loss": 0.8653, "step": 7226 }, { "epoch": 3.2730978260869565, "grad_norm": 0.17898477087976258, "learning_rate": 2.5714885237118453e-05, "loss": 0.8467, "step": 7227 }, { "epoch": 3.273550724637681, "grad_norm": 0.16845408164680548, "learning_rate": 2.570307265865315e-05, "loss": 0.8784, "step": 7228 }, { "epoch": 3.2740036231884058, "grad_norm": 0.139315776088467, "learning_rate": 2.5691261509474325e-05, "loss": 0.879, "step": 7229 }, { "epoch": 3.2744565217391304, "grad_norm": 0.2480000611890058, "learning_rate": 2.5679451790762766e-05, "loss": 0.8629, "step": 7230 }, { "epoch": 3.274909420289855, "grad_norm": 0.2085659578069851, "learning_rate": 2.5667643503699088e-05, "loss": 0.8533, "step": 7231 }, { "epoch": 3.2753623188405796, "grad_norm": 0.16136294149954725, "learning_rate": 2.5655836649463818e-05, "loss": 0.8553, "step": 7232 }, { "epoch": 3.275815217391304, "grad_norm": 0.2141171108343659, "learning_rate": 2.564403122923728e-05, "loss": 0.8597, "step": 7233 }, { "epoch": 3.276268115942029, "grad_norm": 0.23954385191123834, "learning_rate": 2.5632227244199703e-05, "loss": 0.8573, "step": 7234 }, { "epoch": 3.2767210144927534, "grad_norm": 0.20003972893614178, "learning_rate": 2.562042469553111e-05, "loss": 0.8565, "step": 7235 }, { "epoch": 3.2771739130434785, "grad_norm": 0.14322331966896393, "learning_rate": 2.5608623584411466e-05, "loss": 0.847, "step": 7236 }, { "epoch": 3.277626811594203, "grad_norm": 0.20534711607422712, "learning_rate": 2.559682391202052e-05, "loss": 0.8672, "step": 7237 }, { "epoch": 3.2780797101449277, "grad_norm": 0.2507304145565963, "learning_rate": 2.558502567953792e-05, "loss": 0.8525, "step": 7238 }, { "epoch": 3.2785326086956523, "grad_norm": 0.20388262564500187, "learning_rate": 2.557322888814313e-05, "loss": 0.8724, "step": 7239 }, { "epoch": 3.278985507246377, "grad_norm": 0.1311505131458774, "learning_rate": 2.5561433539015524e-05, "loss": 0.8618, "step": 7240 }, { "epoch": 3.2794384057971016, "grad_norm": 0.19022015840248904, "learning_rate": 2.5549639633334277e-05, "loss": 0.8703, "step": 7241 }, { "epoch": 3.279891304347826, "grad_norm": 0.19458155398526816, "learning_rate": 2.5537847172278465e-05, "loss": 0.8435, "step": 7242 }, { "epoch": 3.2803442028985508, "grad_norm": 0.18557683933150146, "learning_rate": 2.552605615702697e-05, "loss": 0.8506, "step": 7243 }, { "epoch": 3.2807971014492754, "grad_norm": 0.17660941004712713, "learning_rate": 2.55142665887586e-05, "loss": 0.8592, "step": 7244 }, { "epoch": 3.28125, "grad_norm": 0.18730316543715755, "learning_rate": 2.550247846865194e-05, "loss": 0.8657, "step": 7245 }, { "epoch": 3.2817028985507246, "grad_norm": 0.3254071816954852, "learning_rate": 2.549069179788549e-05, "loss": 0.8674, "step": 7246 }, { "epoch": 3.2821557971014492, "grad_norm": 0.30751195277976867, "learning_rate": 2.5478906577637555e-05, "loss": 0.8551, "step": 7247 }, { "epoch": 3.282608695652174, "grad_norm": 0.19713757920605568, "learning_rate": 2.5467122809086355e-05, "loss": 0.8745, "step": 7248 }, { "epoch": 3.2830615942028984, "grad_norm": 0.17128323586923838, "learning_rate": 2.5455340493409912e-05, "loss": 0.8544, "step": 7249 }, { "epoch": 3.283514492753623, "grad_norm": 0.1781119894235027, "learning_rate": 2.544355963178614e-05, "loss": 0.8748, "step": 7250 }, { "epoch": 3.2839673913043477, "grad_norm": 0.20871593595388865, "learning_rate": 2.5431780225392758e-05, "loss": 0.8741, "step": 7251 }, { "epoch": 3.2844202898550723, "grad_norm": 0.16029482217941424, "learning_rate": 2.5420002275407402e-05, "loss": 0.8447, "step": 7252 }, { "epoch": 3.284873188405797, "grad_norm": 0.16438326703802467, "learning_rate": 2.5408225783007523e-05, "loss": 0.8793, "step": 7253 }, { "epoch": 3.2853260869565215, "grad_norm": 0.17853036312646262, "learning_rate": 2.539645074937044e-05, "loss": 0.884, "step": 7254 }, { "epoch": 3.2857789855072466, "grad_norm": 0.20615253827655247, "learning_rate": 2.53846771756733e-05, "loss": 0.8707, "step": 7255 }, { "epoch": 3.286231884057971, "grad_norm": 0.17036192706939068, "learning_rate": 2.537290506309316e-05, "loss": 0.8619, "step": 7256 }, { "epoch": 3.286684782608696, "grad_norm": 0.1632960958383494, "learning_rate": 2.536113441280687e-05, "loss": 0.8518, "step": 7257 }, { "epoch": 3.2871376811594204, "grad_norm": 0.21089372781735005, "learning_rate": 2.534936522599118e-05, "loss": 0.847, "step": 7258 }, { "epoch": 3.287590579710145, "grad_norm": 0.21227128848769783, "learning_rate": 2.533759750382265e-05, "loss": 0.8699, "step": 7259 }, { "epoch": 3.2880434782608696, "grad_norm": 0.13681041161594804, "learning_rate": 2.5325831247477747e-05, "loss": 0.8541, "step": 7260 }, { "epoch": 3.2884963768115942, "grad_norm": 0.14947509430616054, "learning_rate": 2.5314066458132737e-05, "loss": 0.8584, "step": 7261 }, { "epoch": 3.288949275362319, "grad_norm": 0.1808691928872809, "learning_rate": 2.5302303136963782e-05, "loss": 0.8552, "step": 7262 }, { "epoch": 3.2894021739130435, "grad_norm": 0.15259574041279833, "learning_rate": 2.5290541285146862e-05, "loss": 0.8436, "step": 7263 }, { "epoch": 3.289855072463768, "grad_norm": 0.13838899038836444, "learning_rate": 2.527878090385785e-05, "loss": 0.859, "step": 7264 }, { "epoch": 3.2903079710144927, "grad_norm": 0.15673465482655616, "learning_rate": 2.5267021994272435e-05, "loss": 0.8655, "step": 7265 }, { "epoch": 3.2907608695652173, "grad_norm": 0.23299707955567917, "learning_rate": 2.5255264557566187e-05, "loss": 0.8522, "step": 7266 }, { "epoch": 3.291213768115942, "grad_norm": 0.22975817972268273, "learning_rate": 2.524350859491448e-05, "loss": 0.8659, "step": 7267 }, { "epoch": 3.2916666666666665, "grad_norm": 0.1371169351041186, "learning_rate": 2.523175410749263e-05, "loss": 0.8547, "step": 7268 }, { "epoch": 3.292119565217391, "grad_norm": 0.17258983373704567, "learning_rate": 2.5220001096475703e-05, "loss": 0.8585, "step": 7269 }, { "epoch": 3.292572463768116, "grad_norm": 0.19319955064108876, "learning_rate": 2.5208249563038712e-05, "loss": 0.8881, "step": 7270 }, { "epoch": 3.293025362318841, "grad_norm": 0.1981860894962162, "learning_rate": 2.5196499508356428e-05, "loss": 0.8685, "step": 7271 }, { "epoch": 3.2934782608695654, "grad_norm": 0.1996556299288955, "learning_rate": 2.518475093360356e-05, "loss": 0.8589, "step": 7272 }, { "epoch": 3.29393115942029, "grad_norm": 0.13983404007199507, "learning_rate": 2.51730038399546e-05, "loss": 0.8507, "step": 7273 }, { "epoch": 3.2943840579710146, "grad_norm": 0.2052813747700037, "learning_rate": 2.5161258228583964e-05, "loss": 0.862, "step": 7274 }, { "epoch": 3.2948369565217392, "grad_norm": 0.18254258722978436, "learning_rate": 2.514951410066585e-05, "loss": 0.8698, "step": 7275 }, { "epoch": 3.295289855072464, "grad_norm": 0.17468431642697416, "learning_rate": 2.5137771457374345e-05, "loss": 0.8495, "step": 7276 }, { "epoch": 3.2957427536231885, "grad_norm": 0.14079860137149294, "learning_rate": 2.5126030299883367e-05, "loss": 0.856, "step": 7277 }, { "epoch": 3.296195652173913, "grad_norm": 0.17436754017736827, "learning_rate": 2.5114290629366727e-05, "loss": 0.8413, "step": 7278 }, { "epoch": 3.2966485507246377, "grad_norm": 0.20952883882867115, "learning_rate": 2.5102552446998027e-05, "loss": 0.854, "step": 7279 }, { "epoch": 3.2971014492753623, "grad_norm": 0.19896527910515957, "learning_rate": 2.5090815753950777e-05, "loss": 0.8849, "step": 7280 }, { "epoch": 3.297554347826087, "grad_norm": 0.1780922073454358, "learning_rate": 2.5079080551398286e-05, "loss": 0.8532, "step": 7281 }, { "epoch": 3.2980072463768115, "grad_norm": 0.26027792141267164, "learning_rate": 2.5067346840513774e-05, "loss": 0.8492, "step": 7282 }, { "epoch": 3.298460144927536, "grad_norm": 0.2744806023182925, "learning_rate": 2.5055614622470252e-05, "loss": 0.8595, "step": 7283 }, { "epoch": 3.2989130434782608, "grad_norm": 0.15182085830673028, "learning_rate": 2.5043883898440626e-05, "loss": 0.8781, "step": 7284 }, { "epoch": 3.2993659420289854, "grad_norm": 0.19492154682131904, "learning_rate": 2.503215466959761e-05, "loss": 0.8824, "step": 7285 }, { "epoch": 3.29981884057971, "grad_norm": 0.21801867004012784, "learning_rate": 2.5020426937113827e-05, "loss": 0.8591, "step": 7286 }, { "epoch": 3.3002717391304346, "grad_norm": 0.22907182241107804, "learning_rate": 2.500870070216169e-05, "loss": 0.8546, "step": 7287 }, { "epoch": 3.300724637681159, "grad_norm": 0.18446126070242155, "learning_rate": 2.499697596591351e-05, "loss": 0.8673, "step": 7288 }, { "epoch": 3.301177536231884, "grad_norm": 0.16204024917313004, "learning_rate": 2.49852527295414e-05, "loss": 0.8669, "step": 7289 }, { "epoch": 3.301630434782609, "grad_norm": 0.2106181092956064, "learning_rate": 2.497353099421738e-05, "loss": 0.86, "step": 7290 }, { "epoch": 3.3020833333333335, "grad_norm": 0.22063568327261301, "learning_rate": 2.4961810761113282e-05, "loss": 0.8507, "step": 7291 }, { "epoch": 3.302536231884058, "grad_norm": 0.2163295413198584, "learning_rate": 2.4950092031400787e-05, "loss": 0.8677, "step": 7292 }, { "epoch": 3.3029891304347827, "grad_norm": 0.1419659312846523, "learning_rate": 2.4938374806251432e-05, "loss": 0.8659, "step": 7293 }, { "epoch": 3.3034420289855073, "grad_norm": 0.1953728912012519, "learning_rate": 2.4926659086836623e-05, "loss": 0.8589, "step": 7294 }, { "epoch": 3.303894927536232, "grad_norm": 0.1570801964194663, "learning_rate": 2.4914944874327593e-05, "loss": 0.8497, "step": 7295 }, { "epoch": 3.3043478260869565, "grad_norm": 0.19778789781162967, "learning_rate": 2.490323216989543e-05, "loss": 0.8721, "step": 7296 }, { "epoch": 3.304800724637681, "grad_norm": 0.14059613305383648, "learning_rate": 2.4891520974711053e-05, "loss": 0.8501, "step": 7297 }, { "epoch": 3.3052536231884058, "grad_norm": 0.17263924676485137, "learning_rate": 2.4879811289945285e-05, "loss": 0.8625, "step": 7298 }, { "epoch": 3.3057065217391304, "grad_norm": 0.200519319647823, "learning_rate": 2.486810311676873e-05, "loss": 0.8762, "step": 7299 }, { "epoch": 3.306159420289855, "grad_norm": 0.21399302701267656, "learning_rate": 2.48563964563519e-05, "loss": 0.8417, "step": 7300 }, { "epoch": 3.3066123188405796, "grad_norm": 0.1471250583134129, "learning_rate": 2.484469130986509e-05, "loss": 0.8627, "step": 7301 }, { "epoch": 3.307065217391304, "grad_norm": 0.18565203473459302, "learning_rate": 2.4832987678478526e-05, "loss": 0.856, "step": 7302 }, { "epoch": 3.307518115942029, "grad_norm": 0.22746600256378377, "learning_rate": 2.482128556336221e-05, "loss": 0.8566, "step": 7303 }, { "epoch": 3.3079710144927534, "grad_norm": 0.18181747348141425, "learning_rate": 2.4809584965686034e-05, "loss": 0.8614, "step": 7304 }, { "epoch": 3.3084239130434785, "grad_norm": 0.1544827767450102, "learning_rate": 2.479788588661971e-05, "loss": 0.8674, "step": 7305 }, { "epoch": 3.308876811594203, "grad_norm": 0.16588552463497677, "learning_rate": 2.4786188327332834e-05, "loss": 0.8634, "step": 7306 }, { "epoch": 3.3093297101449277, "grad_norm": 0.1678313042678839, "learning_rate": 2.4774492288994814e-05, "loss": 0.8441, "step": 7307 }, { "epoch": 3.3097826086956523, "grad_norm": 0.12436608393668343, "learning_rate": 2.4762797772774932e-05, "loss": 0.857, "step": 7308 }, { "epoch": 3.310235507246377, "grad_norm": 0.16884218193538902, "learning_rate": 2.4751104779842288e-05, "loss": 0.8512, "step": 7309 }, { "epoch": 3.3106884057971016, "grad_norm": 0.25720659927818923, "learning_rate": 2.4739413311365877e-05, "loss": 0.8663, "step": 7310 }, { "epoch": 3.311141304347826, "grad_norm": 0.22193553785442902, "learning_rate": 2.4727723368514487e-05, "loss": 0.8706, "step": 7311 }, { "epoch": 3.3115942028985508, "grad_norm": 0.1387931180540384, "learning_rate": 2.4716034952456814e-05, "loss": 0.8415, "step": 7312 }, { "epoch": 3.3120471014492754, "grad_norm": 0.22323688056314842, "learning_rate": 2.4704348064361322e-05, "loss": 0.8546, "step": 7313 }, { "epoch": 3.3125, "grad_norm": 0.20155308089362997, "learning_rate": 2.4692662705396412e-05, "loss": 0.8652, "step": 7314 }, { "epoch": 3.3129528985507246, "grad_norm": 0.2335536503378588, "learning_rate": 2.468097887673027e-05, "loss": 0.8583, "step": 7315 }, { "epoch": 3.3134057971014492, "grad_norm": 0.19075567306645422, "learning_rate": 2.466929657953092e-05, "loss": 0.8476, "step": 7316 }, { "epoch": 3.313858695652174, "grad_norm": 0.20936053593125195, "learning_rate": 2.4657615814966316e-05, "loss": 0.862, "step": 7317 }, { "epoch": 3.3143115942028984, "grad_norm": 0.27695444054189916, "learning_rate": 2.4645936584204142e-05, "loss": 0.8617, "step": 7318 }, { "epoch": 3.314764492753623, "grad_norm": 0.23176332624966758, "learning_rate": 2.4634258888412036e-05, "loss": 0.8604, "step": 7319 }, { "epoch": 3.3152173913043477, "grad_norm": 0.1437560701463925, "learning_rate": 2.46225827287574e-05, "loss": 0.8786, "step": 7320 }, { "epoch": 3.3156702898550723, "grad_norm": 0.2913412808424616, "learning_rate": 2.461090810640756e-05, "loss": 0.8582, "step": 7321 }, { "epoch": 3.316123188405797, "grad_norm": 0.22865473417728166, "learning_rate": 2.459923502252961e-05, "loss": 0.8262, "step": 7322 }, { "epoch": 3.3165760869565215, "grad_norm": 0.23931340619325095, "learning_rate": 2.4587563478290547e-05, "loss": 0.8757, "step": 7323 }, { "epoch": 3.3170289855072466, "grad_norm": 0.21971879242969145, "learning_rate": 2.4575893474857167e-05, "loss": 0.8649, "step": 7324 }, { "epoch": 3.317481884057971, "grad_norm": 0.12958774685525604, "learning_rate": 2.4564225013396178e-05, "loss": 0.8666, "step": 7325 }, { "epoch": 3.317934782608696, "grad_norm": 0.19628653604954543, "learning_rate": 2.4552558095074062e-05, "loss": 0.865, "step": 7326 }, { "epoch": 3.3183876811594204, "grad_norm": 0.15964643741936257, "learning_rate": 2.4540892721057205e-05, "loss": 0.8576, "step": 7327 }, { "epoch": 3.318840579710145, "grad_norm": 0.169330071206692, "learning_rate": 2.452922889251178e-05, "loss": 0.849, "step": 7328 }, { "epoch": 3.3192934782608696, "grad_norm": 0.18787340821462561, "learning_rate": 2.4517566610603877e-05, "loss": 0.8531, "step": 7329 }, { "epoch": 3.3197463768115942, "grad_norm": 0.17176883020212166, "learning_rate": 2.450590587649937e-05, "loss": 0.8403, "step": 7330 }, { "epoch": 3.320199275362319, "grad_norm": 0.1551977692912465, "learning_rate": 2.4494246691364e-05, "loss": 0.8606, "step": 7331 }, { "epoch": 3.3206521739130435, "grad_norm": 0.15532604142262593, "learning_rate": 2.4482589056363356e-05, "loss": 0.8542, "step": 7332 }, { "epoch": 3.321105072463768, "grad_norm": 0.17223903701122717, "learning_rate": 2.4470932972662886e-05, "loss": 0.87, "step": 7333 }, { "epoch": 3.3215579710144927, "grad_norm": 0.15867459091406694, "learning_rate": 2.4459278441427844e-05, "loss": 0.8584, "step": 7334 }, { "epoch": 3.3220108695652173, "grad_norm": 0.16473869613115408, "learning_rate": 2.444762546382337e-05, "loss": 0.8607, "step": 7335 }, { "epoch": 3.322463768115942, "grad_norm": 0.16143052022944998, "learning_rate": 2.443597404101441e-05, "loss": 0.875, "step": 7336 }, { "epoch": 3.3229166666666665, "grad_norm": 0.20571420947631125, "learning_rate": 2.4424324174165808e-05, "loss": 0.8465, "step": 7337 }, { "epoch": 3.323369565217391, "grad_norm": 0.19141596961312737, "learning_rate": 2.4412675864442187e-05, "loss": 0.8537, "step": 7338 }, { "epoch": 3.323822463768116, "grad_norm": 0.1533830900261954, "learning_rate": 2.4401029113008064e-05, "loss": 0.8793, "step": 7339 }, { "epoch": 3.324275362318841, "grad_norm": 0.14423902961326746, "learning_rate": 2.438938392102777e-05, "loss": 0.8535, "step": 7340 }, { "epoch": 3.3247282608695654, "grad_norm": 0.14743741871681104, "learning_rate": 2.4377740289665517e-05, "loss": 0.8614, "step": 7341 }, { "epoch": 3.32518115942029, "grad_norm": 0.16336635335546712, "learning_rate": 2.4366098220085314e-05, "loss": 0.8529, "step": 7342 }, { "epoch": 3.3256340579710146, "grad_norm": 0.18468400879102467, "learning_rate": 2.4354457713451055e-05, "loss": 0.8649, "step": 7343 }, { "epoch": 3.3260869565217392, "grad_norm": 0.16427180251150006, "learning_rate": 2.4342818770926437e-05, "loss": 0.8685, "step": 7344 }, { "epoch": 3.326539855072464, "grad_norm": 0.15243792346333462, "learning_rate": 2.433118139367505e-05, "loss": 0.8572, "step": 7345 }, { "epoch": 3.3269927536231885, "grad_norm": 0.18060727789193798, "learning_rate": 2.4319545582860274e-05, "loss": 0.8471, "step": 7346 }, { "epoch": 3.327445652173913, "grad_norm": 0.19835123600891033, "learning_rate": 2.4307911339645394e-05, "loss": 0.8776, "step": 7347 }, { "epoch": 3.3278985507246377, "grad_norm": 0.1491913337341863, "learning_rate": 2.4296278665193462e-05, "loss": 0.8555, "step": 7348 }, { "epoch": 3.3283514492753623, "grad_norm": 0.14221503415602313, "learning_rate": 2.4284647560667455e-05, "loss": 0.8599, "step": 7349 }, { "epoch": 3.328804347826087, "grad_norm": 0.1567077454085975, "learning_rate": 2.4273018027230126e-05, "loss": 0.8783, "step": 7350 }, { "epoch": 3.3292572463768115, "grad_norm": 0.1925764658791296, "learning_rate": 2.426139006604411e-05, "loss": 0.8688, "step": 7351 }, { "epoch": 3.329710144927536, "grad_norm": 0.1696487251974541, "learning_rate": 2.424976367827186e-05, "loss": 0.8449, "step": 7352 }, { "epoch": 3.3301630434782608, "grad_norm": 0.14548489642973694, "learning_rate": 2.4238138865075703e-05, "loss": 0.8431, "step": 7353 }, { "epoch": 3.3306159420289854, "grad_norm": 0.15907001426188663, "learning_rate": 2.422651562761777e-05, "loss": 0.8534, "step": 7354 }, { "epoch": 3.33106884057971, "grad_norm": 0.1860191236884534, "learning_rate": 2.421489396706008e-05, "loss": 0.843, "step": 7355 }, { "epoch": 3.3315217391304346, "grad_norm": 0.1757612143658465, "learning_rate": 2.4203273884564437e-05, "loss": 0.8602, "step": 7356 }, { "epoch": 3.331974637681159, "grad_norm": 0.15678661556305226, "learning_rate": 2.419165538129255e-05, "loss": 0.8613, "step": 7357 }, { "epoch": 3.332427536231884, "grad_norm": 0.22451629894603675, "learning_rate": 2.4180038458405913e-05, "loss": 0.8721, "step": 7358 }, { "epoch": 3.332880434782609, "grad_norm": 0.1971547406740688, "learning_rate": 2.4168423117065917e-05, "loss": 0.8476, "step": 7359 }, { "epoch": 3.3333333333333335, "grad_norm": 0.1732362610243712, "learning_rate": 2.4156809358433728e-05, "loss": 0.8824, "step": 7360 }, { "epoch": 3.333786231884058, "grad_norm": 0.2196718139696726, "learning_rate": 2.4145197183670433e-05, "loss": 0.861, "step": 7361 }, { "epoch": 3.3342391304347827, "grad_norm": 0.15443340463523636, "learning_rate": 2.413358659393688e-05, "loss": 0.8583, "step": 7362 }, { "epoch": 3.3346920289855073, "grad_norm": 0.1827284459695775, "learning_rate": 2.4121977590393834e-05, "loss": 0.8725, "step": 7363 }, { "epoch": 3.335144927536232, "grad_norm": 0.17275419958954158, "learning_rate": 2.411037017420184e-05, "loss": 0.838, "step": 7364 }, { "epoch": 3.3355978260869565, "grad_norm": 0.1717588251357881, "learning_rate": 2.409876434652133e-05, "loss": 0.8624, "step": 7365 }, { "epoch": 3.336050724637681, "grad_norm": 0.16862173347840875, "learning_rate": 2.4087160108512526e-05, "loss": 0.8385, "step": 7366 }, { "epoch": 3.3365036231884058, "grad_norm": 0.14488108272306408, "learning_rate": 2.4075557461335562e-05, "loss": 0.85, "step": 7367 }, { "epoch": 3.3369565217391304, "grad_norm": 0.13566513219172163, "learning_rate": 2.4063956406150345e-05, "loss": 0.8617, "step": 7368 }, { "epoch": 3.337409420289855, "grad_norm": 0.1883016414908851, "learning_rate": 2.4052356944116665e-05, "loss": 0.8848, "step": 7369 }, { "epoch": 3.3378623188405796, "grad_norm": 0.1592154337687623, "learning_rate": 2.4040759076394115e-05, "loss": 0.8713, "step": 7370 }, { "epoch": 3.338315217391304, "grad_norm": 0.14545986876128303, "learning_rate": 2.402916280414219e-05, "loss": 0.8562, "step": 7371 }, { "epoch": 3.338768115942029, "grad_norm": 0.1500418631224009, "learning_rate": 2.4017568128520158e-05, "loss": 0.8366, "step": 7372 }, { "epoch": 3.3392210144927534, "grad_norm": 0.19856851313475055, "learning_rate": 2.4005975050687172e-05, "loss": 0.8686, "step": 7373 }, { "epoch": 3.3396739130434785, "grad_norm": 0.1562341424502893, "learning_rate": 2.399438357180219e-05, "loss": 0.8674, "step": 7374 }, { "epoch": 3.340126811594203, "grad_norm": 0.1493184395497232, "learning_rate": 2.3982793693024063e-05, "loss": 0.8455, "step": 7375 }, { "epoch": 3.3405797101449277, "grad_norm": 0.15890332153847908, "learning_rate": 2.397120541551142e-05, "loss": 0.8636, "step": 7376 }, { "epoch": 3.3410326086956523, "grad_norm": 0.15663228736097226, "learning_rate": 2.3959618740422785e-05, "loss": 0.8613, "step": 7377 }, { "epoch": 3.341485507246377, "grad_norm": 0.14804712689280186, "learning_rate": 2.3948033668916466e-05, "loss": 0.8662, "step": 7378 }, { "epoch": 3.3419384057971016, "grad_norm": 0.14951221241300475, "learning_rate": 2.393645020215067e-05, "loss": 0.8463, "step": 7379 }, { "epoch": 3.342391304347826, "grad_norm": 0.14620434201567803, "learning_rate": 2.39248683412834e-05, "loss": 0.8619, "step": 7380 }, { "epoch": 3.3428442028985508, "grad_norm": 0.15965232794935488, "learning_rate": 2.3913288087472514e-05, "loss": 0.8743, "step": 7381 }, { "epoch": 3.3432971014492754, "grad_norm": 0.16214206194820338, "learning_rate": 2.3901709441875696e-05, "loss": 0.8458, "step": 7382 }, { "epoch": 3.34375, "grad_norm": 0.14055885650657854, "learning_rate": 2.389013240565052e-05, "loss": 0.8662, "step": 7383 }, { "epoch": 3.3442028985507246, "grad_norm": 0.17406668807466452, "learning_rate": 2.3878556979954316e-05, "loss": 0.8165, "step": 7384 }, { "epoch": 3.3446557971014492, "grad_norm": 0.1617896674910069, "learning_rate": 2.3866983165944336e-05, "loss": 0.8675, "step": 7385 }, { "epoch": 3.345108695652174, "grad_norm": 0.1447009893134207, "learning_rate": 2.3855410964777587e-05, "loss": 0.8714, "step": 7386 }, { "epoch": 3.3455615942028984, "grad_norm": 0.1339630755318677, "learning_rate": 2.3843840377611008e-05, "loss": 0.8572, "step": 7387 }, { "epoch": 3.346014492753623, "grad_norm": 0.16074509206361737, "learning_rate": 2.3832271405601303e-05, "loss": 0.8669, "step": 7388 }, { "epoch": 3.3464673913043477, "grad_norm": 0.1714971730908537, "learning_rate": 2.382070404990505e-05, "loss": 0.8795, "step": 7389 }, { "epoch": 3.3469202898550723, "grad_norm": 0.1447290029940424, "learning_rate": 2.3809138311678637e-05, "loss": 0.886, "step": 7390 }, { "epoch": 3.347373188405797, "grad_norm": 0.15217273367264628, "learning_rate": 2.3797574192078337e-05, "loss": 0.8631, "step": 7391 }, { "epoch": 3.3478260869565215, "grad_norm": 0.2412732216443443, "learning_rate": 2.3786011692260216e-05, "loss": 0.8597, "step": 7392 }, { "epoch": 3.3482789855072466, "grad_norm": 0.14272322570533008, "learning_rate": 2.3774450813380207e-05, "loss": 0.8595, "step": 7393 }, { "epoch": 3.348731884057971, "grad_norm": 0.2096712217121455, "learning_rate": 2.3762891556594045e-05, "loss": 0.8759, "step": 7394 }, { "epoch": 3.349184782608696, "grad_norm": 0.2135555302836928, "learning_rate": 2.3751333923057357e-05, "loss": 0.8589, "step": 7395 }, { "epoch": 3.3496376811594204, "grad_norm": 0.2004930716318937, "learning_rate": 2.373977791392556e-05, "loss": 0.8639, "step": 7396 }, { "epoch": 3.350090579710145, "grad_norm": 0.17941167098153687, "learning_rate": 2.372822353035394e-05, "loss": 0.8631, "step": 7397 }, { "epoch": 3.3505434782608696, "grad_norm": 0.13723413494062955, "learning_rate": 2.371667077349758e-05, "loss": 0.8645, "step": 7398 }, { "epoch": 3.3509963768115942, "grad_norm": 0.19932047260591015, "learning_rate": 2.3705119644511462e-05, "loss": 0.8674, "step": 7399 }, { "epoch": 3.351449275362319, "grad_norm": 0.14752521916010583, "learning_rate": 2.3693570144550348e-05, "loss": 0.8648, "step": 7400 }, { "epoch": 3.3519021739130435, "grad_norm": 0.17708197772323714, "learning_rate": 2.3682022274768874e-05, "loss": 0.8382, "step": 7401 }, { "epoch": 3.352355072463768, "grad_norm": 0.18659436702483054, "learning_rate": 2.3670476036321475e-05, "loss": 0.8552, "step": 7402 }, { "epoch": 3.3528079710144927, "grad_norm": 0.14905679827380525, "learning_rate": 2.3658931430362478e-05, "loss": 0.8574, "step": 7403 }, { "epoch": 3.3532608695652173, "grad_norm": 0.1994817881136365, "learning_rate": 2.364738845804599e-05, "loss": 0.8467, "step": 7404 }, { "epoch": 3.353713768115942, "grad_norm": 0.19813602700389554, "learning_rate": 2.3635847120525995e-05, "loss": 0.8526, "step": 7405 }, { "epoch": 3.3541666666666665, "grad_norm": 0.18300492138044208, "learning_rate": 2.3624307418956298e-05, "loss": 0.8641, "step": 7406 }, { "epoch": 3.354619565217391, "grad_norm": 0.19526196674872637, "learning_rate": 2.3612769354490546e-05, "loss": 0.8561, "step": 7407 }, { "epoch": 3.355072463768116, "grad_norm": 0.18560867575090068, "learning_rate": 2.3601232928282185e-05, "loss": 0.8647, "step": 7408 }, { "epoch": 3.355525362318841, "grad_norm": 0.1734206381053936, "learning_rate": 2.3589698141484575e-05, "loss": 0.8778, "step": 7409 }, { "epoch": 3.3559782608695654, "grad_norm": 0.18896468167696487, "learning_rate": 2.3578164995250837e-05, "loss": 0.8565, "step": 7410 }, { "epoch": 3.35643115942029, "grad_norm": 0.1688791918527589, "learning_rate": 2.3566633490733982e-05, "loss": 0.8743, "step": 7411 }, { "epoch": 3.3568840579710146, "grad_norm": 0.21574729627591527, "learning_rate": 2.355510362908679e-05, "loss": 0.8538, "step": 7412 }, { "epoch": 3.3573369565217392, "grad_norm": 0.14516033846778634, "learning_rate": 2.3543575411461966e-05, "loss": 0.8643, "step": 7413 }, { "epoch": 3.357789855072464, "grad_norm": 0.16566170549493142, "learning_rate": 2.3532048839011985e-05, "loss": 0.8474, "step": 7414 }, { "epoch": 3.3582427536231885, "grad_norm": 0.18388346924467336, "learning_rate": 2.3520523912889185e-05, "loss": 0.8813, "step": 7415 }, { "epoch": 3.358695652173913, "grad_norm": 0.13430564314952512, "learning_rate": 2.35090006342457e-05, "loss": 0.8578, "step": 7416 }, { "epoch": 3.3591485507246377, "grad_norm": 0.1711551100459856, "learning_rate": 2.349747900423357e-05, "loss": 0.8321, "step": 7417 }, { "epoch": 3.3596014492753623, "grad_norm": 0.1707698109776295, "learning_rate": 2.3485959024004612e-05, "loss": 0.8721, "step": 7418 }, { "epoch": 3.360054347826087, "grad_norm": 0.2028509768606936, "learning_rate": 2.3474440694710507e-05, "loss": 0.8357, "step": 7419 }, { "epoch": 3.3605072463768115, "grad_norm": 0.14974213140746634, "learning_rate": 2.3462924017502734e-05, "loss": 0.8616, "step": 7420 }, { "epoch": 3.360960144927536, "grad_norm": 0.14444239483071092, "learning_rate": 2.3451408993532665e-05, "loss": 0.8443, "step": 7421 }, { "epoch": 3.3614130434782608, "grad_norm": 0.17433695062638827, "learning_rate": 2.3439895623951464e-05, "loss": 0.8718, "step": 7422 }, { "epoch": 3.3618659420289854, "grad_norm": 0.17244339607676326, "learning_rate": 2.3428383909910135e-05, "loss": 0.8564, "step": 7423 }, { "epoch": 3.36231884057971, "grad_norm": 0.1451372413184026, "learning_rate": 2.341687385255952e-05, "loss": 0.8925, "step": 7424 }, { "epoch": 3.3627717391304346, "grad_norm": 0.2080687603566139, "learning_rate": 2.3405365453050316e-05, "loss": 0.8553, "step": 7425 }, { "epoch": 3.363224637681159, "grad_norm": 0.15921080574380023, "learning_rate": 2.3393858712533017e-05, "loss": 0.8748, "step": 7426 }, { "epoch": 3.363677536231884, "grad_norm": 0.16050592483639023, "learning_rate": 2.3382353632157985e-05, "loss": 0.8676, "step": 7427 }, { "epoch": 3.364130434782609, "grad_norm": 0.15407476163937994, "learning_rate": 2.3370850213075376e-05, "loss": 0.8663, "step": 7428 }, { "epoch": 3.3645833333333335, "grad_norm": 0.14660648729842599, "learning_rate": 2.3359348456435243e-05, "loss": 0.8877, "step": 7429 }, { "epoch": 3.365036231884058, "grad_norm": 0.17482673484503625, "learning_rate": 2.3347848363387402e-05, "loss": 0.8633, "step": 7430 }, { "epoch": 3.3654891304347827, "grad_norm": 0.1420389678562265, "learning_rate": 2.333634993508156e-05, "loss": 0.8713, "step": 7431 }, { "epoch": 3.3659420289855073, "grad_norm": 0.16733494286303982, "learning_rate": 2.3324853172667198e-05, "loss": 0.8457, "step": 7432 }, { "epoch": 3.366394927536232, "grad_norm": 0.12941602374100933, "learning_rate": 2.331335807729371e-05, "loss": 0.8682, "step": 7433 }, { "epoch": 3.3668478260869565, "grad_norm": 0.2021085568511762, "learning_rate": 2.330186465011024e-05, "loss": 0.8672, "step": 7434 }, { "epoch": 3.367300724637681, "grad_norm": 0.1962275581080014, "learning_rate": 2.3290372892265838e-05, "loss": 0.8626, "step": 7435 }, { "epoch": 3.3677536231884058, "grad_norm": 0.14506563001242162, "learning_rate": 2.3278882804909334e-05, "loss": 0.8588, "step": 7436 }, { "epoch": 3.3682065217391304, "grad_norm": 0.2333844713362742, "learning_rate": 2.3267394389189414e-05, "loss": 0.8517, "step": 7437 }, { "epoch": 3.368659420289855, "grad_norm": 0.15429234268846995, "learning_rate": 2.3255907646254573e-05, "loss": 0.846, "step": 7438 }, { "epoch": 3.3691123188405796, "grad_norm": 0.19156741723246962, "learning_rate": 2.3244422577253198e-05, "loss": 0.859, "step": 7439 }, { "epoch": 3.369565217391304, "grad_norm": 0.17865318833844543, "learning_rate": 2.323293918333345e-05, "loss": 0.846, "step": 7440 }, { "epoch": 3.370018115942029, "grad_norm": 0.164652359625728, "learning_rate": 2.3221457465643322e-05, "loss": 0.8633, "step": 7441 }, { "epoch": 3.3704710144927534, "grad_norm": 0.20342003000350442, "learning_rate": 2.32099774253307e-05, "loss": 0.8524, "step": 7442 }, { "epoch": 3.3709239130434785, "grad_norm": 0.15949972186198036, "learning_rate": 2.319849906354322e-05, "loss": 0.8494, "step": 7443 }, { "epoch": 3.371376811594203, "grad_norm": 0.15981534135281483, "learning_rate": 2.3187022381428433e-05, "loss": 0.8438, "step": 7444 }, { "epoch": 3.3718297101449277, "grad_norm": 0.19734424934056705, "learning_rate": 2.3175547380133656e-05, "loss": 0.8389, "step": 7445 }, { "epoch": 3.3722826086956523, "grad_norm": 0.19152752494547817, "learning_rate": 2.3164074060806076e-05, "loss": 0.8684, "step": 7446 }, { "epoch": 3.372735507246377, "grad_norm": 0.16256530118666687, "learning_rate": 2.3152602424592664e-05, "loss": 0.8696, "step": 7447 }, { "epoch": 3.3731884057971016, "grad_norm": 0.1959607125397238, "learning_rate": 2.3141132472640307e-05, "loss": 0.8781, "step": 7448 }, { "epoch": 3.373641304347826, "grad_norm": 0.20737853414369709, "learning_rate": 2.3129664206095637e-05, "loss": 0.8637, "step": 7449 }, { "epoch": 3.3740942028985508, "grad_norm": 0.18433842421559807, "learning_rate": 2.3118197626105183e-05, "loss": 0.8736, "step": 7450 }, { "epoch": 3.3745471014492754, "grad_norm": 0.1449550484878246, "learning_rate": 2.310673273381525e-05, "loss": 0.858, "step": 7451 }, { "epoch": 3.375, "grad_norm": 0.1739970178016295, "learning_rate": 2.3095269530372032e-05, "loss": 0.8327, "step": 7452 }, { "epoch": 3.3754528985507246, "grad_norm": 0.17640415988153424, "learning_rate": 2.3083808016921485e-05, "loss": 0.8756, "step": 7453 }, { "epoch": 3.3759057971014492, "grad_norm": 0.20888077692509605, "learning_rate": 2.3072348194609476e-05, "loss": 0.8641, "step": 7454 }, { "epoch": 3.376358695652174, "grad_norm": 0.12371698017123771, "learning_rate": 2.3060890064581645e-05, "loss": 0.8388, "step": 7455 }, { "epoch": 3.3768115942028984, "grad_norm": 0.1757163284070835, "learning_rate": 2.304943362798347e-05, "loss": 0.8749, "step": 7456 }, { "epoch": 3.377264492753623, "grad_norm": 0.17486873571017275, "learning_rate": 2.3037978885960266e-05, "loss": 0.8559, "step": 7457 }, { "epoch": 3.3777173913043477, "grad_norm": 0.15048576264818217, "learning_rate": 2.3026525839657207e-05, "loss": 0.8534, "step": 7458 }, { "epoch": 3.3781702898550723, "grad_norm": 0.1240679296171642, "learning_rate": 2.301507449021924e-05, "loss": 0.8482, "step": 7459 }, { "epoch": 3.378623188405797, "grad_norm": 0.1416792224258814, "learning_rate": 2.300362483879121e-05, "loss": 0.8487, "step": 7460 }, { "epoch": 3.3790760869565215, "grad_norm": 0.13144493865306722, "learning_rate": 2.299217688651771e-05, "loss": 0.8494, "step": 7461 }, { "epoch": 3.3795289855072466, "grad_norm": 0.15108458029300437, "learning_rate": 2.2980730634543264e-05, "loss": 0.8675, "step": 7462 }, { "epoch": 3.379981884057971, "grad_norm": 0.1303115054993064, "learning_rate": 2.2969286084012142e-05, "loss": 0.8531, "step": 7463 }, { "epoch": 3.380434782608696, "grad_norm": 0.12544412579439593, "learning_rate": 2.295784323606848e-05, "loss": 0.8508, "step": 7464 }, { "epoch": 3.3808876811594204, "grad_norm": 0.13963230857574524, "learning_rate": 2.294640209185622e-05, "loss": 0.8562, "step": 7465 }, { "epoch": 3.381340579710145, "grad_norm": 0.18153620695291006, "learning_rate": 2.2934962652519182e-05, "loss": 0.8647, "step": 7466 }, { "epoch": 3.3817934782608696, "grad_norm": 0.12554058880027896, "learning_rate": 2.292352491920095e-05, "loss": 0.8774, "step": 7467 }, { "epoch": 3.3822463768115942, "grad_norm": 0.16769629636123085, "learning_rate": 2.2912088893045015e-05, "loss": 0.8598, "step": 7468 }, { "epoch": 3.382699275362319, "grad_norm": 0.14487089792739521, "learning_rate": 2.290065457519461e-05, "loss": 0.8606, "step": 7469 }, { "epoch": 3.3831521739130435, "grad_norm": 0.13783854369387996, "learning_rate": 2.2889221966792877e-05, "loss": 0.868, "step": 7470 }, { "epoch": 3.383605072463768, "grad_norm": 0.14298744929588295, "learning_rate": 2.287779106898274e-05, "loss": 0.8599, "step": 7471 }, { "epoch": 3.3840579710144927, "grad_norm": 0.1356211019603341, "learning_rate": 2.2866361882906966e-05, "loss": 0.8888, "step": 7472 }, { "epoch": 3.3845108695652173, "grad_norm": 0.17642196526462103, "learning_rate": 2.2854934409708126e-05, "loss": 0.8579, "step": 7473 }, { "epoch": 3.384963768115942, "grad_norm": 0.16552084174434056, "learning_rate": 2.284350865052868e-05, "loss": 0.8644, "step": 7474 }, { "epoch": 3.3854166666666665, "grad_norm": 0.1350017526740839, "learning_rate": 2.2832084606510848e-05, "loss": 0.8503, "step": 7475 }, { "epoch": 3.385869565217391, "grad_norm": 0.1594278474554106, "learning_rate": 2.282066227879673e-05, "loss": 0.8694, "step": 7476 }, { "epoch": 3.386322463768116, "grad_norm": 0.15348375451895094, "learning_rate": 2.2809241668528218e-05, "loss": 0.8934, "step": 7477 }, { "epoch": 3.386775362318841, "grad_norm": 0.17053438935813114, "learning_rate": 2.279782277684707e-05, "loss": 0.8731, "step": 7478 }, { "epoch": 3.3872282608695654, "grad_norm": 0.18226520002329982, "learning_rate": 2.2786405604894837e-05, "loss": 0.8493, "step": 7479 }, { "epoch": 3.38768115942029, "grad_norm": 0.19913948917649757, "learning_rate": 2.2774990153812917e-05, "loss": 0.8522, "step": 7480 }, { "epoch": 3.3881340579710146, "grad_norm": 0.18753745044856004, "learning_rate": 2.2763576424742507e-05, "loss": 0.8672, "step": 7481 }, { "epoch": 3.3885869565217392, "grad_norm": 0.22059196739647052, "learning_rate": 2.2752164418824697e-05, "loss": 0.848, "step": 7482 }, { "epoch": 3.389039855072464, "grad_norm": 0.15463207705246723, "learning_rate": 2.274075413720032e-05, "loss": 0.8357, "step": 7483 }, { "epoch": 3.3894927536231885, "grad_norm": 0.17335226825945071, "learning_rate": 2.272934558101012e-05, "loss": 0.8539, "step": 7484 }, { "epoch": 3.389945652173913, "grad_norm": 0.1611504861622186, "learning_rate": 2.2717938751394595e-05, "loss": 0.8527, "step": 7485 }, { "epoch": 3.3903985507246377, "grad_norm": 0.17190306065858316, "learning_rate": 2.2706533649494134e-05, "loss": 0.8356, "step": 7486 }, { "epoch": 3.3908514492753623, "grad_norm": 0.17190819489325396, "learning_rate": 2.2695130276448905e-05, "loss": 0.8567, "step": 7487 }, { "epoch": 3.391304347826087, "grad_norm": 0.12368149290033771, "learning_rate": 2.2683728633398926e-05, "loss": 0.8555, "step": 7488 }, { "epoch": 3.3917572463768115, "grad_norm": 0.14383016220120087, "learning_rate": 2.267232872148402e-05, "loss": 0.8631, "step": 7489 }, { "epoch": 3.392210144927536, "grad_norm": 0.15260540907223466, "learning_rate": 2.2660930541843884e-05, "loss": 0.8546, "step": 7490 }, { "epoch": 3.3926630434782608, "grad_norm": 0.1645645587085306, "learning_rate": 2.2649534095617982e-05, "loss": 0.8781, "step": 7491 }, { "epoch": 3.3931159420289854, "grad_norm": 0.14241099756591138, "learning_rate": 2.263813938394567e-05, "loss": 0.8743, "step": 7492 }, { "epoch": 3.39356884057971, "grad_norm": 0.19297435030994542, "learning_rate": 2.2626746407966057e-05, "loss": 0.834, "step": 7493 }, { "epoch": 3.3940217391304346, "grad_norm": 0.16935806204034978, "learning_rate": 2.2615355168818153e-05, "loss": 0.8717, "step": 7494 }, { "epoch": 3.394474637681159, "grad_norm": 0.1341047985476934, "learning_rate": 2.2603965667640724e-05, "loss": 0.8361, "step": 7495 }, { "epoch": 3.394927536231884, "grad_norm": 0.15912390334546717, "learning_rate": 2.2592577905572433e-05, "loss": 0.869, "step": 7496 }, { "epoch": 3.395380434782609, "grad_norm": 0.2046401937068188, "learning_rate": 2.258119188375171e-05, "loss": 0.8566, "step": 7497 }, { "epoch": 3.3958333333333335, "grad_norm": 0.2243596774421368, "learning_rate": 2.2569807603316836e-05, "loss": 0.8559, "step": 7498 }, { "epoch": 3.396286231884058, "grad_norm": 0.15430463611294312, "learning_rate": 2.2558425065405905e-05, "loss": 0.871, "step": 7499 }, { "epoch": 3.3967391304347827, "grad_norm": 0.1601979422528698, "learning_rate": 2.2547044271156873e-05, "loss": 0.8537, "step": 7500 }, { "epoch": 3.3971920289855073, "grad_norm": 0.16997129455050242, "learning_rate": 2.253566522170746e-05, "loss": 0.8503, "step": 7501 }, { "epoch": 3.397644927536232, "grad_norm": 0.16538191388195891, "learning_rate": 2.2524287918195294e-05, "loss": 0.852, "step": 7502 }, { "epoch": 3.3980978260869565, "grad_norm": 0.15827404825072067, "learning_rate": 2.2512912361757735e-05, "loss": 0.8549, "step": 7503 }, { "epoch": 3.398550724637681, "grad_norm": 0.15732527480818667, "learning_rate": 2.2501538553532055e-05, "loss": 0.8517, "step": 7504 }, { "epoch": 3.3990036231884058, "grad_norm": 0.20797830567738756, "learning_rate": 2.2490166494655295e-05, "loss": 0.8571, "step": 7505 }, { "epoch": 3.3994565217391304, "grad_norm": 0.18819214420656147, "learning_rate": 2.247879618626434e-05, "loss": 0.8428, "step": 7506 }, { "epoch": 3.399909420289855, "grad_norm": 0.12764264866797867, "learning_rate": 2.2467427629495874e-05, "loss": 0.8568, "step": 7507 }, { "epoch": 3.4003623188405796, "grad_norm": 0.1530751628236554, "learning_rate": 2.2456060825486473e-05, "loss": 0.8688, "step": 7508 }, { "epoch": 3.400815217391304, "grad_norm": 0.1580915845207188, "learning_rate": 2.2444695775372454e-05, "loss": 0.8514, "step": 7509 }, { "epoch": 3.401268115942029, "grad_norm": 0.15196470210078922, "learning_rate": 2.243333248029003e-05, "loss": 0.8511, "step": 7510 }, { "epoch": 3.4017210144927534, "grad_norm": 0.19355342954088314, "learning_rate": 2.2421970941375183e-05, "loss": 0.8538, "step": 7511 }, { "epoch": 3.4021739130434785, "grad_norm": 0.18202156168109365, "learning_rate": 2.2410611159763765e-05, "loss": 0.8696, "step": 7512 }, { "epoch": 3.402626811594203, "grad_norm": 0.2030056352807406, "learning_rate": 2.2399253136591427e-05, "loss": 0.8644, "step": 7513 }, { "epoch": 3.4030797101449277, "grad_norm": 0.13332429947223828, "learning_rate": 2.2387896872993645e-05, "loss": 0.8467, "step": 7514 }, { "epoch": 3.4035326086956523, "grad_norm": 0.1613847964926002, "learning_rate": 2.2376542370105706e-05, "loss": 0.841, "step": 7515 }, { "epoch": 3.403985507246377, "grad_norm": 0.1271027336659548, "learning_rate": 2.2365189629062764e-05, "loss": 0.8643, "step": 7516 }, { "epoch": 3.4044384057971016, "grad_norm": 0.2302621994219562, "learning_rate": 2.2353838650999745e-05, "loss": 0.8654, "step": 7517 }, { "epoch": 3.404891304347826, "grad_norm": 0.180167170154584, "learning_rate": 2.2342489437051452e-05, "loss": 0.849, "step": 7518 }, { "epoch": 3.4053442028985508, "grad_norm": 0.1853450580096131, "learning_rate": 2.2331141988352462e-05, "loss": 0.8563, "step": 7519 }, { "epoch": 3.4057971014492754, "grad_norm": 0.21065008735831142, "learning_rate": 2.2319796306037224e-05, "loss": 0.8604, "step": 7520 }, { "epoch": 3.40625, "grad_norm": 0.15864785257983632, "learning_rate": 2.2308452391239958e-05, "loss": 0.8346, "step": 7521 }, { "epoch": 3.4067028985507246, "grad_norm": 0.18933176428044446, "learning_rate": 2.229711024509474e-05, "loss": 0.8497, "step": 7522 }, { "epoch": 3.4071557971014492, "grad_norm": 0.18222772545035487, "learning_rate": 2.228576986873545e-05, "loss": 0.8691, "step": 7523 }, { "epoch": 3.407608695652174, "grad_norm": 0.13851225540378728, "learning_rate": 2.2274431263295834e-05, "loss": 0.8713, "step": 7524 }, { "epoch": 3.4080615942028984, "grad_norm": 0.18238661624007813, "learning_rate": 2.2263094429909398e-05, "loss": 0.8524, "step": 7525 }, { "epoch": 3.408514492753623, "grad_norm": 0.13652131358484332, "learning_rate": 2.2251759369709536e-05, "loss": 0.8435, "step": 7526 }, { "epoch": 3.4089673913043477, "grad_norm": 0.13813868866107393, "learning_rate": 2.2240426083829396e-05, "loss": 0.866, "step": 7527 }, { "epoch": 3.4094202898550723, "grad_norm": 0.1671138518853825, "learning_rate": 2.222909457340202e-05, "loss": 0.861, "step": 7528 }, { "epoch": 3.409873188405797, "grad_norm": 0.12234444872432884, "learning_rate": 2.2217764839560213e-05, "loss": 0.8516, "step": 7529 }, { "epoch": 3.4103260869565215, "grad_norm": 0.14974393503093797, "learning_rate": 2.220643688343664e-05, "loss": 0.8426, "step": 7530 }, { "epoch": 3.4107789855072466, "grad_norm": 0.1382029280329905, "learning_rate": 2.2195110706163753e-05, "loss": 0.8463, "step": 7531 }, { "epoch": 3.411231884057971, "grad_norm": 0.13514039126625688, "learning_rate": 2.218378630887388e-05, "loss": 0.8392, "step": 7532 }, { "epoch": 3.411684782608696, "grad_norm": 0.1461825555726465, "learning_rate": 2.2172463692699103e-05, "loss": 0.8667, "step": 7533 }, { "epoch": 3.4121376811594204, "grad_norm": 0.1224789512424856, "learning_rate": 2.2161142858771393e-05, "loss": 0.8339, "step": 7534 }, { "epoch": 3.412590579710145, "grad_norm": 0.15479842532138904, "learning_rate": 2.2149823808222486e-05, "loss": 0.8455, "step": 7535 }, { "epoch": 3.4130434782608696, "grad_norm": 0.13564649552069447, "learning_rate": 2.2138506542184e-05, "loss": 0.8527, "step": 7536 }, { "epoch": 3.4134963768115942, "grad_norm": 0.14465058460569657, "learning_rate": 2.21271910617873e-05, "loss": 0.8524, "step": 7537 }, { "epoch": 3.413949275362319, "grad_norm": 0.14357348503395148, "learning_rate": 2.2115877368163643e-05, "loss": 0.834, "step": 7538 }, { "epoch": 3.4144021739130435, "grad_norm": 0.1528992708199048, "learning_rate": 2.210456546244407e-05, "loss": 0.8939, "step": 7539 }, { "epoch": 3.414855072463768, "grad_norm": 0.15038297431900935, "learning_rate": 2.2093255345759443e-05, "loss": 0.8393, "step": 7540 }, { "epoch": 3.4153079710144927, "grad_norm": 0.12892910455163853, "learning_rate": 2.2081947019240443e-05, "loss": 0.8731, "step": 7541 }, { "epoch": 3.4157608695652173, "grad_norm": 0.1745256657936443, "learning_rate": 2.2070640484017606e-05, "loss": 0.8508, "step": 7542 }, { "epoch": 3.416213768115942, "grad_norm": 0.15732319044412893, "learning_rate": 2.2059335741221232e-05, "loss": 0.8516, "step": 7543 }, { "epoch": 3.4166666666666665, "grad_norm": 0.1615766824897675, "learning_rate": 2.2048032791981515e-05, "loss": 0.8676, "step": 7544 }, { "epoch": 3.417119565217391, "grad_norm": 0.16620831867623362, "learning_rate": 2.2036731637428393e-05, "loss": 0.8638, "step": 7545 }, { "epoch": 3.417572463768116, "grad_norm": 0.13671751606471436, "learning_rate": 2.2025432278691687e-05, "loss": 0.8571, "step": 7546 }, { "epoch": 3.418025362318841, "grad_norm": 0.1627911460787226, "learning_rate": 2.201413471690099e-05, "loss": 0.8537, "step": 7547 }, { "epoch": 3.4184782608695654, "grad_norm": 0.12905370728660345, "learning_rate": 2.200283895318576e-05, "loss": 0.8596, "step": 7548 }, { "epoch": 3.41893115942029, "grad_norm": 0.16177654596094204, "learning_rate": 2.1991544988675218e-05, "loss": 0.8697, "step": 7549 }, { "epoch": 3.4193840579710146, "grad_norm": 0.14320818290859377, "learning_rate": 2.1980252824498475e-05, "loss": 0.8691, "step": 7550 }, { "epoch": 3.4198369565217392, "grad_norm": 0.12807206606125826, "learning_rate": 2.1968962461784398e-05, "loss": 0.8536, "step": 7551 }, { "epoch": 3.420289855072464, "grad_norm": 0.14255088410612318, "learning_rate": 2.1957673901661735e-05, "loss": 0.8459, "step": 7552 }, { "epoch": 3.4207427536231885, "grad_norm": 0.15323799631848678, "learning_rate": 2.1946387145258976e-05, "loss": 0.8684, "step": 7553 }, { "epoch": 3.421195652173913, "grad_norm": 0.13753067851778777, "learning_rate": 2.193510219370453e-05, "loss": 0.8706, "step": 7554 }, { "epoch": 3.4216485507246377, "grad_norm": 0.1595111579494893, "learning_rate": 2.1923819048126535e-05, "loss": 0.8471, "step": 7555 }, { "epoch": 3.4221014492753623, "grad_norm": 0.14604116423433827, "learning_rate": 2.1912537709653004e-05, "loss": 0.87, "step": 7556 }, { "epoch": 3.422554347826087, "grad_norm": 0.16494474278067933, "learning_rate": 2.1901258179411714e-05, "loss": 0.8593, "step": 7557 }, { "epoch": 3.4230072463768115, "grad_norm": 0.1459170973126281, "learning_rate": 2.1889980458530346e-05, "loss": 0.8513, "step": 7558 }, { "epoch": 3.423460144927536, "grad_norm": 0.1815774209096094, "learning_rate": 2.187870454813631e-05, "loss": 0.8546, "step": 7559 }, { "epoch": 3.4239130434782608, "grad_norm": 0.15638535010471072, "learning_rate": 2.1867430449356917e-05, "loss": 0.8366, "step": 7560 }, { "epoch": 3.4243659420289854, "grad_norm": 0.18172946736277182, "learning_rate": 2.1856158163319215e-05, "loss": 0.8514, "step": 7561 }, { "epoch": 3.42481884057971, "grad_norm": 0.1440020653355666, "learning_rate": 2.184488769115015e-05, "loss": 0.8502, "step": 7562 }, { "epoch": 3.4252717391304346, "grad_norm": 0.16580840107779624, "learning_rate": 2.1833619033976432e-05, "loss": 0.8425, "step": 7563 }, { "epoch": 3.425724637681159, "grad_norm": 0.1481939304905161, "learning_rate": 2.1822352192924603e-05, "loss": 0.8578, "step": 7564 }, { "epoch": 3.426177536231884, "grad_norm": 0.150295241573662, "learning_rate": 2.1811087169121034e-05, "loss": 0.8513, "step": 7565 }, { "epoch": 3.426630434782609, "grad_norm": 0.17181793326432793, "learning_rate": 2.1799823963691892e-05, "loss": 0.8708, "step": 7566 }, { "epoch": 3.4270833333333335, "grad_norm": 0.1545539586647873, "learning_rate": 2.1788562577763192e-05, "loss": 0.8637, "step": 7567 }, { "epoch": 3.427536231884058, "grad_norm": 0.19212461223871416, "learning_rate": 2.1777303012460743e-05, "loss": 0.8599, "step": 7568 }, { "epoch": 3.4279891304347827, "grad_norm": 0.1763308986637828, "learning_rate": 2.17660452689102e-05, "loss": 0.8584, "step": 7569 }, { "epoch": 3.4284420289855073, "grad_norm": 0.16190058342881422, "learning_rate": 2.1754789348236997e-05, "loss": 0.8835, "step": 7570 }, { "epoch": 3.428894927536232, "grad_norm": 0.17099989365777765, "learning_rate": 2.174353525156643e-05, "loss": 0.8543, "step": 7571 }, { "epoch": 3.4293478260869565, "grad_norm": 0.17783239051913138, "learning_rate": 2.1732282980023564e-05, "loss": 0.85, "step": 7572 }, { "epoch": 3.429800724637681, "grad_norm": 0.17334058761179016, "learning_rate": 2.172103253473332e-05, "loss": 0.8806, "step": 7573 }, { "epoch": 3.4302536231884058, "grad_norm": 0.18323242086039757, "learning_rate": 2.1709783916820405e-05, "loss": 0.8686, "step": 7574 }, { "epoch": 3.4307065217391304, "grad_norm": 0.22745484660148413, "learning_rate": 2.1698537127409392e-05, "loss": 0.8469, "step": 7575 }, { "epoch": 3.431159420289855, "grad_norm": 0.16851093768453188, "learning_rate": 2.1687292167624605e-05, "loss": 0.8486, "step": 7576 }, { "epoch": 3.4316123188405796, "grad_norm": 0.22537634030395512, "learning_rate": 2.167604903859026e-05, "loss": 0.8593, "step": 7577 }, { "epoch": 3.432065217391304, "grad_norm": 0.18278080011234846, "learning_rate": 2.1664807741430306e-05, "loss": 0.8654, "step": 7578 }, { "epoch": 3.432518115942029, "grad_norm": 0.17581964132380803, "learning_rate": 2.1653568277268596e-05, "loss": 0.8626, "step": 7579 }, { "epoch": 3.4329710144927534, "grad_norm": 0.1400556464089811, "learning_rate": 2.1642330647228727e-05, "loss": 0.8531, "step": 7580 }, { "epoch": 3.4334239130434785, "grad_norm": 0.21084552900113607, "learning_rate": 2.163109485243418e-05, "loss": 0.8579, "step": 7581 }, { "epoch": 3.433876811594203, "grad_norm": 0.13342151492307633, "learning_rate": 2.1619860894008155e-05, "loss": 0.8503, "step": 7582 }, { "epoch": 3.4343297101449277, "grad_norm": 0.1917784379864494, "learning_rate": 2.1608628773073783e-05, "loss": 0.8561, "step": 7583 }, { "epoch": 3.4347826086956523, "grad_norm": 0.17163738893746222, "learning_rate": 2.1597398490753917e-05, "loss": 0.872, "step": 7584 }, { "epoch": 3.435235507246377, "grad_norm": 0.19230117959488016, "learning_rate": 2.1586170048171304e-05, "loss": 0.8747, "step": 7585 }, { "epoch": 3.4356884057971016, "grad_norm": 0.13507603691117376, "learning_rate": 2.1574943446448434e-05, "loss": 0.8546, "step": 7586 }, { "epoch": 3.436141304347826, "grad_norm": 0.20667510801275235, "learning_rate": 2.1563718686707677e-05, "loss": 0.8678, "step": 7587 }, { "epoch": 3.4365942028985508, "grad_norm": 0.1668565050943327, "learning_rate": 2.1552495770071165e-05, "loss": 0.8778, "step": 7588 }, { "epoch": 3.4370471014492754, "grad_norm": 0.16829682639965057, "learning_rate": 2.15412746976609e-05, "loss": 0.8524, "step": 7589 }, { "epoch": 3.4375, "grad_norm": 0.14471882867799005, "learning_rate": 2.1530055470598654e-05, "loss": 0.8775, "step": 7590 }, { "epoch": 3.4379528985507246, "grad_norm": 0.17117390309560815, "learning_rate": 2.1518838090006025e-05, "loss": 0.8604, "step": 7591 }, { "epoch": 3.4384057971014492, "grad_norm": 0.1366908417836515, "learning_rate": 2.1507622557004425e-05, "loss": 0.8576, "step": 7592 }, { "epoch": 3.438858695652174, "grad_norm": 0.15581370047158208, "learning_rate": 2.1496408872715123e-05, "loss": 0.8746, "step": 7593 }, { "epoch": 3.4393115942028984, "grad_norm": 0.13132798778901159, "learning_rate": 2.1485197038259127e-05, "loss": 0.8606, "step": 7594 }, { "epoch": 3.439764492753623, "grad_norm": 0.1659142883999109, "learning_rate": 2.147398705475734e-05, "loss": 0.8634, "step": 7595 }, { "epoch": 3.4402173913043477, "grad_norm": 0.15379767667175537, "learning_rate": 2.14627789233304e-05, "loss": 0.8572, "step": 7596 }, { "epoch": 3.4406702898550723, "grad_norm": 0.16945681140935875, "learning_rate": 2.1451572645098856e-05, "loss": 0.8639, "step": 7597 }, { "epoch": 3.441123188405797, "grad_norm": 0.1545892241582846, "learning_rate": 2.1440368221182974e-05, "loss": 0.8444, "step": 7598 }, { "epoch": 3.4415760869565215, "grad_norm": 0.13442410127199134, "learning_rate": 2.14291656527029e-05, "loss": 0.8603, "step": 7599 }, { "epoch": 3.4420289855072466, "grad_norm": 0.16122302582704245, "learning_rate": 2.141796494077854e-05, "loss": 0.8692, "step": 7600 }, { "epoch": 3.442481884057971, "grad_norm": 0.1286643678321523, "learning_rate": 2.140676608652969e-05, "loss": 0.8554, "step": 7601 }, { "epoch": 3.442934782608696, "grad_norm": 0.12120977024429888, "learning_rate": 2.1395569091075887e-05, "loss": 0.8496, "step": 7602 }, { "epoch": 3.4433876811594204, "grad_norm": 0.13404678017622898, "learning_rate": 2.1384373955536537e-05, "loss": 0.8654, "step": 7603 }, { "epoch": 3.443840579710145, "grad_norm": 0.1402778071788934, "learning_rate": 2.1373180681030803e-05, "loss": 0.8745, "step": 7604 }, { "epoch": 3.4442934782608696, "grad_norm": 0.1600003255906552, "learning_rate": 2.136198926867773e-05, "loss": 0.8629, "step": 7605 }, { "epoch": 3.4447463768115942, "grad_norm": 0.13371627354826096, "learning_rate": 2.135079971959612e-05, "loss": 0.8537, "step": 7606 }, { "epoch": 3.445199275362319, "grad_norm": 0.14311547043708595, "learning_rate": 2.1339612034904623e-05, "loss": 0.8617, "step": 7607 }, { "epoch": 3.4456521739130435, "grad_norm": 0.13738840602663654, "learning_rate": 2.1328426215721655e-05, "loss": 0.8704, "step": 7608 }, { "epoch": 3.446105072463768, "grad_norm": 0.1391595623716558, "learning_rate": 2.131724226316553e-05, "loss": 0.865, "step": 7609 }, { "epoch": 3.4465579710144927, "grad_norm": 0.14131145490532815, "learning_rate": 2.1306060178354283e-05, "loss": 0.8598, "step": 7610 }, { "epoch": 3.4470108695652173, "grad_norm": 0.13354158490165846, "learning_rate": 2.1294879962405835e-05, "loss": 0.8493, "step": 7611 }, { "epoch": 3.447463768115942, "grad_norm": 0.17657577727609433, "learning_rate": 2.1283701616437867e-05, "loss": 0.8674, "step": 7612 }, { "epoch": 3.4479166666666665, "grad_norm": 0.15105406567643323, "learning_rate": 2.1272525141567925e-05, "loss": 0.875, "step": 7613 }, { "epoch": 3.448369565217391, "grad_norm": 0.13368662617059668, "learning_rate": 2.126135053891331e-05, "loss": 0.8566, "step": 7614 }, { "epoch": 3.448822463768116, "grad_norm": 0.13999856295446855, "learning_rate": 2.1250177809591186e-05, "loss": 0.8499, "step": 7615 }, { "epoch": 3.449275362318841, "grad_norm": 0.15033796667102775, "learning_rate": 2.1239006954718483e-05, "loss": 0.8576, "step": 7616 }, { "epoch": 3.4497282608695654, "grad_norm": 0.1620118119442088, "learning_rate": 2.1227837975411996e-05, "loss": 0.848, "step": 7617 }, { "epoch": 3.45018115942029, "grad_norm": 0.12230621430701738, "learning_rate": 2.1216670872788276e-05, "loss": 0.8599, "step": 7618 }, { "epoch": 3.4506340579710146, "grad_norm": 0.12481476921312216, "learning_rate": 2.1205505647963748e-05, "loss": 0.8474, "step": 7619 }, { "epoch": 3.4510869565217392, "grad_norm": 0.1394687340350263, "learning_rate": 2.1194342302054595e-05, "loss": 0.8775, "step": 7620 }, { "epoch": 3.451539855072464, "grad_norm": 0.12956491250055513, "learning_rate": 2.1183180836176847e-05, "loss": 0.8736, "step": 7621 }, { "epoch": 3.4519927536231885, "grad_norm": 0.14680193318241747, "learning_rate": 2.117202125144634e-05, "loss": 0.8627, "step": 7622 }, { "epoch": 3.452445652173913, "grad_norm": 0.15525368227524866, "learning_rate": 2.1160863548978697e-05, "loss": 0.8399, "step": 7623 }, { "epoch": 3.4528985507246377, "grad_norm": 0.1626965392702238, "learning_rate": 2.1149707729889362e-05, "loss": 0.8472, "step": 7624 }, { "epoch": 3.4533514492753623, "grad_norm": 0.12565645869289951, "learning_rate": 2.1138553795293636e-05, "loss": 0.8552, "step": 7625 }, { "epoch": 3.453804347826087, "grad_norm": 0.16677222255578772, "learning_rate": 2.1127401746306554e-05, "loss": 0.8517, "step": 7626 }, { "epoch": 3.4542572463768115, "grad_norm": 0.14967069918317913, "learning_rate": 2.1116251584043047e-05, "loss": 0.8789, "step": 7627 }, { "epoch": 3.454710144927536, "grad_norm": 0.15671579215559037, "learning_rate": 2.1105103309617773e-05, "loss": 0.8534, "step": 7628 }, { "epoch": 3.4551630434782608, "grad_norm": 0.12909659010070162, "learning_rate": 2.1093956924145283e-05, "loss": 0.8508, "step": 7629 }, { "epoch": 3.4556159420289854, "grad_norm": 0.15563551618693605, "learning_rate": 2.108281242873986e-05, "loss": 0.8512, "step": 7630 }, { "epoch": 3.45606884057971, "grad_norm": 0.1469555929473261, "learning_rate": 2.107166982451567e-05, "loss": 0.8595, "step": 7631 }, { "epoch": 3.4565217391304346, "grad_norm": 0.16317260312859502, "learning_rate": 2.1060529112586642e-05, "loss": 0.8661, "step": 7632 }, { "epoch": 3.456974637681159, "grad_norm": 0.14761485451326162, "learning_rate": 2.1049390294066537e-05, "loss": 0.8507, "step": 7633 }, { "epoch": 3.457427536231884, "grad_norm": 0.17580522621966552, "learning_rate": 2.1038253370068894e-05, "loss": 0.8605, "step": 7634 }, { "epoch": 3.457880434782609, "grad_norm": 0.14655599718677412, "learning_rate": 2.1027118341707124e-05, "loss": 0.8595, "step": 7635 }, { "epoch": 3.4583333333333335, "grad_norm": 0.1475347137426604, "learning_rate": 2.1015985210094385e-05, "loss": 0.8597, "step": 7636 }, { "epoch": 3.458786231884058, "grad_norm": 0.1494442742645475, "learning_rate": 2.1004853976343698e-05, "loss": 0.8845, "step": 7637 }, { "epoch": 3.4592391304347827, "grad_norm": 0.1667816087466907, "learning_rate": 2.099372464156785e-05, "loss": 0.8753, "step": 7638 }, { "epoch": 3.4596920289855073, "grad_norm": 0.1632991933399845, "learning_rate": 2.098259720687948e-05, "loss": 0.8782, "step": 7639 }, { "epoch": 3.460144927536232, "grad_norm": 0.15090699172707395, "learning_rate": 2.0971471673391e-05, "loss": 0.8675, "step": 7640 }, { "epoch": 3.4605978260869565, "grad_norm": 0.1751563792223888, "learning_rate": 2.0960348042214656e-05, "loss": 0.8529, "step": 7641 }, { "epoch": 3.461050724637681, "grad_norm": 0.172178827977916, "learning_rate": 2.094922631446247e-05, "loss": 0.8666, "step": 7642 }, { "epoch": 3.4615036231884058, "grad_norm": 0.18629529804203454, "learning_rate": 2.0938106491246334e-05, "loss": 0.8662, "step": 7643 }, { "epoch": 3.4619565217391304, "grad_norm": 0.14989062004270734, "learning_rate": 2.092698857367788e-05, "loss": 0.8678, "step": 7644 }, { "epoch": 3.462409420289855, "grad_norm": 0.19134634628279634, "learning_rate": 2.0915872562868627e-05, "loss": 0.8801, "step": 7645 }, { "epoch": 3.4628623188405796, "grad_norm": 0.20368794723360528, "learning_rate": 2.0904758459929812e-05, "loss": 0.8611, "step": 7646 }, { "epoch": 3.463315217391304, "grad_norm": 0.16794175055464403, "learning_rate": 2.0893646265972566e-05, "loss": 0.8558, "step": 7647 }, { "epoch": 3.463768115942029, "grad_norm": 0.14346874911397686, "learning_rate": 2.0882535982107785e-05, "loss": 0.8454, "step": 7648 }, { "epoch": 3.4642210144927534, "grad_norm": 0.15296058092113377, "learning_rate": 2.0871427609446174e-05, "loss": 0.8378, "step": 7649 }, { "epoch": 3.4646739130434785, "grad_norm": 0.1393378290419469, "learning_rate": 2.0860321149098244e-05, "loss": 0.8794, "step": 7650 }, { "epoch": 3.465126811594203, "grad_norm": 0.13997918466330372, "learning_rate": 2.0849216602174347e-05, "loss": 0.8709, "step": 7651 }, { "epoch": 3.4655797101449277, "grad_norm": 0.1464393220048813, "learning_rate": 2.0838113969784605e-05, "loss": 0.8478, "step": 7652 }, { "epoch": 3.4660326086956523, "grad_norm": 0.1329279430883919, "learning_rate": 2.0827013253038985e-05, "loss": 0.8643, "step": 7653 }, { "epoch": 3.466485507246377, "grad_norm": 0.17431483930475167, "learning_rate": 2.0815914453047217e-05, "loss": 0.8733, "step": 7654 }, { "epoch": 3.4669384057971016, "grad_norm": 0.1560568095533634, "learning_rate": 2.0804817570918896e-05, "loss": 0.8545, "step": 7655 }, { "epoch": 3.467391304347826, "grad_norm": 0.14682639319637936, "learning_rate": 2.0793722607763383e-05, "loss": 0.8439, "step": 7656 }, { "epoch": 3.4678442028985508, "grad_norm": 0.1442367493856921, "learning_rate": 2.0782629564689855e-05, "loss": 0.855, "step": 7657 }, { "epoch": 3.4682971014492754, "grad_norm": 0.1453347820251005, "learning_rate": 2.0771538442807283e-05, "loss": 0.8729, "step": 7658 }, { "epoch": 3.46875, "grad_norm": 0.14378819428709205, "learning_rate": 2.0760449243224504e-05, "loss": 0.871, "step": 7659 }, { "epoch": 3.4692028985507246, "grad_norm": 0.14294698471304693, "learning_rate": 2.074936196705008e-05, "loss": 0.8612, "step": 7660 }, { "epoch": 3.4696557971014492, "grad_norm": 0.12420245993820593, "learning_rate": 2.0738276615392467e-05, "loss": 0.8528, "step": 7661 }, { "epoch": 3.470108695652174, "grad_norm": 0.12118998057784183, "learning_rate": 2.0727193189359848e-05, "loss": 0.8766, "step": 7662 }, { "epoch": 3.4705615942028984, "grad_norm": 0.13993587782307843, "learning_rate": 2.071611169006028e-05, "loss": 0.8652, "step": 7663 }, { "epoch": 3.471014492753623, "grad_norm": 0.132968288594398, "learning_rate": 2.070503211860158e-05, "loss": 0.8513, "step": 7664 }, { "epoch": 3.4714673913043477, "grad_norm": 0.13874266639601338, "learning_rate": 2.0693954476091398e-05, "loss": 0.863, "step": 7665 }, { "epoch": 3.4719202898550723, "grad_norm": 0.15011695189445343, "learning_rate": 2.0682878763637168e-05, "loss": 0.8552, "step": 7666 }, { "epoch": 3.472373188405797, "grad_norm": 0.1313543053224369, "learning_rate": 2.0671804982346168e-05, "loss": 0.8575, "step": 7667 }, { "epoch": 3.4728260869565215, "grad_norm": 0.13717277261831423, "learning_rate": 2.066073313332544e-05, "loss": 0.8622, "step": 7668 }, { "epoch": 3.4732789855072466, "grad_norm": 0.16837631443305653, "learning_rate": 2.0649663217681882e-05, "loss": 0.8718, "step": 7669 }, { "epoch": 3.473731884057971, "grad_norm": 0.17083587285612237, "learning_rate": 2.0638595236522135e-05, "loss": 0.8815, "step": 7670 }, { "epoch": 3.474184782608696, "grad_norm": 0.20793362293717027, "learning_rate": 2.0627529190952727e-05, "loss": 0.8489, "step": 7671 }, { "epoch": 3.4746376811594204, "grad_norm": 0.14921671602882458, "learning_rate": 2.06164650820799e-05, "loss": 0.8628, "step": 7672 }, { "epoch": 3.475090579710145, "grad_norm": 0.22443323863872577, "learning_rate": 2.0605402911009796e-05, "loss": 0.8412, "step": 7673 }, { "epoch": 3.4755434782608696, "grad_norm": 0.15022949598215843, "learning_rate": 2.0594342678848296e-05, "loss": 0.8552, "step": 7674 }, { "epoch": 3.4759963768115942, "grad_norm": 0.19320409872335734, "learning_rate": 2.0583284386701104e-05, "loss": 0.8509, "step": 7675 }, { "epoch": 3.476449275362319, "grad_norm": 0.15773933561921966, "learning_rate": 2.057222803567373e-05, "loss": 0.8767, "step": 7676 }, { "epoch": 3.4769021739130435, "grad_norm": 0.16768265869119192, "learning_rate": 2.056117362687152e-05, "loss": 0.859, "step": 7677 }, { "epoch": 3.477355072463768, "grad_norm": 0.1624030389890983, "learning_rate": 2.055012116139957e-05, "loss": 0.8676, "step": 7678 }, { "epoch": 3.4778079710144927, "grad_norm": 0.18110014906808464, "learning_rate": 2.0539070640362847e-05, "loss": 0.8494, "step": 7679 }, { "epoch": 3.4782608695652173, "grad_norm": 0.15599846194337938, "learning_rate": 2.0528022064866048e-05, "loss": 0.8618, "step": 7680 }, { "epoch": 3.478713768115942, "grad_norm": 0.20706771712834557, "learning_rate": 2.051697543601376e-05, "loss": 0.8571, "step": 7681 }, { "epoch": 3.4791666666666665, "grad_norm": 0.13516247266419854, "learning_rate": 2.050593075491031e-05, "loss": 0.8491, "step": 7682 }, { "epoch": 3.479619565217391, "grad_norm": 0.15534258381414337, "learning_rate": 2.0494888022659856e-05, "loss": 0.8619, "step": 7683 }, { "epoch": 3.480072463768116, "grad_norm": 0.17969568606597372, "learning_rate": 2.0483847240366335e-05, "loss": 0.8582, "step": 7684 }, { "epoch": 3.480525362318841, "grad_norm": 0.17604744915533382, "learning_rate": 2.047280840913355e-05, "loss": 0.8769, "step": 7685 }, { "epoch": 3.4809782608695654, "grad_norm": 0.12910444282056624, "learning_rate": 2.0461771530065035e-05, "loss": 0.8596, "step": 7686 }, { "epoch": 3.48143115942029, "grad_norm": 0.15180180486055656, "learning_rate": 2.0450736604264193e-05, "loss": 0.872, "step": 7687 }, { "epoch": 3.4818840579710146, "grad_norm": 0.1481552908909907, "learning_rate": 2.0439703632834193e-05, "loss": 0.8745, "step": 7688 }, { "epoch": 3.4823369565217392, "grad_norm": 0.14881078060644878, "learning_rate": 2.042867261687801e-05, "loss": 0.8701, "step": 7689 }, { "epoch": 3.482789855072464, "grad_norm": 0.13798187551744912, "learning_rate": 2.0417643557498444e-05, "loss": 0.8511, "step": 7690 }, { "epoch": 3.4832427536231885, "grad_norm": 0.17643111852436089, "learning_rate": 2.0406616455798084e-05, "loss": 0.8462, "step": 7691 }, { "epoch": 3.483695652173913, "grad_norm": 0.1595897729878256, "learning_rate": 2.0395591312879324e-05, "loss": 0.8631, "step": 7692 }, { "epoch": 3.4841485507246377, "grad_norm": 0.12995156399339908, "learning_rate": 2.0384568129844354e-05, "loss": 0.8649, "step": 7693 }, { "epoch": 3.4846014492753623, "grad_norm": 0.16964578655132692, "learning_rate": 2.037354690779521e-05, "loss": 0.8631, "step": 7694 }, { "epoch": 3.485054347826087, "grad_norm": 0.14685941631905672, "learning_rate": 2.0362527647833662e-05, "loss": 0.8777, "step": 7695 }, { "epoch": 3.4855072463768115, "grad_norm": 0.16197119461391277, "learning_rate": 2.0351510351061357e-05, "loss": 0.8583, "step": 7696 }, { "epoch": 3.485960144927536, "grad_norm": 0.12937142224075615, "learning_rate": 2.034049501857969e-05, "loss": 0.866, "step": 7697 }, { "epoch": 3.4864130434782608, "grad_norm": 0.16111257791245076, "learning_rate": 2.03294816514899e-05, "loss": 0.8743, "step": 7698 }, { "epoch": 3.4868659420289854, "grad_norm": 0.1586170501418957, "learning_rate": 2.0318470250893002e-05, "loss": 0.8802, "step": 7699 }, { "epoch": 3.48731884057971, "grad_norm": 0.16470337435499577, "learning_rate": 2.0307460817889815e-05, "loss": 0.8561, "step": 7700 }, { "epoch": 3.4877717391304346, "grad_norm": 0.17254533527528054, "learning_rate": 2.0296453353580964e-05, "loss": 0.8366, "step": 7701 }, { "epoch": 3.488224637681159, "grad_norm": 0.1901815715527352, "learning_rate": 2.0285447859066913e-05, "loss": 0.8686, "step": 7702 }, { "epoch": 3.488677536231884, "grad_norm": 0.1863470976293391, "learning_rate": 2.027444433544786e-05, "loss": 0.8711, "step": 7703 }, { "epoch": 3.489130434782609, "grad_norm": 0.12514035466686305, "learning_rate": 2.0263442783823884e-05, "loss": 0.8575, "step": 7704 }, { "epoch": 3.4895833333333335, "grad_norm": 0.1376439658702075, "learning_rate": 2.025244320529479e-05, "loss": 0.8564, "step": 7705 }, { "epoch": 3.490036231884058, "grad_norm": 0.13523057767851074, "learning_rate": 2.0241445600960256e-05, "loss": 0.8539, "step": 7706 }, { "epoch": 3.4904891304347827, "grad_norm": 0.16843611395319985, "learning_rate": 2.0230449971919718e-05, "loss": 0.8688, "step": 7707 }, { "epoch": 3.4909420289855073, "grad_norm": 0.11629301519655633, "learning_rate": 2.0219456319272425e-05, "loss": 0.8533, "step": 7708 }, { "epoch": 3.491394927536232, "grad_norm": 0.17688101118507635, "learning_rate": 2.020846464411741e-05, "loss": 0.846, "step": 7709 }, { "epoch": 3.4918478260869565, "grad_norm": 0.13840876598336863, "learning_rate": 2.0197474947553562e-05, "loss": 0.8691, "step": 7710 }, { "epoch": 3.492300724637681, "grad_norm": 0.15312104784129765, "learning_rate": 2.0186487230679508e-05, "loss": 0.8379, "step": 7711 }, { "epoch": 3.4927536231884058, "grad_norm": 0.18603050290438516, "learning_rate": 2.017550149459374e-05, "loss": 0.8637, "step": 7712 }, { "epoch": 3.4932065217391304, "grad_norm": 0.14641345809035794, "learning_rate": 2.0164517740394487e-05, "loss": 0.8502, "step": 7713 }, { "epoch": 3.493659420289855, "grad_norm": 0.14586275269947002, "learning_rate": 2.0153535969179837e-05, "loss": 0.8665, "step": 7714 }, { "epoch": 3.4941123188405796, "grad_norm": 0.20265168062193772, "learning_rate": 2.014255618204764e-05, "loss": 0.8746, "step": 7715 }, { "epoch": 3.494565217391304, "grad_norm": 0.20849425065400998, "learning_rate": 2.0131578380095575e-05, "loss": 0.8635, "step": 7716 }, { "epoch": 3.495018115942029, "grad_norm": 0.13909956391421371, "learning_rate": 2.0120602564421113e-05, "loss": 0.8692, "step": 7717 }, { "epoch": 3.4954710144927534, "grad_norm": 0.21042639363077356, "learning_rate": 2.0109628736121506e-05, "loss": 0.8824, "step": 7718 }, { "epoch": 3.4959239130434785, "grad_norm": 0.24635505528871057, "learning_rate": 2.0098656896293816e-05, "loss": 0.8508, "step": 7719 }, { "epoch": 3.496376811594203, "grad_norm": 0.20120593631851452, "learning_rate": 2.0087687046034955e-05, "loss": 0.849, "step": 7720 }, { "epoch": 3.4968297101449277, "grad_norm": 0.16786402770376344, "learning_rate": 2.007671918644156e-05, "loss": 0.8477, "step": 7721 }, { "epoch": 3.4972826086956523, "grad_norm": 0.259074716106317, "learning_rate": 2.0065753318610123e-05, "loss": 0.8624, "step": 7722 }, { "epoch": 3.497735507246377, "grad_norm": 0.14792685815132836, "learning_rate": 2.005478944363691e-05, "loss": 0.8571, "step": 7723 }, { "epoch": 3.4981884057971016, "grad_norm": 0.2182000895390762, "learning_rate": 2.0043827562618013e-05, "loss": 0.8728, "step": 7724 }, { "epoch": 3.498641304347826, "grad_norm": 0.2428278929162623, "learning_rate": 2.00328676766493e-05, "loss": 0.8587, "step": 7725 }, { "epoch": 3.4990942028985508, "grad_norm": 0.14933466655776095, "learning_rate": 2.0021909786826445e-05, "loss": 0.8711, "step": 7726 }, { "epoch": 3.4995471014492754, "grad_norm": 0.18087064871803382, "learning_rate": 2.0010953894244914e-05, "loss": 0.8635, "step": 7727 }, { "epoch": 3.5, "grad_norm": 0.23498629887265837, "learning_rate": 2.0000000000000012e-05, "loss": 0.8328, "step": 7728 }, { "epoch": 3.5004528985507246, "grad_norm": 0.15267383617971497, "learning_rate": 1.998904810518678e-05, "loss": 0.8406, "step": 7729 }, { "epoch": 3.5009057971014492, "grad_norm": 0.14710664517575764, "learning_rate": 1.9978098210900144e-05, "loss": 0.8464, "step": 7730 }, { "epoch": 3.501358695652174, "grad_norm": 0.17312457402182121, "learning_rate": 1.9967150318234738e-05, "loss": 0.8669, "step": 7731 }, { "epoch": 3.5018115942028984, "grad_norm": 0.17127945478033937, "learning_rate": 1.9956204428285074e-05, "loss": 0.852, "step": 7732 }, { "epoch": 3.502264492753623, "grad_norm": 0.12517360511392628, "learning_rate": 1.994526054214542e-05, "loss": 0.8765, "step": 7733 }, { "epoch": 3.5027173913043477, "grad_norm": 0.15500158570465744, "learning_rate": 1.9934318660909843e-05, "loss": 0.8648, "step": 7734 }, { "epoch": 3.5031702898550723, "grad_norm": 0.1746900892710851, "learning_rate": 1.9923378785672213e-05, "loss": 0.863, "step": 7735 }, { "epoch": 3.503623188405797, "grad_norm": 0.13738163577115908, "learning_rate": 1.9912440917526233e-05, "loss": 0.8534, "step": 7736 }, { "epoch": 3.5040760869565215, "grad_norm": 0.1540760481165102, "learning_rate": 1.9901505057565358e-05, "loss": 0.8489, "step": 7737 }, { "epoch": 3.504528985507246, "grad_norm": 0.15178065816853864, "learning_rate": 1.9890571206882877e-05, "loss": 0.8447, "step": 7738 }, { "epoch": 3.504981884057971, "grad_norm": 0.1372750498022839, "learning_rate": 1.987963936657185e-05, "loss": 0.8677, "step": 7739 }, { "epoch": 3.505434782608696, "grad_norm": 0.12104560840066847, "learning_rate": 1.986870953772517e-05, "loss": 0.8582, "step": 7740 }, { "epoch": 3.5058876811594204, "grad_norm": 0.13268173659335872, "learning_rate": 1.9857781721435504e-05, "loss": 0.8574, "step": 7741 }, { "epoch": 3.506340579710145, "grad_norm": 0.11936960518900508, "learning_rate": 1.9846855918795315e-05, "loss": 0.8644, "step": 7742 }, { "epoch": 3.5067934782608696, "grad_norm": 0.14822551160168124, "learning_rate": 1.9835932130896855e-05, "loss": 0.8691, "step": 7743 }, { "epoch": 3.5072463768115942, "grad_norm": 0.1558468852974401, "learning_rate": 1.982501035883223e-05, "loss": 0.8711, "step": 7744 }, { "epoch": 3.507699275362319, "grad_norm": 0.14486619072415696, "learning_rate": 1.981409060369328e-05, "loss": 0.8532, "step": 7745 }, { "epoch": 3.5081521739130435, "grad_norm": 0.1302721276700119, "learning_rate": 1.980317286657169e-05, "loss": 0.8613, "step": 7746 }, { "epoch": 3.508605072463768, "grad_norm": 0.12590927917668318, "learning_rate": 1.97922571485589e-05, "loss": 0.8612, "step": 7747 }, { "epoch": 3.5090579710144927, "grad_norm": 0.15246532276848337, "learning_rate": 1.97813434507462e-05, "loss": 0.8755, "step": 7748 }, { "epoch": 3.5095108695652173, "grad_norm": 0.136287671485893, "learning_rate": 1.9770431774224632e-05, "loss": 0.8603, "step": 7749 }, { "epoch": 3.509963768115942, "grad_norm": 0.16059815011250975, "learning_rate": 1.9759522120085057e-05, "loss": 0.853, "step": 7750 }, { "epoch": 3.5104166666666665, "grad_norm": 0.11638526821067972, "learning_rate": 1.9748614489418118e-05, "loss": 0.8358, "step": 7751 }, { "epoch": 3.5108695652173916, "grad_norm": 0.15297440602483708, "learning_rate": 1.9737708883314287e-05, "loss": 0.8557, "step": 7752 }, { "epoch": 3.511322463768116, "grad_norm": 0.12700051292974465, "learning_rate": 1.97268053028638e-05, "loss": 0.8705, "step": 7753 }, { "epoch": 3.511775362318841, "grad_norm": 0.163527879604645, "learning_rate": 1.971590374915672e-05, "loss": 0.8517, "step": 7754 }, { "epoch": 3.5122282608695654, "grad_norm": 0.1288371178863507, "learning_rate": 1.9705004223282875e-05, "loss": 0.863, "step": 7755 }, { "epoch": 3.51268115942029, "grad_norm": 0.17068515371116655, "learning_rate": 1.9694106726331933e-05, "loss": 0.8504, "step": 7756 }, { "epoch": 3.5131340579710146, "grad_norm": 0.1890146217691349, "learning_rate": 1.96832112593933e-05, "loss": 0.8638, "step": 7757 }, { "epoch": 3.5135869565217392, "grad_norm": 0.12994824800937008, "learning_rate": 1.9672317823556265e-05, "loss": 0.8623, "step": 7758 }, { "epoch": 3.514039855072464, "grad_norm": 0.1305257435792682, "learning_rate": 1.9661426419909803e-05, "loss": 0.8588, "step": 7759 }, { "epoch": 3.5144927536231885, "grad_norm": 0.12867406618596255, "learning_rate": 1.965053704954278e-05, "loss": 0.8538, "step": 7760 }, { "epoch": 3.514945652173913, "grad_norm": 0.13385216961031515, "learning_rate": 1.9639649713543806e-05, "loss": 0.8673, "step": 7761 }, { "epoch": 3.5153985507246377, "grad_norm": 0.14656755736537422, "learning_rate": 1.962876441300132e-05, "loss": 0.8521, "step": 7762 }, { "epoch": 3.5158514492753623, "grad_norm": 0.16841573287578535, "learning_rate": 1.961788114900353e-05, "loss": 0.8656, "step": 7763 }, { "epoch": 3.516304347826087, "grad_norm": 0.15339108960645054, "learning_rate": 1.960699992263847e-05, "loss": 0.8787, "step": 7764 }, { "epoch": 3.5167572463768115, "grad_norm": 0.16047050394970397, "learning_rate": 1.959612073499393e-05, "loss": 0.8622, "step": 7765 }, { "epoch": 3.517210144927536, "grad_norm": 0.1615242866860115, "learning_rate": 1.9585243587157548e-05, "loss": 0.8714, "step": 7766 }, { "epoch": 3.5176630434782608, "grad_norm": 0.15080456221070784, "learning_rate": 1.9574368480216713e-05, "loss": 0.8665, "step": 7767 }, { "epoch": 3.5181159420289854, "grad_norm": 0.15192884773711282, "learning_rate": 1.9563495415258627e-05, "loss": 0.8416, "step": 7768 }, { "epoch": 3.51856884057971, "grad_norm": 0.1442316456770367, "learning_rate": 1.9552624393370275e-05, "loss": 0.8699, "step": 7769 }, { "epoch": 3.5190217391304346, "grad_norm": 0.16178337227909156, "learning_rate": 1.9541755415638485e-05, "loss": 0.8814, "step": 7770 }, { "epoch": 3.519474637681159, "grad_norm": 0.1405140799265543, "learning_rate": 1.9530888483149802e-05, "loss": 0.8651, "step": 7771 }, { "epoch": 3.519927536231884, "grad_norm": 0.12060008028911741, "learning_rate": 1.952002359699065e-05, "loss": 0.8455, "step": 7772 }, { "epoch": 3.5203804347826084, "grad_norm": 0.1276766798335878, "learning_rate": 1.950916075824718e-05, "loss": 0.8532, "step": 7773 }, { "epoch": 3.5208333333333335, "grad_norm": 0.11617557696377631, "learning_rate": 1.9498299968005393e-05, "loss": 0.8568, "step": 7774 }, { "epoch": 3.521286231884058, "grad_norm": 0.12052174553311912, "learning_rate": 1.9487441227351044e-05, "loss": 0.876, "step": 7775 }, { "epoch": 3.5217391304347827, "grad_norm": 0.13308736405209692, "learning_rate": 1.9476584537369705e-05, "loss": 0.8467, "step": 7776 }, { "epoch": 3.5221920289855073, "grad_norm": 0.11311308034021364, "learning_rate": 1.9465729899146715e-05, "loss": 0.8722, "step": 7777 }, { "epoch": 3.522644927536232, "grad_norm": 0.12572954531330846, "learning_rate": 1.945487731376726e-05, "loss": 0.8812, "step": 7778 }, { "epoch": 3.5230978260869565, "grad_norm": 0.1244921803104392, "learning_rate": 1.9444026782316262e-05, "loss": 0.867, "step": 7779 }, { "epoch": 3.523550724637681, "grad_norm": 0.1300829506780799, "learning_rate": 1.9433178305878497e-05, "loss": 0.8746, "step": 7780 }, { "epoch": 3.5240036231884058, "grad_norm": 0.15821790223897159, "learning_rate": 1.942233188553847e-05, "loss": 0.8687, "step": 7781 }, { "epoch": 3.5244565217391304, "grad_norm": 0.1968664812301664, "learning_rate": 1.9411487522380546e-05, "loss": 0.8689, "step": 7782 }, { "epoch": 3.524909420289855, "grad_norm": 0.173936389463997, "learning_rate": 1.9400645217488843e-05, "loss": 0.8477, "step": 7783 }, { "epoch": 3.5253623188405796, "grad_norm": 0.11718744597984647, "learning_rate": 1.938980497194728e-05, "loss": 0.8495, "step": 7784 }, { "epoch": 3.525815217391304, "grad_norm": 0.18153811063277087, "learning_rate": 1.9378966786839556e-05, "loss": 0.8623, "step": 7785 }, { "epoch": 3.526268115942029, "grad_norm": 0.21730407472826677, "learning_rate": 1.936813066324922e-05, "loss": 0.8677, "step": 7786 }, { "epoch": 3.526721014492754, "grad_norm": 0.12335320901396402, "learning_rate": 1.9357296602259535e-05, "loss": 0.8532, "step": 7787 }, { "epoch": 3.5271739130434785, "grad_norm": 0.15145186008296685, "learning_rate": 1.9346464604953638e-05, "loss": 0.8535, "step": 7788 }, { "epoch": 3.527626811594203, "grad_norm": 0.1391622168054397, "learning_rate": 1.9335634672414386e-05, "loss": 0.8688, "step": 7789 }, { "epoch": 3.5280797101449277, "grad_norm": 0.12850345774027447, "learning_rate": 1.9324806805724497e-05, "loss": 0.8474, "step": 7790 }, { "epoch": 3.5285326086956523, "grad_norm": 0.13386164420977864, "learning_rate": 1.9313981005966432e-05, "loss": 0.8355, "step": 7791 }, { "epoch": 3.528985507246377, "grad_norm": 0.15470549345936724, "learning_rate": 1.930315727422247e-05, "loss": 0.8581, "step": 7792 }, { "epoch": 3.5294384057971016, "grad_norm": 0.12968861546960012, "learning_rate": 1.9292335611574655e-05, "loss": 0.8354, "step": 7793 }, { "epoch": 3.529891304347826, "grad_norm": 0.12813749977974168, "learning_rate": 1.9281516019104874e-05, "loss": 0.8631, "step": 7794 }, { "epoch": 3.5303442028985508, "grad_norm": 0.13209518584129054, "learning_rate": 1.927069849789476e-05, "loss": 0.8492, "step": 7795 }, { "epoch": 3.5307971014492754, "grad_norm": 0.1362091200277809, "learning_rate": 1.9259883049025778e-05, "loss": 0.8915, "step": 7796 }, { "epoch": 3.53125, "grad_norm": 0.1282721325071555, "learning_rate": 1.9249069673579136e-05, "loss": 0.8655, "step": 7797 }, { "epoch": 3.5317028985507246, "grad_norm": 0.14341707417297714, "learning_rate": 1.92382583726359e-05, "loss": 0.8868, "step": 7798 }, { "epoch": 3.5321557971014492, "grad_norm": 0.12789791504839237, "learning_rate": 1.9227449147276873e-05, "loss": 0.8685, "step": 7799 }, { "epoch": 3.532608695652174, "grad_norm": 0.1317195250148104, "learning_rate": 1.9216641998582666e-05, "loss": 0.8629, "step": 7800 }, { "epoch": 3.5330615942028984, "grad_norm": 0.1253627321107785, "learning_rate": 1.920583692763369e-05, "loss": 0.8336, "step": 7801 }, { "epoch": 3.533514492753623, "grad_norm": 0.12442726153952528, "learning_rate": 1.9195033935510152e-05, "loss": 0.8404, "step": 7802 }, { "epoch": 3.5339673913043477, "grad_norm": 0.15831843656394864, "learning_rate": 1.9184233023292026e-05, "loss": 0.8674, "step": 7803 }, { "epoch": 3.5344202898550723, "grad_norm": 0.16831681326276346, "learning_rate": 1.9173434192059125e-05, "loss": 0.8382, "step": 7804 }, { "epoch": 3.534873188405797, "grad_norm": 0.18506193494905904, "learning_rate": 1.9162637442890997e-05, "loss": 0.8456, "step": 7805 }, { "epoch": 3.5353260869565215, "grad_norm": 0.16065894386573412, "learning_rate": 1.915184277686704e-05, "loss": 0.8458, "step": 7806 }, { "epoch": 3.535778985507246, "grad_norm": 0.16209360667769368, "learning_rate": 1.914105019506638e-05, "loss": 0.8467, "step": 7807 }, { "epoch": 3.536231884057971, "grad_norm": 0.1323199358826173, "learning_rate": 1.9130259698568e-05, "loss": 0.861, "step": 7808 }, { "epoch": 3.536684782608696, "grad_norm": 0.14720580176607026, "learning_rate": 1.9119471288450632e-05, "loss": 0.8654, "step": 7809 }, { "epoch": 3.5371376811594204, "grad_norm": 0.12900346683895117, "learning_rate": 1.9108684965792806e-05, "loss": 0.865, "step": 7810 }, { "epoch": 3.537590579710145, "grad_norm": 0.16523799568690334, "learning_rate": 1.909790073167285e-05, "loss": 0.8831, "step": 7811 }, { "epoch": 3.5380434782608696, "grad_norm": 0.12041453331720418, "learning_rate": 1.9087118587168865e-05, "loss": 0.8528, "step": 7812 }, { "epoch": 3.5384963768115942, "grad_norm": 0.18657090868592266, "learning_rate": 1.9076338533358788e-05, "loss": 0.8586, "step": 7813 }, { "epoch": 3.538949275362319, "grad_norm": 0.12944496224639607, "learning_rate": 1.9065560571320287e-05, "loss": 0.8646, "step": 7814 }, { "epoch": 3.5394021739130435, "grad_norm": 0.16374097031193363, "learning_rate": 1.9054784702130888e-05, "loss": 0.8741, "step": 7815 }, { "epoch": 3.539855072463768, "grad_norm": 0.14101152062827474, "learning_rate": 1.904401092686784e-05, "loss": 0.8545, "step": 7816 }, { "epoch": 3.5403079710144927, "grad_norm": 0.15990617463176307, "learning_rate": 1.9033239246608234e-05, "loss": 0.8742, "step": 7817 }, { "epoch": 3.5407608695652173, "grad_norm": 0.17854327972502748, "learning_rate": 1.9022469662428928e-05, "loss": 0.8549, "step": 7818 }, { "epoch": 3.541213768115942, "grad_norm": 0.1307591610833529, "learning_rate": 1.9011702175406576e-05, "loss": 0.8591, "step": 7819 }, { "epoch": 3.5416666666666665, "grad_norm": 0.20639298574629847, "learning_rate": 1.90009367866176e-05, "loss": 0.8572, "step": 7820 }, { "epoch": 3.5421195652173916, "grad_norm": 0.18307082127836663, "learning_rate": 1.8990173497138263e-05, "loss": 0.8483, "step": 7821 }, { "epoch": 3.542572463768116, "grad_norm": 0.14475819840571963, "learning_rate": 1.8979412308044557e-05, "loss": 0.8495, "step": 7822 }, { "epoch": 3.543025362318841, "grad_norm": 0.15425514746000832, "learning_rate": 1.896865322041233e-05, "loss": 0.8622, "step": 7823 }, { "epoch": 3.5434782608695654, "grad_norm": 0.16611269113252244, "learning_rate": 1.8957896235317153e-05, "loss": 0.8699, "step": 7824 }, { "epoch": 3.54393115942029, "grad_norm": 0.14703401872135288, "learning_rate": 1.894714135383445e-05, "loss": 0.8613, "step": 7825 }, { "epoch": 3.5443840579710146, "grad_norm": 0.15209827558790212, "learning_rate": 1.8936388577039384e-05, "loss": 0.8485, "step": 7826 }, { "epoch": 3.5448369565217392, "grad_norm": 0.23906167379935245, "learning_rate": 1.892563790600693e-05, "loss": 0.8558, "step": 7827 }, { "epoch": 3.545289855072464, "grad_norm": 0.15757378257468588, "learning_rate": 1.8914889341811835e-05, "loss": 0.8593, "step": 7828 }, { "epoch": 3.5457427536231885, "grad_norm": 0.14907629531492636, "learning_rate": 1.8904142885528683e-05, "loss": 0.8666, "step": 7829 }, { "epoch": 3.546195652173913, "grad_norm": 0.1625114417251396, "learning_rate": 1.8893398538231778e-05, "loss": 0.8578, "step": 7830 }, { "epoch": 3.5466485507246377, "grad_norm": 0.1306470399107859, "learning_rate": 1.8882656300995283e-05, "loss": 0.8601, "step": 7831 }, { "epoch": 3.5471014492753623, "grad_norm": 0.1359686221690703, "learning_rate": 1.8871916174893085e-05, "loss": 0.862, "step": 7832 }, { "epoch": 3.547554347826087, "grad_norm": 0.15682746147104074, "learning_rate": 1.886117816099892e-05, "loss": 0.8598, "step": 7833 }, { "epoch": 3.5480072463768115, "grad_norm": 0.14855586728861675, "learning_rate": 1.8850442260386276e-05, "loss": 0.8509, "step": 7834 }, { "epoch": 3.548460144927536, "grad_norm": 0.15756642057664533, "learning_rate": 1.8839708474128434e-05, "loss": 0.8502, "step": 7835 }, { "epoch": 3.5489130434782608, "grad_norm": 0.1251953134336173, "learning_rate": 1.8828976803298454e-05, "loss": 0.8498, "step": 7836 }, { "epoch": 3.5493659420289854, "grad_norm": 0.18814494837893805, "learning_rate": 1.881824724896922e-05, "loss": 0.8931, "step": 7837 }, { "epoch": 3.54981884057971, "grad_norm": 0.11801840264527724, "learning_rate": 1.8807519812213365e-05, "loss": 0.8535, "step": 7838 }, { "epoch": 3.5502717391304346, "grad_norm": 0.16975241994320162, "learning_rate": 1.8796794494103344e-05, "loss": 0.8493, "step": 7839 }, { "epoch": 3.550724637681159, "grad_norm": 0.13649993052106177, "learning_rate": 1.878607129571137e-05, "loss": 0.8736, "step": 7840 }, { "epoch": 3.551177536231884, "grad_norm": 0.1402898474658856, "learning_rate": 1.8775350218109473e-05, "loss": 0.8556, "step": 7841 }, { "epoch": 3.5516304347826084, "grad_norm": 0.15233713387191639, "learning_rate": 1.8764631262369453e-05, "loss": 0.8671, "step": 7842 }, { "epoch": 3.5520833333333335, "grad_norm": 0.1422378849898093, "learning_rate": 1.875391442956289e-05, "loss": 0.8532, "step": 7843 }, { "epoch": 3.552536231884058, "grad_norm": 0.12540126959212627, "learning_rate": 1.874319972076116e-05, "loss": 0.8615, "step": 7844 }, { "epoch": 3.5529891304347827, "grad_norm": 0.13865975103099298, "learning_rate": 1.8732487137035452e-05, "loss": 0.8619, "step": 7845 }, { "epoch": 3.5534420289855073, "grad_norm": 0.1220199891645457, "learning_rate": 1.872177667945669e-05, "loss": 0.8492, "step": 7846 }, { "epoch": 3.553894927536232, "grad_norm": 0.1531547816202518, "learning_rate": 1.8711068349095648e-05, "loss": 0.8547, "step": 7847 }, { "epoch": 3.5543478260869565, "grad_norm": 0.13194011436850245, "learning_rate": 1.8700362147022823e-05, "loss": 0.8722, "step": 7848 }, { "epoch": 3.554800724637681, "grad_norm": 0.1318496101355382, "learning_rate": 1.8689658074308557e-05, "loss": 0.8632, "step": 7849 }, { "epoch": 3.5552536231884058, "grad_norm": 0.11284856722733533, "learning_rate": 1.8678956132022934e-05, "loss": 0.8545, "step": 7850 }, { "epoch": 3.5557065217391304, "grad_norm": 0.13573236847145467, "learning_rate": 1.8668256321235865e-05, "loss": 0.858, "step": 7851 }, { "epoch": 3.556159420289855, "grad_norm": 0.1283280323727499, "learning_rate": 1.8657558643017016e-05, "loss": 0.8544, "step": 7852 }, { "epoch": 3.5566123188405796, "grad_norm": 0.11480787886560642, "learning_rate": 1.8646863098435845e-05, "loss": 0.8408, "step": 7853 }, { "epoch": 3.557065217391304, "grad_norm": 0.13516590261852235, "learning_rate": 1.8636169688561595e-05, "loss": 0.8553, "step": 7854 }, { "epoch": 3.557518115942029, "grad_norm": 0.12016569719464969, "learning_rate": 1.862547841446333e-05, "loss": 0.8394, "step": 7855 }, { "epoch": 3.557971014492754, "grad_norm": 0.1270079596623414, "learning_rate": 1.8614789277209845e-05, "loss": 0.8595, "step": 7856 }, { "epoch": 3.5584239130434785, "grad_norm": 0.11705940629824985, "learning_rate": 1.8604102277869777e-05, "loss": 0.8389, "step": 7857 }, { "epoch": 3.558876811594203, "grad_norm": 0.14447696937502105, "learning_rate": 1.859341741751149e-05, "loss": 0.8505, "step": 7858 }, { "epoch": 3.5593297101449277, "grad_norm": 0.11220804197845541, "learning_rate": 1.858273469720321e-05, "loss": 0.8667, "step": 7859 }, { "epoch": 3.5597826086956523, "grad_norm": 0.14782762600641983, "learning_rate": 1.8572054118012874e-05, "loss": 0.8587, "step": 7860 }, { "epoch": 3.560235507246377, "grad_norm": 0.13080413258248325, "learning_rate": 1.856137568100825e-05, "loss": 0.8468, "step": 7861 }, { "epoch": 3.5606884057971016, "grad_norm": 0.13535806428103467, "learning_rate": 1.855069938725685e-05, "loss": 0.8787, "step": 7862 }, { "epoch": 3.561141304347826, "grad_norm": 0.14667633699045451, "learning_rate": 1.8540025237826035e-05, "loss": 0.8427, "step": 7863 }, { "epoch": 3.5615942028985508, "grad_norm": 0.13457537892365773, "learning_rate": 1.8529353233782893e-05, "loss": 0.8658, "step": 7864 }, { "epoch": 3.5620471014492754, "grad_norm": 0.13948465458638418, "learning_rate": 1.8518683376194345e-05, "loss": 0.8578, "step": 7865 }, { "epoch": 3.5625, "grad_norm": 0.13216590665964623, "learning_rate": 1.8508015666127043e-05, "loss": 0.8516, "step": 7866 }, { "epoch": 3.5629528985507246, "grad_norm": 0.13934297363871442, "learning_rate": 1.849735010464749e-05, "loss": 0.8577, "step": 7867 }, { "epoch": 3.5634057971014492, "grad_norm": 0.12819789230618908, "learning_rate": 1.8486686692821914e-05, "loss": 0.8601, "step": 7868 }, { "epoch": 3.563858695652174, "grad_norm": 0.14266400443100322, "learning_rate": 1.8476025431716368e-05, "loss": 0.855, "step": 7869 }, { "epoch": 3.5643115942028984, "grad_norm": 0.15679691038205684, "learning_rate": 1.8465366322396652e-05, "loss": 0.8795, "step": 7870 }, { "epoch": 3.564764492753623, "grad_norm": 0.13857300587942079, "learning_rate": 1.84547093659284e-05, "loss": 0.8712, "step": 7871 }, { "epoch": 3.5652173913043477, "grad_norm": 0.15141616497962052, "learning_rate": 1.844405456337698e-05, "loss": 0.8528, "step": 7872 }, { "epoch": 3.5656702898550723, "grad_norm": 0.14826917299803757, "learning_rate": 1.8433401915807603e-05, "loss": 0.8482, "step": 7873 }, { "epoch": 3.566123188405797, "grad_norm": 0.14663633166675227, "learning_rate": 1.8422751424285192e-05, "loss": 0.8595, "step": 7874 }, { "epoch": 3.5665760869565215, "grad_norm": 0.13971260237648475, "learning_rate": 1.841210308987453e-05, "loss": 0.8593, "step": 7875 }, { "epoch": 3.567028985507246, "grad_norm": 0.1270048798319333, "learning_rate": 1.840145691364013e-05, "loss": 0.8564, "step": 7876 }, { "epoch": 3.567481884057971, "grad_norm": 0.16730517485600654, "learning_rate": 1.839081289664631e-05, "loss": 0.8394, "step": 7877 }, { "epoch": 3.567934782608696, "grad_norm": 0.13513561266697144, "learning_rate": 1.838017103995715e-05, "loss": 0.8462, "step": 7878 }, { "epoch": 3.5683876811594204, "grad_norm": 0.19371277566457468, "learning_rate": 1.8369531344636567e-05, "loss": 0.8552, "step": 7879 }, { "epoch": 3.568840579710145, "grad_norm": 0.19812439691368258, "learning_rate": 1.8358893811748196e-05, "loss": 0.8607, "step": 7880 }, { "epoch": 3.5692934782608696, "grad_norm": 0.142430331258825, "learning_rate": 1.8348258442355523e-05, "loss": 0.8694, "step": 7881 }, { "epoch": 3.5697463768115942, "grad_norm": 0.20107612594507074, "learning_rate": 1.8337625237521744e-05, "loss": 0.8587, "step": 7882 }, { "epoch": 3.570199275362319, "grad_norm": 0.18899689428024943, "learning_rate": 1.8326994198309914e-05, "loss": 0.8298, "step": 7883 }, { "epoch": 3.5706521739130435, "grad_norm": 0.1254629594841244, "learning_rate": 1.8316365325782818e-05, "loss": 0.8735, "step": 7884 }, { "epoch": 3.571105072463768, "grad_norm": 0.1526683754061751, "learning_rate": 1.830573862100304e-05, "loss": 0.8671, "step": 7885 }, { "epoch": 3.5715579710144927, "grad_norm": 0.13589973752651607, "learning_rate": 1.829511408503293e-05, "loss": 0.8502, "step": 7886 }, { "epoch": 3.5720108695652173, "grad_norm": 0.16070450757641094, "learning_rate": 1.8284491718934684e-05, "loss": 0.861, "step": 7887 }, { "epoch": 3.572463768115942, "grad_norm": 0.1408252493420921, "learning_rate": 1.827387152377019e-05, "loss": 0.8758, "step": 7888 }, { "epoch": 3.5729166666666665, "grad_norm": 0.1492231779314868, "learning_rate": 1.826325350060121e-05, "loss": 0.8309, "step": 7889 }, { "epoch": 3.5733695652173916, "grad_norm": 0.13840066934427095, "learning_rate": 1.82526376504892e-05, "loss": 0.863, "step": 7890 }, { "epoch": 3.573822463768116, "grad_norm": 0.1948508631810511, "learning_rate": 1.824202397449548e-05, "loss": 0.869, "step": 7891 }, { "epoch": 3.574275362318841, "grad_norm": 0.13146509435397064, "learning_rate": 1.8231412473681094e-05, "loss": 0.8758, "step": 7892 }, { "epoch": 3.5747282608695654, "grad_norm": 0.16040938058144186, "learning_rate": 1.8220803149106908e-05, "loss": 0.8544, "step": 7893 }, { "epoch": 3.57518115942029, "grad_norm": 0.14101293919065072, "learning_rate": 1.8210196001833545e-05, "loss": 0.8764, "step": 7894 }, { "epoch": 3.5756340579710146, "grad_norm": 0.17049764950192126, "learning_rate": 1.8199591032921416e-05, "loss": 0.876, "step": 7895 }, { "epoch": 3.5760869565217392, "grad_norm": 0.15888369830117716, "learning_rate": 1.8188988243430698e-05, "loss": 0.858, "step": 7896 }, { "epoch": 3.576539855072464, "grad_norm": 0.17981203693140238, "learning_rate": 1.817838763442141e-05, "loss": 0.8613, "step": 7897 }, { "epoch": 3.5769927536231885, "grad_norm": 0.21790586920284982, "learning_rate": 1.816778920695327e-05, "loss": 0.8694, "step": 7898 }, { "epoch": 3.577445652173913, "grad_norm": 0.16469355579884787, "learning_rate": 1.8157192962085854e-05, "loss": 0.8504, "step": 7899 }, { "epoch": 3.5778985507246377, "grad_norm": 0.18126637455682365, "learning_rate": 1.814659890087846e-05, "loss": 0.8464, "step": 7900 }, { "epoch": 3.5783514492753623, "grad_norm": 0.1759749019173258, "learning_rate": 1.8136007024390215e-05, "loss": 0.8642, "step": 7901 }, { "epoch": 3.578804347826087, "grad_norm": 0.15563981193509172, "learning_rate": 1.8125417333679994e-05, "loss": 0.8604, "step": 7902 }, { "epoch": 3.5792572463768115, "grad_norm": 0.13869318901055624, "learning_rate": 1.8114829829806466e-05, "loss": 0.8655, "step": 7903 }, { "epoch": 3.579710144927536, "grad_norm": 0.17818594976738705, "learning_rate": 1.8104244513828067e-05, "loss": 0.8704, "step": 7904 }, { "epoch": 3.5801630434782608, "grad_norm": 0.16374371740417726, "learning_rate": 1.8093661386803053e-05, "loss": 0.8526, "step": 7905 }, { "epoch": 3.5806159420289854, "grad_norm": 0.13463266593937695, "learning_rate": 1.808308044978941e-05, "loss": 0.847, "step": 7906 }, { "epoch": 3.58106884057971, "grad_norm": 0.16747802607548098, "learning_rate": 1.8072501703844958e-05, "loss": 0.8508, "step": 7907 }, { "epoch": 3.5815217391304346, "grad_norm": 0.1187774213032806, "learning_rate": 1.8061925150027244e-05, "loss": 0.8674, "step": 7908 }, { "epoch": 3.581974637681159, "grad_norm": 0.18091672908861148, "learning_rate": 1.8051350789393654e-05, "loss": 0.8489, "step": 7909 }, { "epoch": 3.582427536231884, "grad_norm": 0.14437596803754413, "learning_rate": 1.8040778623001306e-05, "loss": 0.873, "step": 7910 }, { "epoch": 3.5828804347826084, "grad_norm": 0.14538833515950803, "learning_rate": 1.803020865190711e-05, "loss": 0.8585, "step": 7911 }, { "epoch": 3.5833333333333335, "grad_norm": 0.1233121955925946, "learning_rate": 1.8019640877167763e-05, "loss": 0.872, "step": 7912 }, { "epoch": 3.583786231884058, "grad_norm": 0.20185379927789304, "learning_rate": 1.8009075299839755e-05, "loss": 0.8575, "step": 7913 }, { "epoch": 3.5842391304347827, "grad_norm": 0.15708990633334216, "learning_rate": 1.7998511920979327e-05, "loss": 0.8463, "step": 7914 }, { "epoch": 3.5846920289855073, "grad_norm": 0.1712781546877785, "learning_rate": 1.798795074164254e-05, "loss": 0.871, "step": 7915 }, { "epoch": 3.585144927536232, "grad_norm": 0.21648703897173935, "learning_rate": 1.7977391762885185e-05, "loss": 0.8585, "step": 7916 }, { "epoch": 3.5855978260869565, "grad_norm": 0.143674446514942, "learning_rate": 1.7966834985762885e-05, "loss": 0.8615, "step": 7917 }, { "epoch": 3.586050724637681, "grad_norm": 0.15112954444827556, "learning_rate": 1.795628041133101e-05, "loss": 0.8433, "step": 7918 }, { "epoch": 3.5865036231884058, "grad_norm": 0.17867512288283238, "learning_rate": 1.7945728040644713e-05, "loss": 0.8488, "step": 7919 }, { "epoch": 3.5869565217391304, "grad_norm": 0.13908629193848782, "learning_rate": 1.793517787475891e-05, "loss": 0.8501, "step": 7920 }, { "epoch": 3.587409420289855, "grad_norm": 0.18536311506726075, "learning_rate": 1.7924629914728363e-05, "loss": 0.8445, "step": 7921 }, { "epoch": 3.5878623188405796, "grad_norm": 0.1565202784190657, "learning_rate": 1.7914084161607528e-05, "loss": 0.8656, "step": 7922 }, { "epoch": 3.588315217391304, "grad_norm": 0.1637267784054023, "learning_rate": 1.7903540616450713e-05, "loss": 0.8677, "step": 7923 }, { "epoch": 3.588768115942029, "grad_norm": 0.1443231485255609, "learning_rate": 1.789299928031194e-05, "loss": 0.8417, "step": 7924 }, { "epoch": 3.589221014492754, "grad_norm": 0.17389314935505704, "learning_rate": 1.7882460154245077e-05, "loss": 0.8654, "step": 7925 }, { "epoch": 3.5896739130434785, "grad_norm": 0.2035233117851019, "learning_rate": 1.7871923239303718e-05, "loss": 0.8683, "step": 7926 }, { "epoch": 3.590126811594203, "grad_norm": 0.13170923308718435, "learning_rate": 1.786138853654126e-05, "loss": 0.8439, "step": 7927 }, { "epoch": 3.5905797101449277, "grad_norm": 0.17691187742017325, "learning_rate": 1.785085604701086e-05, "loss": 0.8379, "step": 7928 }, { "epoch": 3.5910326086956523, "grad_norm": 0.15754788331132222, "learning_rate": 1.784032577176549e-05, "loss": 0.8648, "step": 7929 }, { "epoch": 3.591485507246377, "grad_norm": 0.1970970156815833, "learning_rate": 1.782979771185785e-05, "loss": 0.8768, "step": 7930 }, { "epoch": 3.5919384057971016, "grad_norm": 0.17122949508771237, "learning_rate": 1.7819271868340482e-05, "loss": 0.882, "step": 7931 }, { "epoch": 3.592391304347826, "grad_norm": 0.18207199150976874, "learning_rate": 1.780874824226564e-05, "loss": 0.8587, "step": 7932 }, { "epoch": 3.5928442028985508, "grad_norm": 0.13810229057292772, "learning_rate": 1.7798226834685416e-05, "loss": 0.8673, "step": 7933 }, { "epoch": 3.5932971014492754, "grad_norm": 0.14299993868176034, "learning_rate": 1.778770764665162e-05, "loss": 0.8663, "step": 7934 }, { "epoch": 3.59375, "grad_norm": 0.13769172018867884, "learning_rate": 1.7777190679215923e-05, "loss": 0.8476, "step": 7935 }, { "epoch": 3.5942028985507246, "grad_norm": 0.12225303865192459, "learning_rate": 1.776667593342967e-05, "loss": 0.851, "step": 7936 }, { "epoch": 3.5946557971014492, "grad_norm": 0.12403320649557623, "learning_rate": 1.7756163410344044e-05, "loss": 0.8478, "step": 7937 }, { "epoch": 3.595108695652174, "grad_norm": 0.1674110140124677, "learning_rate": 1.7745653111010024e-05, "loss": 0.8622, "step": 7938 }, { "epoch": 3.5955615942028984, "grad_norm": 0.1411111901734779, "learning_rate": 1.7735145036478313e-05, "loss": 0.8678, "step": 7939 }, { "epoch": 3.596014492753623, "grad_norm": 0.1248671357300354, "learning_rate": 1.772463918779945e-05, "loss": 0.8603, "step": 7940 }, { "epoch": 3.5964673913043477, "grad_norm": 0.2304606898074725, "learning_rate": 1.7714135566023688e-05, "loss": 0.8778, "step": 7941 }, { "epoch": 3.5969202898550723, "grad_norm": 0.23451410294634437, "learning_rate": 1.770363417220112e-05, "loss": 0.874, "step": 7942 }, { "epoch": 3.597373188405797, "grad_norm": 0.13317754572581017, "learning_rate": 1.7693135007381568e-05, "loss": 0.8591, "step": 7943 }, { "epoch": 3.5978260869565215, "grad_norm": 0.31183960682012996, "learning_rate": 1.7682638072614667e-05, "loss": 0.8698, "step": 7944 }, { "epoch": 3.598278985507246, "grad_norm": 0.16850135737151273, "learning_rate": 1.7672143368949806e-05, "loss": 0.8476, "step": 7945 }, { "epoch": 3.598731884057971, "grad_norm": 0.2479383148438272, "learning_rate": 1.7661650897436155e-05, "loss": 0.8618, "step": 7946 }, { "epoch": 3.599184782608696, "grad_norm": 0.21226936682511954, "learning_rate": 1.7651160659122647e-05, "loss": 0.872, "step": 7947 }, { "epoch": 3.5996376811594204, "grad_norm": 0.18902837948735532, "learning_rate": 1.764067265505803e-05, "loss": 0.8764, "step": 7948 }, { "epoch": 3.600090579710145, "grad_norm": 0.25491077356417224, "learning_rate": 1.7630186886290794e-05, "loss": 0.8535, "step": 7949 }, { "epoch": 3.6005434782608696, "grad_norm": 0.12697166916133268, "learning_rate": 1.7619703353869235e-05, "loss": 0.866, "step": 7950 }, { "epoch": 3.6009963768115942, "grad_norm": 0.25918887270435587, "learning_rate": 1.7609222058841377e-05, "loss": 0.871, "step": 7951 }, { "epoch": 3.601449275362319, "grad_norm": 0.1925060507509067, "learning_rate": 1.759874300225509e-05, "loss": 0.8723, "step": 7952 }, { "epoch": 3.6019021739130435, "grad_norm": 0.22172360164899255, "learning_rate": 1.758826618515796e-05, "loss": 0.8552, "step": 7953 }, { "epoch": 3.602355072463768, "grad_norm": 0.1606544283524565, "learning_rate": 1.757779160859738e-05, "loss": 0.8711, "step": 7954 }, { "epoch": 3.6028079710144927, "grad_norm": 0.21345018917307534, "learning_rate": 1.7567319273620484e-05, "loss": 0.8555, "step": 7955 }, { "epoch": 3.6032608695652173, "grad_norm": 0.1933042846636257, "learning_rate": 1.7556849181274244e-05, "loss": 0.8646, "step": 7956 }, { "epoch": 3.603713768115942, "grad_norm": 0.17487206478865738, "learning_rate": 1.7546381332605342e-05, "loss": 0.8385, "step": 7957 }, { "epoch": 3.6041666666666665, "grad_norm": 0.26732287323453263, "learning_rate": 1.753591572866029e-05, "loss": 0.8497, "step": 7958 }, { "epoch": 3.6046195652173916, "grad_norm": 0.11712623103902856, "learning_rate": 1.752545237048533e-05, "loss": 0.8472, "step": 7959 }, { "epoch": 3.605072463768116, "grad_norm": 0.20495059726746312, "learning_rate": 1.7514991259126518e-05, "loss": 0.861, "step": 7960 }, { "epoch": 3.605525362318841, "grad_norm": 0.15076350512579362, "learning_rate": 1.7504532395629668e-05, "loss": 0.8635, "step": 7961 }, { "epoch": 3.6059782608695654, "grad_norm": 0.21863201348864408, "learning_rate": 1.749407578104036e-05, "loss": 0.8339, "step": 7962 }, { "epoch": 3.60643115942029, "grad_norm": 0.16760849552448076, "learning_rate": 1.7483621416403945e-05, "loss": 0.8657, "step": 7963 }, { "epoch": 3.6068840579710146, "grad_norm": 0.26349794727413356, "learning_rate": 1.7473169302765587e-05, "loss": 0.841, "step": 7964 }, { "epoch": 3.6073369565217392, "grad_norm": 0.13316933094447467, "learning_rate": 1.7462719441170182e-05, "loss": 0.8551, "step": 7965 }, { "epoch": 3.607789855072464, "grad_norm": 0.25794920268571364, "learning_rate": 1.745227183266244e-05, "loss": 0.8633, "step": 7966 }, { "epoch": 3.6082427536231885, "grad_norm": 0.1493828449565377, "learning_rate": 1.7441826478286797e-05, "loss": 0.8442, "step": 7967 }, { "epoch": 3.608695652173913, "grad_norm": 0.219105086669696, "learning_rate": 1.7431383379087523e-05, "loss": 0.8707, "step": 7968 }, { "epoch": 3.6091485507246377, "grad_norm": 0.16321033607122765, "learning_rate": 1.7420942536108616e-05, "loss": 0.8651, "step": 7969 }, { "epoch": 3.6096014492753623, "grad_norm": 0.17777552153495876, "learning_rate": 1.7410503950393867e-05, "loss": 0.8583, "step": 7970 }, { "epoch": 3.610054347826087, "grad_norm": 0.15594315153484914, "learning_rate": 1.7400067622986813e-05, "loss": 0.8209, "step": 7971 }, { "epoch": 3.6105072463768115, "grad_norm": 0.1432094330337803, "learning_rate": 1.738963355493083e-05, "loss": 0.8714, "step": 7972 }, { "epoch": 3.610960144927536, "grad_norm": 0.14459347077982476, "learning_rate": 1.737920174726899e-05, "loss": 0.8712, "step": 7973 }, { "epoch": 3.6114130434782608, "grad_norm": 0.17415434749165895, "learning_rate": 1.736877220104422e-05, "loss": 0.8452, "step": 7974 }, { "epoch": 3.6118659420289854, "grad_norm": 0.1267481403350883, "learning_rate": 1.7358344917299134e-05, "loss": 0.8446, "step": 7975 }, { "epoch": 3.61231884057971, "grad_norm": 0.18301607858113597, "learning_rate": 1.7347919897076205e-05, "loss": 0.8569, "step": 7976 }, { "epoch": 3.6127717391304346, "grad_norm": 0.13778143006811, "learning_rate": 1.733749714141762e-05, "loss": 0.879, "step": 7977 }, { "epoch": 3.613224637681159, "grad_norm": 0.15047031557518634, "learning_rate": 1.7327076651365355e-05, "loss": 0.8552, "step": 7978 }, { "epoch": 3.613677536231884, "grad_norm": 0.14405008022769336, "learning_rate": 1.731665842796115e-05, "loss": 0.8682, "step": 7979 }, { "epoch": 3.6141304347826084, "grad_norm": 0.14845362627845743, "learning_rate": 1.7306242472246556e-05, "loss": 0.8517, "step": 7980 }, { "epoch": 3.6145833333333335, "grad_norm": 0.12051388765460931, "learning_rate": 1.7295828785262857e-05, "loss": 0.8623, "step": 7981 }, { "epoch": 3.615036231884058, "grad_norm": 0.1514437148666081, "learning_rate": 1.7285417368051142e-05, "loss": 0.8628, "step": 7982 }, { "epoch": 3.6154891304347827, "grad_norm": 0.12229466266296043, "learning_rate": 1.7275008221652237e-05, "loss": 0.8654, "step": 7983 }, { "epoch": 3.6159420289855073, "grad_norm": 0.13085523908703475, "learning_rate": 1.7264601347106783e-05, "loss": 0.8563, "step": 7984 }, { "epoch": 3.616394927536232, "grad_norm": 0.12194283727157312, "learning_rate": 1.7254196745455144e-05, "loss": 0.8487, "step": 7985 }, { "epoch": 3.6168478260869565, "grad_norm": 0.14807593742906902, "learning_rate": 1.724379441773752e-05, "loss": 0.8709, "step": 7986 }, { "epoch": 3.617300724637681, "grad_norm": 0.12516775341650788, "learning_rate": 1.7233394364993822e-05, "loss": 0.8563, "step": 7987 }, { "epoch": 3.6177536231884058, "grad_norm": 0.1528942421837745, "learning_rate": 1.722299658826377e-05, "loss": 0.8644, "step": 7988 }, { "epoch": 3.6182065217391304, "grad_norm": 0.12927742669055883, "learning_rate": 1.7212601088586823e-05, "loss": 0.8471, "step": 7989 }, { "epoch": 3.618659420289855, "grad_norm": 0.1750713914256908, "learning_rate": 1.7202207867002273e-05, "loss": 0.8494, "step": 7990 }, { "epoch": 3.6191123188405796, "grad_norm": 0.13919240847053496, "learning_rate": 1.719181692454911e-05, "loss": 0.8402, "step": 7991 }, { "epoch": 3.619565217391304, "grad_norm": 0.15874335680684065, "learning_rate": 1.7181428262266172e-05, "loss": 0.8583, "step": 7992 }, { "epoch": 3.620018115942029, "grad_norm": 0.14212937102511194, "learning_rate": 1.7171041881191995e-05, "loss": 0.8342, "step": 7993 }, { "epoch": 3.620471014492754, "grad_norm": 0.13196778289063113, "learning_rate": 1.716065778236495e-05, "loss": 0.851, "step": 7994 }, { "epoch": 3.6209239130434785, "grad_norm": 0.15941610736265216, "learning_rate": 1.715027596682314e-05, "loss": 0.8568, "step": 7995 }, { "epoch": 3.621376811594203, "grad_norm": 0.10896075414339317, "learning_rate": 1.7139896435604448e-05, "loss": 0.8579, "step": 7996 }, { "epoch": 3.6218297101449277, "grad_norm": 0.17696572834345653, "learning_rate": 1.7129519189746518e-05, "loss": 0.846, "step": 7997 }, { "epoch": 3.6222826086956523, "grad_norm": 0.1672987324896743, "learning_rate": 1.7119144230286812e-05, "loss": 0.8668, "step": 7998 }, { "epoch": 3.622735507246377, "grad_norm": 0.1254284465930488, "learning_rate": 1.7108771558262504e-05, "loss": 0.856, "step": 7999 }, { "epoch": 3.6231884057971016, "grad_norm": 0.18218997166655698, "learning_rate": 1.7098401174710586e-05, "loss": 0.8635, "step": 8000 }, { "epoch": 3.623641304347826, "grad_norm": 0.16029349521959552, "learning_rate": 1.7088033080667782e-05, "loss": 0.8633, "step": 8001 }, { "epoch": 3.6240942028985508, "grad_norm": 0.11958928355993094, "learning_rate": 1.7077667277170635e-05, "loss": 0.8442, "step": 8002 }, { "epoch": 3.6245471014492754, "grad_norm": 0.15492562322478418, "learning_rate": 1.706730376525541e-05, "loss": 0.8717, "step": 8003 }, { "epoch": 3.625, "grad_norm": 0.14619659476522792, "learning_rate": 1.7056942545958167e-05, "loss": 0.86, "step": 8004 }, { "epoch": 3.6254528985507246, "grad_norm": 0.13482988844192856, "learning_rate": 1.7046583620314723e-05, "loss": 0.8817, "step": 8005 }, { "epoch": 3.6259057971014492, "grad_norm": 0.16790167037022508, "learning_rate": 1.70362269893607e-05, "loss": 0.8704, "step": 8006 }, { "epoch": 3.626358695652174, "grad_norm": 0.16406486365112405, "learning_rate": 1.702587265413144e-05, "loss": 0.8545, "step": 8007 }, { "epoch": 3.6268115942028984, "grad_norm": 0.1236816882690762, "learning_rate": 1.7015520615662108e-05, "loss": 0.8747, "step": 8008 }, { "epoch": 3.627264492753623, "grad_norm": 0.1557080695477006, "learning_rate": 1.700517087498759e-05, "loss": 0.8512, "step": 8009 }, { "epoch": 3.6277173913043477, "grad_norm": 0.14719049951044372, "learning_rate": 1.6994823433142595e-05, "loss": 0.8388, "step": 8010 }, { "epoch": 3.6281702898550723, "grad_norm": 0.12415791939826025, "learning_rate": 1.6984478291161558e-05, "loss": 0.8758, "step": 8011 }, { "epoch": 3.628623188405797, "grad_norm": 0.16937767400327183, "learning_rate": 1.697413545007869e-05, "loss": 0.8731, "step": 8012 }, { "epoch": 3.6290760869565215, "grad_norm": 0.13709400548653902, "learning_rate": 1.696379491092798e-05, "loss": 0.847, "step": 8013 }, { "epoch": 3.629528985507246, "grad_norm": 0.13843654385846824, "learning_rate": 1.6953456674743206e-05, "loss": 0.8663, "step": 8014 }, { "epoch": 3.629981884057971, "grad_norm": 0.15671827517941495, "learning_rate": 1.694312074255788e-05, "loss": 0.862, "step": 8015 }, { "epoch": 3.630434782608696, "grad_norm": 0.1387727833639926, "learning_rate": 1.6932787115405318e-05, "loss": 0.8515, "step": 8016 }, { "epoch": 3.6308876811594204, "grad_norm": 0.1426084417592196, "learning_rate": 1.692245579431857e-05, "loss": 0.8632, "step": 8017 }, { "epoch": 3.631340579710145, "grad_norm": 0.14568357449457536, "learning_rate": 1.6912126780330503e-05, "loss": 0.858, "step": 8018 }, { "epoch": 3.6317934782608696, "grad_norm": 0.1375739747878222, "learning_rate": 1.6901800074473706e-05, "loss": 0.8877, "step": 8019 }, { "epoch": 3.6322463768115942, "grad_norm": 0.16162593861934127, "learning_rate": 1.6891475677780556e-05, "loss": 0.8705, "step": 8020 }, { "epoch": 3.632699275362319, "grad_norm": 0.16151337495576532, "learning_rate": 1.6881153591283185e-05, "loss": 0.8662, "step": 8021 }, { "epoch": 3.6331521739130435, "grad_norm": 0.1236445618686073, "learning_rate": 1.6870833816013545e-05, "loss": 0.8676, "step": 8022 }, { "epoch": 3.633605072463768, "grad_norm": 0.19017523141042617, "learning_rate": 1.686051635300328e-05, "loss": 0.8562, "step": 8023 }, { "epoch": 3.6340579710144927, "grad_norm": 0.12040306697724097, "learning_rate": 1.685020120328388e-05, "loss": 0.8544, "step": 8024 }, { "epoch": 3.6345108695652173, "grad_norm": 0.14654705796653056, "learning_rate": 1.6839888367886533e-05, "loss": 0.8512, "step": 8025 }, { "epoch": 3.634963768115942, "grad_norm": 0.1194402599509789, "learning_rate": 1.6829577847842263e-05, "loss": 0.8571, "step": 8026 }, { "epoch": 3.6354166666666665, "grad_norm": 0.1112171133628566, "learning_rate": 1.68192696441818e-05, "loss": 0.8563, "step": 8027 }, { "epoch": 3.6358695652173916, "grad_norm": 0.11931360907686046, "learning_rate": 1.6808963757935696e-05, "loss": 0.8761, "step": 8028 }, { "epoch": 3.636322463768116, "grad_norm": 0.11625290786657841, "learning_rate": 1.679866019013423e-05, "loss": 0.8774, "step": 8029 }, { "epoch": 3.636775362318841, "grad_norm": 0.1272381649325411, "learning_rate": 1.6788358941807472e-05, "loss": 0.846, "step": 8030 }, { "epoch": 3.6372282608695654, "grad_norm": 0.13758022716432705, "learning_rate": 1.677806001398524e-05, "loss": 0.866, "step": 8031 }, { "epoch": 3.63768115942029, "grad_norm": 0.11339326817855719, "learning_rate": 1.6767763407697165e-05, "loss": 0.8666, "step": 8032 }, { "epoch": 3.6381340579710146, "grad_norm": 0.1200959635674345, "learning_rate": 1.6757469123972578e-05, "loss": 0.8587, "step": 8033 }, { "epoch": 3.6385869565217392, "grad_norm": 0.10578031407740728, "learning_rate": 1.674717716384065e-05, "loss": 0.8538, "step": 8034 }, { "epoch": 3.639039855072464, "grad_norm": 0.13057091859786832, "learning_rate": 1.6736887528330248e-05, "loss": 0.8701, "step": 8035 }, { "epoch": 3.6394927536231885, "grad_norm": 0.1180895376589924, "learning_rate": 1.6726600218470085e-05, "loss": 0.8603, "step": 8036 }, { "epoch": 3.639945652173913, "grad_norm": 0.15302802362801482, "learning_rate": 1.6716315235288574e-05, "loss": 0.8842, "step": 8037 }, { "epoch": 3.6403985507246377, "grad_norm": 0.11477462423034024, "learning_rate": 1.6706032579813927e-05, "loss": 0.8648, "step": 8038 }, { "epoch": 3.6408514492753623, "grad_norm": 0.12823552872377472, "learning_rate": 1.6695752253074097e-05, "loss": 0.8501, "step": 8039 }, { "epoch": 3.641304347826087, "grad_norm": 0.15035730975660314, "learning_rate": 1.6685474256096865e-05, "loss": 0.8774, "step": 8040 }, { "epoch": 3.6417572463768115, "grad_norm": 0.1480243481387624, "learning_rate": 1.6675198589909696e-05, "loss": 0.8708, "step": 8041 }, { "epoch": 3.642210144927536, "grad_norm": 0.12320709635727464, "learning_rate": 1.66649252555399e-05, "loss": 0.8722, "step": 8042 }, { "epoch": 3.6426630434782608, "grad_norm": 0.16344128240094102, "learning_rate": 1.6654654254014494e-05, "loss": 0.8571, "step": 8043 }, { "epoch": 3.6431159420289854, "grad_norm": 0.1525895999301456, "learning_rate": 1.6644385586360304e-05, "loss": 0.8732, "step": 8044 }, { "epoch": 3.64356884057971, "grad_norm": 0.12549208482442367, "learning_rate": 1.6634119253603903e-05, "loss": 0.857, "step": 8045 }, { "epoch": 3.6440217391304346, "grad_norm": 0.12381232510476282, "learning_rate": 1.6623855256771623e-05, "loss": 0.8717, "step": 8046 }, { "epoch": 3.644474637681159, "grad_norm": 0.1134996657524412, "learning_rate": 1.6613593596889566e-05, "loss": 0.8515, "step": 8047 }, { "epoch": 3.644927536231884, "grad_norm": 0.11964706801099984, "learning_rate": 1.6603334274983626e-05, "loss": 0.8576, "step": 8048 }, { "epoch": 3.6453804347826084, "grad_norm": 0.1223492901105095, "learning_rate": 1.659307729207942e-05, "loss": 0.8282, "step": 8049 }, { "epoch": 3.6458333333333335, "grad_norm": 0.12626164140414672, "learning_rate": 1.6582822649202382e-05, "loss": 0.8646, "step": 8050 }, { "epoch": 3.646286231884058, "grad_norm": 0.1222099030951564, "learning_rate": 1.657257034737766e-05, "loss": 0.8639, "step": 8051 }, { "epoch": 3.6467391304347827, "grad_norm": 0.13315350391842434, "learning_rate": 1.6562320387630216e-05, "loss": 0.8692, "step": 8052 }, { "epoch": 3.6471920289855073, "grad_norm": 0.1178102951670237, "learning_rate": 1.6552072770984744e-05, "loss": 0.8662, "step": 8053 }, { "epoch": 3.647644927536232, "grad_norm": 0.1245180874969702, "learning_rate": 1.654182749846571e-05, "loss": 0.8656, "step": 8054 }, { "epoch": 3.6480978260869565, "grad_norm": 0.11615521172301876, "learning_rate": 1.6531584571097344e-05, "loss": 0.8621, "step": 8055 }, { "epoch": 3.648550724637681, "grad_norm": 0.144275738046079, "learning_rate": 1.6521343989903673e-05, "loss": 0.8767, "step": 8056 }, { "epoch": 3.6490036231884058, "grad_norm": 0.12377470817181505, "learning_rate": 1.6511105755908434e-05, "loss": 0.8739, "step": 8057 }, { "epoch": 3.6494565217391304, "grad_norm": 0.13048513991865268, "learning_rate": 1.6500869870135186e-05, "loss": 0.8558, "step": 8058 }, { "epoch": 3.649909420289855, "grad_norm": 0.1301440684089599, "learning_rate": 1.64906363336072e-05, "loss": 0.8605, "step": 8059 }, { "epoch": 3.6503623188405796, "grad_norm": 0.1276106264620708, "learning_rate": 1.6480405147347567e-05, "loss": 0.8483, "step": 8060 }, { "epoch": 3.650815217391304, "grad_norm": 0.12683158867872912, "learning_rate": 1.64701763123791e-05, "loss": 0.8634, "step": 8061 }, { "epoch": 3.651268115942029, "grad_norm": 0.13549564473691125, "learning_rate": 1.6459949829724397e-05, "loss": 0.8637, "step": 8062 }, { "epoch": 3.651721014492754, "grad_norm": 0.13497891506317622, "learning_rate": 1.6449725700405813e-05, "loss": 0.8633, "step": 8063 }, { "epoch": 3.6521739130434785, "grad_norm": 0.14063953714179814, "learning_rate": 1.6439503925445453e-05, "loss": 0.8645, "step": 8064 }, { "epoch": 3.652626811594203, "grad_norm": 0.1525981597625576, "learning_rate": 1.6429284505865232e-05, "loss": 0.8465, "step": 8065 }, { "epoch": 3.6530797101449277, "grad_norm": 0.12064439664287879, "learning_rate": 1.6419067442686777e-05, "loss": 0.8567, "step": 8066 }, { "epoch": 3.6535326086956523, "grad_norm": 0.14985308735838251, "learning_rate": 1.6408852736931532e-05, "loss": 0.8642, "step": 8067 }, { "epoch": 3.653985507246377, "grad_norm": 0.12335214332918926, "learning_rate": 1.639864038962065e-05, "loss": 0.849, "step": 8068 }, { "epoch": 3.6544384057971016, "grad_norm": 0.13683855870714132, "learning_rate": 1.63884304017751e-05, "loss": 0.8501, "step": 8069 }, { "epoch": 3.654891304347826, "grad_norm": 0.1384472632178436, "learning_rate": 1.6378222774415565e-05, "loss": 0.8855, "step": 8070 }, { "epoch": 3.6553442028985508, "grad_norm": 0.14942443066994146, "learning_rate": 1.6368017508562545e-05, "loss": 0.8706, "step": 8071 }, { "epoch": 3.6557971014492754, "grad_norm": 0.13646830659302472, "learning_rate": 1.6357814605236262e-05, "loss": 0.8431, "step": 8072 }, { "epoch": 3.65625, "grad_norm": 0.1224618046502797, "learning_rate": 1.6347614065456715e-05, "loss": 0.8661, "step": 8073 }, { "epoch": 3.6567028985507246, "grad_norm": 0.12528607671789924, "learning_rate": 1.633741589024366e-05, "loss": 0.8504, "step": 8074 }, { "epoch": 3.6571557971014492, "grad_norm": 0.1224873701542631, "learning_rate": 1.6327220080616652e-05, "loss": 0.8624, "step": 8075 }, { "epoch": 3.657608695652174, "grad_norm": 0.12809512663521322, "learning_rate": 1.6317026637594945e-05, "loss": 0.8417, "step": 8076 }, { "epoch": 3.6580615942028984, "grad_norm": 0.1396321963734148, "learning_rate": 1.630683556219763e-05, "loss": 0.8543, "step": 8077 }, { "epoch": 3.658514492753623, "grad_norm": 0.12938835676280358, "learning_rate": 1.6296646855443495e-05, "loss": 0.858, "step": 8078 }, { "epoch": 3.6589673913043477, "grad_norm": 0.12298837361581504, "learning_rate": 1.6286460518351148e-05, "loss": 0.8499, "step": 8079 }, { "epoch": 3.6594202898550723, "grad_norm": 0.11963929340641835, "learning_rate": 1.6276276551938926e-05, "loss": 0.856, "step": 8080 }, { "epoch": 3.659873188405797, "grad_norm": 0.13618821830822153, "learning_rate": 1.6266094957224924e-05, "loss": 0.8717, "step": 8081 }, { "epoch": 3.6603260869565215, "grad_norm": 0.11396668030269065, "learning_rate": 1.6255915735227006e-05, "loss": 0.8476, "step": 8082 }, { "epoch": 3.660778985507246, "grad_norm": 0.12939613009362652, "learning_rate": 1.6245738886962832e-05, "loss": 0.8495, "step": 8083 }, { "epoch": 3.661231884057971, "grad_norm": 0.4647679207447464, "learning_rate": 1.623556441344977e-05, "loss": 0.8615, "step": 8084 }, { "epoch": 3.661684782608696, "grad_norm": 0.1392303128128472, "learning_rate": 1.6225392315705006e-05, "loss": 0.8764, "step": 8085 }, { "epoch": 3.6621376811594204, "grad_norm": 0.14836733900123963, "learning_rate": 1.621522259474543e-05, "loss": 0.8562, "step": 8086 }, { "epoch": 3.662590579710145, "grad_norm": 0.12969937785085103, "learning_rate": 1.6205055251587757e-05, "loss": 0.865, "step": 8087 }, { "epoch": 3.6630434782608696, "grad_norm": 0.14680570548060637, "learning_rate": 1.619489028724842e-05, "loss": 0.8464, "step": 8088 }, { "epoch": 3.6634963768115942, "grad_norm": 0.1508125137889278, "learning_rate": 1.618472770274362e-05, "loss": 0.8662, "step": 8089 }, { "epoch": 3.663949275362319, "grad_norm": 0.1542245010248938, "learning_rate": 1.6174567499089318e-05, "loss": 0.8703, "step": 8090 }, { "epoch": 3.6644021739130435, "grad_norm": 0.16355382820247885, "learning_rate": 1.616440967730127e-05, "loss": 0.8522, "step": 8091 }, { "epoch": 3.664855072463768, "grad_norm": 0.11012843445330214, "learning_rate": 1.615425423839494e-05, "loss": 0.8623, "step": 8092 }, { "epoch": 3.6653079710144927, "grad_norm": 0.16029370383006228, "learning_rate": 1.614410118338562e-05, "loss": 0.8713, "step": 8093 }, { "epoch": 3.6657608695652173, "grad_norm": 0.16752702190577914, "learning_rate": 1.613395051328829e-05, "loss": 0.8827, "step": 8094 }, { "epoch": 3.666213768115942, "grad_norm": 0.15088490089588744, "learning_rate": 1.6123802229117762e-05, "loss": 0.8647, "step": 8095 }, { "epoch": 3.6666666666666665, "grad_norm": 0.16340379006112457, "learning_rate": 1.6113656331888563e-05, "loss": 0.8614, "step": 8096 }, { "epoch": 3.6671195652173916, "grad_norm": 0.1569842837351639, "learning_rate": 1.610351282261499e-05, "loss": 0.854, "step": 8097 }, { "epoch": 3.667572463768116, "grad_norm": 0.12596030463671062, "learning_rate": 1.6093371702311088e-05, "loss": 0.8772, "step": 8098 }, { "epoch": 3.668025362318841, "grad_norm": 0.1548643277843078, "learning_rate": 1.608323297199072e-05, "loss": 0.8569, "step": 8099 }, { "epoch": 3.6684782608695654, "grad_norm": 0.13665376019103473, "learning_rate": 1.6073096632667437e-05, "loss": 0.8434, "step": 8100 }, { "epoch": 3.66893115942029, "grad_norm": 0.13249605025585454, "learning_rate": 1.606296268535461e-05, "loss": 0.8593, "step": 8101 }, { "epoch": 3.6693840579710146, "grad_norm": 0.1606016456272206, "learning_rate": 1.6052831131065323e-05, "loss": 0.8477, "step": 8102 }, { "epoch": 3.6698369565217392, "grad_norm": 0.13243055448299046, "learning_rate": 1.6042701970812468e-05, "loss": 0.8799, "step": 8103 }, { "epoch": 3.670289855072464, "grad_norm": 0.13711815534816651, "learning_rate": 1.6032575205608664e-05, "loss": 0.847, "step": 8104 }, { "epoch": 3.6707427536231885, "grad_norm": 0.16407062346594037, "learning_rate": 1.6022450836466294e-05, "loss": 0.8605, "step": 8105 }, { "epoch": 3.671195652173913, "grad_norm": 0.16569683309058433, "learning_rate": 1.6012328864397494e-05, "loss": 0.8528, "step": 8106 }, { "epoch": 3.6716485507246377, "grad_norm": 0.14522213346033916, "learning_rate": 1.6002209290414203e-05, "loss": 0.8638, "step": 8107 }, { "epoch": 3.6721014492753623, "grad_norm": 0.18705027468173044, "learning_rate": 1.5992092115528067e-05, "loss": 0.8714, "step": 8108 }, { "epoch": 3.672554347826087, "grad_norm": 0.14066039073880637, "learning_rate": 1.5981977340750537e-05, "loss": 0.8727, "step": 8109 }, { "epoch": 3.6730072463768115, "grad_norm": 0.18404275159398076, "learning_rate": 1.5971864967092776e-05, "loss": 0.8613, "step": 8110 }, { "epoch": 3.673460144927536, "grad_norm": 0.13137960746895458, "learning_rate": 1.596175499556577e-05, "loss": 0.8581, "step": 8111 }, { "epoch": 3.6739130434782608, "grad_norm": 0.15318549148666627, "learning_rate": 1.595164742718019e-05, "loss": 0.8525, "step": 8112 }, { "epoch": 3.6743659420289854, "grad_norm": 0.16035999973241383, "learning_rate": 1.5941542262946553e-05, "loss": 0.8492, "step": 8113 }, { "epoch": 3.67481884057971, "grad_norm": 0.11677185198008323, "learning_rate": 1.5931439503875036e-05, "loss": 0.8597, "step": 8114 }, { "epoch": 3.6752717391304346, "grad_norm": 0.14892119865112644, "learning_rate": 1.592133915097566e-05, "loss": 0.8656, "step": 8115 }, { "epoch": 3.675724637681159, "grad_norm": 0.1318392751919621, "learning_rate": 1.5911241205258153e-05, "loss": 0.8707, "step": 8116 }, { "epoch": 3.676177536231884, "grad_norm": 0.1505951873420978, "learning_rate": 1.5901145667732046e-05, "loss": 0.8573, "step": 8117 }, { "epoch": 3.6766304347826084, "grad_norm": 0.12769062296980474, "learning_rate": 1.589105253940659e-05, "loss": 0.8505, "step": 8118 }, { "epoch": 3.6770833333333335, "grad_norm": 0.1574207493937818, "learning_rate": 1.588096182129082e-05, "loss": 0.8541, "step": 8119 }, { "epoch": 3.677536231884058, "grad_norm": 0.11451747857193552, "learning_rate": 1.5870873514393507e-05, "loss": 0.8711, "step": 8120 }, { "epoch": 3.6779891304347827, "grad_norm": 0.14806724414145758, "learning_rate": 1.586078761972322e-05, "loss": 0.8554, "step": 8121 }, { "epoch": 3.6784420289855073, "grad_norm": 0.1291566182353367, "learning_rate": 1.5850704138288244e-05, "loss": 0.8636, "step": 8122 }, { "epoch": 3.678894927536232, "grad_norm": 0.14285950049204524, "learning_rate": 1.584062307109664e-05, "loss": 0.8428, "step": 8123 }, { "epoch": 3.6793478260869565, "grad_norm": 0.1363601090237014, "learning_rate": 1.5830544419156223e-05, "loss": 0.8451, "step": 8124 }, { "epoch": 3.679800724637681, "grad_norm": 0.10705261837645397, "learning_rate": 1.5820468183474588e-05, "loss": 0.8657, "step": 8125 }, { "epoch": 3.6802536231884058, "grad_norm": 0.16105410819002397, "learning_rate": 1.5810394365059058e-05, "loss": 0.8503, "step": 8126 }, { "epoch": 3.6807065217391304, "grad_norm": 0.13954037310937284, "learning_rate": 1.5800322964916747e-05, "loss": 0.8587, "step": 8127 }, { "epoch": 3.681159420289855, "grad_norm": 0.15993129584038715, "learning_rate": 1.579025398405448e-05, "loss": 0.8604, "step": 8128 }, { "epoch": 3.6816123188405796, "grad_norm": 0.1581588627997275, "learning_rate": 1.5780187423478902e-05, "loss": 0.8764, "step": 8129 }, { "epoch": 3.682065217391304, "grad_norm": 0.12858598461113982, "learning_rate": 1.5770123284196365e-05, "loss": 0.8553, "step": 8130 }, { "epoch": 3.682518115942029, "grad_norm": 0.1349071999711203, "learning_rate": 1.5760061567212996e-05, "loss": 0.8708, "step": 8131 }, { "epoch": 3.682971014492754, "grad_norm": 0.13670843739244812, "learning_rate": 1.5750002273534676e-05, "loss": 0.8629, "step": 8132 }, { "epoch": 3.6834239130434785, "grad_norm": 0.12545963075862426, "learning_rate": 1.5739945404167064e-05, "loss": 0.8441, "step": 8133 }, { "epoch": 3.683876811594203, "grad_norm": 0.14038931570758018, "learning_rate": 1.5729890960115545e-05, "loss": 0.8553, "step": 8134 }, { "epoch": 3.6843297101449277, "grad_norm": 0.12943537738713137, "learning_rate": 1.5719838942385298e-05, "loss": 0.8497, "step": 8135 }, { "epoch": 3.6847826086956523, "grad_norm": 0.12399740930714509, "learning_rate": 1.5709789351981214e-05, "loss": 0.8668, "step": 8136 }, { "epoch": 3.685235507246377, "grad_norm": 0.13649415011484103, "learning_rate": 1.569974218990799e-05, "loss": 0.8587, "step": 8137 }, { "epoch": 3.6856884057971016, "grad_norm": 0.14761395401014357, "learning_rate": 1.5689697457170053e-05, "loss": 0.8676, "step": 8138 }, { "epoch": 3.686141304347826, "grad_norm": 0.13813814963953028, "learning_rate": 1.567965515477158e-05, "loss": 0.8442, "step": 8139 }, { "epoch": 3.6865942028985508, "grad_norm": 0.12730302576298097, "learning_rate": 1.5669615283716506e-05, "loss": 0.8634, "step": 8140 }, { "epoch": 3.6870471014492754, "grad_norm": 0.14183750249181395, "learning_rate": 1.565957784500856e-05, "loss": 0.8674, "step": 8141 }, { "epoch": 3.6875, "grad_norm": 0.15007163555378294, "learning_rate": 1.5649542839651175e-05, "loss": 0.8615, "step": 8142 }, { "epoch": 3.6879528985507246, "grad_norm": 0.13583521264534926, "learning_rate": 1.5639510268647594e-05, "loss": 0.8488, "step": 8143 }, { "epoch": 3.6884057971014492, "grad_norm": 0.14217142949169106, "learning_rate": 1.5629480133000753e-05, "loss": 0.8537, "step": 8144 }, { "epoch": 3.688858695652174, "grad_norm": 0.13729565168267993, "learning_rate": 1.5619452433713426e-05, "loss": 0.8922, "step": 8145 }, { "epoch": 3.6893115942028984, "grad_norm": 0.14293112254080811, "learning_rate": 1.5609427171788064e-05, "loss": 0.8578, "step": 8146 }, { "epoch": 3.689764492753623, "grad_norm": 0.12101116143249975, "learning_rate": 1.5599404348226923e-05, "loss": 0.8595, "step": 8147 }, { "epoch": 3.6902173913043477, "grad_norm": 0.14237497469800503, "learning_rate": 1.5589383964031982e-05, "loss": 0.8539, "step": 8148 }, { "epoch": 3.6906702898550723, "grad_norm": 0.1218275395209722, "learning_rate": 1.557936602020502e-05, "loss": 0.85, "step": 8149 }, { "epoch": 3.691123188405797, "grad_norm": 0.17231858531436142, "learning_rate": 1.5569350517747524e-05, "loss": 0.8535, "step": 8150 }, { "epoch": 3.6915760869565215, "grad_norm": 0.13230209684218128, "learning_rate": 1.5559337457660778e-05, "loss": 0.861, "step": 8151 }, { "epoch": 3.692028985507246, "grad_norm": 0.12809758740405677, "learning_rate": 1.5549326840945787e-05, "loss": 0.8507, "step": 8152 }, { "epoch": 3.692481884057971, "grad_norm": 0.1290531294020583, "learning_rate": 1.553931866860335e-05, "loss": 0.858, "step": 8153 }, { "epoch": 3.692934782608696, "grad_norm": 0.12692679681863014, "learning_rate": 1.5529312941633988e-05, "loss": 0.8629, "step": 8154 }, { "epoch": 3.6933876811594204, "grad_norm": 0.13309647754447604, "learning_rate": 1.551930966103799e-05, "loss": 0.8684, "step": 8155 }, { "epoch": 3.693840579710145, "grad_norm": 0.1279301327221806, "learning_rate": 1.550930882781538e-05, "loss": 0.8745, "step": 8156 }, { "epoch": 3.6942934782608696, "grad_norm": 0.12084645584119959, "learning_rate": 1.549931044296599e-05, "loss": 0.855, "step": 8157 }, { "epoch": 3.6947463768115942, "grad_norm": 0.11655421046383202, "learning_rate": 1.5489314507489344e-05, "loss": 0.8713, "step": 8158 }, { "epoch": 3.695199275362319, "grad_norm": 0.10860360645001368, "learning_rate": 1.5479321022384787e-05, "loss": 0.8365, "step": 8159 }, { "epoch": 3.6956521739130435, "grad_norm": 0.11333652692596916, "learning_rate": 1.5469329988651344e-05, "loss": 0.8557, "step": 8160 }, { "epoch": 3.696105072463768, "grad_norm": 0.12128600766040636, "learning_rate": 1.5459341407287873e-05, "loss": 0.8793, "step": 8161 }, { "epoch": 3.6965579710144927, "grad_norm": 0.1165482120662447, "learning_rate": 1.5449355279292914e-05, "loss": 0.8454, "step": 8162 }, { "epoch": 3.6970108695652173, "grad_norm": 0.10268887647261735, "learning_rate": 1.543937160566482e-05, "loss": 0.8617, "step": 8163 }, { "epoch": 3.697463768115942, "grad_norm": 0.12561296000863922, "learning_rate": 1.5429390387401673e-05, "loss": 0.8649, "step": 8164 }, { "epoch": 3.6979166666666665, "grad_norm": 0.1233994817221698, "learning_rate": 1.5419411625501302e-05, "loss": 0.8708, "step": 8165 }, { "epoch": 3.6983695652173916, "grad_norm": 0.11974887744408834, "learning_rate": 1.540943532096129e-05, "loss": 0.8566, "step": 8166 }, { "epoch": 3.698822463768116, "grad_norm": 0.12803494660550008, "learning_rate": 1.5399461474779006e-05, "loss": 0.8778, "step": 8167 }, { "epoch": 3.699275362318841, "grad_norm": 0.13793067689623992, "learning_rate": 1.5389490087951532e-05, "loss": 0.8487, "step": 8168 }, { "epoch": 3.6997282608695654, "grad_norm": 0.12258557867114596, "learning_rate": 1.537952116147574e-05, "loss": 0.8561, "step": 8169 }, { "epoch": 3.70018115942029, "grad_norm": 0.1376453337110925, "learning_rate": 1.5369554696348226e-05, "loss": 0.8553, "step": 8170 }, { "epoch": 3.7006340579710146, "grad_norm": 0.12091253570138098, "learning_rate": 1.5359590693565367e-05, "loss": 0.8669, "step": 8171 }, { "epoch": 3.7010869565217392, "grad_norm": 0.13281204798517227, "learning_rate": 1.5349629154123275e-05, "loss": 0.8517, "step": 8172 }, { "epoch": 3.701539855072464, "grad_norm": 0.12705393635160306, "learning_rate": 1.5339670079017823e-05, "loss": 0.8487, "step": 8173 }, { "epoch": 3.7019927536231885, "grad_norm": 0.11877956512200812, "learning_rate": 1.532971346924461e-05, "loss": 0.8365, "step": 8174 }, { "epoch": 3.702445652173913, "grad_norm": 0.13104740700304973, "learning_rate": 1.531975932579905e-05, "loss": 0.8611, "step": 8175 }, { "epoch": 3.7028985507246377, "grad_norm": 0.11003747914432255, "learning_rate": 1.530980764967625e-05, "loss": 0.8604, "step": 8176 }, { "epoch": 3.7033514492753623, "grad_norm": 0.1310331710521898, "learning_rate": 1.5299858441871113e-05, "loss": 0.8741, "step": 8177 }, { "epoch": 3.703804347826087, "grad_norm": 0.11322332527033166, "learning_rate": 1.5289911703378258e-05, "loss": 0.8545, "step": 8178 }, { "epoch": 3.7042572463768115, "grad_norm": 0.11040078397583042, "learning_rate": 1.5279967435192097e-05, "loss": 0.8694, "step": 8179 }, { "epoch": 3.704710144927536, "grad_norm": 0.12220622390859391, "learning_rate": 1.5270025638306766e-05, "loss": 0.8699, "step": 8180 }, { "epoch": 3.7051630434782608, "grad_norm": 0.11157568117832568, "learning_rate": 1.5260086313716157e-05, "loss": 0.8598, "step": 8181 }, { "epoch": 3.7056159420289854, "grad_norm": 0.12315748258484399, "learning_rate": 1.5250149462413913e-05, "loss": 0.852, "step": 8182 }, { "epoch": 3.70606884057971, "grad_norm": 0.10193097467933024, "learning_rate": 1.5240215085393457e-05, "loss": 0.8445, "step": 8183 }, { "epoch": 3.7065217391304346, "grad_norm": 0.14397716162742974, "learning_rate": 1.5230283183647938e-05, "loss": 0.8539, "step": 8184 }, { "epoch": 3.706974637681159, "grad_norm": 0.09535312814047757, "learning_rate": 1.522035375817024e-05, "loss": 0.8589, "step": 8185 }, { "epoch": 3.707427536231884, "grad_norm": 0.14591645901794462, "learning_rate": 1.5210426809953066e-05, "loss": 0.8687, "step": 8186 }, { "epoch": 3.7078804347826084, "grad_norm": 0.10854279886146401, "learning_rate": 1.5200502339988784e-05, "loss": 0.8727, "step": 8187 }, { "epoch": 3.7083333333333335, "grad_norm": 0.12648379279801264, "learning_rate": 1.5190580349269604e-05, "loss": 0.8691, "step": 8188 }, { "epoch": 3.708786231884058, "grad_norm": 0.10894794242103722, "learning_rate": 1.5180660838787416e-05, "loss": 0.8517, "step": 8189 }, { "epoch": 3.7092391304347827, "grad_norm": 0.11275513011024026, "learning_rate": 1.517074380953389e-05, "loss": 0.8753, "step": 8190 }, { "epoch": 3.7096920289855073, "grad_norm": 0.11552035390557239, "learning_rate": 1.5160829262500438e-05, "loss": 0.8436, "step": 8191 }, { "epoch": 3.710144927536232, "grad_norm": 0.16205588466838775, "learning_rate": 1.5150917198678263e-05, "loss": 0.8571, "step": 8192 }, { "epoch": 3.7105978260869565, "grad_norm": 0.1127356330314698, "learning_rate": 1.5141007619058257e-05, "loss": 0.8573, "step": 8193 }, { "epoch": 3.711050724637681, "grad_norm": 0.13133318525437987, "learning_rate": 1.5131100524631128e-05, "loss": 0.8746, "step": 8194 }, { "epoch": 3.7115036231884058, "grad_norm": 0.1584894667740957, "learning_rate": 1.5121195916387273e-05, "loss": 0.846, "step": 8195 }, { "epoch": 3.7119565217391304, "grad_norm": 0.11223905704539316, "learning_rate": 1.5111293795316896e-05, "loss": 0.8674, "step": 8196 }, { "epoch": 3.712409420289855, "grad_norm": 0.12781156195748994, "learning_rate": 1.5101394162409922e-05, "loss": 0.8521, "step": 8197 }, { "epoch": 3.7128623188405796, "grad_norm": 0.13459622581582095, "learning_rate": 1.5091497018656026e-05, "loss": 0.8793, "step": 8198 }, { "epoch": 3.713315217391304, "grad_norm": 0.12341818958925502, "learning_rate": 1.5081602365044629e-05, "loss": 0.8553, "step": 8199 }, { "epoch": 3.713768115942029, "grad_norm": 0.13641303082187892, "learning_rate": 1.5071710202564949e-05, "loss": 0.8591, "step": 8200 }, { "epoch": 3.714221014492754, "grad_norm": 0.12198275720548057, "learning_rate": 1.5061820532205885e-05, "loss": 0.8558, "step": 8201 }, { "epoch": 3.7146739130434785, "grad_norm": 0.14347041128518664, "learning_rate": 1.5051933354956152e-05, "loss": 0.8526, "step": 8202 }, { "epoch": 3.715126811594203, "grad_norm": 0.15309613172929792, "learning_rate": 1.5042048671804166e-05, "loss": 0.8785, "step": 8203 }, { "epoch": 3.7155797101449277, "grad_norm": 0.13542930953484048, "learning_rate": 1.5032166483738136e-05, "loss": 0.8639, "step": 8204 }, { "epoch": 3.7160326086956523, "grad_norm": 0.13127379323556715, "learning_rate": 1.5022286791745972e-05, "loss": 0.846, "step": 8205 }, { "epoch": 3.716485507246377, "grad_norm": 0.13892837839482652, "learning_rate": 1.5012409596815393e-05, "loss": 0.8597, "step": 8206 }, { "epoch": 3.7169384057971016, "grad_norm": 0.11064315551755456, "learning_rate": 1.5002534899933823e-05, "loss": 0.876, "step": 8207 }, { "epoch": 3.717391304347826, "grad_norm": 0.11580745548167949, "learning_rate": 1.4992662702088451e-05, "loss": 0.8684, "step": 8208 }, { "epoch": 3.7178442028985508, "grad_norm": 0.12409402098265014, "learning_rate": 1.4982793004266203e-05, "loss": 0.8792, "step": 8209 }, { "epoch": 3.7182971014492754, "grad_norm": 0.10731408695292176, "learning_rate": 1.4972925807453793e-05, "loss": 0.8699, "step": 8210 }, { "epoch": 3.71875, "grad_norm": 0.1240834791089001, "learning_rate": 1.4963061112637637e-05, "loss": 0.8651, "step": 8211 }, { "epoch": 3.7192028985507246, "grad_norm": 0.10827859139961228, "learning_rate": 1.495319892080395e-05, "loss": 0.8543, "step": 8212 }, { "epoch": 3.7196557971014492, "grad_norm": 0.1103039191430379, "learning_rate": 1.4943339232938643e-05, "loss": 0.8807, "step": 8213 }, { "epoch": 3.720108695652174, "grad_norm": 0.11691369150382959, "learning_rate": 1.493348205002743e-05, "loss": 0.8624, "step": 8214 }, { "epoch": 3.7205615942028984, "grad_norm": 0.11785301441017106, "learning_rate": 1.492362737305574e-05, "loss": 0.8541, "step": 8215 }, { "epoch": 3.721014492753623, "grad_norm": 0.09918383045859326, "learning_rate": 1.4913775203008759e-05, "loss": 0.8685, "step": 8216 }, { "epoch": 3.7214673913043477, "grad_norm": 0.12514250621369855, "learning_rate": 1.4903925540871406e-05, "loss": 0.8612, "step": 8217 }, { "epoch": 3.7219202898550723, "grad_norm": 0.12078811034796413, "learning_rate": 1.4894078387628406e-05, "loss": 0.8609, "step": 8218 }, { "epoch": 3.722373188405797, "grad_norm": 0.18896351315326002, "learning_rate": 1.4884233744264153e-05, "loss": 0.8648, "step": 8219 }, { "epoch": 3.7228260869565215, "grad_norm": 0.11310527256489006, "learning_rate": 1.4874391611762868e-05, "loss": 0.849, "step": 8220 }, { "epoch": 3.723278985507246, "grad_norm": 0.1495265794281604, "learning_rate": 1.4864551991108455e-05, "loss": 0.8629, "step": 8221 }, { "epoch": 3.723731884057971, "grad_norm": 0.12482402294309179, "learning_rate": 1.4854714883284623e-05, "loss": 0.868, "step": 8222 }, { "epoch": 3.724184782608696, "grad_norm": 0.12226362897481939, "learning_rate": 1.484488028927479e-05, "loss": 0.8729, "step": 8223 }, { "epoch": 3.7246376811594204, "grad_norm": 0.13559973193202013, "learning_rate": 1.4835048210062137e-05, "loss": 0.8656, "step": 8224 }, { "epoch": 3.725090579710145, "grad_norm": 0.11692641926861236, "learning_rate": 1.4825218646629576e-05, "loss": 0.8566, "step": 8225 }, { "epoch": 3.7255434782608696, "grad_norm": 0.13564327181699762, "learning_rate": 1.4815391599959816e-05, "loss": 0.8974, "step": 8226 }, { "epoch": 3.7259963768115942, "grad_norm": 0.13868106475879138, "learning_rate": 1.4805567071035252e-05, "loss": 0.8651, "step": 8227 }, { "epoch": 3.726449275362319, "grad_norm": 0.1266112440468834, "learning_rate": 1.4795745060838083e-05, "loss": 0.8585, "step": 8228 }, { "epoch": 3.7269021739130435, "grad_norm": 0.12326850552807617, "learning_rate": 1.4785925570350208e-05, "loss": 0.8455, "step": 8229 }, { "epoch": 3.727355072463768, "grad_norm": 0.11694579674244128, "learning_rate": 1.4776108600553323e-05, "loss": 0.8593, "step": 8230 }, { "epoch": 3.7278079710144927, "grad_norm": 0.12105792577003043, "learning_rate": 1.4766294152428832e-05, "loss": 0.8449, "step": 8231 }, { "epoch": 3.7282608695652173, "grad_norm": 0.1299496213447257, "learning_rate": 1.47564822269579e-05, "loss": 0.8648, "step": 8232 }, { "epoch": 3.728713768115942, "grad_norm": 0.12596780649070544, "learning_rate": 1.4746672825121429e-05, "loss": 0.8586, "step": 8233 }, { "epoch": 3.7291666666666665, "grad_norm": 0.1367200935487819, "learning_rate": 1.4736865947900106e-05, "loss": 0.8554, "step": 8234 }, { "epoch": 3.7296195652173916, "grad_norm": 0.13672543384069005, "learning_rate": 1.4727061596274315e-05, "loss": 0.8716, "step": 8235 }, { "epoch": 3.730072463768116, "grad_norm": 0.11087010681632795, "learning_rate": 1.4717259771224237e-05, "loss": 0.861, "step": 8236 }, { "epoch": 3.730525362318841, "grad_norm": 0.13293689063202518, "learning_rate": 1.4707460473729747e-05, "loss": 0.8467, "step": 8237 }, { "epoch": 3.7309782608695654, "grad_norm": 0.11804625834474057, "learning_rate": 1.469766370477053e-05, "loss": 0.8301, "step": 8238 }, { "epoch": 3.73143115942029, "grad_norm": 0.13113669432585837, "learning_rate": 1.4687869465325966e-05, "loss": 0.8546, "step": 8239 }, { "epoch": 3.7318840579710146, "grad_norm": 0.11263959443241492, "learning_rate": 1.4678077756375207e-05, "loss": 0.8609, "step": 8240 }, { "epoch": 3.7323369565217392, "grad_norm": 0.12697414364078385, "learning_rate": 1.466828857889712e-05, "loss": 0.8581, "step": 8241 }, { "epoch": 3.732789855072464, "grad_norm": 0.1241608647019097, "learning_rate": 1.4658501933870382e-05, "loss": 0.8396, "step": 8242 }, { "epoch": 3.7332427536231885, "grad_norm": 0.12211361103998869, "learning_rate": 1.464871782227335e-05, "loss": 0.8872, "step": 8243 }, { "epoch": 3.733695652173913, "grad_norm": 0.11184481295265611, "learning_rate": 1.463893624508418e-05, "loss": 0.8378, "step": 8244 }, { "epoch": 3.7341485507246377, "grad_norm": 0.11928737791720397, "learning_rate": 1.462915720328073e-05, "loss": 0.8718, "step": 8245 }, { "epoch": 3.7346014492753623, "grad_norm": 0.11100438578599037, "learning_rate": 1.4619380697840653e-05, "loss": 0.87, "step": 8246 }, { "epoch": 3.735054347826087, "grad_norm": 0.13381651575471779, "learning_rate": 1.4609606729741291e-05, "loss": 0.8504, "step": 8247 }, { "epoch": 3.7355072463768115, "grad_norm": 0.12115815954755894, "learning_rate": 1.4599835299959794e-05, "loss": 0.8404, "step": 8248 }, { "epoch": 3.735960144927536, "grad_norm": 0.13382739719653652, "learning_rate": 1.4590066409473011e-05, "loss": 0.8587, "step": 8249 }, { "epoch": 3.7364130434782608, "grad_norm": 0.12658026770445685, "learning_rate": 1.4580300059257554e-05, "loss": 0.8617, "step": 8250 }, { "epoch": 3.7368659420289854, "grad_norm": 0.14314313650880095, "learning_rate": 1.4570536250289768e-05, "loss": 0.8573, "step": 8251 }, { "epoch": 3.73731884057971, "grad_norm": 0.13502675438868367, "learning_rate": 1.4560774983545787e-05, "loss": 0.865, "step": 8252 }, { "epoch": 3.7377717391304346, "grad_norm": 0.13868822426518385, "learning_rate": 1.4551016260001424e-05, "loss": 0.8643, "step": 8253 }, { "epoch": 3.738224637681159, "grad_norm": 0.10889906604103035, "learning_rate": 1.4541260080632307e-05, "loss": 0.8561, "step": 8254 }, { "epoch": 3.738677536231884, "grad_norm": 0.1290479987140222, "learning_rate": 1.453150644641375e-05, "loss": 0.8566, "step": 8255 }, { "epoch": 3.7391304347826084, "grad_norm": 0.10910066879494172, "learning_rate": 1.4521755358320868e-05, "loss": 0.8393, "step": 8256 }, { "epoch": 3.7395833333333335, "grad_norm": 0.11943510863108589, "learning_rate": 1.4512006817328472e-05, "loss": 0.8587, "step": 8257 }, { "epoch": 3.740036231884058, "grad_norm": 0.12751668137721558, "learning_rate": 1.4502260824411147e-05, "loss": 0.8764, "step": 8258 }, { "epoch": 3.7404891304347827, "grad_norm": 0.10813693848620742, "learning_rate": 1.4492517380543198e-05, "loss": 0.8545, "step": 8259 }, { "epoch": 3.7409420289855073, "grad_norm": 0.14162278430670647, "learning_rate": 1.4482776486698717e-05, "loss": 0.8476, "step": 8260 }, { "epoch": 3.741394927536232, "grad_norm": 0.15158321152231216, "learning_rate": 1.4473038143851494e-05, "loss": 0.8519, "step": 8261 }, { "epoch": 3.7418478260869565, "grad_norm": 0.12376179586085272, "learning_rate": 1.4463302352975115e-05, "loss": 0.8604, "step": 8262 }, { "epoch": 3.742300724637681, "grad_norm": 0.14579275224836694, "learning_rate": 1.445356911504285e-05, "loss": 0.8954, "step": 8263 }, { "epoch": 3.7427536231884058, "grad_norm": 0.16704727390101676, "learning_rate": 1.4443838431027777e-05, "loss": 0.8591, "step": 8264 }, { "epoch": 3.7432065217391304, "grad_norm": 0.1308251395118993, "learning_rate": 1.4434110301902675e-05, "loss": 0.8658, "step": 8265 }, { "epoch": 3.743659420289855, "grad_norm": 0.1556505291922239, "learning_rate": 1.4424384728640077e-05, "loss": 0.8747, "step": 8266 }, { "epoch": 3.7441123188405796, "grad_norm": 0.1735934990795526, "learning_rate": 1.441466171221225e-05, "loss": 0.8501, "step": 8267 }, { "epoch": 3.744565217391304, "grad_norm": 0.17947405006110623, "learning_rate": 1.4404941253591251e-05, "loss": 0.858, "step": 8268 }, { "epoch": 3.745018115942029, "grad_norm": 0.1412163471748644, "learning_rate": 1.439522335374882e-05, "loss": 0.8593, "step": 8269 }, { "epoch": 3.745471014492754, "grad_norm": 0.15701898247771726, "learning_rate": 1.4385508013656497e-05, "loss": 0.8831, "step": 8270 }, { "epoch": 3.7459239130434785, "grad_norm": 0.1249261733685694, "learning_rate": 1.4375795234285512e-05, "loss": 0.8744, "step": 8271 }, { "epoch": 3.746376811594203, "grad_norm": 0.14226346269898477, "learning_rate": 1.4366085016606901e-05, "loss": 0.8768, "step": 8272 }, { "epoch": 3.7468297101449277, "grad_norm": 0.12178631092619838, "learning_rate": 1.435637736159139e-05, "loss": 0.8532, "step": 8273 }, { "epoch": 3.7472826086956523, "grad_norm": 0.15521219957076526, "learning_rate": 1.4346672270209468e-05, "loss": 0.8893, "step": 8274 }, { "epoch": 3.747735507246377, "grad_norm": 0.1350201327374929, "learning_rate": 1.4336969743431358e-05, "loss": 0.8668, "step": 8275 }, { "epoch": 3.7481884057971016, "grad_norm": 0.14340803100340793, "learning_rate": 1.4327269782227062e-05, "loss": 0.8793, "step": 8276 }, { "epoch": 3.748641304347826, "grad_norm": 0.15567992445684242, "learning_rate": 1.4317572387566276e-05, "loss": 0.8545, "step": 8277 }, { "epoch": 3.7490942028985508, "grad_norm": 0.1438613459003997, "learning_rate": 1.4307877560418488e-05, "loss": 0.8523, "step": 8278 }, { "epoch": 3.7495471014492754, "grad_norm": 0.11731724802320227, "learning_rate": 1.4298185301752879e-05, "loss": 0.8361, "step": 8279 }, { "epoch": 3.75, "grad_norm": 0.14855748608512728, "learning_rate": 1.4288495612538427e-05, "loss": 0.8615, "step": 8280 }, { "epoch": 3.7504528985507246, "grad_norm": 0.11003742678237502, "learning_rate": 1.4278808493743812e-05, "loss": 0.8621, "step": 8281 }, { "epoch": 3.7509057971014492, "grad_norm": 0.1195278136130136, "learning_rate": 1.4269123946337474e-05, "loss": 0.8558, "step": 8282 }, { "epoch": 3.751358695652174, "grad_norm": 0.1273290879417686, "learning_rate": 1.4259441971287573e-05, "loss": 0.8464, "step": 8283 }, { "epoch": 3.7518115942028984, "grad_norm": 0.11700539404950358, "learning_rate": 1.4249762569562058e-05, "loss": 0.8651, "step": 8284 }, { "epoch": 3.752264492753623, "grad_norm": 0.11797162481206873, "learning_rate": 1.4240085742128576e-05, "loss": 0.8828, "step": 8285 }, { "epoch": 3.7527173913043477, "grad_norm": 0.14222064038461887, "learning_rate": 1.4230411489954552e-05, "loss": 0.8507, "step": 8286 }, { "epoch": 3.7531702898550723, "grad_norm": 0.15351243154107938, "learning_rate": 1.4220739814007112e-05, "loss": 0.8454, "step": 8287 }, { "epoch": 3.753623188405797, "grad_norm": 0.12987732528226406, "learning_rate": 1.4211070715253179e-05, "loss": 0.8448, "step": 8288 }, { "epoch": 3.7540760869565215, "grad_norm": 0.15668792043575355, "learning_rate": 1.4201404194659359e-05, "loss": 0.859, "step": 8289 }, { "epoch": 3.754528985507246, "grad_norm": 0.15282928198335202, "learning_rate": 1.4191740253192073e-05, "loss": 0.8607, "step": 8290 }, { "epoch": 3.754981884057971, "grad_norm": 0.1591705787905065, "learning_rate": 1.4182078891817379e-05, "loss": 0.8572, "step": 8291 }, { "epoch": 3.755434782608696, "grad_norm": 0.1631692329377474, "learning_rate": 1.4172420111501186e-05, "loss": 0.8591, "step": 8292 }, { "epoch": 3.7558876811594204, "grad_norm": 0.15321644643511537, "learning_rate": 1.4162763913209068e-05, "loss": 0.8445, "step": 8293 }, { "epoch": 3.756340579710145, "grad_norm": 0.1586428728111053, "learning_rate": 1.4153110297906398e-05, "loss": 0.8674, "step": 8294 }, { "epoch": 3.7567934782608696, "grad_norm": 0.14060431762014705, "learning_rate": 1.4143459266558237e-05, "loss": 0.8631, "step": 8295 }, { "epoch": 3.7572463768115942, "grad_norm": 0.12201945418460779, "learning_rate": 1.4133810820129438e-05, "loss": 0.8523, "step": 8296 }, { "epoch": 3.757699275362319, "grad_norm": 0.1248132193618527, "learning_rate": 1.4124164959584548e-05, "loss": 0.8627, "step": 8297 }, { "epoch": 3.7581521739130435, "grad_norm": 0.12508421902736594, "learning_rate": 1.4114521685887898e-05, "loss": 0.848, "step": 8298 }, { "epoch": 3.758605072463768, "grad_norm": 0.12065458418692142, "learning_rate": 1.4104881000003543e-05, "loss": 0.8708, "step": 8299 }, { "epoch": 3.7590579710144927, "grad_norm": 0.12264044242118077, "learning_rate": 1.4095242902895265e-05, "loss": 0.867, "step": 8300 }, { "epoch": 3.7595108695652173, "grad_norm": 0.13241196820958256, "learning_rate": 1.4085607395526589e-05, "loss": 0.8726, "step": 8301 }, { "epoch": 3.759963768115942, "grad_norm": 0.11339491315091609, "learning_rate": 1.4075974478860821e-05, "loss": 0.8626, "step": 8302 }, { "epoch": 3.7604166666666665, "grad_norm": 0.14207761121923768, "learning_rate": 1.406634415386095e-05, "loss": 0.8565, "step": 8303 }, { "epoch": 3.7608695652173916, "grad_norm": 0.11307140360855038, "learning_rate": 1.4056716421489762e-05, "loss": 0.835, "step": 8304 }, { "epoch": 3.761322463768116, "grad_norm": 0.11946783456679519, "learning_rate": 1.4047091282709735e-05, "loss": 0.8784, "step": 8305 }, { "epoch": 3.761775362318841, "grad_norm": 0.12423065292506567, "learning_rate": 1.4037468738483129e-05, "loss": 0.8621, "step": 8306 }, { "epoch": 3.7622282608695654, "grad_norm": 0.11623818781541466, "learning_rate": 1.4027848789771917e-05, "loss": 0.8464, "step": 8307 }, { "epoch": 3.76268115942029, "grad_norm": 0.13101245221920443, "learning_rate": 1.4018231437537812e-05, "loss": 0.8668, "step": 8308 }, { "epoch": 3.7631340579710146, "grad_norm": 0.14119720515197498, "learning_rate": 1.4008616682742285e-05, "loss": 0.8778, "step": 8309 }, { "epoch": 3.7635869565217392, "grad_norm": 0.1250536970046679, "learning_rate": 1.3999004526346523e-05, "loss": 0.8683, "step": 8310 }, { "epoch": 3.764039855072464, "grad_norm": 0.1726865800969299, "learning_rate": 1.3989394969311491e-05, "loss": 0.8568, "step": 8311 }, { "epoch": 3.7644927536231885, "grad_norm": 0.10874542216395712, "learning_rate": 1.397978801259785e-05, "loss": 0.8603, "step": 8312 }, { "epoch": 3.764945652173913, "grad_norm": 0.1476989080507943, "learning_rate": 1.3970183657166047e-05, "loss": 0.8574, "step": 8313 }, { "epoch": 3.7653985507246377, "grad_norm": 0.148249451165051, "learning_rate": 1.3960581903976217e-05, "loss": 0.848, "step": 8314 }, { "epoch": 3.7658514492753623, "grad_norm": 0.12070665390400319, "learning_rate": 1.3950982753988296e-05, "loss": 0.8605, "step": 8315 }, { "epoch": 3.766304347826087, "grad_norm": 0.14980176213348664, "learning_rate": 1.39413862081619e-05, "loss": 0.8449, "step": 8316 }, { "epoch": 3.7667572463768115, "grad_norm": 0.1458145245207288, "learning_rate": 1.3931792267456424e-05, "loss": 0.857, "step": 8317 }, { "epoch": 3.767210144927536, "grad_norm": 0.11536062527210414, "learning_rate": 1.3922200932830965e-05, "loss": 0.8511, "step": 8318 }, { "epoch": 3.7676630434782608, "grad_norm": 0.10438678925200923, "learning_rate": 1.391261220524442e-05, "loss": 0.8747, "step": 8319 }, { "epoch": 3.7681159420289854, "grad_norm": 0.12824006589673131, "learning_rate": 1.3903026085655356e-05, "loss": 0.8687, "step": 8320 }, { "epoch": 3.76856884057971, "grad_norm": 0.11958216054229581, "learning_rate": 1.3893442575022143e-05, "loss": 0.8597, "step": 8321 }, { "epoch": 3.7690217391304346, "grad_norm": 0.12980582457600096, "learning_rate": 1.3883861674302832e-05, "loss": 0.8677, "step": 8322 }, { "epoch": 3.769474637681159, "grad_norm": 0.11888087623185153, "learning_rate": 1.3874283384455267e-05, "loss": 0.8603, "step": 8323 }, { "epoch": 3.769927536231884, "grad_norm": 0.1315117727818703, "learning_rate": 1.3864707706436993e-05, "loss": 0.8374, "step": 8324 }, { "epoch": 3.7703804347826084, "grad_norm": 0.126365813392794, "learning_rate": 1.3855134641205302e-05, "loss": 0.8797, "step": 8325 }, { "epoch": 3.7708333333333335, "grad_norm": 0.11887044981062912, "learning_rate": 1.3845564189717218e-05, "loss": 0.8555, "step": 8326 }, { "epoch": 3.771286231884058, "grad_norm": 0.11833456805991485, "learning_rate": 1.3835996352929542e-05, "loss": 0.8716, "step": 8327 }, { "epoch": 3.7717391304347827, "grad_norm": 0.12092842105470203, "learning_rate": 1.3826431131798757e-05, "loss": 0.8595, "step": 8328 }, { "epoch": 3.7721920289855073, "grad_norm": 0.11355644820379476, "learning_rate": 1.3816868527281143e-05, "loss": 0.8495, "step": 8329 }, { "epoch": 3.772644927536232, "grad_norm": 0.11914921845228812, "learning_rate": 1.3807308540332659e-05, "loss": 0.8494, "step": 8330 }, { "epoch": 3.7730978260869565, "grad_norm": 0.124222553844004, "learning_rate": 1.3797751171909055e-05, "loss": 0.8514, "step": 8331 }, { "epoch": 3.773550724637681, "grad_norm": 0.11133678727643617, "learning_rate": 1.3788196422965792e-05, "loss": 0.8762, "step": 8332 }, { "epoch": 3.7740036231884058, "grad_norm": 0.1303478798042645, "learning_rate": 1.3778644294458064e-05, "loss": 0.8713, "step": 8333 }, { "epoch": 3.7744565217391304, "grad_norm": 0.11623236242270515, "learning_rate": 1.376909478734081e-05, "loss": 0.8607, "step": 8334 }, { "epoch": 3.774909420289855, "grad_norm": 0.14693263079908292, "learning_rate": 1.3759547902568726e-05, "loss": 0.8452, "step": 8335 }, { "epoch": 3.7753623188405796, "grad_norm": 0.12421049776794829, "learning_rate": 1.3750003641096207e-05, "loss": 0.8537, "step": 8336 }, { "epoch": 3.775815217391304, "grad_norm": 0.1395711736378473, "learning_rate": 1.3740462003877433e-05, "loss": 0.8478, "step": 8337 }, { "epoch": 3.776268115942029, "grad_norm": 0.11807687255747956, "learning_rate": 1.3730922991866273e-05, "loss": 0.8467, "step": 8338 }, { "epoch": 3.776721014492754, "grad_norm": 0.13190340632518194, "learning_rate": 1.372138660601638e-05, "loss": 0.8614, "step": 8339 }, { "epoch": 3.7771739130434785, "grad_norm": 0.1403096655444127, "learning_rate": 1.3711852847281098e-05, "loss": 0.8497, "step": 8340 }, { "epoch": 3.777626811594203, "grad_norm": 0.1232350626170014, "learning_rate": 1.3702321716613556e-05, "loss": 0.8499, "step": 8341 }, { "epoch": 3.7780797101449277, "grad_norm": 0.12758015191695551, "learning_rate": 1.3692793214966584e-05, "loss": 0.8603, "step": 8342 }, { "epoch": 3.7785326086956523, "grad_norm": 0.13581475253613265, "learning_rate": 1.3683267343292763e-05, "loss": 0.8685, "step": 8343 }, { "epoch": 3.778985507246377, "grad_norm": 0.10308315626444986, "learning_rate": 1.3673744102544397e-05, "loss": 0.8358, "step": 8344 }, { "epoch": 3.7794384057971016, "grad_norm": 0.12198098974374903, "learning_rate": 1.3664223493673561e-05, "loss": 0.8619, "step": 8345 }, { "epoch": 3.779891304347826, "grad_norm": 0.12203148979698132, "learning_rate": 1.365470551763202e-05, "loss": 0.8582, "step": 8346 }, { "epoch": 3.7803442028985508, "grad_norm": 0.13127931432916598, "learning_rate": 1.3645190175371331e-05, "loss": 0.844, "step": 8347 }, { "epoch": 3.7807971014492754, "grad_norm": 0.13771984305196733, "learning_rate": 1.3635677467842734e-05, "loss": 0.8549, "step": 8348 }, { "epoch": 3.78125, "grad_norm": 0.16113447334532094, "learning_rate": 1.3626167395997247e-05, "loss": 0.8595, "step": 8349 }, { "epoch": 3.7817028985507246, "grad_norm": 0.12648162011509181, "learning_rate": 1.36166599607856e-05, "loss": 0.8574, "step": 8350 }, { "epoch": 3.7821557971014492, "grad_norm": 0.15158438462208515, "learning_rate": 1.360715516315826e-05, "loss": 0.8665, "step": 8351 }, { "epoch": 3.782608695652174, "grad_norm": 0.10775976260318659, "learning_rate": 1.3597653004065436e-05, "loss": 0.8445, "step": 8352 }, { "epoch": 3.7830615942028984, "grad_norm": 0.13773315572934725, "learning_rate": 1.3588153484457087e-05, "loss": 0.8333, "step": 8353 }, { "epoch": 3.783514492753623, "grad_norm": 0.11019211592643267, "learning_rate": 1.3578656605282875e-05, "loss": 0.8555, "step": 8354 }, { "epoch": 3.7839673913043477, "grad_norm": 0.14553896564337074, "learning_rate": 1.3569162367492247e-05, "loss": 0.8616, "step": 8355 }, { "epoch": 3.7844202898550723, "grad_norm": 0.14108567437666025, "learning_rate": 1.3559670772034324e-05, "loss": 0.8433, "step": 8356 }, { "epoch": 3.784873188405797, "grad_norm": 0.1282697154376824, "learning_rate": 1.3550181819858024e-05, "loss": 0.8689, "step": 8357 }, { "epoch": 3.7853260869565215, "grad_norm": 0.12921834432933102, "learning_rate": 1.3540695511911963e-05, "loss": 0.8619, "step": 8358 }, { "epoch": 3.785778985507246, "grad_norm": 0.15820263886243965, "learning_rate": 1.3531211849144499e-05, "loss": 0.8877, "step": 8359 }, { "epoch": 3.786231884057971, "grad_norm": 0.12208536980481267, "learning_rate": 1.3521730832503722e-05, "loss": 0.8631, "step": 8360 }, { "epoch": 3.786684782608696, "grad_norm": 0.15832898193639502, "learning_rate": 1.3512252462937477e-05, "loss": 0.8707, "step": 8361 }, { "epoch": 3.7871376811594204, "grad_norm": 0.11339891760249046, "learning_rate": 1.3502776741393317e-05, "loss": 0.8806, "step": 8362 }, { "epoch": 3.787590579710145, "grad_norm": 0.1639383341823003, "learning_rate": 1.3493303668818567e-05, "loss": 0.8607, "step": 8363 }, { "epoch": 3.7880434782608696, "grad_norm": 0.1061178514938774, "learning_rate": 1.3483833246160241e-05, "loss": 0.8558, "step": 8364 }, { "epoch": 3.7884963768115942, "grad_norm": 0.1515860171793517, "learning_rate": 1.3474365474365136e-05, "loss": 0.8396, "step": 8365 }, { "epoch": 3.788949275362319, "grad_norm": 0.11559919762848268, "learning_rate": 1.3464900354379751e-05, "loss": 0.8595, "step": 8366 }, { "epoch": 3.7894021739130435, "grad_norm": 0.14506812758027804, "learning_rate": 1.3455437887150323e-05, "loss": 0.8633, "step": 8367 }, { "epoch": 3.789855072463768, "grad_norm": 0.12182637088605776, "learning_rate": 1.344597807362282e-05, "loss": 0.8639, "step": 8368 }, { "epoch": 3.7903079710144927, "grad_norm": 0.12583678440213328, "learning_rate": 1.3436520914742981e-05, "loss": 0.8526, "step": 8369 }, { "epoch": 3.7907608695652173, "grad_norm": 0.10720387848971305, "learning_rate": 1.3427066411456226e-05, "loss": 0.8728, "step": 8370 }, { "epoch": 3.791213768115942, "grad_norm": 0.1274343998312889, "learning_rate": 1.3417614564707764e-05, "loss": 0.8643, "step": 8371 }, { "epoch": 3.7916666666666665, "grad_norm": 0.11219591338013869, "learning_rate": 1.3408165375442486e-05, "loss": 0.8794, "step": 8372 }, { "epoch": 3.7921195652173916, "grad_norm": 0.13422103765784602, "learning_rate": 1.3398718844605063e-05, "loss": 0.8587, "step": 8373 }, { "epoch": 3.792572463768116, "grad_norm": 0.10909522143270009, "learning_rate": 1.3389274973139874e-05, "loss": 0.8586, "step": 8374 }, { "epoch": 3.793025362318841, "grad_norm": 0.11595150533623454, "learning_rate": 1.337983376199103e-05, "loss": 0.857, "step": 8375 }, { "epoch": 3.7934782608695654, "grad_norm": 0.10467605538400739, "learning_rate": 1.337039521210238e-05, "loss": 0.8802, "step": 8376 }, { "epoch": 3.79393115942029, "grad_norm": 0.14508480036428578, "learning_rate": 1.3360959324417526e-05, "loss": 0.8421, "step": 8377 }, { "epoch": 3.7943840579710146, "grad_norm": 0.11368753864041443, "learning_rate": 1.3351526099879774e-05, "loss": 0.8682, "step": 8378 }, { "epoch": 3.7948369565217392, "grad_norm": 0.1443499156975603, "learning_rate": 1.3342095539432197e-05, "loss": 0.8565, "step": 8379 }, { "epoch": 3.795289855072464, "grad_norm": 0.11816013219836274, "learning_rate": 1.333266764401756e-05, "loss": 0.849, "step": 8380 }, { "epoch": 3.7957427536231885, "grad_norm": 0.18303507566106592, "learning_rate": 1.3323242414578407e-05, "loss": 0.8556, "step": 8381 }, { "epoch": 3.796195652173913, "grad_norm": 0.11314416269323307, "learning_rate": 1.331381985205697e-05, "loss": 0.8473, "step": 8382 }, { "epoch": 3.7966485507246377, "grad_norm": 0.1598186195560102, "learning_rate": 1.3304399957395267e-05, "loss": 0.87, "step": 8383 }, { "epoch": 3.7971014492753623, "grad_norm": 0.13514025046410952, "learning_rate": 1.3294982731535e-05, "loss": 0.8635, "step": 8384 }, { "epoch": 3.797554347826087, "grad_norm": 0.1352705751528191, "learning_rate": 1.3285568175417628e-05, "loss": 0.8442, "step": 8385 }, { "epoch": 3.7980072463768115, "grad_norm": 0.12225323699202396, "learning_rate": 1.3276156289984323e-05, "loss": 0.8676, "step": 8386 }, { "epoch": 3.798460144927536, "grad_norm": 0.12463520614043982, "learning_rate": 1.326674707617604e-05, "loss": 0.8647, "step": 8387 }, { "epoch": 3.7989130434782608, "grad_norm": 0.10739410734472121, "learning_rate": 1.3257340534933394e-05, "loss": 0.8409, "step": 8388 }, { "epoch": 3.7993659420289854, "grad_norm": 0.12580744632634536, "learning_rate": 1.324793666719681e-05, "loss": 0.8642, "step": 8389 }, { "epoch": 3.79981884057971, "grad_norm": 0.11834284458280928, "learning_rate": 1.3238535473906375e-05, "loss": 0.8558, "step": 8390 }, { "epoch": 3.8002717391304346, "grad_norm": 0.12318601150008364, "learning_rate": 1.322913695600197e-05, "loss": 0.8591, "step": 8391 }, { "epoch": 3.800724637681159, "grad_norm": 0.11975128484427419, "learning_rate": 1.3219741114423164e-05, "loss": 0.8508, "step": 8392 }, { "epoch": 3.801177536231884, "grad_norm": 0.12370840154160379, "learning_rate": 1.3210347950109274e-05, "loss": 0.8718, "step": 8393 }, { "epoch": 3.8016304347826084, "grad_norm": 0.11872379797559685, "learning_rate": 1.3200957463999343e-05, "loss": 0.8586, "step": 8394 }, { "epoch": 3.8020833333333335, "grad_norm": 0.13391906880086887, "learning_rate": 1.319156965703217e-05, "loss": 0.877, "step": 8395 }, { "epoch": 3.802536231884058, "grad_norm": 0.12273772690075796, "learning_rate": 1.318218453014625e-05, "loss": 0.8534, "step": 8396 }, { "epoch": 3.8029891304347827, "grad_norm": 0.1274114603073028, "learning_rate": 1.317280208427985e-05, "loss": 0.8322, "step": 8397 }, { "epoch": 3.8034420289855073, "grad_norm": 0.11750670251235695, "learning_rate": 1.3163422320370924e-05, "loss": 0.8735, "step": 8398 }, { "epoch": 3.803894927536232, "grad_norm": 0.1396496977923721, "learning_rate": 1.315404523935721e-05, "loss": 0.8658, "step": 8399 }, { "epoch": 3.8043478260869565, "grad_norm": 0.10615238495181005, "learning_rate": 1.3144670842176134e-05, "loss": 0.8345, "step": 8400 }, { "epoch": 3.804800724637681, "grad_norm": 0.1421224236331934, "learning_rate": 1.313529912976487e-05, "loss": 0.8563, "step": 8401 }, { "epoch": 3.8052536231884058, "grad_norm": 0.12162208176661009, "learning_rate": 1.3125930103060304e-05, "loss": 0.8542, "step": 8402 }, { "epoch": 3.8057065217391304, "grad_norm": 0.1407126474225629, "learning_rate": 1.3116563762999111e-05, "loss": 0.8425, "step": 8403 }, { "epoch": 3.806159420289855, "grad_norm": 0.13800239163234665, "learning_rate": 1.3107200110517622e-05, "loss": 0.8624, "step": 8404 }, { "epoch": 3.8066123188405796, "grad_norm": 0.14489955876812796, "learning_rate": 1.3097839146551969e-05, "loss": 0.86, "step": 8405 }, { "epoch": 3.807065217391304, "grad_norm": 0.12031739882275395, "learning_rate": 1.3088480872037952e-05, "loss": 0.8698, "step": 8406 }, { "epoch": 3.807518115942029, "grad_norm": 0.1360291889583435, "learning_rate": 1.3079125287911162e-05, "loss": 0.8615, "step": 8407 }, { "epoch": 3.807971014492754, "grad_norm": 0.10648299966249983, "learning_rate": 1.3069772395106877e-05, "loss": 0.873, "step": 8408 }, { "epoch": 3.8084239130434785, "grad_norm": 0.12728003187635376, "learning_rate": 1.3060422194560119e-05, "loss": 0.8746, "step": 8409 }, { "epoch": 3.808876811594203, "grad_norm": 0.1270817047765427, "learning_rate": 1.305107468720563e-05, "loss": 0.8722, "step": 8410 }, { "epoch": 3.8093297101449277, "grad_norm": 0.10402571567797413, "learning_rate": 1.3041729873977924e-05, "loss": 0.8712, "step": 8411 }, { "epoch": 3.8097826086956523, "grad_norm": 0.1299686124869803, "learning_rate": 1.3032387755811184e-05, "loss": 0.8643, "step": 8412 }, { "epoch": 3.810235507246377, "grad_norm": 0.11991177312227358, "learning_rate": 1.3023048333639392e-05, "loss": 0.8659, "step": 8413 }, { "epoch": 3.8106884057971016, "grad_norm": 0.13734241147515755, "learning_rate": 1.301371160839619e-05, "loss": 0.8522, "step": 8414 }, { "epoch": 3.811141304347826, "grad_norm": 0.10636549795617413, "learning_rate": 1.3004377581015017e-05, "loss": 0.8644, "step": 8415 }, { "epoch": 3.8115942028985508, "grad_norm": 0.15775263214184526, "learning_rate": 1.2995046252428996e-05, "loss": 0.8659, "step": 8416 }, { "epoch": 3.8120471014492754, "grad_norm": 0.11618003463686918, "learning_rate": 1.2985717623570992e-05, "loss": 0.854, "step": 8417 }, { "epoch": 3.8125, "grad_norm": 0.13192391542471302, "learning_rate": 1.297639169537359e-05, "loss": 0.8709, "step": 8418 }, { "epoch": 3.8129528985507246, "grad_norm": 0.14940424705154595, "learning_rate": 1.2967068468769154e-05, "loss": 0.8663, "step": 8419 }, { "epoch": 3.8134057971014492, "grad_norm": 0.13111439245687845, "learning_rate": 1.29577479446897e-05, "loss": 0.853, "step": 8420 }, { "epoch": 3.813858695652174, "grad_norm": 0.1516814936182635, "learning_rate": 1.2948430124067054e-05, "loss": 0.8537, "step": 8421 }, { "epoch": 3.8143115942028984, "grad_norm": 0.11364454027151744, "learning_rate": 1.2939115007832697e-05, "loss": 0.8664, "step": 8422 }, { "epoch": 3.814764492753623, "grad_norm": 0.17608001492693295, "learning_rate": 1.292980259691791e-05, "loss": 0.8621, "step": 8423 }, { "epoch": 3.8152173913043477, "grad_norm": 0.1075798702818306, "learning_rate": 1.2920492892253642e-05, "loss": 0.8318, "step": 8424 }, { "epoch": 3.8156702898550723, "grad_norm": 0.14624038464931388, "learning_rate": 1.2911185894770623e-05, "loss": 0.8772, "step": 8425 }, { "epoch": 3.816123188405797, "grad_norm": 0.12434208158126196, "learning_rate": 1.2901881605399273e-05, "loss": 0.8389, "step": 8426 }, { "epoch": 3.8165760869565215, "grad_norm": 0.14706420494362313, "learning_rate": 1.2892580025069762e-05, "loss": 0.8618, "step": 8427 }, { "epoch": 3.817028985507246, "grad_norm": 0.11949404210936052, "learning_rate": 1.2883281154711971e-05, "loss": 0.8605, "step": 8428 }, { "epoch": 3.817481884057971, "grad_norm": 0.14875119984957183, "learning_rate": 1.2873984995255545e-05, "loss": 0.8553, "step": 8429 }, { "epoch": 3.817934782608696, "grad_norm": 0.1558200460326225, "learning_rate": 1.2864691547629807e-05, "loss": 0.8693, "step": 8430 }, { "epoch": 3.8183876811594204, "grad_norm": 0.13968285176640352, "learning_rate": 1.2855400812763872e-05, "loss": 0.864, "step": 8431 }, { "epoch": 3.818840579710145, "grad_norm": 0.13490819353561584, "learning_rate": 1.2846112791586526e-05, "loss": 0.8716, "step": 8432 }, { "epoch": 3.8192934782608696, "grad_norm": 0.1646777008777057, "learning_rate": 1.2836827485026308e-05, "loss": 0.8653, "step": 8433 }, { "epoch": 3.8197463768115942, "grad_norm": 0.13301187695001818, "learning_rate": 1.2827544894011496e-05, "loss": 0.86, "step": 8434 }, { "epoch": 3.820199275362319, "grad_norm": 0.14357562481794212, "learning_rate": 1.2818265019470087e-05, "loss": 0.8542, "step": 8435 }, { "epoch": 3.8206521739130435, "grad_norm": 0.16147921808617705, "learning_rate": 1.280898786232979e-05, "loss": 0.8628, "step": 8436 }, { "epoch": 3.821105072463768, "grad_norm": 0.13054523297981388, "learning_rate": 1.2799713423518054e-05, "loss": 0.8699, "step": 8437 }, { "epoch": 3.8215579710144927, "grad_norm": 0.17406930481728347, "learning_rate": 1.2790441703962082e-05, "loss": 0.8467, "step": 8438 }, { "epoch": 3.8220108695652173, "grad_norm": 0.11814681707852258, "learning_rate": 1.2781172704588757e-05, "loss": 0.8788, "step": 8439 }, { "epoch": 3.822463768115942, "grad_norm": 0.15316127985917685, "learning_rate": 1.2771906426324746e-05, "loss": 0.8709, "step": 8440 }, { "epoch": 3.8229166666666665, "grad_norm": 0.12685028911502413, "learning_rate": 1.2762642870096377e-05, "loss": 0.8438, "step": 8441 }, { "epoch": 3.8233695652173916, "grad_norm": 0.1329459886068805, "learning_rate": 1.2753382036829778e-05, "loss": 0.8628, "step": 8442 }, { "epoch": 3.823822463768116, "grad_norm": 0.11862715549169787, "learning_rate": 1.2744123927450755e-05, "loss": 0.8802, "step": 8443 }, { "epoch": 3.824275362318841, "grad_norm": 0.13209382271356923, "learning_rate": 1.273486854288485e-05, "loss": 0.8607, "step": 8444 }, { "epoch": 3.8247282608695654, "grad_norm": 0.13222123413502399, "learning_rate": 1.2725615884057328e-05, "loss": 0.8635, "step": 8445 }, { "epoch": 3.82518115942029, "grad_norm": 0.4126677071714713, "learning_rate": 1.2716365951893224e-05, "loss": 0.8598, "step": 8446 }, { "epoch": 3.8256340579710146, "grad_norm": 0.13769703426958282, "learning_rate": 1.270711874731723e-05, "loss": 0.8687, "step": 8447 }, { "epoch": 3.8260869565217392, "grad_norm": 0.14912126557178168, "learning_rate": 1.2697874271253844e-05, "loss": 0.8651, "step": 8448 }, { "epoch": 3.826539855072464, "grad_norm": 0.12398974058108543, "learning_rate": 1.2688632524627215e-05, "loss": 0.8735, "step": 8449 }, { "epoch": 3.8269927536231885, "grad_norm": 0.1369375803494553, "learning_rate": 1.2679393508361284e-05, "loss": 0.8715, "step": 8450 }, { "epoch": 3.827445652173913, "grad_norm": 0.14547191420655528, "learning_rate": 1.2670157223379675e-05, "loss": 0.8491, "step": 8451 }, { "epoch": 3.8278985507246377, "grad_norm": 0.14055178142780367, "learning_rate": 1.2660923670605763e-05, "loss": 0.8733, "step": 8452 }, { "epoch": 3.8283514492753623, "grad_norm": 0.14711752284952823, "learning_rate": 1.2651692850962615e-05, "loss": 0.8538, "step": 8453 }, { "epoch": 3.828804347826087, "grad_norm": 0.14275102036252463, "learning_rate": 1.2642464765373084e-05, "loss": 0.8466, "step": 8454 }, { "epoch": 3.8292572463768115, "grad_norm": 0.1253667893489217, "learning_rate": 1.263323941475969e-05, "loss": 0.8685, "step": 8455 }, { "epoch": 3.829710144927536, "grad_norm": 0.15356770872334324, "learning_rate": 1.2624016800044733e-05, "loss": 0.8612, "step": 8456 }, { "epoch": 3.8301630434782608, "grad_norm": 0.11086456496250105, "learning_rate": 1.2614796922150183e-05, "loss": 0.8423, "step": 8457 }, { "epoch": 3.8306159420289854, "grad_norm": 0.16171455317070313, "learning_rate": 1.2605579781997789e-05, "loss": 0.8672, "step": 8458 }, { "epoch": 3.83106884057971, "grad_norm": 0.12968875841930474, "learning_rate": 1.2596365380508999e-05, "loss": 0.8561, "step": 8459 }, { "epoch": 3.8315217391304346, "grad_norm": 0.15187897409880285, "learning_rate": 1.2587153718604985e-05, "loss": 0.8648, "step": 8460 }, { "epoch": 3.831974637681159, "grad_norm": 0.1328460113215322, "learning_rate": 1.2577944797206643e-05, "loss": 0.8475, "step": 8461 }, { "epoch": 3.832427536231884, "grad_norm": 0.15489165645856554, "learning_rate": 1.2568738617234626e-05, "loss": 0.8493, "step": 8462 }, { "epoch": 3.8328804347826084, "grad_norm": 0.12393512493391871, "learning_rate": 1.2559535179609262e-05, "loss": 0.8824, "step": 8463 }, { "epoch": 3.8333333333333335, "grad_norm": 0.1523816093388345, "learning_rate": 1.2550334485250661e-05, "loss": 0.8487, "step": 8464 }, { "epoch": 3.833786231884058, "grad_norm": 0.15843905452494217, "learning_rate": 1.254113653507861e-05, "loss": 0.8527, "step": 8465 }, { "epoch": 3.8342391304347827, "grad_norm": 0.13652284872660436, "learning_rate": 1.2531941330012667e-05, "loss": 0.858, "step": 8466 }, { "epoch": 3.8346920289855073, "grad_norm": 0.1669516161220841, "learning_rate": 1.252274887097206e-05, "loss": 0.8549, "step": 8467 }, { "epoch": 3.835144927536232, "grad_norm": 0.11893837661676959, "learning_rate": 1.2513559158875813e-05, "loss": 0.8473, "step": 8468 }, { "epoch": 3.8355978260869565, "grad_norm": 0.1668056483485591, "learning_rate": 1.2504372194642591e-05, "loss": 0.884, "step": 8469 }, { "epoch": 3.836050724637681, "grad_norm": 0.1380523760568813, "learning_rate": 1.2495187979190861e-05, "loss": 0.8622, "step": 8470 }, { "epoch": 3.8365036231884058, "grad_norm": 0.14915384723180533, "learning_rate": 1.2486006513438764e-05, "loss": 0.8735, "step": 8471 }, { "epoch": 3.8369565217391304, "grad_norm": 0.1761392208365565, "learning_rate": 1.247682779830421e-05, "loss": 0.8557, "step": 8472 }, { "epoch": 3.837409420289855, "grad_norm": 0.16968383937631498, "learning_rate": 1.246765183470478e-05, "loss": 0.8489, "step": 8473 }, { "epoch": 3.8378623188405796, "grad_norm": 0.12958063457629448, "learning_rate": 1.2458478623557833e-05, "loss": 0.859, "step": 8474 }, { "epoch": 3.838315217391304, "grad_norm": 0.1884113001863462, "learning_rate": 1.2449308165780414e-05, "loss": 0.87, "step": 8475 }, { "epoch": 3.838768115942029, "grad_norm": 0.11406958585642174, "learning_rate": 1.2440140462289328e-05, "loss": 0.8588, "step": 8476 }, { "epoch": 3.839221014492754, "grad_norm": 0.1274184700989375, "learning_rate": 1.2430975514001071e-05, "loss": 0.8629, "step": 8477 }, { "epoch": 3.8396739130434785, "grad_norm": 0.12216809190588233, "learning_rate": 1.2421813321831877e-05, "loss": 0.8438, "step": 8478 }, { "epoch": 3.840126811594203, "grad_norm": 0.1167313291471366, "learning_rate": 1.2412653886697696e-05, "loss": 0.8582, "step": 8479 }, { "epoch": 3.8405797101449277, "grad_norm": 0.16827324974258437, "learning_rate": 1.2403497209514237e-05, "loss": 0.8821, "step": 8480 }, { "epoch": 3.8410326086956523, "grad_norm": 0.12682541202413614, "learning_rate": 1.2394343291196878e-05, "loss": 0.8651, "step": 8481 }, { "epoch": 3.841485507246377, "grad_norm": 0.13673562775758802, "learning_rate": 1.2385192132660775e-05, "loss": 0.8505, "step": 8482 }, { "epoch": 3.8419384057971016, "grad_norm": 0.16298417389969325, "learning_rate": 1.2376043734820762e-05, "loss": 0.8764, "step": 8483 }, { "epoch": 3.842391304347826, "grad_norm": 0.12912328326149422, "learning_rate": 1.2366898098591445e-05, "loss": 0.8507, "step": 8484 }, { "epoch": 3.8428442028985508, "grad_norm": 0.14346759180825566, "learning_rate": 1.2357755224887113e-05, "loss": 0.8436, "step": 8485 }, { "epoch": 3.8432971014492754, "grad_norm": 0.1414349733722445, "learning_rate": 1.2348615114621794e-05, "loss": 0.8391, "step": 8486 }, { "epoch": 3.84375, "grad_norm": 0.14422960185124323, "learning_rate": 1.233947776870923e-05, "loss": 0.8609, "step": 8487 }, { "epoch": 3.8442028985507246, "grad_norm": 0.12115010346637088, "learning_rate": 1.2330343188062912e-05, "loss": 0.8347, "step": 8488 }, { "epoch": 3.8446557971014492, "grad_norm": 0.13642752013184278, "learning_rate": 1.2321211373596023e-05, "loss": 0.8508, "step": 8489 }, { "epoch": 3.845108695652174, "grad_norm": 0.11439399398839072, "learning_rate": 1.2312082326221507e-05, "loss": 0.8557, "step": 8490 }, { "epoch": 3.8455615942028984, "grad_norm": 0.12027530910560598, "learning_rate": 1.230295604685198e-05, "loss": 0.8767, "step": 8491 }, { "epoch": 3.846014492753623, "grad_norm": 0.12732556980243223, "learning_rate": 1.2293832536399841e-05, "loss": 0.847, "step": 8492 }, { "epoch": 3.8464673913043477, "grad_norm": 0.12635166462266276, "learning_rate": 1.2284711795777167e-05, "loss": 0.8697, "step": 8493 }, { "epoch": 3.8469202898550723, "grad_norm": 0.1268704843956641, "learning_rate": 1.2275593825895778e-05, "loss": 0.8593, "step": 8494 }, { "epoch": 3.847373188405797, "grad_norm": 0.11286597820153485, "learning_rate": 1.2266478627667189e-05, "loss": 0.8463, "step": 8495 }, { "epoch": 3.8478260869565215, "grad_norm": 0.1181101876035346, "learning_rate": 1.2257366202002693e-05, "loss": 0.8634, "step": 8496 }, { "epoch": 3.848278985507246, "grad_norm": 0.11596327445038923, "learning_rate": 1.2248256549813249e-05, "loss": 0.8605, "step": 8497 }, { "epoch": 3.848731884057971, "grad_norm": 0.11958055856520948, "learning_rate": 1.2239149672009583e-05, "loss": 0.8519, "step": 8498 }, { "epoch": 3.849184782608696, "grad_norm": 0.1069599833838891, "learning_rate": 1.2230045569502105e-05, "loss": 0.871, "step": 8499 }, { "epoch": 3.8496376811594204, "grad_norm": 0.12935169395752344, "learning_rate": 1.222094424320099e-05, "loss": 0.8579, "step": 8500 }, { "epoch": 3.850090579710145, "grad_norm": 0.11636187643598002, "learning_rate": 1.22118456940161e-05, "loss": 0.872, "step": 8501 }, { "epoch": 3.8505434782608696, "grad_norm": 0.12157157936348557, "learning_rate": 1.2202749922857024e-05, "loss": 0.874, "step": 8502 }, { "epoch": 3.8509963768115942, "grad_norm": 0.11596906517534324, "learning_rate": 1.219365693063308e-05, "loss": 0.8563, "step": 8503 }, { "epoch": 3.851449275362319, "grad_norm": 0.11504695165717967, "learning_rate": 1.2184566718253326e-05, "loss": 0.8658, "step": 8504 }, { "epoch": 3.8519021739130435, "grad_norm": 0.12272025069068707, "learning_rate": 1.2175479286626501e-05, "loss": 0.8515, "step": 8505 }, { "epoch": 3.852355072463768, "grad_norm": 0.11439768103243393, "learning_rate": 1.216639463666112e-05, "loss": 0.8618, "step": 8506 }, { "epoch": 3.8528079710144927, "grad_norm": 0.13444070803409594, "learning_rate": 1.2157312769265359e-05, "loss": 0.8432, "step": 8507 }, { "epoch": 3.8532608695652173, "grad_norm": 0.09786045052279203, "learning_rate": 1.214823368534717e-05, "loss": 0.8597, "step": 8508 }, { "epoch": 3.853713768115942, "grad_norm": 0.14772752801225214, "learning_rate": 1.2139157385814198e-05, "loss": 0.8713, "step": 8509 }, { "epoch": 3.8541666666666665, "grad_norm": 0.13112151181104034, "learning_rate": 1.2130083871573812e-05, "loss": 0.8396, "step": 8510 }, { "epoch": 3.8546195652173916, "grad_norm": 0.1494970506780253, "learning_rate": 1.2121013143533089e-05, "loss": 0.8632, "step": 8511 }, { "epoch": 3.855072463768116, "grad_norm": 0.1624894122237799, "learning_rate": 1.2111945202598876e-05, "loss": 0.8583, "step": 8512 }, { "epoch": 3.855525362318841, "grad_norm": 0.12394710415250579, "learning_rate": 1.2102880049677679e-05, "loss": 0.8704, "step": 8513 }, { "epoch": 3.8559782608695654, "grad_norm": 0.11559190720448162, "learning_rate": 1.2093817685675782e-05, "loss": 0.8749, "step": 8514 }, { "epoch": 3.85643115942029, "grad_norm": 0.14984367971773, "learning_rate": 1.208475811149914e-05, "loss": 0.8429, "step": 8515 }, { "epoch": 3.8568840579710146, "grad_norm": 0.11623363288353857, "learning_rate": 1.2075701328053482e-05, "loss": 0.8587, "step": 8516 }, { "epoch": 3.8573369565217392, "grad_norm": 0.1343426460688325, "learning_rate": 1.2066647336244195e-05, "loss": 0.8613, "step": 8517 }, { "epoch": 3.857789855072464, "grad_norm": 0.14639204376163528, "learning_rate": 1.2057596136976453e-05, "loss": 0.8532, "step": 8518 }, { "epoch": 3.8582427536231885, "grad_norm": 0.10130131149911686, "learning_rate": 1.20485477311551e-05, "loss": 0.8854, "step": 8519 }, { "epoch": 3.858695652173913, "grad_norm": 0.13331433638288037, "learning_rate": 1.2039502119684725e-05, "loss": 0.8347, "step": 8520 }, { "epoch": 3.8591485507246377, "grad_norm": 0.09484576172337957, "learning_rate": 1.2030459303469617e-05, "loss": 0.8485, "step": 8521 }, { "epoch": 3.8596014492753623, "grad_norm": 0.14756222838171326, "learning_rate": 1.2021419283413831e-05, "loss": 0.8514, "step": 8522 }, { "epoch": 3.860054347826087, "grad_norm": 0.12294002492782803, "learning_rate": 1.2012382060421079e-05, "loss": 0.8317, "step": 8523 }, { "epoch": 3.8605072463768115, "grad_norm": 0.12415500568367707, "learning_rate": 1.2003347635394853e-05, "loss": 0.8705, "step": 8524 }, { "epoch": 3.860960144927536, "grad_norm": 0.14429351033986518, "learning_rate": 1.1994316009238317e-05, "loss": 0.8388, "step": 8525 }, { "epoch": 3.8614130434782608, "grad_norm": 0.12945588076101733, "learning_rate": 1.1985287182854406e-05, "loss": 0.8779, "step": 8526 }, { "epoch": 3.8618659420289854, "grad_norm": 0.17127161047661926, "learning_rate": 1.1976261157145724e-05, "loss": 0.8527, "step": 8527 }, { "epoch": 3.86231884057971, "grad_norm": 0.1303405141560109, "learning_rate": 1.1967237933014619e-05, "loss": 0.8566, "step": 8528 }, { "epoch": 3.8627717391304346, "grad_norm": 0.1563463398743931, "learning_rate": 1.195821751136315e-05, "loss": 0.8471, "step": 8529 }, { "epoch": 3.863224637681159, "grad_norm": 0.12481154041254722, "learning_rate": 1.1949199893093128e-05, "loss": 0.8629, "step": 8530 }, { "epoch": 3.863677536231884, "grad_norm": 0.1466766922228988, "learning_rate": 1.1940185079106028e-05, "loss": 0.8417, "step": 8531 }, { "epoch": 3.8641304347826084, "grad_norm": 0.1237994556140674, "learning_rate": 1.19311730703031e-05, "loss": 0.858, "step": 8532 }, { "epoch": 3.8645833333333335, "grad_norm": 0.10966050831057547, "learning_rate": 1.1922163867585268e-05, "loss": 0.8638, "step": 8533 }, { "epoch": 3.865036231884058, "grad_norm": 0.1153670842447332, "learning_rate": 1.191315747185322e-05, "loss": 0.8526, "step": 8534 }, { "epoch": 3.8654891304347827, "grad_norm": 0.14122644812343035, "learning_rate": 1.1904153884007333e-05, "loss": 0.8587, "step": 8535 }, { "epoch": 3.8659420289855073, "grad_norm": 0.12980519630240026, "learning_rate": 1.1895153104947696e-05, "loss": 0.8487, "step": 8536 }, { "epoch": 3.866394927536232, "grad_norm": 0.13304890898722468, "learning_rate": 1.1886155135574132e-05, "loss": 0.8586, "step": 8537 }, { "epoch": 3.8668478260869565, "grad_norm": 0.16117991818395574, "learning_rate": 1.1877159976786197e-05, "loss": 0.8643, "step": 8538 }, { "epoch": 3.867300724637681, "grad_norm": 0.11382980810554527, "learning_rate": 1.1868167629483134e-05, "loss": 0.8642, "step": 8539 }, { "epoch": 3.8677536231884058, "grad_norm": 0.13644877650005266, "learning_rate": 1.1859178094563939e-05, "loss": 0.8656, "step": 8540 }, { "epoch": 3.8682065217391304, "grad_norm": 0.11763793847104034, "learning_rate": 1.1850191372927298e-05, "loss": 0.8709, "step": 8541 }, { "epoch": 3.868659420289855, "grad_norm": 0.11219713031970047, "learning_rate": 1.1841207465471642e-05, "loss": 0.8446, "step": 8542 }, { "epoch": 3.8691123188405796, "grad_norm": 0.1510202102592016, "learning_rate": 1.1832226373095095e-05, "loss": 0.8513, "step": 8543 }, { "epoch": 3.869565217391304, "grad_norm": 0.1298775186507516, "learning_rate": 1.1823248096695515e-05, "loss": 0.8523, "step": 8544 }, { "epoch": 3.870018115942029, "grad_norm": 0.1358252990285185, "learning_rate": 1.1814272637170458e-05, "loss": 0.8744, "step": 8545 }, { "epoch": 3.870471014492754, "grad_norm": 0.13672301185250943, "learning_rate": 1.1805299995417245e-05, "loss": 0.8324, "step": 8546 }, { "epoch": 3.8709239130434785, "grad_norm": 0.19319618602482921, "learning_rate": 1.179633017233286e-05, "loss": 0.8624, "step": 8547 }, { "epoch": 3.871376811594203, "grad_norm": 0.15775922433108325, "learning_rate": 1.1787363168814046e-05, "loss": 0.8541, "step": 8548 }, { "epoch": 3.8718297101449277, "grad_norm": 0.1371593041309119, "learning_rate": 1.1778398985757234e-05, "loss": 0.8519, "step": 8549 }, { "epoch": 3.8722826086956523, "grad_norm": 0.22103263395186692, "learning_rate": 1.1769437624058608e-05, "loss": 0.8494, "step": 8550 }, { "epoch": 3.872735507246377, "grad_norm": 0.12407769639730071, "learning_rate": 1.176047908461404e-05, "loss": 0.8361, "step": 8551 }, { "epoch": 3.8731884057971016, "grad_norm": 0.24184692131546318, "learning_rate": 1.1751523368319125e-05, "loss": 0.8462, "step": 8552 }, { "epoch": 3.873641304347826, "grad_norm": 0.2209491475712937, "learning_rate": 1.174257047606917e-05, "loss": 0.8477, "step": 8553 }, { "epoch": 3.8740942028985508, "grad_norm": 0.11782337723879509, "learning_rate": 1.1733620408759237e-05, "loss": 0.8557, "step": 8554 }, { "epoch": 3.8745471014492754, "grad_norm": 0.3084192404403008, "learning_rate": 1.1724673167284046e-05, "loss": 0.8545, "step": 8555 }, { "epoch": 3.875, "grad_norm": 0.22922506915836696, "learning_rate": 1.1715728752538103e-05, "loss": 0.8741, "step": 8556 }, { "epoch": 3.8754528985507246, "grad_norm": 0.16413646166551377, "learning_rate": 1.1706787165415574e-05, "loss": 0.8789, "step": 8557 }, { "epoch": 3.8759057971014492, "grad_norm": 0.28908369324032734, "learning_rate": 1.1697848406810354e-05, "loss": 0.8478, "step": 8558 }, { "epoch": 3.876358695652174, "grad_norm": 0.14178071670446332, "learning_rate": 1.168891247761609e-05, "loss": 0.8658, "step": 8559 }, { "epoch": 3.8768115942028984, "grad_norm": 0.1844142097260412, "learning_rate": 1.1679979378726096e-05, "loss": 0.8679, "step": 8560 }, { "epoch": 3.877264492753623, "grad_norm": 0.24329722838972512, "learning_rate": 1.1671049111033454e-05, "loss": 0.8642, "step": 8561 }, { "epoch": 3.8777173913043477, "grad_norm": 0.13304145603796663, "learning_rate": 1.1662121675430922e-05, "loss": 0.8429, "step": 8562 }, { "epoch": 3.8781702898550723, "grad_norm": 0.1491809695642831, "learning_rate": 1.1653197072810993e-05, "loss": 0.847, "step": 8563 }, { "epoch": 3.878623188405797, "grad_norm": 0.17924104714511485, "learning_rate": 1.1644275304065857e-05, "loss": 0.8432, "step": 8564 }, { "epoch": 3.8790760869565215, "grad_norm": 0.1201357967402993, "learning_rate": 1.1635356370087467e-05, "loss": 0.8578, "step": 8565 }, { "epoch": 3.879528985507246, "grad_norm": 0.11362101008749023, "learning_rate": 1.1626440271767433e-05, "loss": 0.8844, "step": 8566 }, { "epoch": 3.879981884057971, "grad_norm": 0.1320374093465243, "learning_rate": 1.1617527009997143e-05, "loss": 0.8561, "step": 8567 }, { "epoch": 3.880434782608696, "grad_norm": 0.21044078979190511, "learning_rate": 1.160861658566764e-05, "loss": 0.8724, "step": 8568 }, { "epoch": 3.8808876811594204, "grad_norm": 0.15743395555763853, "learning_rate": 1.1599708999669738e-05, "loss": 0.8794, "step": 8569 }, { "epoch": 3.881340579710145, "grad_norm": 0.16643266396285505, "learning_rate": 1.159080425289393e-05, "loss": 0.8656, "step": 8570 }, { "epoch": 3.8817934782608696, "grad_norm": 0.12576582080357843, "learning_rate": 1.1581902346230445e-05, "loss": 0.8543, "step": 8571 }, { "epoch": 3.8822463768115942, "grad_norm": 0.24357044722156668, "learning_rate": 1.1573003280569197e-05, "loss": 0.8516, "step": 8572 }, { "epoch": 3.882699275362319, "grad_norm": 0.21373993272183528, "learning_rate": 1.1564107056799872e-05, "loss": 0.8581, "step": 8573 }, { "epoch": 3.8831521739130435, "grad_norm": 0.16048564981792757, "learning_rate": 1.155521367581181e-05, "loss": 0.8664, "step": 8574 }, { "epoch": 3.883605072463768, "grad_norm": 0.1314234755252817, "learning_rate": 1.1546323138494127e-05, "loss": 0.8453, "step": 8575 }, { "epoch": 3.8840579710144927, "grad_norm": 0.11576613300940312, "learning_rate": 1.1537435445735596e-05, "loss": 0.8515, "step": 8576 }, { "epoch": 3.8845108695652173, "grad_norm": 0.15674190124593562, "learning_rate": 1.1528550598424757e-05, "loss": 0.8727, "step": 8577 }, { "epoch": 3.884963768115942, "grad_norm": 0.12936372279560604, "learning_rate": 1.151966859744983e-05, "loss": 0.8738, "step": 8578 }, { "epoch": 3.8854166666666665, "grad_norm": 0.14059369022559298, "learning_rate": 1.1510789443698772e-05, "loss": 0.8553, "step": 8579 }, { "epoch": 3.8858695652173916, "grad_norm": 0.11481777848145307, "learning_rate": 1.1501913138059218e-05, "loss": 0.8724, "step": 8580 }, { "epoch": 3.886322463768116, "grad_norm": 0.11062321534026454, "learning_rate": 1.149303968141858e-05, "loss": 0.8705, "step": 8581 }, { "epoch": 3.886775362318841, "grad_norm": 0.13845442805631133, "learning_rate": 1.148416907466393e-05, "loss": 0.8635, "step": 8582 }, { "epoch": 3.8872282608695654, "grad_norm": 0.09804463314282683, "learning_rate": 1.1475301318682095e-05, "loss": 0.862, "step": 8583 }, { "epoch": 3.88768115942029, "grad_norm": 0.13990578374632562, "learning_rate": 1.1466436414359574e-05, "loss": 0.8673, "step": 8584 }, { "epoch": 3.8881340579710146, "grad_norm": 0.10439091334849335, "learning_rate": 1.1457574362582635e-05, "loss": 0.8512, "step": 8585 }, { "epoch": 3.8885869565217392, "grad_norm": 0.1409198208756014, "learning_rate": 1.1448715164237219e-05, "loss": 0.8597, "step": 8586 }, { "epoch": 3.889039855072464, "grad_norm": 0.12393295158149428, "learning_rate": 1.1439858820208985e-05, "loss": 0.8528, "step": 8587 }, { "epoch": 3.8894927536231885, "grad_norm": 0.12748051905817137, "learning_rate": 1.1431005331383314e-05, "loss": 0.8665, "step": 8588 }, { "epoch": 3.889945652173913, "grad_norm": 0.11690019353512611, "learning_rate": 1.142215469864532e-05, "loss": 0.8402, "step": 8589 }, { "epoch": 3.8903985507246377, "grad_norm": 0.10585034619700082, "learning_rate": 1.1413306922879799e-05, "loss": 0.8409, "step": 8590 }, { "epoch": 3.8908514492753623, "grad_norm": 0.13267845657821947, "learning_rate": 1.1404462004971296e-05, "loss": 0.8821, "step": 8591 }, { "epoch": 3.891304347826087, "grad_norm": 0.10998155436316244, "learning_rate": 1.1395619945804021e-05, "loss": 0.8386, "step": 8592 }, { "epoch": 3.8917572463768115, "grad_norm": 0.10157684875943501, "learning_rate": 1.1386780746261965e-05, "loss": 0.8383, "step": 8593 }, { "epoch": 3.892210144927536, "grad_norm": 0.10934231197886354, "learning_rate": 1.1377944407228783e-05, "loss": 0.8465, "step": 8594 }, { "epoch": 3.8926630434782608, "grad_norm": 0.10804499796722102, "learning_rate": 1.1369110929587848e-05, "loss": 0.8667, "step": 8595 }, { "epoch": 3.8931159420289854, "grad_norm": 0.10501948930091914, "learning_rate": 1.1360280314222255e-05, "loss": 0.845, "step": 8596 }, { "epoch": 3.89356884057971, "grad_norm": 0.11306518646472936, "learning_rate": 1.135145256201483e-05, "loss": 0.8381, "step": 8597 }, { "epoch": 3.8940217391304346, "grad_norm": 0.11506860445553645, "learning_rate": 1.1342627673848079e-05, "loss": 0.8568, "step": 8598 }, { "epoch": 3.894474637681159, "grad_norm": 0.12566862628445116, "learning_rate": 1.1333805650604264e-05, "loss": 0.8515, "step": 8599 }, { "epoch": 3.894927536231884, "grad_norm": 0.11992079665285064, "learning_rate": 1.1324986493165314e-05, "loss": 0.8581, "step": 8600 }, { "epoch": 3.8953804347826084, "grad_norm": 0.11217101698152211, "learning_rate": 1.1316170202412912e-05, "loss": 0.8632, "step": 8601 }, { "epoch": 3.8958333333333335, "grad_norm": 0.12164035892509953, "learning_rate": 1.130735677922842e-05, "loss": 0.8561, "step": 8602 }, { "epoch": 3.896286231884058, "grad_norm": 0.1338822902929974, "learning_rate": 1.1298546224492948e-05, "loss": 0.8712, "step": 8603 }, { "epoch": 3.8967391304347827, "grad_norm": 0.12586980971279235, "learning_rate": 1.1289738539087289e-05, "loss": 0.867, "step": 8604 }, { "epoch": 3.8971920289855073, "grad_norm": 0.1358638826344929, "learning_rate": 1.1280933723891968e-05, "loss": 0.8422, "step": 8605 }, { "epoch": 3.897644927536232, "grad_norm": 0.11546389891532531, "learning_rate": 1.1272131779787196e-05, "loss": 0.8554, "step": 8606 }, { "epoch": 3.8980978260869565, "grad_norm": 0.14504802166558678, "learning_rate": 1.126333270765295e-05, "loss": 0.8572, "step": 8607 }, { "epoch": 3.898550724637681, "grad_norm": 0.11474363591386312, "learning_rate": 1.1254536508368852e-05, "loss": 0.873, "step": 8608 }, { "epoch": 3.8990036231884058, "grad_norm": 0.15221088892028065, "learning_rate": 1.1245743182814305e-05, "loss": 0.8449, "step": 8609 }, { "epoch": 3.8994565217391304, "grad_norm": 0.108319618891882, "learning_rate": 1.1236952731868369e-05, "loss": 0.8537, "step": 8610 }, { "epoch": 3.899909420289855, "grad_norm": 0.10653927893478506, "learning_rate": 1.1228165156409854e-05, "loss": 0.8764, "step": 8611 }, { "epoch": 3.9003623188405796, "grad_norm": 0.12208417856379652, "learning_rate": 1.1219380457317266e-05, "loss": 0.8488, "step": 8612 }, { "epoch": 3.900815217391304, "grad_norm": 0.10400889225973825, "learning_rate": 1.1210598635468815e-05, "loss": 0.8402, "step": 8613 }, { "epoch": 3.901268115942029, "grad_norm": 0.12587329516487597, "learning_rate": 1.1201819691742433e-05, "loss": 0.8611, "step": 8614 }, { "epoch": 3.901721014492754, "grad_norm": 0.09927636320057698, "learning_rate": 1.1193043627015779e-05, "loss": 0.8504, "step": 8615 }, { "epoch": 3.9021739130434785, "grad_norm": 0.1292211247840729, "learning_rate": 1.118427044216619e-05, "loss": 0.853, "step": 8616 }, { "epoch": 3.902626811594203, "grad_norm": 0.10359860290041993, "learning_rate": 1.1175500138070761e-05, "loss": 0.8462, "step": 8617 }, { "epoch": 3.9030797101449277, "grad_norm": 0.09826633763574459, "learning_rate": 1.1166732715606248e-05, "loss": 0.8684, "step": 8618 }, { "epoch": 3.9035326086956523, "grad_norm": 0.11267872661907989, "learning_rate": 1.1157968175649163e-05, "loss": 0.8581, "step": 8619 }, { "epoch": 3.903985507246377, "grad_norm": 0.1156126105070513, "learning_rate": 1.1149206519075708e-05, "loss": 0.8703, "step": 8620 }, { "epoch": 3.9044384057971016, "grad_norm": 0.11361684018785477, "learning_rate": 1.1140447746761796e-05, "loss": 0.8435, "step": 8621 }, { "epoch": 3.904891304347826, "grad_norm": 0.09495390903553368, "learning_rate": 1.113169185958304e-05, "loss": 0.8711, "step": 8622 }, { "epoch": 3.9053442028985508, "grad_norm": 0.11440193817615495, "learning_rate": 1.1122938858414809e-05, "loss": 0.858, "step": 8623 }, { "epoch": 3.9057971014492754, "grad_norm": 0.10790815800306457, "learning_rate": 1.1114188744132125e-05, "loss": 0.8695, "step": 8624 }, { "epoch": 3.90625, "grad_norm": 0.10977429526591599, "learning_rate": 1.110544151760978e-05, "loss": 0.8629, "step": 8625 }, { "epoch": 3.9067028985507246, "grad_norm": 0.10278296331545306, "learning_rate": 1.109669717972222e-05, "loss": 0.8675, "step": 8626 }, { "epoch": 3.9071557971014492, "grad_norm": 0.1150444125099604, "learning_rate": 1.108795573134366e-05, "loss": 0.8562, "step": 8627 }, { "epoch": 3.907608695652174, "grad_norm": 0.09435610830010745, "learning_rate": 1.1079217173347976e-05, "loss": 0.8808, "step": 8628 }, { "epoch": 3.9080615942028984, "grad_norm": 0.13501460443492147, "learning_rate": 1.1070481506608783e-05, "loss": 0.8592, "step": 8629 }, { "epoch": 3.908514492753623, "grad_norm": 0.10059975006422568, "learning_rate": 1.1061748731999384e-05, "loss": 0.8409, "step": 8630 }, { "epoch": 3.9089673913043477, "grad_norm": 0.11403536536947789, "learning_rate": 1.1053018850392832e-05, "loss": 0.8389, "step": 8631 }, { "epoch": 3.9094202898550723, "grad_norm": 0.10525552575845283, "learning_rate": 1.1044291862661844e-05, "loss": 0.8488, "step": 8632 }, { "epoch": 3.909873188405797, "grad_norm": 0.10797274975117392, "learning_rate": 1.1035567769678895e-05, "loss": 0.8576, "step": 8633 }, { "epoch": 3.9103260869565215, "grad_norm": 0.12013330403134001, "learning_rate": 1.1026846572316124e-05, "loss": 0.8685, "step": 8634 }, { "epoch": 3.910778985507246, "grad_norm": 0.10365711300749503, "learning_rate": 1.1018128271445425e-05, "loss": 0.8634, "step": 8635 }, { "epoch": 3.911231884057971, "grad_norm": 0.10925573245203206, "learning_rate": 1.1009412867938365e-05, "loss": 0.8599, "step": 8636 }, { "epoch": 3.911684782608696, "grad_norm": 0.10691738931869586, "learning_rate": 1.1000700362666241e-05, "loss": 0.8723, "step": 8637 }, { "epoch": 3.9121376811594204, "grad_norm": 0.11615035666150267, "learning_rate": 1.0991990756500047e-05, "loss": 0.8684, "step": 8638 }, { "epoch": 3.912590579710145, "grad_norm": 0.11755831797520976, "learning_rate": 1.098328405031051e-05, "loss": 0.8832, "step": 8639 }, { "epoch": 3.9130434782608696, "grad_norm": 0.11302751770576919, "learning_rate": 1.0974580244968039e-05, "loss": 0.8566, "step": 8640 }, { "epoch": 3.9134963768115942, "grad_norm": 0.10843430207038951, "learning_rate": 1.0965879341342789e-05, "loss": 0.8679, "step": 8641 }, { "epoch": 3.913949275362319, "grad_norm": 0.10834412606585672, "learning_rate": 1.0957181340304577e-05, "loss": 0.8505, "step": 8642 }, { "epoch": 3.9144021739130435, "grad_norm": 0.11422062841944176, "learning_rate": 1.0948486242722978e-05, "loss": 0.8598, "step": 8643 }, { "epoch": 3.914855072463768, "grad_norm": 0.10485140227101213, "learning_rate": 1.0939794049467234e-05, "loss": 0.8505, "step": 8644 }, { "epoch": 3.9153079710144927, "grad_norm": 0.12496351606166145, "learning_rate": 1.0931104761406353e-05, "loss": 0.8877, "step": 8645 }, { "epoch": 3.9157608695652173, "grad_norm": 0.09545678289617578, "learning_rate": 1.0922418379408973e-05, "loss": 0.8793, "step": 8646 }, { "epoch": 3.916213768115942, "grad_norm": 0.09844617255099151, "learning_rate": 1.091373490434351e-05, "loss": 0.851, "step": 8647 }, { "epoch": 3.9166666666666665, "grad_norm": 0.09141782729195347, "learning_rate": 1.0905054337078051e-05, "loss": 0.854, "step": 8648 }, { "epoch": 3.9171195652173916, "grad_norm": 0.10799942460532701, "learning_rate": 1.0896376678480428e-05, "loss": 0.8572, "step": 8649 }, { "epoch": 3.917572463768116, "grad_norm": 0.09162438108921069, "learning_rate": 1.0887701929418135e-05, "loss": 0.8532, "step": 8650 }, { "epoch": 3.918025362318841, "grad_norm": 0.13924110453068944, "learning_rate": 1.0879030090758422e-05, "loss": 0.862, "step": 8651 }, { "epoch": 3.9184782608695654, "grad_norm": 0.0974121415694807, "learning_rate": 1.0870361163368205e-05, "loss": 0.8517, "step": 8652 }, { "epoch": 3.91893115942029, "grad_norm": 0.12433589924264785, "learning_rate": 1.0861695148114157e-05, "loss": 0.8682, "step": 8653 }, { "epoch": 3.9193840579710146, "grad_norm": 0.12486180408937522, "learning_rate": 1.085303204586262e-05, "loss": 0.8633, "step": 8654 }, { "epoch": 3.9198369565217392, "grad_norm": 0.10603325781229589, "learning_rate": 1.0844371857479653e-05, "loss": 0.8487, "step": 8655 }, { "epoch": 3.920289855072464, "grad_norm": 0.11282863465255598, "learning_rate": 1.0835714583831024e-05, "loss": 0.8591, "step": 8656 }, { "epoch": 3.9207427536231885, "grad_norm": 0.12906949115467126, "learning_rate": 1.0827060225782228e-05, "loss": 0.8532, "step": 8657 }, { "epoch": 3.921195652173913, "grad_norm": 0.11383879683621906, "learning_rate": 1.0818408784198447e-05, "loss": 0.8568, "step": 8658 }, { "epoch": 3.9216485507246377, "grad_norm": 0.13550170814912318, "learning_rate": 1.0809760259944588e-05, "loss": 0.8428, "step": 8659 }, { "epoch": 3.9221014492753623, "grad_norm": 0.12137947542078006, "learning_rate": 1.0801114653885238e-05, "loss": 0.8678, "step": 8660 }, { "epoch": 3.922554347826087, "grad_norm": 0.13136921744301244, "learning_rate": 1.0792471966884741e-05, "loss": 0.8465, "step": 8661 }, { "epoch": 3.9230072463768115, "grad_norm": 0.13553877505556355, "learning_rate": 1.0783832199807108e-05, "loss": 0.854, "step": 8662 }, { "epoch": 3.923460144927536, "grad_norm": 0.11584281810584096, "learning_rate": 1.0775195353516064e-05, "loss": 0.8323, "step": 8663 }, { "epoch": 3.9239130434782608, "grad_norm": 0.12213794320934498, "learning_rate": 1.0766561428875036e-05, "loss": 0.8578, "step": 8664 }, { "epoch": 3.9243659420289854, "grad_norm": 0.12190957446404449, "learning_rate": 1.0757930426747202e-05, "loss": 0.8608, "step": 8665 }, { "epoch": 3.92481884057971, "grad_norm": 0.09854963613594779, "learning_rate": 1.0749302347995388e-05, "loss": 0.8544, "step": 8666 }, { "epoch": 3.9252717391304346, "grad_norm": 0.14021730938865484, "learning_rate": 1.0740677193482184e-05, "loss": 0.8574, "step": 8667 }, { "epoch": 3.925724637681159, "grad_norm": 0.10727619237345372, "learning_rate": 1.0732054964069828e-05, "loss": 0.8575, "step": 8668 }, { "epoch": 3.926177536231884, "grad_norm": 0.11183191824685444, "learning_rate": 1.0723435660620334e-05, "loss": 0.8575, "step": 8669 }, { "epoch": 3.9266304347826084, "grad_norm": 0.11225481074456509, "learning_rate": 1.0714819283995368e-05, "loss": 0.8353, "step": 8670 }, { "epoch": 3.9270833333333335, "grad_norm": 0.10617371047746497, "learning_rate": 1.0706205835056326e-05, "loss": 0.8441, "step": 8671 }, { "epoch": 3.927536231884058, "grad_norm": 0.09902872652917422, "learning_rate": 1.0697595314664295e-05, "loss": 0.8336, "step": 8672 }, { "epoch": 3.9279891304347827, "grad_norm": 0.11399916427417582, "learning_rate": 1.0688987723680105e-05, "loss": 0.8622, "step": 8673 }, { "epoch": 3.9284420289855073, "grad_norm": 0.10106311789210905, "learning_rate": 1.0680383062964249e-05, "loss": 0.8505, "step": 8674 }, { "epoch": 3.928894927536232, "grad_norm": 0.09932524395661309, "learning_rate": 1.0671781333376967e-05, "loss": 0.8678, "step": 8675 }, { "epoch": 3.9293478260869565, "grad_norm": 0.10356147439447361, "learning_rate": 1.0663182535778169e-05, "loss": 0.8434, "step": 8676 }, { "epoch": 3.929800724637681, "grad_norm": 0.10809402522115213, "learning_rate": 1.0654586671027513e-05, "loss": 0.8502, "step": 8677 }, { "epoch": 3.9302536231884058, "grad_norm": 0.10468809897806361, "learning_rate": 1.064599373998433e-05, "loss": 0.8508, "step": 8678 }, { "epoch": 3.9307065217391304, "grad_norm": 0.1033034610133108, "learning_rate": 1.063740374350767e-05, "loss": 0.8494, "step": 8679 }, { "epoch": 3.931159420289855, "grad_norm": 0.13021337297941288, "learning_rate": 1.0628816682456273e-05, "loss": 0.8493, "step": 8680 }, { "epoch": 3.9316123188405796, "grad_norm": 0.11244738633628282, "learning_rate": 1.062023255768863e-05, "loss": 0.864, "step": 8681 }, { "epoch": 3.932065217391304, "grad_norm": 0.1363233134606835, "learning_rate": 1.061165137006289e-05, "loss": 0.8457, "step": 8682 }, { "epoch": 3.932518115942029, "grad_norm": 0.12087682879482757, "learning_rate": 1.0603073120436923e-05, "loss": 0.8568, "step": 8683 }, { "epoch": 3.932971014492754, "grad_norm": 0.13752304412224636, "learning_rate": 1.059449780966833e-05, "loss": 0.8608, "step": 8684 }, { "epoch": 3.9334239130434785, "grad_norm": 0.13235089231989877, "learning_rate": 1.0585925438614377e-05, "loss": 0.8621, "step": 8685 }, { "epoch": 3.933876811594203, "grad_norm": 0.13138170781162997, "learning_rate": 1.0577356008132083e-05, "loss": 0.8703, "step": 8686 }, { "epoch": 3.9343297101449277, "grad_norm": 0.1834439418818248, "learning_rate": 1.0568789519078129e-05, "loss": 0.8775, "step": 8687 }, { "epoch": 3.9347826086956523, "grad_norm": 0.11983517387745635, "learning_rate": 1.0560225972308925e-05, "loss": 0.8528, "step": 8688 }, { "epoch": 3.935235507246377, "grad_norm": 0.1282602128249133, "learning_rate": 1.055166536868057e-05, "loss": 0.8623, "step": 8689 }, { "epoch": 3.9356884057971016, "grad_norm": 0.10932609360939186, "learning_rate": 1.0543107709048908e-05, "loss": 0.8632, "step": 8690 }, { "epoch": 3.936141304347826, "grad_norm": 0.12749071973628115, "learning_rate": 1.053455299426943e-05, "loss": 0.8573, "step": 8691 }, { "epoch": 3.9365942028985508, "grad_norm": 0.11255280414001337, "learning_rate": 1.0526001225197397e-05, "loss": 0.8651, "step": 8692 }, { "epoch": 3.9370471014492754, "grad_norm": 0.1268663322702014, "learning_rate": 1.051745240268771e-05, "loss": 0.8612, "step": 8693 }, { "epoch": 3.9375, "grad_norm": 0.14282712636973754, "learning_rate": 1.0508906527595042e-05, "loss": 0.853, "step": 8694 }, { "epoch": 3.9379528985507246, "grad_norm": 0.10790141774336941, "learning_rate": 1.050036360077371e-05, "loss": 0.8466, "step": 8695 }, { "epoch": 3.9384057971014492, "grad_norm": 0.12483459860356393, "learning_rate": 1.0491823623077786e-05, "loss": 0.8751, "step": 8696 }, { "epoch": 3.938858695652174, "grad_norm": 0.11390807200601864, "learning_rate": 1.048328659536101e-05, "loss": 0.8633, "step": 8697 }, { "epoch": 3.9393115942028984, "grad_norm": 0.11634413925486556, "learning_rate": 1.047475251847685e-05, "loss": 0.8292, "step": 8698 }, { "epoch": 3.939764492753623, "grad_norm": 0.11152362406102165, "learning_rate": 1.0466221393278455e-05, "loss": 0.8583, "step": 8699 }, { "epoch": 3.9402173913043477, "grad_norm": 0.1174881659977935, "learning_rate": 1.0457693220618719e-05, "loss": 0.8857, "step": 8700 }, { "epoch": 3.9406702898550723, "grad_norm": 0.09705348951899652, "learning_rate": 1.0449168001350193e-05, "loss": 0.8381, "step": 8701 }, { "epoch": 3.941123188405797, "grad_norm": 0.10711824105066768, "learning_rate": 1.0440645736325186e-05, "loss": 0.8466, "step": 8702 }, { "epoch": 3.9415760869565215, "grad_norm": 0.09232237522569488, "learning_rate": 1.0432126426395648e-05, "loss": 0.8586, "step": 8703 }, { "epoch": 3.942028985507246, "grad_norm": 0.09216581173431483, "learning_rate": 1.04236100724133e-05, "loss": 0.8533, "step": 8704 }, { "epoch": 3.942481884057971, "grad_norm": 0.09675000808331055, "learning_rate": 1.0415096675229521e-05, "loss": 0.8719, "step": 8705 }, { "epoch": 3.942934782608696, "grad_norm": 0.09772932722246028, "learning_rate": 1.0406586235695406e-05, "loss": 0.8402, "step": 8706 }, { "epoch": 3.9433876811594204, "grad_norm": 0.10905479253139522, "learning_rate": 1.0398078754661754e-05, "loss": 0.8424, "step": 8707 }, { "epoch": 3.943840579710145, "grad_norm": 0.1215286780926707, "learning_rate": 1.0389574232979082e-05, "loss": 0.853, "step": 8708 }, { "epoch": 3.9442934782608696, "grad_norm": 0.11766680591281567, "learning_rate": 1.0381072671497585e-05, "loss": 0.8571, "step": 8709 }, { "epoch": 3.9447463768115942, "grad_norm": 0.10197223191162716, "learning_rate": 1.0372574071067203e-05, "loss": 0.8598, "step": 8710 }, { "epoch": 3.945199275362319, "grad_norm": 0.12589539851931678, "learning_rate": 1.0364078432537528e-05, "loss": 0.8557, "step": 8711 }, { "epoch": 3.9456521739130435, "grad_norm": 0.09390815190328825, "learning_rate": 1.0355585756757907e-05, "loss": 0.8622, "step": 8712 }, { "epoch": 3.946105072463768, "grad_norm": 0.11487431201785389, "learning_rate": 1.0347096044577349e-05, "loss": 0.8432, "step": 8713 }, { "epoch": 3.9465579710144927, "grad_norm": 0.1224563099109468, "learning_rate": 1.0338609296844591e-05, "loss": 0.8465, "step": 8714 }, { "epoch": 3.9470108695652173, "grad_norm": 0.11918023887446666, "learning_rate": 1.0330125514408053e-05, "loss": 0.8573, "step": 8715 }, { "epoch": 3.947463768115942, "grad_norm": 0.10607062517000002, "learning_rate": 1.0321644698115892e-05, "loss": 0.8603, "step": 8716 }, { "epoch": 3.9479166666666665, "grad_norm": 0.12599555888110808, "learning_rate": 1.0313166848815931e-05, "loss": 0.8649, "step": 8717 }, { "epoch": 3.9483695652173916, "grad_norm": 0.12108354732282478, "learning_rate": 1.0304691967355733e-05, "loss": 0.8459, "step": 8718 }, { "epoch": 3.948822463768116, "grad_norm": 0.10253725986936328, "learning_rate": 1.0296220054582525e-05, "loss": 0.8443, "step": 8719 }, { "epoch": 3.949275362318841, "grad_norm": 0.1307712618739725, "learning_rate": 1.028775111134328e-05, "loss": 0.8338, "step": 8720 }, { "epoch": 3.9497282608695654, "grad_norm": 0.11772805298717094, "learning_rate": 1.0279285138484641e-05, "loss": 0.858, "step": 8721 }, { "epoch": 3.95018115942029, "grad_norm": 0.08906699430809725, "learning_rate": 1.0270822136852959e-05, "loss": 0.8614, "step": 8722 }, { "epoch": 3.9506340579710146, "grad_norm": 0.12933325218263117, "learning_rate": 1.0262362107294286e-05, "loss": 0.8422, "step": 8723 }, { "epoch": 3.9510869565217392, "grad_norm": 0.11519017720461525, "learning_rate": 1.0253905050654409e-05, "loss": 0.8651, "step": 8724 }, { "epoch": 3.951539855072464, "grad_norm": 0.12733229645768296, "learning_rate": 1.0245450967778772e-05, "loss": 0.8706, "step": 8725 }, { "epoch": 3.9519927536231885, "grad_norm": 0.11562060338096987, "learning_rate": 1.0236999859512564e-05, "loss": 0.8458, "step": 8726 }, { "epoch": 3.952445652173913, "grad_norm": 0.11189609411292056, "learning_rate": 1.0228551726700626e-05, "loss": 0.874, "step": 8727 }, { "epoch": 3.9528985507246377, "grad_norm": 0.125554986078543, "learning_rate": 1.0220106570187568e-05, "loss": 0.8613, "step": 8728 }, { "epoch": 3.9533514492753623, "grad_norm": 0.09886228550250804, "learning_rate": 1.021166439081764e-05, "loss": 0.8699, "step": 8729 }, { "epoch": 3.953804347826087, "grad_norm": 0.14670141650775376, "learning_rate": 1.0203225189434832e-05, "loss": 0.8634, "step": 8730 }, { "epoch": 3.9542572463768115, "grad_norm": 0.10437600110134182, "learning_rate": 1.01947889668828e-05, "loss": 0.852, "step": 8731 }, { "epoch": 3.954710144927536, "grad_norm": 0.1272393560593215, "learning_rate": 1.0186355724004962e-05, "loss": 0.8571, "step": 8732 }, { "epoch": 3.9551630434782608, "grad_norm": 0.10909518428755921, "learning_rate": 1.0177925461644374e-05, "loss": 0.8536, "step": 8733 }, { "epoch": 3.9556159420289854, "grad_norm": 0.10601127680526816, "learning_rate": 1.0169498180643846e-05, "loss": 0.8729, "step": 8734 }, { "epoch": 3.95606884057971, "grad_norm": 0.12456583436849303, "learning_rate": 1.0161073881845845e-05, "loss": 0.8564, "step": 8735 }, { "epoch": 3.9565217391304346, "grad_norm": 0.12327497247937974, "learning_rate": 1.0152652566092582e-05, "loss": 0.8552, "step": 8736 }, { "epoch": 3.956974637681159, "grad_norm": 0.11717688723412106, "learning_rate": 1.014423423422593e-05, "loss": 0.8498, "step": 8737 }, { "epoch": 3.957427536231884, "grad_norm": 0.12390441202570712, "learning_rate": 1.01358188870875e-05, "loss": 0.8392, "step": 8738 }, { "epoch": 3.9578804347826084, "grad_norm": 0.11497402666064856, "learning_rate": 1.0127406525518584e-05, "loss": 0.8766, "step": 8739 }, { "epoch": 3.9583333333333335, "grad_norm": 0.12606124593702156, "learning_rate": 1.0118997150360169e-05, "loss": 0.8753, "step": 8740 }, { "epoch": 3.958786231884058, "grad_norm": 0.10521818703563818, "learning_rate": 1.011059076245295e-05, "loss": 0.8478, "step": 8741 }, { "epoch": 3.9592391304347827, "grad_norm": 0.09860586166478796, "learning_rate": 1.010218736263735e-05, "loss": 0.8773, "step": 8742 }, { "epoch": 3.9596920289855073, "grad_norm": 0.11987482133935075, "learning_rate": 1.0093786951753443e-05, "loss": 0.8794, "step": 8743 }, { "epoch": 3.960144927536232, "grad_norm": 0.12474338835094134, "learning_rate": 1.0085389530641057e-05, "loss": 0.8472, "step": 8744 }, { "epoch": 3.9605978260869565, "grad_norm": 0.1266355831369556, "learning_rate": 1.0076995100139672e-05, "loss": 0.8396, "step": 8745 }, { "epoch": 3.961050724637681, "grad_norm": 0.11131573523830537, "learning_rate": 1.006860366108852e-05, "loss": 0.8667, "step": 8746 }, { "epoch": 3.9615036231884058, "grad_norm": 0.1361574737234047, "learning_rate": 1.0060215214326483e-05, "loss": 0.8719, "step": 8747 }, { "epoch": 3.9619565217391304, "grad_norm": 0.0950472973950577, "learning_rate": 1.005182976069218e-05, "loss": 0.8517, "step": 8748 }, { "epoch": 3.962409420289855, "grad_norm": 0.11960700562357615, "learning_rate": 1.0043447301023898e-05, "loss": 0.8483, "step": 8749 }, { "epoch": 3.9628623188405796, "grad_norm": 0.10252541273751567, "learning_rate": 1.0035067836159671e-05, "loss": 0.8555, "step": 8750 }, { "epoch": 3.963315217391304, "grad_norm": 0.11662314600862281, "learning_rate": 1.0026691366937187e-05, "loss": 0.8551, "step": 8751 }, { "epoch": 3.963768115942029, "grad_norm": 0.10748350305122079, "learning_rate": 1.001831789419387e-05, "loss": 0.8641, "step": 8752 }, { "epoch": 3.964221014492754, "grad_norm": 0.09911554324099911, "learning_rate": 1.0009947418766818e-05, "loss": 0.8544, "step": 8753 }, { "epoch": 3.9646739130434785, "grad_norm": 0.1339661883301657, "learning_rate": 1.0001579941492854e-05, "loss": 0.8549, "step": 8754 }, { "epoch": 3.965126811594203, "grad_norm": 0.09652585224511774, "learning_rate": 9.993215463208479e-06, "loss": 0.8642, "step": 8755 }, { "epoch": 3.9655797101449277, "grad_norm": 0.11072282317584298, "learning_rate": 9.984853984749904e-06, "loss": 0.8606, "step": 8756 }, { "epoch": 3.9660326086956523, "grad_norm": 0.12016283999390912, "learning_rate": 9.976495506953032e-06, "loss": 0.8445, "step": 8757 }, { "epoch": 3.966485507246377, "grad_norm": 0.08598317285009817, "learning_rate": 9.96814003065349e-06, "loss": 0.8618, "step": 8758 }, { "epoch": 3.9669384057971016, "grad_norm": 0.11745478018576906, "learning_rate": 9.959787556686567e-06, "loss": 0.8497, "step": 8759 }, { "epoch": 3.967391304347826, "grad_norm": 0.09966908686377067, "learning_rate": 9.951438085887294e-06, "loss": 0.8596, "step": 8760 }, { "epoch": 3.9678442028985508, "grad_norm": 0.089763488932501, "learning_rate": 9.943091619090359e-06, "loss": 0.8421, "step": 8761 }, { "epoch": 3.9682971014492754, "grad_norm": 0.10558687137349428, "learning_rate": 9.934748157130203e-06, "loss": 0.8582, "step": 8762 }, { "epoch": 3.96875, "grad_norm": 0.09997575963434958, "learning_rate": 9.92640770084091e-06, "loss": 0.8581, "step": 8763 }, { "epoch": 3.9692028985507246, "grad_norm": 0.1063960148901854, "learning_rate": 9.918070251056297e-06, "loss": 0.8527, "step": 8764 }, { "epoch": 3.9696557971014492, "grad_norm": 0.10986116428575535, "learning_rate": 9.909735808609856e-06, "loss": 0.8527, "step": 8765 }, { "epoch": 3.970108695652174, "grad_norm": 0.10246136506190409, "learning_rate": 9.901404374334823e-06, "loss": 0.868, "step": 8766 }, { "epoch": 3.9705615942028984, "grad_norm": 0.11715756542191208, "learning_rate": 9.893075949064075e-06, "loss": 0.8503, "step": 8767 }, { "epoch": 3.971014492753623, "grad_norm": 0.104192440228047, "learning_rate": 9.884750533630249e-06, "loss": 0.867, "step": 8768 }, { "epoch": 3.9714673913043477, "grad_norm": 0.1261760929087164, "learning_rate": 9.876428128865618e-06, "loss": 0.8816, "step": 8769 }, { "epoch": 3.9719202898550723, "grad_norm": 0.11533962616273384, "learning_rate": 9.868108735602217e-06, "loss": 0.8563, "step": 8770 }, { "epoch": 3.972373188405797, "grad_norm": 0.10055183372609083, "learning_rate": 9.85979235467173e-06, "loss": 0.8721, "step": 8771 }, { "epoch": 3.9728260869565215, "grad_norm": 0.12408879100224704, "learning_rate": 9.851478986905567e-06, "loss": 0.8525, "step": 8772 }, { "epoch": 3.973278985507246, "grad_norm": 0.12620024705971336, "learning_rate": 9.84316863313481e-06, "loss": 0.8747, "step": 8773 }, { "epoch": 3.973731884057971, "grad_norm": 0.12031203219186874, "learning_rate": 9.834861294190282e-06, "loss": 0.8564, "step": 8774 }, { "epoch": 3.974184782608696, "grad_norm": 0.12295441932090573, "learning_rate": 9.826556970902464e-06, "loss": 0.8598, "step": 8775 }, { "epoch": 3.9746376811594204, "grad_norm": 0.09917650328041212, "learning_rate": 9.818255664101568e-06, "loss": 0.8561, "step": 8776 }, { "epoch": 3.975090579710145, "grad_norm": 0.11389720357922288, "learning_rate": 9.80995737461747e-06, "loss": 0.8562, "step": 8777 }, { "epoch": 3.9755434782608696, "grad_norm": 0.10639870392267427, "learning_rate": 9.801662103279783e-06, "loss": 0.8788, "step": 8778 }, { "epoch": 3.9759963768115942, "grad_norm": 0.09833061552378837, "learning_rate": 9.79336985091778e-06, "loss": 0.848, "step": 8779 }, { "epoch": 3.976449275362319, "grad_norm": 0.1308678690447716, "learning_rate": 9.785080618360468e-06, "loss": 0.8447, "step": 8780 }, { "epoch": 3.9769021739130435, "grad_norm": 0.11514779156832168, "learning_rate": 9.77679440643653e-06, "loss": 0.8784, "step": 8781 }, { "epoch": 3.977355072463768, "grad_norm": 0.10267982720877983, "learning_rate": 9.768511215974343e-06, "loss": 0.86, "step": 8782 }, { "epoch": 3.9778079710144927, "grad_norm": 0.10581726684344155, "learning_rate": 9.76023104780199e-06, "loss": 0.8423, "step": 8783 }, { "epoch": 3.9782608695652173, "grad_norm": 0.1531634813573062, "learning_rate": 9.751953902747267e-06, "loss": 0.8446, "step": 8784 }, { "epoch": 3.978713768115942, "grad_norm": 0.10719409587654966, "learning_rate": 9.743679781637634e-06, "loss": 0.8531, "step": 8785 }, { "epoch": 3.9791666666666665, "grad_norm": 0.1400733010261118, "learning_rate": 9.735408685300287e-06, "loss": 0.8638, "step": 8786 }, { "epoch": 3.9796195652173916, "grad_norm": 0.10256776541199979, "learning_rate": 9.72714061456208e-06, "loss": 0.8509, "step": 8787 }, { "epoch": 3.980072463768116, "grad_norm": 0.18697831967496256, "learning_rate": 9.718875570249615e-06, "loss": 0.868, "step": 8788 }, { "epoch": 3.980525362318841, "grad_norm": 0.11725773179484991, "learning_rate": 9.710613553189141e-06, "loss": 0.8817, "step": 8789 }, { "epoch": 3.9809782608695654, "grad_norm": 0.14978525221472758, "learning_rate": 9.702354564206625e-06, "loss": 0.8516, "step": 8790 }, { "epoch": 3.98143115942029, "grad_norm": 0.10057817021227006, "learning_rate": 9.694098604127724e-06, "loss": 0.8559, "step": 8791 }, { "epoch": 3.9818840579710146, "grad_norm": 0.1652766793498433, "learning_rate": 9.68584567377782e-06, "loss": 0.8733, "step": 8792 }, { "epoch": 3.9823369565217392, "grad_norm": 0.09982499757068965, "learning_rate": 9.677595773981951e-06, "loss": 0.8569, "step": 8793 }, { "epoch": 3.982789855072464, "grad_norm": 0.12544891531191768, "learning_rate": 9.669348905564893e-06, "loss": 0.8623, "step": 8794 }, { "epoch": 3.9832427536231885, "grad_norm": 0.11198722025971822, "learning_rate": 9.661105069351078e-06, "loss": 0.8575, "step": 8795 }, { "epoch": 3.983695652173913, "grad_norm": 0.09838160201437109, "learning_rate": 9.652864266164674e-06, "loss": 0.8536, "step": 8796 }, { "epoch": 3.9841485507246377, "grad_norm": 0.13186469933270248, "learning_rate": 9.644626496829522e-06, "loss": 0.8424, "step": 8797 }, { "epoch": 3.9846014492753623, "grad_norm": 0.10312812340128454, "learning_rate": 9.636391762169159e-06, "loss": 0.8613, "step": 8798 }, { "epoch": 3.985054347826087, "grad_norm": 0.101290343380651, "learning_rate": 9.628160063006814e-06, "loss": 0.8689, "step": 8799 }, { "epoch": 3.9855072463768115, "grad_norm": 0.09400173714257352, "learning_rate": 9.619931400165451e-06, "loss": 0.848, "step": 8800 }, { "epoch": 3.985960144927536, "grad_norm": 0.11061491631600621, "learning_rate": 9.611705774467674e-06, "loss": 0.8723, "step": 8801 }, { "epoch": 3.9864130434782608, "grad_norm": 0.10152733775111367, "learning_rate": 9.603483186735837e-06, "loss": 0.8398, "step": 8802 }, { "epoch": 3.9868659420289854, "grad_norm": 0.12585445781726562, "learning_rate": 9.595263637791938e-06, "loss": 0.8441, "step": 8803 }, { "epoch": 3.98731884057971, "grad_norm": 0.09853467640239083, "learning_rate": 9.587047128457732e-06, "loss": 0.8552, "step": 8804 }, { "epoch": 3.9877717391304346, "grad_norm": 0.09390615338444673, "learning_rate": 9.578833659554614e-06, "loss": 0.8582, "step": 8805 }, { "epoch": 3.988224637681159, "grad_norm": 0.09909641520423422, "learning_rate": 9.570623231903702e-06, "loss": 0.856, "step": 8806 }, { "epoch": 3.988677536231884, "grad_norm": 0.09705177815204998, "learning_rate": 9.562415846325805e-06, "loss": 0.8652, "step": 8807 }, { "epoch": 3.9891304347826084, "grad_norm": 0.09929808715200662, "learning_rate": 9.554211503641416e-06, "loss": 0.8577, "step": 8808 }, { "epoch": 3.9895833333333335, "grad_norm": 0.09529814653780888, "learning_rate": 9.546010204670759e-06, "loss": 0.8613, "step": 8809 }, { "epoch": 3.990036231884058, "grad_norm": 0.0983804951607374, "learning_rate": 9.53781195023371e-06, "loss": 0.8801, "step": 8810 }, { "epoch": 3.9904891304347827, "grad_norm": 0.0863651110658744, "learning_rate": 9.529616741149885e-06, "loss": 0.8454, "step": 8811 }, { "epoch": 3.9909420289855073, "grad_norm": 0.09464367686155845, "learning_rate": 9.521424578238547e-06, "loss": 0.8586, "step": 8812 }, { "epoch": 3.991394927536232, "grad_norm": 0.08945137823305427, "learning_rate": 9.5132354623187e-06, "loss": 0.853, "step": 8813 }, { "epoch": 3.9918478260869565, "grad_norm": 0.09862527690100995, "learning_rate": 9.505049394209015e-06, "loss": 0.8557, "step": 8814 }, { "epoch": 3.992300724637681, "grad_norm": 0.1152002859135843, "learning_rate": 9.49686637472787e-06, "loss": 0.866, "step": 8815 }, { "epoch": 3.9927536231884058, "grad_norm": 0.0953141556484831, "learning_rate": 9.488686404693314e-06, "loss": 0.855, "step": 8816 }, { "epoch": 3.9932065217391304, "grad_norm": 0.13045073287758904, "learning_rate": 9.48050948492314e-06, "loss": 0.8571, "step": 8817 }, { "epoch": 3.993659420289855, "grad_norm": 0.11335468729472113, "learning_rate": 9.472335616234782e-06, "loss": 0.8435, "step": 8818 }, { "epoch": 3.9941123188405796, "grad_norm": 0.11870911910190757, "learning_rate": 9.464164799445421e-06, "loss": 0.8553, "step": 8819 }, { "epoch": 3.994565217391304, "grad_norm": 0.1192137237320295, "learning_rate": 9.455997035371881e-06, "loss": 0.8555, "step": 8820 }, { "epoch": 3.995018115942029, "grad_norm": 0.09696370867287545, "learning_rate": 9.44783232483073e-06, "loss": 0.8603, "step": 8821 }, { "epoch": 3.995471014492754, "grad_norm": 0.08904958110364046, "learning_rate": 9.439670668638188e-06, "loss": 0.8549, "step": 8822 }, { "epoch": 3.9959239130434785, "grad_norm": 0.09744173193034167, "learning_rate": 9.431512067610216e-06, "loss": 0.8457, "step": 8823 }, { "epoch": 3.996376811594203, "grad_norm": 0.08637889964801407, "learning_rate": 9.423356522562398e-06, "loss": 0.8581, "step": 8824 }, { "epoch": 3.9968297101449277, "grad_norm": 0.10426694709211803, "learning_rate": 9.415204034310093e-06, "loss": 0.8583, "step": 8825 }, { "epoch": 3.9972826086956523, "grad_norm": 0.10108290424403346, "learning_rate": 9.407054603668295e-06, "loss": 0.8524, "step": 8826 }, { "epoch": 3.997735507246377, "grad_norm": 0.10554567818552438, "learning_rate": 9.39890823145174e-06, "loss": 0.8703, "step": 8827 }, { "epoch": 3.9981884057971016, "grad_norm": 0.09330840599258398, "learning_rate": 9.390764918474811e-06, "loss": 0.8749, "step": 8828 }, { "epoch": 3.998641304347826, "grad_norm": 0.10503835101916761, "learning_rate": 9.382624665551625e-06, "loss": 0.8429, "step": 8829 }, { "epoch": 3.9990942028985508, "grad_norm": 0.09193311812137804, "learning_rate": 9.374487473495959e-06, "loss": 0.8481, "step": 8830 }, { "epoch": 3.9995471014492754, "grad_norm": 0.09816381946111956, "learning_rate": 9.366353343121321e-06, "loss": 0.8596, "step": 8831 }, { "epoch": 4.0, "grad_norm": 0.10641097728257871, "learning_rate": 9.358222275240884e-06, "loss": 0.8507, "step": 8832 }, { "epoch": 4.000452898550725, "grad_norm": 0.10812071117182635, "learning_rate": 9.350094270667522e-06, "loss": 0.8387, "step": 8833 }, { "epoch": 4.000905797101449, "grad_norm": 0.10150428550917545, "learning_rate": 9.341969330213794e-06, "loss": 0.8644, "step": 8834 }, { "epoch": 4.001358695652174, "grad_norm": 0.12211619521104515, "learning_rate": 9.333847454691983e-06, "loss": 0.8556, "step": 8835 }, { "epoch": 4.0018115942028984, "grad_norm": 0.11535894949714373, "learning_rate": 9.325728644914024e-06, "loss": 0.8408, "step": 8836 }, { "epoch": 4.002264492753623, "grad_norm": 0.1094194889779793, "learning_rate": 9.317612901691597e-06, "loss": 0.8319, "step": 8837 }, { "epoch": 4.002717391304348, "grad_norm": 0.13317439406822645, "learning_rate": 9.309500225836014e-06, "loss": 0.8366, "step": 8838 }, { "epoch": 4.003170289855072, "grad_norm": 0.10074590572228012, "learning_rate": 9.301390618158334e-06, "loss": 0.8502, "step": 8839 }, { "epoch": 4.003623188405797, "grad_norm": 0.10826875658927802, "learning_rate": 9.293284079469288e-06, "loss": 0.8627, "step": 8840 }, { "epoch": 4.0040760869565215, "grad_norm": 0.13894523742204723, "learning_rate": 9.285180610579284e-06, "loss": 0.8284, "step": 8841 }, { "epoch": 4.004528985507246, "grad_norm": 0.09540667524472372, "learning_rate": 9.277080212298442e-06, "loss": 0.8296, "step": 8842 }, { "epoch": 4.004981884057971, "grad_norm": 0.10459955761550033, "learning_rate": 9.268982885436584e-06, "loss": 0.8406, "step": 8843 }, { "epoch": 4.005434782608695, "grad_norm": 0.14849642792311543, "learning_rate": 9.260888630803188e-06, "loss": 0.8464, "step": 8844 }, { "epoch": 4.00588768115942, "grad_norm": 0.09538122564076006, "learning_rate": 9.252797449207485e-06, "loss": 0.8232, "step": 8845 }, { "epoch": 4.006340579710145, "grad_norm": 0.13756952940772316, "learning_rate": 9.244709341458327e-06, "loss": 0.8592, "step": 8846 }, { "epoch": 4.006793478260869, "grad_norm": 0.1237729085655787, "learning_rate": 9.236624308364326e-06, "loss": 0.8717, "step": 8847 }, { "epoch": 4.007246376811594, "grad_norm": 0.12096333522090248, "learning_rate": 9.228542350733743e-06, "loss": 0.8346, "step": 8848 }, { "epoch": 4.007699275362318, "grad_norm": 0.14588113155866306, "learning_rate": 9.220463469374544e-06, "loss": 0.8489, "step": 8849 }, { "epoch": 4.008152173913044, "grad_norm": 0.1130610307528198, "learning_rate": 9.212387665094371e-06, "loss": 0.8502, "step": 8850 }, { "epoch": 4.0086050724637685, "grad_norm": 0.12298119971231021, "learning_rate": 9.204314938700602e-06, "loss": 0.8639, "step": 8851 }, { "epoch": 4.009057971014493, "grad_norm": 0.09661415036367883, "learning_rate": 9.19624529100026e-06, "loss": 0.8329, "step": 8852 }, { "epoch": 4.009510869565218, "grad_norm": 0.10216994436832912, "learning_rate": 9.188178722800095e-06, "loss": 0.8451, "step": 8853 }, { "epoch": 4.009963768115942, "grad_norm": 0.12586517254033824, "learning_rate": 9.180115234906522e-06, "loss": 0.868, "step": 8854 }, { "epoch": 4.010416666666667, "grad_norm": 0.0973612227243833, "learning_rate": 9.172054828125678e-06, "loss": 0.8594, "step": 8855 }, { "epoch": 4.010869565217392, "grad_norm": 0.14533544575553742, "learning_rate": 9.163997503263364e-06, "loss": 0.8678, "step": 8856 }, { "epoch": 4.011322463768116, "grad_norm": 0.11943889213167758, "learning_rate": 9.155943261125081e-06, "loss": 0.8657, "step": 8857 }, { "epoch": 4.011775362318841, "grad_norm": 0.11173646990709532, "learning_rate": 9.147892102516019e-06, "loss": 0.8486, "step": 8858 }, { "epoch": 4.012228260869565, "grad_norm": 0.11469257549165181, "learning_rate": 9.13984402824108e-06, "loss": 0.8692, "step": 8859 }, { "epoch": 4.01268115942029, "grad_norm": 0.09129878871163112, "learning_rate": 9.131799039104825e-06, "loss": 0.8446, "step": 8860 }, { "epoch": 4.013134057971015, "grad_norm": 0.13310109614819024, "learning_rate": 9.12375713591155e-06, "loss": 0.847, "step": 8861 }, { "epoch": 4.013586956521739, "grad_norm": 0.09831292229975112, "learning_rate": 9.115718319465187e-06, "loss": 0.8618, "step": 8862 }, { "epoch": 4.014039855072464, "grad_norm": 0.11062876976405613, "learning_rate": 9.107682590569418e-06, "loss": 0.8336, "step": 8863 }, { "epoch": 4.0144927536231885, "grad_norm": 0.11319110560377048, "learning_rate": 9.099649950027572e-06, "loss": 0.87, "step": 8864 }, { "epoch": 4.014945652173913, "grad_norm": 0.10302922447280499, "learning_rate": 9.091620398642686e-06, "loss": 0.8399, "step": 8865 }, { "epoch": 4.015398550724638, "grad_norm": 0.12300426269322526, "learning_rate": 9.083593937217477e-06, "loss": 0.8641, "step": 8866 }, { "epoch": 4.015851449275362, "grad_norm": 0.0977228313974661, "learning_rate": 9.075570566554383e-06, "loss": 0.8425, "step": 8867 }, { "epoch": 4.016304347826087, "grad_norm": 0.10829615615976342, "learning_rate": 9.067550287455495e-06, "loss": 0.8538, "step": 8868 }, { "epoch": 4.0167572463768115, "grad_norm": 0.12545118056160717, "learning_rate": 9.059533100722628e-06, "loss": 0.8324, "step": 8869 }, { "epoch": 4.017210144927536, "grad_norm": 0.11706914292549836, "learning_rate": 9.051519007157257e-06, "loss": 0.8552, "step": 8870 }, { "epoch": 4.017663043478261, "grad_norm": 0.10491967145237562, "learning_rate": 9.043508007560588e-06, "loss": 0.8416, "step": 8871 }, { "epoch": 4.018115942028985, "grad_norm": 0.11177997900866397, "learning_rate": 9.035500102733463e-06, "loss": 0.8443, "step": 8872 }, { "epoch": 4.01856884057971, "grad_norm": 0.10899692833933687, "learning_rate": 9.027495293476471e-06, "loss": 0.851, "step": 8873 }, { "epoch": 4.019021739130435, "grad_norm": 0.12567614959268392, "learning_rate": 9.019493580589858e-06, "loss": 0.8754, "step": 8874 }, { "epoch": 4.019474637681159, "grad_norm": 0.10515526532801933, "learning_rate": 9.011494964873564e-06, "loss": 0.8518, "step": 8875 }, { "epoch": 4.019927536231884, "grad_norm": 0.116707404220969, "learning_rate": 9.003499447127209e-06, "loss": 0.8524, "step": 8876 }, { "epoch": 4.020380434782608, "grad_norm": 0.09367410281329598, "learning_rate": 8.995507028150147e-06, "loss": 0.8341, "step": 8877 }, { "epoch": 4.020833333333333, "grad_norm": 0.11673692937263203, "learning_rate": 8.987517708741364e-06, "loss": 0.8764, "step": 8878 }, { "epoch": 4.021286231884058, "grad_norm": 0.09834679256218098, "learning_rate": 8.979531489699594e-06, "loss": 0.8602, "step": 8879 }, { "epoch": 4.021739130434782, "grad_norm": 0.11085823478305495, "learning_rate": 8.971548371823205e-06, "loss": 0.8469, "step": 8880 }, { "epoch": 4.022192028985507, "grad_norm": 0.10618921959098696, "learning_rate": 8.963568355910301e-06, "loss": 0.8594, "step": 8881 }, { "epoch": 4.0226449275362315, "grad_norm": 0.10911402042948654, "learning_rate": 8.955591442758655e-06, "loss": 0.8314, "step": 8882 }, { "epoch": 4.023097826086956, "grad_norm": 0.09119335600955895, "learning_rate": 8.947617633165726e-06, "loss": 0.832, "step": 8883 }, { "epoch": 4.023550724637682, "grad_norm": 0.12877887409802222, "learning_rate": 8.939646927928659e-06, "loss": 0.8319, "step": 8884 }, { "epoch": 4.024003623188406, "grad_norm": 0.09770192346724514, "learning_rate": 8.931679327844316e-06, "loss": 0.8585, "step": 8885 }, { "epoch": 4.024456521739131, "grad_norm": 0.12644214837096932, "learning_rate": 8.92371483370921e-06, "loss": 0.8413, "step": 8886 }, { "epoch": 4.024909420289855, "grad_norm": 0.09270490322803848, "learning_rate": 8.91575344631959e-06, "loss": 0.8489, "step": 8887 }, { "epoch": 4.02536231884058, "grad_norm": 0.12541118735135867, "learning_rate": 8.907795166471342e-06, "loss": 0.8642, "step": 8888 }, { "epoch": 4.025815217391305, "grad_norm": 0.10991977053164051, "learning_rate": 8.899839994960095e-06, "loss": 0.859, "step": 8889 }, { "epoch": 4.026268115942029, "grad_norm": 0.10465408075404262, "learning_rate": 8.891887932581128e-06, "loss": 0.8409, "step": 8890 }, { "epoch": 4.026721014492754, "grad_norm": 0.1088769221370972, "learning_rate": 8.883938980129417e-06, "loss": 0.8485, "step": 8891 }, { "epoch": 4.0271739130434785, "grad_norm": 0.10658464467781935, "learning_rate": 8.875993138399624e-06, "loss": 0.8435, "step": 8892 }, { "epoch": 4.027626811594203, "grad_norm": 0.11630692874818807, "learning_rate": 8.868050408186125e-06, "loss": 0.8312, "step": 8893 }, { "epoch": 4.028079710144928, "grad_norm": 0.12088182829473283, "learning_rate": 8.86011079028295e-06, "loss": 0.8472, "step": 8894 }, { "epoch": 4.028532608695652, "grad_norm": 0.10306640578105815, "learning_rate": 8.852174285483856e-06, "loss": 0.8608, "step": 8895 }, { "epoch": 4.028985507246377, "grad_norm": 0.11313823509223213, "learning_rate": 8.844240894582245e-06, "loss": 0.8652, "step": 8896 }, { "epoch": 4.0294384057971016, "grad_norm": 0.1010628839129015, "learning_rate": 8.836310618371256e-06, "loss": 0.8722, "step": 8897 }, { "epoch": 4.029891304347826, "grad_norm": 0.10898963988493458, "learning_rate": 8.828383457643675e-06, "loss": 0.8637, "step": 8898 }, { "epoch": 4.030344202898551, "grad_norm": 0.11224219348165553, "learning_rate": 8.820459413191993e-06, "loss": 0.8523, "step": 8899 }, { "epoch": 4.030797101449275, "grad_norm": 0.1246139127775101, "learning_rate": 8.812538485808386e-06, "loss": 0.8418, "step": 8900 }, { "epoch": 4.03125, "grad_norm": 0.11070116122044553, "learning_rate": 8.804620676284736e-06, "loss": 0.8311, "step": 8901 }, { "epoch": 4.031702898550725, "grad_norm": 0.1112190389242049, "learning_rate": 8.79670598541258e-06, "loss": 0.8561, "step": 8902 }, { "epoch": 4.032155797101449, "grad_norm": 0.10758584880589765, "learning_rate": 8.788794413983179e-06, "loss": 0.8276, "step": 8903 }, { "epoch": 4.032608695652174, "grad_norm": 0.0990195591063918, "learning_rate": 8.780885962787452e-06, "loss": 0.8484, "step": 8904 }, { "epoch": 4.0330615942028984, "grad_norm": 0.10481472284770944, "learning_rate": 8.772980632616037e-06, "loss": 0.8554, "step": 8905 }, { "epoch": 4.033514492753623, "grad_norm": 0.10384372356742821, "learning_rate": 8.76507842425923e-06, "loss": 0.8348, "step": 8906 }, { "epoch": 4.033967391304348, "grad_norm": 0.09335946345737046, "learning_rate": 8.757179338507033e-06, "loss": 0.8254, "step": 8907 }, { "epoch": 4.034420289855072, "grad_norm": 0.13085648650304274, "learning_rate": 8.749283376149108e-06, "loss": 0.8702, "step": 8908 }, { "epoch": 4.034873188405797, "grad_norm": 0.08941464693556588, "learning_rate": 8.741390537974865e-06, "loss": 0.8448, "step": 8909 }, { "epoch": 4.0353260869565215, "grad_norm": 0.10154265632972032, "learning_rate": 8.733500824773325e-06, "loss": 0.8444, "step": 8910 }, { "epoch": 4.035778985507246, "grad_norm": 0.12149080435365568, "learning_rate": 8.725614237333268e-06, "loss": 0.8507, "step": 8911 }, { "epoch": 4.036231884057971, "grad_norm": 0.10714396739221282, "learning_rate": 8.717730776443104e-06, "loss": 0.8304, "step": 8912 }, { "epoch": 4.036684782608695, "grad_norm": 0.09496500443244973, "learning_rate": 8.709850442890974e-06, "loss": 0.8443, "step": 8913 }, { "epoch": 4.03713768115942, "grad_norm": 0.10227426231448344, "learning_rate": 8.701973237464667e-06, "loss": 0.8543, "step": 8914 }, { "epoch": 4.037590579710145, "grad_norm": 0.10657213845698915, "learning_rate": 8.694099160951706e-06, "loss": 0.8448, "step": 8915 }, { "epoch": 4.038043478260869, "grad_norm": 0.09492597486036246, "learning_rate": 8.686228214139256e-06, "loss": 0.8637, "step": 8916 }, { "epoch": 4.038496376811594, "grad_norm": 0.11037379995920758, "learning_rate": 8.678360397814196e-06, "loss": 0.8575, "step": 8917 }, { "epoch": 4.038949275362318, "grad_norm": 0.1037179047273683, "learning_rate": 8.670495712763066e-06, "loss": 0.8636, "step": 8918 }, { "epoch": 4.039402173913044, "grad_norm": 0.09533923192412404, "learning_rate": 8.66263415977214e-06, "loss": 0.8623, "step": 8919 }, { "epoch": 4.0398550724637685, "grad_norm": 0.1233653123180593, "learning_rate": 8.654775739627324e-06, "loss": 0.8499, "step": 8920 }, { "epoch": 4.040307971014493, "grad_norm": 0.10284127450435368, "learning_rate": 8.646920453114256e-06, "loss": 0.8705, "step": 8921 }, { "epoch": 4.040760869565218, "grad_norm": 0.11941142495045273, "learning_rate": 8.639068301018226e-06, "loss": 0.8729, "step": 8922 }, { "epoch": 4.041213768115942, "grad_norm": 0.11312066720806964, "learning_rate": 8.631219284124239e-06, "loss": 0.8527, "step": 8923 }, { "epoch": 4.041666666666667, "grad_norm": 0.10691660506917093, "learning_rate": 8.623373403216972e-06, "loss": 0.8737, "step": 8924 }, { "epoch": 4.042119565217392, "grad_norm": 0.09833191033928969, "learning_rate": 8.615530659080788e-06, "loss": 0.8712, "step": 8925 }, { "epoch": 4.042572463768116, "grad_norm": 0.12738311778716235, "learning_rate": 8.607691052499726e-06, "loss": 0.8334, "step": 8926 }, { "epoch": 4.043025362318841, "grad_norm": 0.10118835611273354, "learning_rate": 8.599854584257543e-06, "loss": 0.8294, "step": 8927 }, { "epoch": 4.043478260869565, "grad_norm": 0.126297937988198, "learning_rate": 8.592021255137646e-06, "loss": 0.8485, "step": 8928 }, { "epoch": 4.04393115942029, "grad_norm": 0.11298853464650141, "learning_rate": 8.584191065923163e-06, "loss": 0.8597, "step": 8929 }, { "epoch": 4.044384057971015, "grad_norm": 0.09881808034180857, "learning_rate": 8.576364017396886e-06, "loss": 0.8427, "step": 8930 }, { "epoch": 4.044836956521739, "grad_norm": 0.0921107353124904, "learning_rate": 8.568540110341282e-06, "loss": 0.8391, "step": 8931 }, { "epoch": 4.045289855072464, "grad_norm": 0.09712445914326226, "learning_rate": 8.560719345538535e-06, "loss": 0.8576, "step": 8932 }, { "epoch": 4.0457427536231885, "grad_norm": 0.10922700063805445, "learning_rate": 8.552901723770502e-06, "loss": 0.8558, "step": 8933 }, { "epoch": 4.046195652173913, "grad_norm": 0.09295273873179893, "learning_rate": 8.54508724581871e-06, "loss": 0.8403, "step": 8934 }, { "epoch": 4.046648550724638, "grad_norm": 0.11438017827330843, "learning_rate": 8.537275912464383e-06, "loss": 0.8508, "step": 8935 }, { "epoch": 4.047101449275362, "grad_norm": 0.1223498322854627, "learning_rate": 8.529467724488452e-06, "loss": 0.8502, "step": 8936 }, { "epoch": 4.047554347826087, "grad_norm": 0.09789033537804591, "learning_rate": 8.52166268267149e-06, "loss": 0.846, "step": 8937 }, { "epoch": 4.0480072463768115, "grad_norm": 0.0941807361235538, "learning_rate": 8.513860787793801e-06, "loss": 0.8467, "step": 8938 }, { "epoch": 4.048460144927536, "grad_norm": 0.10834975867684066, "learning_rate": 8.506062040635332e-06, "loss": 0.8358, "step": 8939 }, { "epoch": 4.048913043478261, "grad_norm": 0.10792267743919434, "learning_rate": 8.49826644197576e-06, "loss": 0.858, "step": 8940 }, { "epoch": 4.049365942028985, "grad_norm": 0.1040618243620686, "learning_rate": 8.490473992594413e-06, "loss": 0.8677, "step": 8941 }, { "epoch": 4.04981884057971, "grad_norm": 0.13719904215430795, "learning_rate": 8.482684693270306e-06, "loss": 0.8557, "step": 8942 }, { "epoch": 4.050271739130435, "grad_norm": 0.09447737967930533, "learning_rate": 8.474898544782144e-06, "loss": 0.8312, "step": 8943 }, { "epoch": 4.050724637681159, "grad_norm": 0.1128666707666903, "learning_rate": 8.467115547908348e-06, "loss": 0.8315, "step": 8944 }, { "epoch": 4.051177536231884, "grad_norm": 0.1251677538352005, "learning_rate": 8.459335703426959e-06, "loss": 0.8569, "step": 8945 }, { "epoch": 4.051630434782608, "grad_norm": 0.11582469327352536, "learning_rate": 8.451559012115775e-06, "loss": 0.8622, "step": 8946 }, { "epoch": 4.052083333333333, "grad_norm": 0.11697097793006934, "learning_rate": 8.44378547475222e-06, "loss": 0.8611, "step": 8947 }, { "epoch": 4.052536231884058, "grad_norm": 0.1296384236759848, "learning_rate": 8.436015092113443e-06, "loss": 0.8545, "step": 8948 }, { "epoch": 4.052989130434782, "grad_norm": 0.0976497304059224, "learning_rate": 8.42824786497626e-06, "loss": 0.8535, "step": 8949 }, { "epoch": 4.053442028985507, "grad_norm": 0.09879079081024708, "learning_rate": 8.420483794117165e-06, "loss": 0.8634, "step": 8950 }, { "epoch": 4.0538949275362315, "grad_norm": 0.11702184232929158, "learning_rate": 8.412722880312331e-06, "loss": 0.8707, "step": 8951 }, { "epoch": 4.054347826086956, "grad_norm": 0.08976438327862242, "learning_rate": 8.404965124337665e-06, "loss": 0.8407, "step": 8952 }, { "epoch": 4.054800724637682, "grad_norm": 0.08823066333730303, "learning_rate": 8.397210526968687e-06, "loss": 0.8597, "step": 8953 }, { "epoch": 4.055253623188406, "grad_norm": 0.09646988606712303, "learning_rate": 8.38945908898066e-06, "loss": 0.8477, "step": 8954 }, { "epoch": 4.055706521739131, "grad_norm": 0.09515580847877236, "learning_rate": 8.381710811148492e-06, "loss": 0.8627, "step": 8955 }, { "epoch": 4.056159420289855, "grad_norm": 0.09789791349614982, "learning_rate": 8.373965694246809e-06, "loss": 0.8562, "step": 8956 }, { "epoch": 4.05661231884058, "grad_norm": 0.09122595613743929, "learning_rate": 8.366223739049876e-06, "loss": 0.8381, "step": 8957 }, { "epoch": 4.057065217391305, "grad_norm": 0.13463974729607597, "learning_rate": 8.3584849463317e-06, "loss": 0.8498, "step": 8958 }, { "epoch": 4.057518115942029, "grad_norm": 0.08275575466337289, "learning_rate": 8.35074931686592e-06, "loss": 0.849, "step": 8959 }, { "epoch": 4.057971014492754, "grad_norm": 0.12762745497442488, "learning_rate": 8.343016851425885e-06, "loss": 0.8585, "step": 8960 }, { "epoch": 4.0584239130434785, "grad_norm": 0.09523279371739388, "learning_rate": 8.33528755078461e-06, "loss": 0.8508, "step": 8961 }, { "epoch": 4.058876811594203, "grad_norm": 0.11969991572495614, "learning_rate": 8.327561415714824e-06, "loss": 0.8539, "step": 8962 }, { "epoch": 4.059329710144928, "grad_norm": 0.09510365171096126, "learning_rate": 8.319838446988902e-06, "loss": 0.8555, "step": 8963 }, { "epoch": 4.059782608695652, "grad_norm": 0.09702145961224086, "learning_rate": 8.312118645378943e-06, "loss": 0.8679, "step": 8964 }, { "epoch": 4.060235507246377, "grad_norm": 0.11473693006647898, "learning_rate": 8.30440201165669e-06, "loss": 0.8528, "step": 8965 }, { "epoch": 4.0606884057971016, "grad_norm": 0.09839325337567234, "learning_rate": 8.2966885465936e-06, "loss": 0.8618, "step": 8966 }, { "epoch": 4.061141304347826, "grad_norm": 0.1168247304746081, "learning_rate": 8.2889782509608e-06, "loss": 0.8269, "step": 8967 }, { "epoch": 4.061594202898551, "grad_norm": 0.08490938243110582, "learning_rate": 8.281271125529087e-06, "loss": 0.8467, "step": 8968 }, { "epoch": 4.062047101449275, "grad_norm": 0.10293613472243991, "learning_rate": 8.273567171068957e-06, "loss": 0.8608, "step": 8969 }, { "epoch": 4.0625, "grad_norm": 0.11593468265191992, "learning_rate": 8.265866388350598e-06, "loss": 0.8389, "step": 8970 }, { "epoch": 4.062952898550725, "grad_norm": 0.0979751719494509, "learning_rate": 8.258168778143858e-06, "loss": 0.8437, "step": 8971 }, { "epoch": 4.063405797101449, "grad_norm": 0.10737490386957042, "learning_rate": 8.250474341218292e-06, "loss": 0.8476, "step": 8972 }, { "epoch": 4.063858695652174, "grad_norm": 0.09370915969375222, "learning_rate": 8.242783078343106e-06, "loss": 0.8551, "step": 8973 }, { "epoch": 4.0643115942028984, "grad_norm": 0.0979902362361842, "learning_rate": 8.235094990287234e-06, "loss": 0.8435, "step": 8974 }, { "epoch": 4.064764492753623, "grad_norm": 0.0973826685900336, "learning_rate": 8.227410077819247e-06, "loss": 0.8429, "step": 8975 }, { "epoch": 4.065217391304348, "grad_norm": 0.1025868108889731, "learning_rate": 8.219728341707425e-06, "loss": 0.8557, "step": 8976 }, { "epoch": 4.065670289855072, "grad_norm": 0.08503865414512102, "learning_rate": 8.212049782719708e-06, "loss": 0.8372, "step": 8977 }, { "epoch": 4.066123188405797, "grad_norm": 0.1085301293174432, "learning_rate": 8.20437440162376e-06, "loss": 0.8632, "step": 8978 }, { "epoch": 4.0665760869565215, "grad_norm": 0.09104737798493504, "learning_rate": 8.196702199186873e-06, "loss": 0.8612, "step": 8979 }, { "epoch": 4.067028985507246, "grad_norm": 0.09305773761320912, "learning_rate": 8.18903317617608e-06, "loss": 0.8615, "step": 8980 }, { "epoch": 4.067481884057971, "grad_norm": 0.10226698417889991, "learning_rate": 8.181367333358032e-06, "loss": 0.8349, "step": 8981 }, { "epoch": 4.067934782608695, "grad_norm": 0.09252066661985871, "learning_rate": 8.17370467149913e-06, "loss": 0.8549, "step": 8982 }, { "epoch": 4.06838768115942, "grad_norm": 0.10404881508676872, "learning_rate": 8.1660451913654e-06, "loss": 0.836, "step": 8983 }, { "epoch": 4.068840579710145, "grad_norm": 0.09111178031663378, "learning_rate": 8.15838889372258e-06, "loss": 0.8518, "step": 8984 }, { "epoch": 4.069293478260869, "grad_norm": 0.11269842509043564, "learning_rate": 8.150735779336067e-06, "loss": 0.8412, "step": 8985 }, { "epoch": 4.069746376811594, "grad_norm": 0.09251968857523953, "learning_rate": 8.14308584897098e-06, "loss": 0.8417, "step": 8986 }, { "epoch": 4.070199275362318, "grad_norm": 0.10941861493876297, "learning_rate": 8.135439103392074e-06, "loss": 0.8382, "step": 8987 }, { "epoch": 4.070652173913044, "grad_norm": 0.0887241647143654, "learning_rate": 8.127795543363821e-06, "loss": 0.8692, "step": 8988 }, { "epoch": 4.0711050724637685, "grad_norm": 0.09965169390897458, "learning_rate": 8.120155169650346e-06, "loss": 0.841, "step": 8989 }, { "epoch": 4.071557971014493, "grad_norm": 0.1054135693498461, "learning_rate": 8.112517983015485e-06, "loss": 0.8538, "step": 8990 }, { "epoch": 4.072010869565218, "grad_norm": 0.09263037391551399, "learning_rate": 8.10488398422273e-06, "loss": 0.84, "step": 8991 }, { "epoch": 4.072463768115942, "grad_norm": 0.1225896001960304, "learning_rate": 8.097253174035268e-06, "loss": 0.8737, "step": 8992 }, { "epoch": 4.072916666666667, "grad_norm": 0.09696603077286148, "learning_rate": 8.089625553215947e-06, "loss": 0.839, "step": 8993 }, { "epoch": 4.073369565217392, "grad_norm": 0.110344952847156, "learning_rate": 8.08200112252734e-06, "loss": 0.844, "step": 8994 }, { "epoch": 4.073822463768116, "grad_norm": 0.09551896008975944, "learning_rate": 8.074379882731645e-06, "loss": 0.8367, "step": 8995 }, { "epoch": 4.074275362318841, "grad_norm": 0.12225990544501594, "learning_rate": 8.066761834590794e-06, "loss": 0.8567, "step": 8996 }, { "epoch": 4.074728260869565, "grad_norm": 0.09338190401665025, "learning_rate": 8.059146978866358e-06, "loss": 0.8563, "step": 8997 }, { "epoch": 4.07518115942029, "grad_norm": 0.11812948595439055, "learning_rate": 8.051535316319622e-06, "loss": 0.8596, "step": 8998 }, { "epoch": 4.075634057971015, "grad_norm": 0.10621323600276605, "learning_rate": 8.043926847711518e-06, "loss": 0.8558, "step": 8999 }, { "epoch": 4.076086956521739, "grad_norm": 0.09094007246224024, "learning_rate": 8.036321573802706e-06, "loss": 0.8635, "step": 9000 }, { "epoch": 4.076539855072464, "grad_norm": 0.11937560570025155, "learning_rate": 8.028719495353457e-06, "loss": 0.8574, "step": 9001 }, { "epoch": 4.0769927536231885, "grad_norm": 0.09915005674125213, "learning_rate": 8.021120613123789e-06, "loss": 0.8443, "step": 9002 }, { "epoch": 4.077445652173913, "grad_norm": 0.1101189605525426, "learning_rate": 8.01352492787336e-06, "loss": 0.8573, "step": 9003 }, { "epoch": 4.077898550724638, "grad_norm": 0.09955720223802121, "learning_rate": 8.005932440361541e-06, "loss": 0.8611, "step": 9004 }, { "epoch": 4.078351449275362, "grad_norm": 0.08531963837818866, "learning_rate": 7.998343151347346e-06, "loss": 0.8523, "step": 9005 }, { "epoch": 4.078804347826087, "grad_norm": 0.0905002426637731, "learning_rate": 7.990757061589508e-06, "loss": 0.8557, "step": 9006 }, { "epoch": 4.0792572463768115, "grad_norm": 0.08640899605814217, "learning_rate": 7.983174171846397e-06, "loss": 0.8318, "step": 9007 }, { "epoch": 4.079710144927536, "grad_norm": 0.0874415446590857, "learning_rate": 7.975594482876113e-06, "loss": 0.8481, "step": 9008 }, { "epoch": 4.080163043478261, "grad_norm": 0.08306937184847134, "learning_rate": 7.968017995436393e-06, "loss": 0.8387, "step": 9009 }, { "epoch": 4.080615942028985, "grad_norm": 0.09376160508025205, "learning_rate": 7.960444710284671e-06, "loss": 0.8593, "step": 9010 }, { "epoch": 4.08106884057971, "grad_norm": 0.0825752021054964, "learning_rate": 7.952874628178055e-06, "loss": 0.8619, "step": 9011 }, { "epoch": 4.081521739130435, "grad_norm": 0.09929192005028346, "learning_rate": 7.945307749873356e-06, "loss": 0.8383, "step": 9012 }, { "epoch": 4.081974637681159, "grad_norm": 0.08570988152619102, "learning_rate": 7.937744076127022e-06, "loss": 0.8585, "step": 9013 }, { "epoch": 4.082427536231884, "grad_norm": 0.10242337492821267, "learning_rate": 7.930183607695237e-06, "loss": 0.8558, "step": 9014 }, { "epoch": 4.082880434782608, "grad_norm": 0.09217737283517241, "learning_rate": 7.922626345333797e-06, "loss": 0.8647, "step": 9015 }, { "epoch": 4.083333333333333, "grad_norm": 0.10783101098289191, "learning_rate": 7.915072289798247e-06, "loss": 0.8511, "step": 9016 }, { "epoch": 4.083786231884058, "grad_norm": 0.11258530287108579, "learning_rate": 7.907521441843764e-06, "loss": 0.8582, "step": 9017 }, { "epoch": 4.084239130434782, "grad_norm": 0.09985093529122396, "learning_rate": 7.899973802225216e-06, "loss": 0.8381, "step": 9018 }, { "epoch": 4.084692028985507, "grad_norm": 0.08612259054020847, "learning_rate": 7.892429371697146e-06, "loss": 0.8642, "step": 9019 }, { "epoch": 4.0851449275362315, "grad_norm": 0.13199590589590737, "learning_rate": 7.8848881510138e-06, "loss": 0.85, "step": 9020 }, { "epoch": 4.085597826086956, "grad_norm": 0.09659359320851041, "learning_rate": 7.87735014092906e-06, "loss": 0.8592, "step": 9021 }, { "epoch": 4.086050724637682, "grad_norm": 0.1412756747360551, "learning_rate": 7.869815342196547e-06, "loss": 0.8385, "step": 9022 }, { "epoch": 4.086503623188406, "grad_norm": 0.07771146238385142, "learning_rate": 7.862283755569496e-06, "loss": 0.8461, "step": 9023 }, { "epoch": 4.086956521739131, "grad_norm": 0.11825364916678116, "learning_rate": 7.854755381800872e-06, "loss": 0.8428, "step": 9024 }, { "epoch": 4.087409420289855, "grad_norm": 0.12965890210090153, "learning_rate": 7.847230221643296e-06, "loss": 0.8392, "step": 9025 }, { "epoch": 4.08786231884058, "grad_norm": 0.11615407404464566, "learning_rate": 7.83970827584906e-06, "loss": 0.8613, "step": 9026 }, { "epoch": 4.088315217391305, "grad_norm": 0.11708395813028508, "learning_rate": 7.832189545170142e-06, "loss": 0.836, "step": 9027 }, { "epoch": 4.088768115942029, "grad_norm": 0.1260561638973568, "learning_rate": 7.824674030358222e-06, "loss": 0.8467, "step": 9028 }, { "epoch": 4.089221014492754, "grad_norm": 0.13636243457437006, "learning_rate": 7.81716173216461e-06, "loss": 0.8399, "step": 9029 }, { "epoch": 4.0896739130434785, "grad_norm": 0.1252403622467168, "learning_rate": 7.809652651340353e-06, "loss": 0.8472, "step": 9030 }, { "epoch": 4.090126811594203, "grad_norm": 0.124835908218541, "learning_rate": 7.802146788636121e-06, "loss": 0.8398, "step": 9031 }, { "epoch": 4.090579710144928, "grad_norm": 0.11070445366447512, "learning_rate": 7.794644144802302e-06, "loss": 0.8561, "step": 9032 }, { "epoch": 4.091032608695652, "grad_norm": 0.13449658373942422, "learning_rate": 7.787144720588942e-06, "loss": 0.8497, "step": 9033 }, { "epoch": 4.091485507246377, "grad_norm": 0.13683028549721435, "learning_rate": 7.779648516745775e-06, "loss": 0.8659, "step": 9034 }, { "epoch": 4.0919384057971016, "grad_norm": 0.1532754319646102, "learning_rate": 7.772155534022188e-06, "loss": 0.8562, "step": 9035 }, { "epoch": 4.092391304347826, "grad_norm": 0.13535555734696042, "learning_rate": 7.764665773167293e-06, "loss": 0.856, "step": 9036 }, { "epoch": 4.092844202898551, "grad_norm": 0.17398254452659376, "learning_rate": 7.757179234929833e-06, "loss": 0.8475, "step": 9037 }, { "epoch": 4.093297101449275, "grad_norm": 0.15685051203289777, "learning_rate": 7.749695920058266e-06, "loss": 0.8545, "step": 9038 }, { "epoch": 4.09375, "grad_norm": 0.1825133036005155, "learning_rate": 7.742215829300695e-06, "loss": 0.8452, "step": 9039 }, { "epoch": 4.094202898550725, "grad_norm": 0.14044215015379624, "learning_rate": 7.734738963404932e-06, "loss": 0.8545, "step": 9040 }, { "epoch": 4.094655797101449, "grad_norm": 0.1624151700508136, "learning_rate": 7.727265323118445e-06, "loss": 0.8494, "step": 9041 }, { "epoch": 4.095108695652174, "grad_norm": 0.18387666479995146, "learning_rate": 7.719794909188385e-06, "loss": 0.8511, "step": 9042 }, { "epoch": 4.0955615942028984, "grad_norm": 0.12959773893732865, "learning_rate": 7.712327722361567e-06, "loss": 0.8325, "step": 9043 }, { "epoch": 4.096014492753623, "grad_norm": 0.12330386996614644, "learning_rate": 7.704863763384526e-06, "loss": 0.8619, "step": 9044 }, { "epoch": 4.096467391304348, "grad_norm": 0.1301146928093126, "learning_rate": 7.697403033003419e-06, "loss": 0.8659, "step": 9045 }, { "epoch": 4.096920289855072, "grad_norm": 0.17453518272208787, "learning_rate": 7.689945531964125e-06, "loss": 0.8354, "step": 9046 }, { "epoch": 4.097373188405797, "grad_norm": 0.1270512099077989, "learning_rate": 7.682491261012171e-06, "loss": 0.843, "step": 9047 }, { "epoch": 4.0978260869565215, "grad_norm": 0.1936174146860638, "learning_rate": 7.675040220892782e-06, "loss": 0.8417, "step": 9048 }, { "epoch": 4.098278985507246, "grad_norm": 0.14862619413771436, "learning_rate": 7.667592412350839e-06, "loss": 0.8461, "step": 9049 }, { "epoch": 4.098731884057971, "grad_norm": 0.19661878117540452, "learning_rate": 7.660147836130928e-06, "loss": 0.8586, "step": 9050 }, { "epoch": 4.099184782608695, "grad_norm": 0.14085801012762675, "learning_rate": 7.652706492977282e-06, "loss": 0.8429, "step": 9051 }, { "epoch": 4.09963768115942, "grad_norm": 0.17730387555284002, "learning_rate": 7.645268383633828e-06, "loss": 0.8573, "step": 9052 }, { "epoch": 4.100090579710145, "grad_norm": 0.15975527420969107, "learning_rate": 7.637833508844164e-06, "loss": 0.8563, "step": 9053 }, { "epoch": 4.100543478260869, "grad_norm": 0.2350979909625802, "learning_rate": 7.630401869351556e-06, "loss": 0.8642, "step": 9054 }, { "epoch": 4.100996376811594, "grad_norm": 0.15343466468622768, "learning_rate": 7.622973465898979e-06, "loss": 0.86, "step": 9055 }, { "epoch": 4.101449275362318, "grad_norm": 0.26710695958356, "learning_rate": 7.615548299229036e-06, "loss": 0.8405, "step": 9056 }, { "epoch": 4.101902173913044, "grad_norm": 0.11263617350033955, "learning_rate": 7.608126370084061e-06, "loss": 0.8459, "step": 9057 }, { "epoch": 4.1023550724637685, "grad_norm": 0.268970768082865, "learning_rate": 7.600707679206012e-06, "loss": 0.8578, "step": 9058 }, { "epoch": 4.102807971014493, "grad_norm": 0.10817298831338142, "learning_rate": 7.5932922273365665e-06, "loss": 0.8424, "step": 9059 }, { "epoch": 4.103260869565218, "grad_norm": 0.19748228990904926, "learning_rate": 7.585880015217051e-06, "loss": 0.8467, "step": 9060 }, { "epoch": 4.103713768115942, "grad_norm": 0.09915308201708638, "learning_rate": 7.578471043588473e-06, "loss": 0.8362, "step": 9061 }, { "epoch": 4.104166666666667, "grad_norm": 0.24937877547351514, "learning_rate": 7.5710653131915125e-06, "loss": 0.8613, "step": 9062 }, { "epoch": 4.104619565217392, "grad_norm": 0.11060614417747729, "learning_rate": 7.563662824766549e-06, "loss": 0.8459, "step": 9063 }, { "epoch": 4.105072463768116, "grad_norm": 0.1889222392991746, "learning_rate": 7.556263579053605e-06, "loss": 0.8359, "step": 9064 }, { "epoch": 4.105525362318841, "grad_norm": 0.1027483389262317, "learning_rate": 7.548867576792411e-06, "loss": 0.8484, "step": 9065 }, { "epoch": 4.105978260869565, "grad_norm": 0.16018319472351017, "learning_rate": 7.541474818722338e-06, "loss": 0.8567, "step": 9066 }, { "epoch": 4.10643115942029, "grad_norm": 0.2017263842209221, "learning_rate": 7.534085305582475e-06, "loss": 0.8508, "step": 9067 }, { "epoch": 4.106884057971015, "grad_norm": 0.2142755598815399, "learning_rate": 7.526699038111549e-06, "loss": 0.8647, "step": 9068 }, { "epoch": 4.107336956521739, "grad_norm": 0.16237661764856123, "learning_rate": 7.5193160170479795e-06, "loss": 0.8391, "step": 9069 }, { "epoch": 4.107789855072464, "grad_norm": 0.18214631949658147, "learning_rate": 7.511936243129847e-06, "loss": 0.8581, "step": 9070 }, { "epoch": 4.1082427536231885, "grad_norm": 0.21910689361137287, "learning_rate": 7.504559717094939e-06, "loss": 0.8618, "step": 9071 }, { "epoch": 4.108695652173913, "grad_norm": 0.17085571168564817, "learning_rate": 7.4971864396806795e-06, "loss": 0.8529, "step": 9072 }, { "epoch": 4.109148550724638, "grad_norm": 0.1860260728747064, "learning_rate": 7.489816411624207e-06, "loss": 0.836, "step": 9073 }, { "epoch": 4.109601449275362, "grad_norm": 0.13286393337087604, "learning_rate": 7.482449633662296e-06, "loss": 0.836, "step": 9074 }, { "epoch": 4.110054347826087, "grad_norm": 0.19068823105580585, "learning_rate": 7.475086106531431e-06, "loss": 0.8446, "step": 9075 }, { "epoch": 4.1105072463768115, "grad_norm": 0.09013105575073571, "learning_rate": 7.467725830967753e-06, "loss": 0.8359, "step": 9076 }, { "epoch": 4.110960144927536, "grad_norm": 0.16862958816650123, "learning_rate": 7.460368807707072e-06, "loss": 0.8477, "step": 9077 }, { "epoch": 4.111413043478261, "grad_norm": 0.08246508105700141, "learning_rate": 7.4530150374848745e-06, "loss": 0.8417, "step": 9078 }, { "epoch": 4.111865942028985, "grad_norm": 0.16318532554537307, "learning_rate": 7.445664521036349e-06, "loss": 0.8461, "step": 9079 }, { "epoch": 4.11231884057971, "grad_norm": 0.09113593278193315, "learning_rate": 7.438317259096317e-06, "loss": 0.8396, "step": 9080 }, { "epoch": 4.112771739130435, "grad_norm": 0.14704684795929257, "learning_rate": 7.430973252399321e-06, "loss": 0.8635, "step": 9081 }, { "epoch": 4.113224637681159, "grad_norm": 0.09136832492331333, "learning_rate": 7.4236325016795275e-06, "loss": 0.8509, "step": 9082 }, { "epoch": 4.113677536231884, "grad_norm": 0.1362873718003191, "learning_rate": 7.416295007670822e-06, "loss": 0.8445, "step": 9083 }, { "epoch": 4.114130434782608, "grad_norm": 0.10245001830593697, "learning_rate": 7.408960771106741e-06, "loss": 0.8491, "step": 9084 }, { "epoch": 4.114583333333333, "grad_norm": 0.10573320874201592, "learning_rate": 7.401629792720495e-06, "loss": 0.8569, "step": 9085 }, { "epoch": 4.115036231884058, "grad_norm": 0.0993366055168205, "learning_rate": 7.3943020732449635e-06, "loss": 0.8415, "step": 9086 }, { "epoch": 4.115489130434782, "grad_norm": 0.09355999834071574, "learning_rate": 7.3869776134127335e-06, "loss": 0.8564, "step": 9087 }, { "epoch": 4.115942028985507, "grad_norm": 0.1119989732009352, "learning_rate": 7.37965641395602e-06, "loss": 0.852, "step": 9088 }, { "epoch": 4.1163949275362315, "grad_norm": 0.09004587210516157, "learning_rate": 7.372338475606757e-06, "loss": 0.8688, "step": 9089 }, { "epoch": 4.116847826086956, "grad_norm": 0.10805330641633602, "learning_rate": 7.365023799096511e-06, "loss": 0.858, "step": 9090 }, { "epoch": 4.117300724637682, "grad_norm": 0.10931868837631627, "learning_rate": 7.357712385156559e-06, "loss": 0.8484, "step": 9091 }, { "epoch": 4.117753623188406, "grad_norm": 0.11377035055687479, "learning_rate": 7.350404234517818e-06, "loss": 0.856, "step": 9092 }, { "epoch": 4.118206521739131, "grad_norm": 0.10203052813786492, "learning_rate": 7.343099347910914e-06, "loss": 0.8345, "step": 9093 }, { "epoch": 4.118659420289855, "grad_norm": 0.10556124131086583, "learning_rate": 7.335797726066114e-06, "loss": 0.8617, "step": 9094 }, { "epoch": 4.11911231884058, "grad_norm": 0.11969716989560364, "learning_rate": 7.328499369713378e-06, "loss": 0.8602, "step": 9095 }, { "epoch": 4.119565217391305, "grad_norm": 0.11820409628294823, "learning_rate": 7.321204279582321e-06, "loss": 0.8326, "step": 9096 }, { "epoch": 4.120018115942029, "grad_norm": 0.12151410073075022, "learning_rate": 7.313912456402268e-06, "loss": 0.8638, "step": 9097 }, { "epoch": 4.120471014492754, "grad_norm": 0.10811481976560755, "learning_rate": 7.306623900902173e-06, "loss": 0.8591, "step": 9098 }, { "epoch": 4.1209239130434785, "grad_norm": 0.1265025423514144, "learning_rate": 7.2993386138107e-06, "loss": 0.8728, "step": 9099 }, { "epoch": 4.121376811594203, "grad_norm": 0.09393781767172844, "learning_rate": 7.292056595856158e-06, "loss": 0.859, "step": 9100 }, { "epoch": 4.121829710144928, "grad_norm": 0.12518011939971502, "learning_rate": 7.284777847766555e-06, "loss": 0.8463, "step": 9101 }, { "epoch": 4.122282608695652, "grad_norm": 0.09350747597125711, "learning_rate": 7.277502370269553e-06, "loss": 0.8439, "step": 9102 }, { "epoch": 4.122735507246377, "grad_norm": 0.09508525047665264, "learning_rate": 7.270230164092495e-06, "loss": 0.8452, "step": 9103 }, { "epoch": 4.1231884057971016, "grad_norm": 0.09782047593578284, "learning_rate": 7.262961229962382e-06, "loss": 0.8408, "step": 9104 }, { "epoch": 4.123641304347826, "grad_norm": 0.09788598142887181, "learning_rate": 7.255695568605921e-06, "loss": 0.8379, "step": 9105 }, { "epoch": 4.124094202898551, "grad_norm": 0.08745911548966573, "learning_rate": 7.24843318074945e-06, "loss": 0.861, "step": 9106 }, { "epoch": 4.124547101449275, "grad_norm": 0.09473678367222226, "learning_rate": 7.241174067119021e-06, "loss": 0.8568, "step": 9107 }, { "epoch": 4.125, "grad_norm": 0.10067300757502516, "learning_rate": 7.233918228440324e-06, "loss": 0.8489, "step": 9108 }, { "epoch": 4.125452898550725, "grad_norm": 0.08951070292853226, "learning_rate": 7.226665665438757e-06, "loss": 0.8598, "step": 9109 }, { "epoch": 4.125905797101449, "grad_norm": 0.09245030757176777, "learning_rate": 7.219416378839353e-06, "loss": 0.8423, "step": 9110 }, { "epoch": 4.126358695652174, "grad_norm": 0.09270081299381724, "learning_rate": 7.212170369366842e-06, "loss": 0.8458, "step": 9111 }, { "epoch": 4.1268115942028984, "grad_norm": 0.10877674941413795, "learning_rate": 7.204927637745602e-06, "loss": 0.8493, "step": 9112 }, { "epoch": 4.127264492753623, "grad_norm": 0.09386859624224116, "learning_rate": 7.197688184699729e-06, "loss": 0.836, "step": 9113 }, { "epoch": 4.127717391304348, "grad_norm": 0.09214441776700903, "learning_rate": 7.190452010952933e-06, "loss": 0.8386, "step": 9114 }, { "epoch": 4.128170289855072, "grad_norm": 0.10390434989704456, "learning_rate": 7.183219117228652e-06, "loss": 0.8773, "step": 9115 }, { "epoch": 4.128623188405797, "grad_norm": 0.10990770598068984, "learning_rate": 7.175989504249954e-06, "loss": 0.8535, "step": 9116 }, { "epoch": 4.1290760869565215, "grad_norm": 0.09525731408887424, "learning_rate": 7.168763172739601e-06, "loss": 0.8376, "step": 9117 }, { "epoch": 4.129528985507246, "grad_norm": 0.09713168970840255, "learning_rate": 7.161540123420026e-06, "loss": 0.83, "step": 9118 }, { "epoch": 4.129981884057971, "grad_norm": 0.09762741207153464, "learning_rate": 7.15432035701332e-06, "loss": 0.8601, "step": 9119 }, { "epoch": 4.130434782608695, "grad_norm": 0.09440489570476562, "learning_rate": 7.147103874241246e-06, "loss": 0.8444, "step": 9120 }, { "epoch": 4.13088768115942, "grad_norm": 0.10550302137989241, "learning_rate": 7.139890675825265e-06, "loss": 0.8482, "step": 9121 }, { "epoch": 4.131340579710145, "grad_norm": 0.10802771702785033, "learning_rate": 7.132680762486477e-06, "loss": 0.8529, "step": 9122 }, { "epoch": 4.131793478260869, "grad_norm": 0.09701122013858739, "learning_rate": 7.125474134945682e-06, "loss": 0.8424, "step": 9123 }, { "epoch": 4.132246376811594, "grad_norm": 0.12561608100145497, "learning_rate": 7.118270793923327e-06, "loss": 0.8527, "step": 9124 }, { "epoch": 4.132699275362318, "grad_norm": 0.08594152441958877, "learning_rate": 7.111070740139556e-06, "loss": 0.8588, "step": 9125 }, { "epoch": 4.133152173913044, "grad_norm": 0.10073532141229802, "learning_rate": 7.1038739743141574e-06, "loss": 0.8713, "step": 9126 }, { "epoch": 4.1336050724637685, "grad_norm": 0.11617972103153305, "learning_rate": 7.096680497166608e-06, "loss": 0.8648, "step": 9127 }, { "epoch": 4.134057971014493, "grad_norm": 0.09773226670465003, "learning_rate": 7.089490309416036e-06, "loss": 0.8542, "step": 9128 }, { "epoch": 4.134510869565218, "grad_norm": 0.09689256264392349, "learning_rate": 7.082303411781284e-06, "loss": 0.8525, "step": 9129 }, { "epoch": 4.134963768115942, "grad_norm": 0.08550688190200795, "learning_rate": 7.075119804980808e-06, "loss": 0.8471, "step": 9130 }, { "epoch": 4.135416666666667, "grad_norm": 0.10185544488401793, "learning_rate": 7.067939489732794e-06, "loss": 0.8735, "step": 9131 }, { "epoch": 4.135869565217392, "grad_norm": 0.0974917036009281, "learning_rate": 7.06076246675504e-06, "loss": 0.8648, "step": 9132 }, { "epoch": 4.136322463768116, "grad_norm": 0.08604642580803669, "learning_rate": 7.053588736765067e-06, "loss": 0.8306, "step": 9133 }, { "epoch": 4.136775362318841, "grad_norm": 0.09842288151604826, "learning_rate": 7.046418300480029e-06, "loss": 0.8612, "step": 9134 }, { "epoch": 4.137228260869565, "grad_norm": 0.10394317162233209, "learning_rate": 7.0392511586167846e-06, "loss": 0.8566, "step": 9135 }, { "epoch": 4.13768115942029, "grad_norm": 0.09791604738693997, "learning_rate": 7.032087311891831e-06, "loss": 0.8501, "step": 9136 }, { "epoch": 4.138134057971015, "grad_norm": 0.1047154314461727, "learning_rate": 7.0249267610213514e-06, "loss": 0.8576, "step": 9137 }, { "epoch": 4.138586956521739, "grad_norm": 0.09988561420526634, "learning_rate": 7.017769506721186e-06, "loss": 0.8773, "step": 9138 }, { "epoch": 4.139039855072464, "grad_norm": 0.09324870785057236, "learning_rate": 7.0106155497068785e-06, "loss": 0.8426, "step": 9139 }, { "epoch": 4.1394927536231885, "grad_norm": 0.09294260917382012, "learning_rate": 7.003464890693603e-06, "loss": 0.8815, "step": 9140 }, { "epoch": 4.139945652173913, "grad_norm": 0.11059987551455412, "learning_rate": 6.996317530396237e-06, "loss": 0.8458, "step": 9141 }, { "epoch": 4.140398550724638, "grad_norm": 0.09891209872108743, "learning_rate": 6.9891734695292975e-06, "loss": 0.857, "step": 9142 }, { "epoch": 4.140851449275362, "grad_norm": 0.09324510704379724, "learning_rate": 6.982032708807005e-06, "loss": 0.8767, "step": 9143 }, { "epoch": 4.141304347826087, "grad_norm": 0.10438984295703142, "learning_rate": 6.974895248943223e-06, "loss": 0.8233, "step": 9144 }, { "epoch": 4.1417572463768115, "grad_norm": 0.112871470406054, "learning_rate": 6.967761090651501e-06, "loss": 0.8464, "step": 9145 }, { "epoch": 4.142210144927536, "grad_norm": 0.10469676686471642, "learning_rate": 6.9606302346450335e-06, "loss": 0.8369, "step": 9146 }, { "epoch": 4.142663043478261, "grad_norm": 0.09363743684225394, "learning_rate": 6.953502681636726e-06, "loss": 0.8485, "step": 9147 }, { "epoch": 4.143115942028985, "grad_norm": 0.09787366460167995, "learning_rate": 6.946378432339114e-06, "loss": 0.8451, "step": 9148 }, { "epoch": 4.14356884057971, "grad_norm": 0.1061002521115684, "learning_rate": 6.939257487464437e-06, "loss": 0.8518, "step": 9149 }, { "epoch": 4.144021739130435, "grad_norm": 0.09586965712690718, "learning_rate": 6.932139847724566e-06, "loss": 0.8425, "step": 9150 }, { "epoch": 4.144474637681159, "grad_norm": 0.09544119879379652, "learning_rate": 6.925025513831087e-06, "loss": 0.846, "step": 9151 }, { "epoch": 4.144927536231884, "grad_norm": 0.09243410887264243, "learning_rate": 6.917914486495218e-06, "loss": 0.8403, "step": 9152 }, { "epoch": 4.145380434782608, "grad_norm": 0.08316280000593328, "learning_rate": 6.910806766427862e-06, "loss": 0.853, "step": 9153 }, { "epoch": 4.145833333333333, "grad_norm": 0.09741335459351422, "learning_rate": 6.903702354339578e-06, "loss": 0.8507, "step": 9154 }, { "epoch": 4.146286231884058, "grad_norm": 0.10469244443018222, "learning_rate": 6.896601250940627e-06, "loss": 0.845, "step": 9155 }, { "epoch": 4.146739130434782, "grad_norm": 0.08026588397952329, "learning_rate": 6.8895034569408915e-06, "loss": 0.8245, "step": 9156 }, { "epoch": 4.147192028985507, "grad_norm": 0.09394804426275563, "learning_rate": 6.882408973049979e-06, "loss": 0.8361, "step": 9157 }, { "epoch": 4.1476449275362315, "grad_norm": 0.09365042016689712, "learning_rate": 6.875317799977108e-06, "loss": 0.8551, "step": 9158 }, { "epoch": 4.148097826086956, "grad_norm": 0.09913694560024293, "learning_rate": 6.868229938431215e-06, "loss": 0.8535, "step": 9159 }, { "epoch": 4.148550724637682, "grad_norm": 0.0859462403893112, "learning_rate": 6.861145389120878e-06, "loss": 0.8579, "step": 9160 }, { "epoch": 4.149003623188406, "grad_norm": 0.12088403587468767, "learning_rate": 6.8540641527543495e-06, "loss": 0.8518, "step": 9161 }, { "epoch": 4.149456521739131, "grad_norm": 0.09022333191684305, "learning_rate": 6.846986230039543e-06, "loss": 0.8424, "step": 9162 }, { "epoch": 4.149909420289855, "grad_norm": 0.11447728072342743, "learning_rate": 6.83991162168407e-06, "loss": 0.8412, "step": 9163 }, { "epoch": 4.15036231884058, "grad_norm": 0.08462879068034701, "learning_rate": 6.832840328395165e-06, "loss": 0.8582, "step": 9164 }, { "epoch": 4.150815217391305, "grad_norm": 0.102954630701684, "learning_rate": 6.8257723508797824e-06, "loss": 0.8275, "step": 9165 }, { "epoch": 4.151268115942029, "grad_norm": 0.11242550943142673, "learning_rate": 6.818707689844499e-06, "loss": 0.8309, "step": 9166 }, { "epoch": 4.151721014492754, "grad_norm": 0.09007154677680936, "learning_rate": 6.811646345995595e-06, "loss": 0.8262, "step": 9167 }, { "epoch": 4.1521739130434785, "grad_norm": 0.10366230928033679, "learning_rate": 6.804588320038998e-06, "loss": 0.8556, "step": 9168 }, { "epoch": 4.152626811594203, "grad_norm": 0.08679765165254864, "learning_rate": 6.797533612680309e-06, "loss": 0.8789, "step": 9169 }, { "epoch": 4.153079710144928, "grad_norm": 0.09797935783581856, "learning_rate": 6.7904822246247905e-06, "loss": 0.8549, "step": 9170 }, { "epoch": 4.153532608695652, "grad_norm": 0.10216420307622205, "learning_rate": 6.783434156577398e-06, "loss": 0.843, "step": 9171 }, { "epoch": 4.153985507246377, "grad_norm": 0.09212299382199649, "learning_rate": 6.776389409242718e-06, "loss": 0.856, "step": 9172 }, { "epoch": 4.1544384057971016, "grad_norm": 0.08507332855487372, "learning_rate": 6.769347983325043e-06, "loss": 0.8675, "step": 9173 }, { "epoch": 4.154891304347826, "grad_norm": 0.10833020716803346, "learning_rate": 6.762309879528306e-06, "loss": 0.8365, "step": 9174 }, { "epoch": 4.155344202898551, "grad_norm": 0.09507305430283855, "learning_rate": 6.755275098556127e-06, "loss": 0.8436, "step": 9175 }, { "epoch": 4.155797101449275, "grad_norm": 0.09076519983284406, "learning_rate": 6.7482436411117646e-06, "loss": 0.8394, "step": 9176 }, { "epoch": 4.15625, "grad_norm": 0.08879564534156352, "learning_rate": 6.7412155078981865e-06, "loss": 0.8499, "step": 9177 }, { "epoch": 4.156702898550725, "grad_norm": 0.09640159848734746, "learning_rate": 6.734190699618013e-06, "loss": 0.8505, "step": 9178 }, { "epoch": 4.157155797101449, "grad_norm": 0.09711809552213685, "learning_rate": 6.7271692169734855e-06, "loss": 0.8479, "step": 9179 }, { "epoch": 4.157608695652174, "grad_norm": 0.10166360432849011, "learning_rate": 6.7201510606665865e-06, "loss": 0.852, "step": 9180 }, { "epoch": 4.1580615942028984, "grad_norm": 0.10166219977944767, "learning_rate": 6.7131362313989135e-06, "loss": 0.8373, "step": 9181 }, { "epoch": 4.158514492753623, "grad_norm": 0.09562986775234754, "learning_rate": 6.706124729871768e-06, "loss": 0.8289, "step": 9182 }, { "epoch": 4.158967391304348, "grad_norm": 0.10700740988375561, "learning_rate": 6.6991165567860824e-06, "loss": 0.8558, "step": 9183 }, { "epoch": 4.159420289855072, "grad_norm": 0.11983205621963867, "learning_rate": 6.692111712842497e-06, "loss": 0.8537, "step": 9184 }, { "epoch": 4.159873188405797, "grad_norm": 0.09742777285673028, "learning_rate": 6.6851101987412695e-06, "loss": 0.8592, "step": 9185 }, { "epoch": 4.1603260869565215, "grad_norm": 0.11847566885394789, "learning_rate": 6.678112015182381e-06, "loss": 0.8615, "step": 9186 }, { "epoch": 4.160778985507246, "grad_norm": 0.08490798597954317, "learning_rate": 6.671117162865437e-06, "loss": 0.8487, "step": 9187 }, { "epoch": 4.161231884057971, "grad_norm": 0.0967155420041607, "learning_rate": 6.664125642489723e-06, "loss": 0.8474, "step": 9188 }, { "epoch": 4.161684782608695, "grad_norm": 0.10391270554225271, "learning_rate": 6.657137454754185e-06, "loss": 0.8408, "step": 9189 }, { "epoch": 4.16213768115942, "grad_norm": 0.10853874467497482, "learning_rate": 6.650152600357463e-06, "loss": 0.8532, "step": 9190 }, { "epoch": 4.162590579710145, "grad_norm": 0.1341015636183611, "learning_rate": 6.643171079997821e-06, "loss": 0.8541, "step": 9191 }, { "epoch": 4.163043478260869, "grad_norm": 0.08945905695535349, "learning_rate": 6.636192894373232e-06, "loss": 0.8734, "step": 9192 }, { "epoch": 4.163496376811594, "grad_norm": 0.13651785589365875, "learning_rate": 6.629218044181303e-06, "loss": 0.8433, "step": 9193 }, { "epoch": 4.163949275362318, "grad_norm": 0.09244703031207666, "learning_rate": 6.622246530119336e-06, "loss": 0.8501, "step": 9194 }, { "epoch": 4.164402173913044, "grad_norm": 0.09191246881315088, "learning_rate": 6.615278352884274e-06, "loss": 0.8616, "step": 9195 }, { "epoch": 4.1648550724637685, "grad_norm": 0.09992760863824589, "learning_rate": 6.60831351317274e-06, "loss": 0.8478, "step": 9196 }, { "epoch": 4.165307971014493, "grad_norm": 0.08704963083197222, "learning_rate": 6.601352011681004e-06, "loss": 0.8501, "step": 9197 }, { "epoch": 4.165760869565218, "grad_norm": 0.09314427171442131, "learning_rate": 6.594393849105043e-06, "loss": 0.8606, "step": 9198 }, { "epoch": 4.166213768115942, "grad_norm": 0.09098491061271341, "learning_rate": 6.587439026140456e-06, "loss": 0.8641, "step": 9199 }, { "epoch": 4.166666666666667, "grad_norm": 0.11406493113195758, "learning_rate": 6.58048754348255e-06, "loss": 0.8454, "step": 9200 }, { "epoch": 4.167119565217392, "grad_norm": 0.08274240879141674, "learning_rate": 6.5735394018262475e-06, "loss": 0.8565, "step": 9201 }, { "epoch": 4.167572463768116, "grad_norm": 0.10707226567512954, "learning_rate": 6.56659460186619e-06, "loss": 0.8492, "step": 9202 }, { "epoch": 4.168025362318841, "grad_norm": 0.08726895814168084, "learning_rate": 6.559653144296647e-06, "loss": 0.8279, "step": 9203 }, { "epoch": 4.168478260869565, "grad_norm": 0.09185715541378527, "learning_rate": 6.552715029811576e-06, "loss": 0.8497, "step": 9204 }, { "epoch": 4.16893115942029, "grad_norm": 0.12853302516406975, "learning_rate": 6.545780259104573e-06, "loss": 0.8406, "step": 9205 }, { "epoch": 4.169384057971015, "grad_norm": 0.08386189847422888, "learning_rate": 6.538848832868936e-06, "loss": 0.8614, "step": 9206 }, { "epoch": 4.169836956521739, "grad_norm": 0.10864280940991679, "learning_rate": 6.531920751797601e-06, "loss": 0.8599, "step": 9207 }, { "epoch": 4.170289855072464, "grad_norm": 0.08976387275507695, "learning_rate": 6.524996016583189e-06, "loss": 0.8332, "step": 9208 }, { "epoch": 4.1707427536231885, "grad_norm": 0.1031215159670482, "learning_rate": 6.518074627917963e-06, "loss": 0.8463, "step": 9209 }, { "epoch": 4.171195652173913, "grad_norm": 0.10494036945237752, "learning_rate": 6.511156586493883e-06, "loss": 0.8469, "step": 9210 }, { "epoch": 4.171648550724638, "grad_norm": 0.10224965880036167, "learning_rate": 6.504241893002548e-06, "loss": 0.8644, "step": 9211 }, { "epoch": 4.172101449275362, "grad_norm": 0.08848460930444803, "learning_rate": 6.497330548135225e-06, "loss": 0.8306, "step": 9212 }, { "epoch": 4.172554347826087, "grad_norm": 0.1091075359730467, "learning_rate": 6.490422552582854e-06, "loss": 0.8434, "step": 9213 }, { "epoch": 4.1730072463768115, "grad_norm": 0.09569903293485009, "learning_rate": 6.4835179070360435e-06, "loss": 0.8481, "step": 9214 }, { "epoch": 4.173460144927536, "grad_norm": 0.09041262482741483, "learning_rate": 6.476616612185052e-06, "loss": 0.8629, "step": 9215 }, { "epoch": 4.173913043478261, "grad_norm": 0.08966138689305764, "learning_rate": 6.4697186687198285e-06, "loss": 0.8303, "step": 9216 }, { "epoch": 4.174365942028985, "grad_norm": 0.09907187255294789, "learning_rate": 6.462824077329957e-06, "loss": 0.8526, "step": 9217 }, { "epoch": 4.17481884057971, "grad_norm": 0.08166393537884062, "learning_rate": 6.455932838704715e-06, "loss": 0.8449, "step": 9218 }, { "epoch": 4.175271739130435, "grad_norm": 0.09953406609305133, "learning_rate": 6.449044953533024e-06, "loss": 0.8496, "step": 9219 }, { "epoch": 4.175724637681159, "grad_norm": 0.10207600168649031, "learning_rate": 6.442160422503469e-06, "loss": 0.8593, "step": 9220 }, { "epoch": 4.176177536231884, "grad_norm": 0.09422312976789775, "learning_rate": 6.435279246304307e-06, "loss": 0.8319, "step": 9221 }, { "epoch": 4.176630434782608, "grad_norm": 0.10474984192787899, "learning_rate": 6.428401425623479e-06, "loss": 0.8525, "step": 9222 }, { "epoch": 4.177083333333333, "grad_norm": 0.10270330981959083, "learning_rate": 6.421526961148545e-06, "loss": 0.8432, "step": 9223 }, { "epoch": 4.177536231884058, "grad_norm": 0.10071490085010686, "learning_rate": 6.414655853566785e-06, "loss": 0.849, "step": 9224 }, { "epoch": 4.177989130434782, "grad_norm": 0.09595116399641346, "learning_rate": 6.407788103565087e-06, "loss": 0.8433, "step": 9225 }, { "epoch": 4.178442028985507, "grad_norm": 0.11090227629517238, "learning_rate": 6.4009237118300535e-06, "loss": 0.8622, "step": 9226 }, { "epoch": 4.1788949275362315, "grad_norm": 0.09328739380598607, "learning_rate": 6.39406267904791e-06, "loss": 0.8636, "step": 9227 }, { "epoch": 4.179347826086956, "grad_norm": 0.10446548521812565, "learning_rate": 6.387205005904582e-06, "loss": 0.8409, "step": 9228 }, { "epoch": 4.179800724637682, "grad_norm": 0.08443031437838802, "learning_rate": 6.380350693085637e-06, "loss": 0.8515, "step": 9229 }, { "epoch": 4.180253623188406, "grad_norm": 0.12144638289323442, "learning_rate": 6.373499741276305e-06, "loss": 0.8458, "step": 9230 }, { "epoch": 4.180706521739131, "grad_norm": 0.0798570569596023, "learning_rate": 6.3666521511614785e-06, "loss": 0.8519, "step": 9231 }, { "epoch": 4.181159420289855, "grad_norm": 0.10069664634148026, "learning_rate": 6.359807923425747e-06, "loss": 0.8732, "step": 9232 }, { "epoch": 4.18161231884058, "grad_norm": 0.08880492248499042, "learning_rate": 6.352967058753314e-06, "loss": 0.8449, "step": 9233 }, { "epoch": 4.182065217391305, "grad_norm": 0.09234991065014447, "learning_rate": 6.346129557828091e-06, "loss": 0.8451, "step": 9234 }, { "epoch": 4.182518115942029, "grad_norm": 0.08834982310510871, "learning_rate": 6.339295421333616e-06, "loss": 0.8449, "step": 9235 }, { "epoch": 4.182971014492754, "grad_norm": 0.10105076779374694, "learning_rate": 6.332464649953127e-06, "loss": 0.8218, "step": 9236 }, { "epoch": 4.1834239130434785, "grad_norm": 0.10112909564000949, "learning_rate": 6.325637244369498e-06, "loss": 0.8401, "step": 9237 }, { "epoch": 4.183876811594203, "grad_norm": 0.09507682382428682, "learning_rate": 6.318813205265275e-06, "loss": 0.8726, "step": 9238 }, { "epoch": 4.184329710144928, "grad_norm": 0.11176960175843637, "learning_rate": 6.311992533322655e-06, "loss": 0.8608, "step": 9239 }, { "epoch": 4.184782608695652, "grad_norm": 0.11624667477780651, "learning_rate": 6.305175229223541e-06, "loss": 0.8506, "step": 9240 }, { "epoch": 4.185235507246377, "grad_norm": 0.09192044592360123, "learning_rate": 6.298361293649438e-06, "loss": 0.8692, "step": 9241 }, { "epoch": 4.1856884057971016, "grad_norm": 0.12760919133329193, "learning_rate": 6.291550727281577e-06, "loss": 0.854, "step": 9242 }, { "epoch": 4.186141304347826, "grad_norm": 0.10809015788346936, "learning_rate": 6.284743530800792e-06, "loss": 0.8454, "step": 9243 }, { "epoch": 4.186594202898551, "grad_norm": 0.09809024097694173, "learning_rate": 6.277939704887633e-06, "loss": 0.854, "step": 9244 }, { "epoch": 4.187047101449275, "grad_norm": 0.1039387064717963, "learning_rate": 6.271139250222278e-06, "loss": 0.8424, "step": 9245 }, { "epoch": 4.1875, "grad_norm": 0.09662901848959854, "learning_rate": 6.26434216748458e-06, "loss": 0.8284, "step": 9246 }, { "epoch": 4.187952898550725, "grad_norm": 0.09365944203902203, "learning_rate": 6.257548457354046e-06, "loss": 0.8422, "step": 9247 }, { "epoch": 4.188405797101449, "grad_norm": 0.09347541949545209, "learning_rate": 6.250758120509872e-06, "loss": 0.8452, "step": 9248 }, { "epoch": 4.188858695652174, "grad_norm": 0.09673232829282993, "learning_rate": 6.243971157630877e-06, "loss": 0.8615, "step": 9249 }, { "epoch": 4.1893115942028984, "grad_norm": 0.08670894691652933, "learning_rate": 6.237187569395588e-06, "loss": 0.8383, "step": 9250 }, { "epoch": 4.189764492753623, "grad_norm": 0.10817954121818667, "learning_rate": 6.230407356482149e-06, "loss": 0.8642, "step": 9251 }, { "epoch": 4.190217391304348, "grad_norm": 0.08429755228407053, "learning_rate": 6.223630519568411e-06, "loss": 0.8495, "step": 9252 }, { "epoch": 4.190670289855072, "grad_norm": 0.10005765080311174, "learning_rate": 6.216857059331851e-06, "loss": 0.8664, "step": 9253 }, { "epoch": 4.191123188405797, "grad_norm": 0.10372366528171346, "learning_rate": 6.210086976449625e-06, "loss": 0.8321, "step": 9254 }, { "epoch": 4.1915760869565215, "grad_norm": 0.08477994332676514, "learning_rate": 6.203320271598539e-06, "loss": 0.8662, "step": 9255 }, { "epoch": 4.192028985507246, "grad_norm": 0.11230996269676236, "learning_rate": 6.19655694545509e-06, "loss": 0.8657, "step": 9256 }, { "epoch": 4.192481884057971, "grad_norm": 0.0913203221560745, "learning_rate": 6.189796998695401e-06, "loss": 0.8411, "step": 9257 }, { "epoch": 4.192934782608695, "grad_norm": 0.1007611387406871, "learning_rate": 6.1830404319952906e-06, "loss": 0.8371, "step": 9258 }, { "epoch": 4.19338768115942, "grad_norm": 0.09837572454556241, "learning_rate": 6.176287246030206e-06, "loss": 0.8391, "step": 9259 }, { "epoch": 4.193840579710145, "grad_norm": 0.08936581833619685, "learning_rate": 6.169537441475295e-06, "loss": 0.8321, "step": 9260 }, { "epoch": 4.194293478260869, "grad_norm": 0.08671236341459002, "learning_rate": 6.162791019005334e-06, "loss": 0.8586, "step": 9261 }, { "epoch": 4.194746376811594, "grad_norm": 0.09546232982941835, "learning_rate": 6.156047979294775e-06, "loss": 0.8533, "step": 9262 }, { "epoch": 4.195199275362318, "grad_norm": 0.08952408962243846, "learning_rate": 6.149308323017722e-06, "loss": 0.8481, "step": 9263 }, { "epoch": 4.195652173913044, "grad_norm": 0.09357355109681724, "learning_rate": 6.142572050847966e-06, "loss": 0.8526, "step": 9264 }, { "epoch": 4.1961050724637685, "grad_norm": 0.08640884672897349, "learning_rate": 6.135839163458923e-06, "loss": 0.8474, "step": 9265 }, { "epoch": 4.196557971014493, "grad_norm": 0.09273369107186973, "learning_rate": 6.129109661523709e-06, "loss": 0.8455, "step": 9266 }, { "epoch": 4.197010869565218, "grad_norm": 0.08347045676057399, "learning_rate": 6.122383545715069e-06, "loss": 0.8544, "step": 9267 }, { "epoch": 4.197463768115942, "grad_norm": 0.09002155068628152, "learning_rate": 6.115660816705435e-06, "loss": 0.85, "step": 9268 }, { "epoch": 4.197916666666667, "grad_norm": 0.08688062950665544, "learning_rate": 6.108941475166879e-06, "loss": 0.8274, "step": 9269 }, { "epoch": 4.198369565217392, "grad_norm": 0.08963573927140911, "learning_rate": 6.102225521771154e-06, "loss": 0.844, "step": 9270 }, { "epoch": 4.198822463768116, "grad_norm": 0.09836786106705689, "learning_rate": 6.095512957189664e-06, "loss": 0.8677, "step": 9271 }, { "epoch": 4.199275362318841, "grad_norm": 0.08882975811558994, "learning_rate": 6.088803782093466e-06, "loss": 0.8554, "step": 9272 }, { "epoch": 4.199728260869565, "grad_norm": 0.09108305120794664, "learning_rate": 6.0820979971532825e-06, "loss": 0.8483, "step": 9273 }, { "epoch": 4.20018115942029, "grad_norm": 0.08570265294615707, "learning_rate": 6.075395603039518e-06, "loss": 0.8581, "step": 9274 }, { "epoch": 4.200634057971015, "grad_norm": 0.08319275119926763, "learning_rate": 6.068696600422206e-06, "loss": 0.8445, "step": 9275 }, { "epoch": 4.201086956521739, "grad_norm": 0.08555859432631344, "learning_rate": 6.062000989971073e-06, "loss": 0.8455, "step": 9276 }, { "epoch": 4.201539855072464, "grad_norm": 0.08311639567420198, "learning_rate": 6.055308772355468e-06, "loss": 0.849, "step": 9277 }, { "epoch": 4.2019927536231885, "grad_norm": 0.0923945985311739, "learning_rate": 6.048619948244448e-06, "loss": 0.8264, "step": 9278 }, { "epoch": 4.202445652173913, "grad_norm": 0.08986973901340743, "learning_rate": 6.041934518306689e-06, "loss": 0.8359, "step": 9279 }, { "epoch": 4.202898550724638, "grad_norm": 0.10009376476741888, "learning_rate": 6.035252483210547e-06, "loss": 0.8779, "step": 9280 }, { "epoch": 4.203351449275362, "grad_norm": 0.08972492680367813, "learning_rate": 6.028573843624031e-06, "loss": 0.8524, "step": 9281 }, { "epoch": 4.203804347826087, "grad_norm": 0.08986219685178995, "learning_rate": 6.021898600214826e-06, "loss": 0.8448, "step": 9282 }, { "epoch": 4.2042572463768115, "grad_norm": 0.09321745480502731, "learning_rate": 6.015226753650254e-06, "loss": 0.8568, "step": 9283 }, { "epoch": 4.204710144927536, "grad_norm": 0.09208675955866528, "learning_rate": 6.0085583045973224e-06, "loss": 0.8639, "step": 9284 }, { "epoch": 4.205163043478261, "grad_norm": 0.08831794215093228, "learning_rate": 6.001893253722677e-06, "loss": 0.8631, "step": 9285 }, { "epoch": 4.205615942028985, "grad_norm": 0.09289410697875256, "learning_rate": 5.995231601692641e-06, "loss": 0.8211, "step": 9286 }, { "epoch": 4.20606884057971, "grad_norm": 0.09781208087581265, "learning_rate": 5.988573349173195e-06, "loss": 0.8378, "step": 9287 }, { "epoch": 4.206521739130435, "grad_norm": 0.08682831186627231, "learning_rate": 5.981918496829959e-06, "loss": 0.8387, "step": 9288 }, { "epoch": 4.206974637681159, "grad_norm": 0.09608530871942851, "learning_rate": 5.97526704532823e-06, "loss": 0.8581, "step": 9289 }, { "epoch": 4.207427536231884, "grad_norm": 0.10426860243078095, "learning_rate": 5.968618995332982e-06, "loss": 0.8569, "step": 9290 }, { "epoch": 4.207880434782608, "grad_norm": 0.08613328301098679, "learning_rate": 5.961974347508812e-06, "loss": 0.8583, "step": 9291 }, { "epoch": 4.208333333333333, "grad_norm": 0.090718260808174, "learning_rate": 5.955333102520011e-06, "loss": 0.8575, "step": 9292 }, { "epoch": 4.208786231884058, "grad_norm": 0.10903653725765793, "learning_rate": 5.948695261030502e-06, "loss": 0.8576, "step": 9293 }, { "epoch": 4.209239130434782, "grad_norm": 0.0830997267825627, "learning_rate": 5.942060823703895e-06, "loss": 0.862, "step": 9294 }, { "epoch": 4.209692028985507, "grad_norm": 0.08698096301069329, "learning_rate": 5.935429791203437e-06, "loss": 0.8292, "step": 9295 }, { "epoch": 4.2101449275362315, "grad_norm": 0.09631416186620097, "learning_rate": 5.92880216419204e-06, "loss": 0.8675, "step": 9296 }, { "epoch": 4.210597826086956, "grad_norm": 0.08386048253048428, "learning_rate": 5.922177943332275e-06, "loss": 0.8565, "step": 9297 }, { "epoch": 4.211050724637682, "grad_norm": 0.10382806695410926, "learning_rate": 5.915557129286389e-06, "loss": 0.8557, "step": 9298 }, { "epoch": 4.211503623188406, "grad_norm": 0.09813607419699821, "learning_rate": 5.908939722716258e-06, "loss": 0.8571, "step": 9299 }, { "epoch": 4.211956521739131, "grad_norm": 0.11985723164467876, "learning_rate": 5.9023257242834595e-06, "loss": 0.8491, "step": 9300 }, { "epoch": 4.212409420289855, "grad_norm": 0.0825908752427582, "learning_rate": 5.8957151346491756e-06, "loss": 0.861, "step": 9301 }, { "epoch": 4.21286231884058, "grad_norm": 0.11636556424825262, "learning_rate": 5.8891079544743045e-06, "loss": 0.8441, "step": 9302 }, { "epoch": 4.213315217391305, "grad_norm": 0.1051716791664082, "learning_rate": 5.882504184419367e-06, "loss": 0.8644, "step": 9303 }, { "epoch": 4.213768115942029, "grad_norm": 0.08839288999636594, "learning_rate": 5.875903825144544e-06, "loss": 0.8628, "step": 9304 }, { "epoch": 4.214221014492754, "grad_norm": 0.11164018714804007, "learning_rate": 5.869306877309693e-06, "loss": 0.8621, "step": 9305 }, { "epoch": 4.2146739130434785, "grad_norm": 0.09507979860416932, "learning_rate": 5.862713341574311e-06, "loss": 0.8726, "step": 9306 }, { "epoch": 4.215126811594203, "grad_norm": 0.10875255979537606, "learning_rate": 5.856123218597578e-06, "loss": 0.8365, "step": 9307 }, { "epoch": 4.215579710144928, "grad_norm": 0.11092068156592359, "learning_rate": 5.849536509038305e-06, "loss": 0.8452, "step": 9308 }, { "epoch": 4.216032608695652, "grad_norm": 0.104127910580047, "learning_rate": 5.8429532135549915e-06, "loss": 0.8401, "step": 9309 }, { "epoch": 4.216485507246377, "grad_norm": 0.09759675787082063, "learning_rate": 5.83637333280576e-06, "loss": 0.837, "step": 9310 }, { "epoch": 4.2169384057971016, "grad_norm": 0.09301279233579032, "learning_rate": 5.8297968674484364e-06, "loss": 0.8678, "step": 9311 }, { "epoch": 4.217391304347826, "grad_norm": 0.09785906605360777, "learning_rate": 5.823223818140458e-06, "loss": 0.8558, "step": 9312 }, { "epoch": 4.217844202898551, "grad_norm": 0.08822353055255125, "learning_rate": 5.816654185538957e-06, "loss": 0.841, "step": 9313 }, { "epoch": 4.218297101449275, "grad_norm": 0.11262870822729461, "learning_rate": 5.810087970300711e-06, "loss": 0.8616, "step": 9314 }, { "epoch": 4.21875, "grad_norm": 0.08555419212115921, "learning_rate": 5.803525173082145e-06, "loss": 0.8598, "step": 9315 }, { "epoch": 4.219202898550725, "grad_norm": 0.08860874765769929, "learning_rate": 5.796965794539349e-06, "loss": 0.8583, "step": 9316 }, { "epoch": 4.219655797101449, "grad_norm": 0.09956720944071558, "learning_rate": 5.7904098353280904e-06, "loss": 0.8335, "step": 9317 }, { "epoch": 4.220108695652174, "grad_norm": 0.0912935597883182, "learning_rate": 5.783857296103761e-06, "loss": 0.8506, "step": 9318 }, { "epoch": 4.2205615942028984, "grad_norm": 0.08414695377347, "learning_rate": 5.777308177521445e-06, "loss": 0.8278, "step": 9319 }, { "epoch": 4.221014492753623, "grad_norm": 0.0976421714727342, "learning_rate": 5.7707624802358565e-06, "loss": 0.8611, "step": 9320 }, { "epoch": 4.221467391304348, "grad_norm": 0.08683615239909434, "learning_rate": 5.764220204901389e-06, "loss": 0.8428, "step": 9321 }, { "epoch": 4.221920289855072, "grad_norm": 0.09908116119668052, "learning_rate": 5.7576813521720824e-06, "loss": 0.8495, "step": 9322 }, { "epoch": 4.222373188405797, "grad_norm": 0.09298207570351687, "learning_rate": 5.751145922701629e-06, "loss": 0.8639, "step": 9323 }, { "epoch": 4.2228260869565215, "grad_norm": 0.10412442351968762, "learning_rate": 5.744613917143382e-06, "loss": 0.8306, "step": 9324 }, { "epoch": 4.223278985507246, "grad_norm": 0.08518364387873939, "learning_rate": 5.738085336150377e-06, "loss": 0.8563, "step": 9325 }, { "epoch": 4.223731884057971, "grad_norm": 0.1039517003466703, "learning_rate": 5.731560180375257e-06, "loss": 0.8538, "step": 9326 }, { "epoch": 4.224184782608695, "grad_norm": 0.0884430019886903, "learning_rate": 5.725038450470383e-06, "loss": 0.8611, "step": 9327 }, { "epoch": 4.22463768115942, "grad_norm": 0.09764432715380471, "learning_rate": 5.718520147087719e-06, "loss": 0.853, "step": 9328 }, { "epoch": 4.225090579710145, "grad_norm": 0.11921673464856769, "learning_rate": 5.712005270878927e-06, "loss": 0.8549, "step": 9329 }, { "epoch": 4.225543478260869, "grad_norm": 0.09036927852201918, "learning_rate": 5.705493822495305e-06, "loss": 0.8414, "step": 9330 }, { "epoch": 4.225996376811594, "grad_norm": 0.11601920811704027, "learning_rate": 5.698985802587804e-06, "loss": 0.8478, "step": 9331 }, { "epoch": 4.226449275362318, "grad_norm": 0.10976412594549924, "learning_rate": 5.692481211807046e-06, "loss": 0.8523, "step": 9332 }, { "epoch": 4.226902173913044, "grad_norm": 0.08767788863608067, "learning_rate": 5.6859800508033105e-06, "loss": 0.8577, "step": 9333 }, { "epoch": 4.2273550724637685, "grad_norm": 0.09372820936057882, "learning_rate": 5.6794823202265125e-06, "loss": 0.8643, "step": 9334 }, { "epoch": 4.227807971014493, "grad_norm": 0.09629118312999668, "learning_rate": 5.672988020726266e-06, "loss": 0.8708, "step": 9335 }, { "epoch": 4.228260869565218, "grad_norm": 0.08080270505186063, "learning_rate": 5.666497152951791e-06, "loss": 0.8612, "step": 9336 }, { "epoch": 4.228713768115942, "grad_norm": 0.08986488533328123, "learning_rate": 5.660009717552015e-06, "loss": 0.8529, "step": 9337 }, { "epoch": 4.229166666666667, "grad_norm": 0.09771653573501132, "learning_rate": 5.653525715175483e-06, "loss": 0.8548, "step": 9338 }, { "epoch": 4.229619565217392, "grad_norm": 0.08144890234503435, "learning_rate": 5.647045146470409e-06, "loss": 0.8262, "step": 9339 }, { "epoch": 4.230072463768116, "grad_norm": 0.07402317248788394, "learning_rate": 5.640568012084662e-06, "loss": 0.8448, "step": 9340 }, { "epoch": 4.230525362318841, "grad_norm": 0.10271449903133363, "learning_rate": 5.634094312665785e-06, "loss": 0.8273, "step": 9341 }, { "epoch": 4.230978260869565, "grad_norm": 0.1044787832483076, "learning_rate": 5.6276240488609464e-06, "loss": 0.8385, "step": 9342 }, { "epoch": 4.23143115942029, "grad_norm": 0.08717646870784768, "learning_rate": 5.621157221317011e-06, "loss": 0.8291, "step": 9343 }, { "epoch": 4.231884057971015, "grad_norm": 0.10582972217800925, "learning_rate": 5.61469383068046e-06, "loss": 0.8594, "step": 9344 }, { "epoch": 4.232336956521739, "grad_norm": 0.09710606627187081, "learning_rate": 5.608233877597458e-06, "loss": 0.8421, "step": 9345 }, { "epoch": 4.232789855072464, "grad_norm": 0.08296461972985827, "learning_rate": 5.601777362713816e-06, "loss": 0.8685, "step": 9346 }, { "epoch": 4.2332427536231885, "grad_norm": 0.10437011955239216, "learning_rate": 5.595324286675001e-06, "loss": 0.839, "step": 9347 }, { "epoch": 4.233695652173913, "grad_norm": 0.08603421128674531, "learning_rate": 5.588874650126128e-06, "loss": 0.8538, "step": 9348 }, { "epoch": 4.234148550724638, "grad_norm": 0.07997504231418555, "learning_rate": 5.582428453711992e-06, "loss": 0.8477, "step": 9349 }, { "epoch": 4.234601449275362, "grad_norm": 0.0833752890220431, "learning_rate": 5.575985698077016e-06, "loss": 0.8448, "step": 9350 }, { "epoch": 4.235054347826087, "grad_norm": 0.0946592595066834, "learning_rate": 5.569546383865305e-06, "loss": 0.856, "step": 9351 }, { "epoch": 4.2355072463768115, "grad_norm": 0.0879149979112413, "learning_rate": 5.563110511720595e-06, "loss": 0.8529, "step": 9352 }, { "epoch": 4.235960144927536, "grad_norm": 0.09091566815329677, "learning_rate": 5.556678082286304e-06, "loss": 0.837, "step": 9353 }, { "epoch": 4.236413043478261, "grad_norm": 0.08576594962523822, "learning_rate": 5.5502490962054775e-06, "loss": 0.8477, "step": 9354 }, { "epoch": 4.236865942028985, "grad_norm": 0.08890346004043091, "learning_rate": 5.54382355412086e-06, "loss": 0.8515, "step": 9355 }, { "epoch": 4.23731884057971, "grad_norm": 0.08747077603093048, "learning_rate": 5.53740145667478e-06, "loss": 0.8458, "step": 9356 }, { "epoch": 4.237771739130435, "grad_norm": 0.10662403346682996, "learning_rate": 5.5309828045093e-06, "loss": 0.8381, "step": 9357 }, { "epoch": 4.238224637681159, "grad_norm": 0.08683638537474782, "learning_rate": 5.524567598266078e-06, "loss": 0.8412, "step": 9358 }, { "epoch": 4.238677536231884, "grad_norm": 0.07927753575606199, "learning_rate": 5.518155838586472e-06, "loss": 0.8479, "step": 9359 }, { "epoch": 4.239130434782608, "grad_norm": 0.10649366698655784, "learning_rate": 5.511747526111459e-06, "loss": 0.8522, "step": 9360 }, { "epoch": 4.239583333333333, "grad_norm": 0.08725087485003609, "learning_rate": 5.5053426614817094e-06, "loss": 0.8349, "step": 9361 }, { "epoch": 4.240036231884058, "grad_norm": 0.09944004944260265, "learning_rate": 5.498941245337506e-06, "loss": 0.8684, "step": 9362 }, { "epoch": 4.240489130434782, "grad_norm": 0.10397412616725574, "learning_rate": 5.492543278318825e-06, "loss": 0.8647, "step": 9363 }, { "epoch": 4.240942028985507, "grad_norm": 0.08708142884032247, "learning_rate": 5.486148761065275e-06, "loss": 0.8513, "step": 9364 }, { "epoch": 4.2413949275362315, "grad_norm": 0.08061528877650664, "learning_rate": 5.479757694216128e-06, "loss": 0.8425, "step": 9365 }, { "epoch": 4.241847826086956, "grad_norm": 0.08976466672146256, "learning_rate": 5.4733700784102935e-06, "loss": 0.8446, "step": 9366 }, { "epoch": 4.242300724637682, "grad_norm": 0.09377376852772201, "learning_rate": 5.466985914286373e-06, "loss": 0.867, "step": 9367 }, { "epoch": 4.242753623188406, "grad_norm": 0.0840254712673827, "learning_rate": 5.460605202482585e-06, "loss": 0.8548, "step": 9368 }, { "epoch": 4.243206521739131, "grad_norm": 0.09755834942542678, "learning_rate": 5.454227943636836e-06, "loss": 0.8383, "step": 9369 }, { "epoch": 4.243659420289855, "grad_norm": 0.087580743206258, "learning_rate": 5.447854138386656e-06, "loss": 0.8635, "step": 9370 }, { "epoch": 4.24411231884058, "grad_norm": 0.08889538891359268, "learning_rate": 5.441483787369257e-06, "loss": 0.8629, "step": 9371 }, { "epoch": 4.244565217391305, "grad_norm": 0.0894219862365475, "learning_rate": 5.435116891221488e-06, "loss": 0.8581, "step": 9372 }, { "epoch": 4.245018115942029, "grad_norm": 0.09430937905739419, "learning_rate": 5.428753450579853e-06, "loss": 0.8559, "step": 9373 }, { "epoch": 4.245471014492754, "grad_norm": 0.10296899970612669, "learning_rate": 5.422393466080511e-06, "loss": 0.8637, "step": 9374 }, { "epoch": 4.2459239130434785, "grad_norm": 0.08822413375087879, "learning_rate": 5.416036938359295e-06, "loss": 0.8448, "step": 9375 }, { "epoch": 4.246376811594203, "grad_norm": 0.08277775706943506, "learning_rate": 5.409683868051661e-06, "loss": 0.8429, "step": 9376 }, { "epoch": 4.246829710144928, "grad_norm": 0.09637797041678214, "learning_rate": 5.403334255792754e-06, "loss": 0.8512, "step": 9377 }, { "epoch": 4.247282608695652, "grad_norm": 0.08255902751235125, "learning_rate": 5.396988102217333e-06, "loss": 0.8685, "step": 9378 }, { "epoch": 4.247735507246377, "grad_norm": 0.09157497966126842, "learning_rate": 5.390645407959856e-06, "loss": 0.8548, "step": 9379 }, { "epoch": 4.2481884057971016, "grad_norm": 0.08981235780051651, "learning_rate": 5.384306173654401e-06, "loss": 0.8628, "step": 9380 }, { "epoch": 4.248641304347826, "grad_norm": 0.09808441117021358, "learning_rate": 5.377970399934711e-06, "loss": 0.8363, "step": 9381 }, { "epoch": 4.249094202898551, "grad_norm": 0.08294859005218393, "learning_rate": 5.371638087434172e-06, "loss": 0.8409, "step": 9382 }, { "epoch": 4.249547101449275, "grad_norm": 0.08698206004458943, "learning_rate": 5.365309236785856e-06, "loss": 0.8586, "step": 9383 }, { "epoch": 4.25, "grad_norm": 0.09094140949208049, "learning_rate": 5.358983848622452e-06, "loss": 0.8192, "step": 9384 }, { "epoch": 4.250452898550725, "grad_norm": 0.08473419681800787, "learning_rate": 5.352661923576334e-06, "loss": 0.8528, "step": 9385 }, { "epoch": 4.250905797101449, "grad_norm": 0.09211143472915623, "learning_rate": 5.3463434622795e-06, "loss": 0.873, "step": 9386 }, { "epoch": 4.251358695652174, "grad_norm": 0.09275536506497127, "learning_rate": 5.3400284653636295e-06, "loss": 0.8357, "step": 9387 }, { "epoch": 4.2518115942028984, "grad_norm": 0.07228628413410704, "learning_rate": 5.33371693346004e-06, "loss": 0.8443, "step": 9388 }, { "epoch": 4.252264492753623, "grad_norm": 0.07761767378393807, "learning_rate": 5.327408867199699e-06, "loss": 0.8462, "step": 9389 }, { "epoch": 4.252717391304348, "grad_norm": 0.09846907444325674, "learning_rate": 5.3211042672132305e-06, "loss": 0.8486, "step": 9390 }, { "epoch": 4.253170289855072, "grad_norm": 0.08337062339452937, "learning_rate": 5.31480313413093e-06, "loss": 0.8554, "step": 9391 }, { "epoch": 4.253623188405797, "grad_norm": 0.0894991374671169, "learning_rate": 5.308505468582712e-06, "loss": 0.8649, "step": 9392 }, { "epoch": 4.2540760869565215, "grad_norm": 0.09140689922170035, "learning_rate": 5.302211271198188e-06, "loss": 0.8292, "step": 9393 }, { "epoch": 4.254528985507246, "grad_norm": 0.08863121312742335, "learning_rate": 5.2959205426065744e-06, "loss": 0.8441, "step": 9394 }, { "epoch": 4.254981884057971, "grad_norm": 0.09528074780267813, "learning_rate": 5.28963328343679e-06, "loss": 0.8555, "step": 9395 }, { "epoch": 4.255434782608695, "grad_norm": 0.08731232441683003, "learning_rate": 5.283349494317369e-06, "loss": 0.8333, "step": 9396 }, { "epoch": 4.25588768115942, "grad_norm": 0.09574068000758415, "learning_rate": 5.27706917587651e-06, "loss": 0.8319, "step": 9397 }, { "epoch": 4.256340579710145, "grad_norm": 0.09562689536931364, "learning_rate": 5.270792328742062e-06, "loss": 0.8358, "step": 9398 }, { "epoch": 4.256793478260869, "grad_norm": 0.08658551207316942, "learning_rate": 5.264518953541546e-06, "loss": 0.8442, "step": 9399 }, { "epoch": 4.257246376811594, "grad_norm": 0.08260614634793442, "learning_rate": 5.258249050902104e-06, "loss": 0.8381, "step": 9400 }, { "epoch": 4.257699275362318, "grad_norm": 0.10771555201138781, "learning_rate": 5.251982621450564e-06, "loss": 0.84, "step": 9401 }, { "epoch": 4.258152173913043, "grad_norm": 0.09855327973453742, "learning_rate": 5.2457196658133804e-06, "loss": 0.8598, "step": 9402 }, { "epoch": 4.2586050724637685, "grad_norm": 0.09914189147198048, "learning_rate": 5.239460184616682e-06, "loss": 0.8618, "step": 9403 }, { "epoch": 4.259057971014493, "grad_norm": 0.08657711323243339, "learning_rate": 5.23320417848622e-06, "loss": 0.8406, "step": 9404 }, { "epoch": 4.259510869565218, "grad_norm": 0.0998498062791056, "learning_rate": 5.226951648047442e-06, "loss": 0.8531, "step": 9405 }, { "epoch": 4.259963768115942, "grad_norm": 0.09088060450729708, "learning_rate": 5.220702593925407e-06, "loss": 0.8234, "step": 9406 }, { "epoch": 4.260416666666667, "grad_norm": 0.1083955197977849, "learning_rate": 5.2144570167448475e-06, "loss": 0.8561, "step": 9407 }, { "epoch": 4.260869565217392, "grad_norm": 0.0891293400157334, "learning_rate": 5.208214917130132e-06, "loss": 0.8603, "step": 9408 }, { "epoch": 4.261322463768116, "grad_norm": 0.09214922714357977, "learning_rate": 5.201976295705318e-06, "loss": 0.8517, "step": 9409 }, { "epoch": 4.261775362318841, "grad_norm": 0.09543036627151688, "learning_rate": 5.19574115309406e-06, "loss": 0.8596, "step": 9410 }, { "epoch": 4.262228260869565, "grad_norm": 0.07734733371383833, "learning_rate": 5.189509489919724e-06, "loss": 0.832, "step": 9411 }, { "epoch": 4.26268115942029, "grad_norm": 0.09672314884305953, "learning_rate": 5.1832813068052765e-06, "loss": 0.8729, "step": 9412 }, { "epoch": 4.263134057971015, "grad_norm": 0.0891701816459389, "learning_rate": 5.177056604373376e-06, "loss": 0.8425, "step": 9413 }, { "epoch": 4.263586956521739, "grad_norm": 0.07739460890045838, "learning_rate": 5.170835383246311e-06, "loss": 0.8549, "step": 9414 }, { "epoch": 4.264039855072464, "grad_norm": 0.10017368074576752, "learning_rate": 5.164617644046024e-06, "loss": 0.851, "step": 9415 }, { "epoch": 4.2644927536231885, "grad_norm": 0.08454345557688053, "learning_rate": 5.158403387394102e-06, "loss": 0.8426, "step": 9416 }, { "epoch": 4.264945652173913, "grad_norm": 0.08163587007402097, "learning_rate": 5.152192613911813e-06, "loss": 0.8568, "step": 9417 }, { "epoch": 4.265398550724638, "grad_norm": 0.08423671926185418, "learning_rate": 5.145985324220042e-06, "loss": 0.8556, "step": 9418 }, { "epoch": 4.265851449275362, "grad_norm": 0.08215042992476893, "learning_rate": 5.13978151893936e-06, "loss": 0.8728, "step": 9419 }, { "epoch": 4.266304347826087, "grad_norm": 0.07673210542288325, "learning_rate": 5.133581198689946e-06, "loss": 0.8653, "step": 9420 }, { "epoch": 4.2667572463768115, "grad_norm": 0.08223010751256794, "learning_rate": 5.127384364091685e-06, "loss": 0.8532, "step": 9421 }, { "epoch": 4.267210144927536, "grad_norm": 0.07841113950813022, "learning_rate": 5.121191015764066e-06, "loss": 0.8545, "step": 9422 }, { "epoch": 4.267663043478261, "grad_norm": 0.08564396298825783, "learning_rate": 5.1150011543262516e-06, "loss": 0.8455, "step": 9423 }, { "epoch": 4.268115942028985, "grad_norm": 0.08116175021669696, "learning_rate": 5.108814780397042e-06, "loss": 0.8636, "step": 9424 }, { "epoch": 4.26856884057971, "grad_norm": 0.08277306485814721, "learning_rate": 5.102631894594919e-06, "loss": 0.8684, "step": 9425 }, { "epoch": 4.269021739130435, "grad_norm": 0.09040080775258462, "learning_rate": 5.096452497537985e-06, "loss": 0.8602, "step": 9426 }, { "epoch": 4.269474637681159, "grad_norm": 0.07679447008180486, "learning_rate": 5.090276589843996e-06, "loss": 0.8293, "step": 9427 }, { "epoch": 4.269927536231884, "grad_norm": 0.0807875156577604, "learning_rate": 5.084104172130384e-06, "loss": 0.8422, "step": 9428 }, { "epoch": 4.270380434782608, "grad_norm": 0.09174790538066095, "learning_rate": 5.077935245014196e-06, "loss": 0.8451, "step": 9429 }, { "epoch": 4.270833333333333, "grad_norm": 0.07791721967898599, "learning_rate": 5.07176980911217e-06, "loss": 0.8436, "step": 9430 }, { "epoch": 4.271286231884058, "grad_norm": 0.09263679329000618, "learning_rate": 5.065607865040667e-06, "loss": 0.8357, "step": 9431 }, { "epoch": 4.271739130434782, "grad_norm": 0.09589197210303986, "learning_rate": 5.059449413415705e-06, "loss": 0.8403, "step": 9432 }, { "epoch": 4.272192028985507, "grad_norm": 0.08161520523175991, "learning_rate": 5.053294454852942e-06, "loss": 0.8355, "step": 9433 }, { "epoch": 4.2726449275362315, "grad_norm": 0.08982077837290646, "learning_rate": 5.047142989967717e-06, "loss": 0.8667, "step": 9434 }, { "epoch": 4.273097826086957, "grad_norm": 0.09354155171450607, "learning_rate": 5.040995019374988e-06, "loss": 0.8609, "step": 9435 }, { "epoch": 4.273550724637682, "grad_norm": 0.09128370528502706, "learning_rate": 5.034850543689391e-06, "loss": 0.8307, "step": 9436 }, { "epoch": 4.274003623188406, "grad_norm": 0.07618739333196554, "learning_rate": 5.028709563525187e-06, "loss": 0.8422, "step": 9437 }, { "epoch": 4.274456521739131, "grad_norm": 0.08915019701254856, "learning_rate": 5.0225720794963105e-06, "loss": 0.839, "step": 9438 }, { "epoch": 4.274909420289855, "grad_norm": 0.09768654797708953, "learning_rate": 5.0164380922163335e-06, "loss": 0.8391, "step": 9439 }, { "epoch": 4.27536231884058, "grad_norm": 0.09126174402158671, "learning_rate": 5.010307602298472e-06, "loss": 0.8434, "step": 9440 }, { "epoch": 4.275815217391305, "grad_norm": 0.09169465359036204, "learning_rate": 5.0041806103556e-06, "loss": 0.8713, "step": 9441 }, { "epoch": 4.276268115942029, "grad_norm": 0.09545379367797803, "learning_rate": 4.998057117000255e-06, "loss": 0.8565, "step": 9442 }, { "epoch": 4.276721014492754, "grad_norm": 0.08497762261428882, "learning_rate": 4.991937122844599e-06, "loss": 0.8455, "step": 9443 }, { "epoch": 4.2771739130434785, "grad_norm": 0.0789215750887753, "learning_rate": 4.985820628500473e-06, "loss": 0.8667, "step": 9444 }, { "epoch": 4.277626811594203, "grad_norm": 0.09387468134425798, "learning_rate": 4.979707634579329e-06, "loss": 0.8406, "step": 9445 }, { "epoch": 4.278079710144928, "grad_norm": 0.09118748912315046, "learning_rate": 4.97359814169232e-06, "loss": 0.8536, "step": 9446 }, { "epoch": 4.278532608695652, "grad_norm": 0.10047617569072256, "learning_rate": 4.9674921504502e-06, "loss": 0.8448, "step": 9447 }, { "epoch": 4.278985507246377, "grad_norm": 0.0826141750538381, "learning_rate": 4.9613896614634136e-06, "loss": 0.8354, "step": 9448 }, { "epoch": 4.2794384057971016, "grad_norm": 0.1008842247558892, "learning_rate": 4.955290675342022e-06, "loss": 0.8511, "step": 9449 }, { "epoch": 4.279891304347826, "grad_norm": 0.08411389517655611, "learning_rate": 4.9491951926957574e-06, "loss": 0.861, "step": 9450 }, { "epoch": 4.280344202898551, "grad_norm": 0.09341896582121362, "learning_rate": 4.943103214133986e-06, "loss": 0.8442, "step": 9451 }, { "epoch": 4.280797101449275, "grad_norm": 0.08750568355854015, "learning_rate": 4.937014740265742e-06, "loss": 0.8425, "step": 9452 }, { "epoch": 4.28125, "grad_norm": 0.08841655054472146, "learning_rate": 4.930929771699693e-06, "loss": 0.8553, "step": 9453 }, { "epoch": 4.281702898550725, "grad_norm": 0.10427548781891441, "learning_rate": 4.92484830904417e-06, "loss": 0.8443, "step": 9454 }, { "epoch": 4.282155797101449, "grad_norm": 0.11579805528507915, "learning_rate": 4.918770352907136e-06, "loss": 0.8511, "step": 9455 }, { "epoch": 4.282608695652174, "grad_norm": 0.08282573367649516, "learning_rate": 4.91269590389623e-06, "loss": 0.869, "step": 9456 }, { "epoch": 4.2830615942028984, "grad_norm": 0.09210330764635517, "learning_rate": 4.9066249626187155e-06, "loss": 0.8318, "step": 9457 }, { "epoch": 4.283514492753623, "grad_norm": 0.1057557067061972, "learning_rate": 4.900557529681513e-06, "loss": 0.8667, "step": 9458 }, { "epoch": 4.283967391304348, "grad_norm": 0.09167004585757932, "learning_rate": 4.894493605691186e-06, "loss": 0.8492, "step": 9459 }, { "epoch": 4.284420289855072, "grad_norm": 0.11892122461884798, "learning_rate": 4.888433191253974e-06, "loss": 0.863, "step": 9460 }, { "epoch": 4.284873188405797, "grad_norm": 0.10020342186005582, "learning_rate": 4.882376286975721e-06, "loss": 0.8456, "step": 9461 }, { "epoch": 4.2853260869565215, "grad_norm": 0.09674075220286689, "learning_rate": 4.876322893461974e-06, "loss": 0.8591, "step": 9462 }, { "epoch": 4.285778985507246, "grad_norm": 0.08519864063637089, "learning_rate": 4.87027301131787e-06, "loss": 0.8361, "step": 9463 }, { "epoch": 4.286231884057971, "grad_norm": 0.09509681908764067, "learning_rate": 4.864226641148256e-06, "loss": 0.8331, "step": 9464 }, { "epoch": 4.286684782608695, "grad_norm": 0.08984040471102457, "learning_rate": 4.858183783557575e-06, "loss": 0.843, "step": 9465 }, { "epoch": 4.28713768115942, "grad_norm": 0.09105976809382468, "learning_rate": 4.852144439149955e-06, "loss": 0.8571, "step": 9466 }, { "epoch": 4.287590579710145, "grad_norm": 0.08624807164049647, "learning_rate": 4.846108608529139e-06, "loss": 0.8472, "step": 9467 }, { "epoch": 4.288043478260869, "grad_norm": 0.11060599845027398, "learning_rate": 4.8400762922985635e-06, "loss": 0.8612, "step": 9468 }, { "epoch": 4.288496376811594, "grad_norm": 0.10117055349019237, "learning_rate": 4.834047491061267e-06, "loss": 0.8313, "step": 9469 }, { "epoch": 4.288949275362318, "grad_norm": 0.10050677237616881, "learning_rate": 4.828022205419975e-06, "loss": 0.8392, "step": 9470 }, { "epoch": 4.289402173913043, "grad_norm": 0.09343494344160881, "learning_rate": 4.82200043597703e-06, "loss": 0.8354, "step": 9471 }, { "epoch": 4.2898550724637685, "grad_norm": 0.10824630045176058, "learning_rate": 4.815982183334456e-06, "loss": 0.85, "step": 9472 }, { "epoch": 4.290307971014493, "grad_norm": 0.09605601815703035, "learning_rate": 4.8099674480938955e-06, "loss": 0.8543, "step": 9473 }, { "epoch": 4.290760869565218, "grad_norm": 0.08169915266512416, "learning_rate": 4.80395623085665e-06, "loss": 0.8561, "step": 9474 }, { "epoch": 4.291213768115942, "grad_norm": 0.0962837968899622, "learning_rate": 4.797948532223666e-06, "loss": 0.8437, "step": 9475 }, { "epoch": 4.291666666666667, "grad_norm": 0.10072243332876592, "learning_rate": 4.791944352795561e-06, "loss": 0.8704, "step": 9476 }, { "epoch": 4.292119565217392, "grad_norm": 0.07883246223934169, "learning_rate": 4.785943693172561e-06, "loss": 0.8524, "step": 9477 }, { "epoch": 4.292572463768116, "grad_norm": 0.09449864961087948, "learning_rate": 4.779946553954578e-06, "loss": 0.8589, "step": 9478 }, { "epoch": 4.293025362318841, "grad_norm": 0.0927188193635233, "learning_rate": 4.773952935741144e-06, "loss": 0.8648, "step": 9479 }, { "epoch": 4.293478260869565, "grad_norm": 0.10146548907515758, "learning_rate": 4.7679628391314615e-06, "loss": 0.8636, "step": 9480 }, { "epoch": 4.29393115942029, "grad_norm": 0.082327596286561, "learning_rate": 4.761976264724366e-06, "loss": 0.844, "step": 9481 }, { "epoch": 4.294384057971015, "grad_norm": 0.09487683637480096, "learning_rate": 4.7559932131183395e-06, "loss": 0.8353, "step": 9482 }, { "epoch": 4.294836956521739, "grad_norm": 0.09483944031745173, "learning_rate": 4.750013684911512e-06, "loss": 0.8591, "step": 9483 }, { "epoch": 4.295289855072464, "grad_norm": 0.09446839067977708, "learning_rate": 4.744037680701686e-06, "loss": 0.8401, "step": 9484 }, { "epoch": 4.2957427536231885, "grad_norm": 0.09639915174124498, "learning_rate": 4.73806520108627e-06, "loss": 0.8639, "step": 9485 }, { "epoch": 4.296195652173913, "grad_norm": 0.0826931303996065, "learning_rate": 4.732096246662368e-06, "loss": 0.8472, "step": 9486 }, { "epoch": 4.296648550724638, "grad_norm": 0.0821535538039084, "learning_rate": 4.7261308180266774e-06, "loss": 0.8529, "step": 9487 }, { "epoch": 4.297101449275362, "grad_norm": 0.09475029986390472, "learning_rate": 4.7201689157755935e-06, "loss": 0.85, "step": 9488 }, { "epoch": 4.297554347826087, "grad_norm": 0.10455484033114198, "learning_rate": 4.71421054050512e-06, "loss": 0.8365, "step": 9489 }, { "epoch": 4.2980072463768115, "grad_norm": 0.09665183465332727, "learning_rate": 4.7082556928109435e-06, "loss": 0.8419, "step": 9490 }, { "epoch": 4.298460144927536, "grad_norm": 0.09816718839969246, "learning_rate": 4.702304373288367e-06, "loss": 0.8578, "step": 9491 }, { "epoch": 4.298913043478261, "grad_norm": 0.09431323065346171, "learning_rate": 4.6963565825323576e-06, "loss": 0.8472, "step": 9492 }, { "epoch": 4.299365942028985, "grad_norm": 0.09827715618265763, "learning_rate": 4.690412321137521e-06, "loss": 0.8615, "step": 9493 }, { "epoch": 4.29981884057971, "grad_norm": 0.104939554605521, "learning_rate": 4.684471589698119e-06, "loss": 0.8431, "step": 9494 }, { "epoch": 4.300271739130435, "grad_norm": 0.08752918392847063, "learning_rate": 4.678534388808049e-06, "loss": 0.8619, "step": 9495 }, { "epoch": 4.300724637681159, "grad_norm": 0.08593625993592972, "learning_rate": 4.672600719060878e-06, "loss": 0.8532, "step": 9496 }, { "epoch": 4.301177536231884, "grad_norm": 0.09498518671195848, "learning_rate": 4.666670581049784e-06, "loss": 0.8479, "step": 9497 }, { "epoch": 4.301630434782608, "grad_norm": 0.08674634783935152, "learning_rate": 4.66074397536763e-06, "loss": 0.8382, "step": 9498 }, { "epoch": 4.302083333333333, "grad_norm": 0.09734920502691492, "learning_rate": 4.654820902606898e-06, "loss": 0.8386, "step": 9499 }, { "epoch": 4.302536231884058, "grad_norm": 0.09226799712978498, "learning_rate": 4.648901363359737e-06, "loss": 0.8563, "step": 9500 }, { "epoch": 4.302989130434782, "grad_norm": 0.08788852145009117, "learning_rate": 4.642985358217913e-06, "loss": 0.8545, "step": 9501 }, { "epoch": 4.303442028985507, "grad_norm": 0.09047354868286901, "learning_rate": 4.637072887772877e-06, "loss": 0.8746, "step": 9502 }, { "epoch": 4.3038949275362315, "grad_norm": 0.09948483543317978, "learning_rate": 4.631163952615696e-06, "loss": 0.8245, "step": 9503 }, { "epoch": 4.304347826086957, "grad_norm": 0.10015946061574518, "learning_rate": 4.625258553337108e-06, "loss": 0.8624, "step": 9504 }, { "epoch": 4.304800724637682, "grad_norm": 0.07767870190998725, "learning_rate": 4.61935669052747e-06, "loss": 0.8567, "step": 9505 }, { "epoch": 4.305253623188406, "grad_norm": 0.10109594890284018, "learning_rate": 4.613458364776819e-06, "loss": 0.8342, "step": 9506 }, { "epoch": 4.305706521739131, "grad_norm": 0.08834721773217485, "learning_rate": 4.60756357667481e-06, "loss": 0.8256, "step": 9507 }, { "epoch": 4.306159420289855, "grad_norm": 0.08531627983203945, "learning_rate": 4.6016723268107555e-06, "loss": 0.8397, "step": 9508 }, { "epoch": 4.30661231884058, "grad_norm": 0.09118955288959939, "learning_rate": 4.595784615773599e-06, "loss": 0.8507, "step": 9509 }, { "epoch": 4.307065217391305, "grad_norm": 0.09832414417866281, "learning_rate": 4.589900444151973e-06, "loss": 0.841, "step": 9510 }, { "epoch": 4.307518115942029, "grad_norm": 0.10522731841601958, "learning_rate": 4.5840198125341e-06, "loss": 0.8363, "step": 9511 }, { "epoch": 4.307971014492754, "grad_norm": 0.09859917471551402, "learning_rate": 4.578142721507894e-06, "loss": 0.8427, "step": 9512 }, { "epoch": 4.3084239130434785, "grad_norm": 0.1025138587282903, "learning_rate": 4.572269171660883e-06, "loss": 0.8562, "step": 9513 }, { "epoch": 4.308876811594203, "grad_norm": 0.08611700918048416, "learning_rate": 4.566399163580273e-06, "loss": 0.8527, "step": 9514 }, { "epoch": 4.309329710144928, "grad_norm": 0.08967504588207839, "learning_rate": 4.560532697852886e-06, "loss": 0.8438, "step": 9515 }, { "epoch": 4.309782608695652, "grad_norm": 0.09926784506170518, "learning_rate": 4.554669775065206e-06, "loss": 0.8643, "step": 9516 }, { "epoch": 4.310235507246377, "grad_norm": 0.08802364447011783, "learning_rate": 4.548810395803345e-06, "loss": 0.8137, "step": 9517 }, { "epoch": 4.3106884057971016, "grad_norm": 0.09179878577769003, "learning_rate": 4.542954560653097e-06, "loss": 0.8634, "step": 9518 }, { "epoch": 4.311141304347826, "grad_norm": 0.0922874775529859, "learning_rate": 4.53710227019986e-06, "loss": 0.8671, "step": 9519 }, { "epoch": 4.311594202898551, "grad_norm": 0.08005291730271367, "learning_rate": 4.5312535250287136e-06, "loss": 0.8607, "step": 9520 }, { "epoch": 4.312047101449275, "grad_norm": 0.09950687989183402, "learning_rate": 4.525408325724346e-06, "loss": 0.8311, "step": 9521 }, { "epoch": 4.3125, "grad_norm": 0.09034023282651446, "learning_rate": 4.519566672871132e-06, "loss": 0.8495, "step": 9522 }, { "epoch": 4.312952898550725, "grad_norm": 0.08173332142417859, "learning_rate": 4.513728567053064e-06, "loss": 0.857, "step": 9523 }, { "epoch": 4.313405797101449, "grad_norm": 0.08624967864153073, "learning_rate": 4.507894008853781e-06, "loss": 0.8516, "step": 9524 }, { "epoch": 4.313858695652174, "grad_norm": 0.08346161668516587, "learning_rate": 4.502062998856569e-06, "loss": 0.8424, "step": 9525 }, { "epoch": 4.3143115942028984, "grad_norm": 0.08664448188802233, "learning_rate": 4.496235537644378e-06, "loss": 0.8613, "step": 9526 }, { "epoch": 4.314764492753623, "grad_norm": 0.08447240225184804, "learning_rate": 4.490411625799769e-06, "loss": 0.8588, "step": 9527 }, { "epoch": 4.315217391304348, "grad_norm": 0.0893556496818855, "learning_rate": 4.484591263904992e-06, "loss": 0.8428, "step": 9528 }, { "epoch": 4.315670289855072, "grad_norm": 0.09911413849989414, "learning_rate": 4.478774452541896e-06, "loss": 0.8423, "step": 9529 }, { "epoch": 4.316123188405797, "grad_norm": 0.07649289961067954, "learning_rate": 4.472961192292018e-06, "loss": 0.8283, "step": 9530 }, { "epoch": 4.3165760869565215, "grad_norm": 0.08200189564631137, "learning_rate": 4.467151483736496e-06, "loss": 0.8579, "step": 9531 }, { "epoch": 4.317028985507246, "grad_norm": 0.09038341273607473, "learning_rate": 4.461345327456159e-06, "loss": 0.8472, "step": 9532 }, { "epoch": 4.317481884057971, "grad_norm": 0.08955105709325284, "learning_rate": 4.455542724031436e-06, "loss": 0.8481, "step": 9533 }, { "epoch": 4.317934782608695, "grad_norm": 0.08567853323619237, "learning_rate": 4.449743674042437e-06, "loss": 0.8336, "step": 9534 }, { "epoch": 4.31838768115942, "grad_norm": 0.10202130918104349, "learning_rate": 4.443948178068888e-06, "loss": 0.8664, "step": 9535 }, { "epoch": 4.318840579710145, "grad_norm": 0.08331003133914948, "learning_rate": 4.438156236690194e-06, "loss": 0.846, "step": 9536 }, { "epoch": 4.319293478260869, "grad_norm": 0.0752974584536794, "learning_rate": 4.432367850485362e-06, "loss": 0.8548, "step": 9537 }, { "epoch": 4.319746376811594, "grad_norm": 0.0941032301380768, "learning_rate": 4.426583020033088e-06, "loss": 0.8455, "step": 9538 }, { "epoch": 4.320199275362318, "grad_norm": 0.08554245267531978, "learning_rate": 4.420801745911672e-06, "loss": 0.8394, "step": 9539 }, { "epoch": 4.320652173913043, "grad_norm": 0.08855800892825803, "learning_rate": 4.415024028699093e-06, "loss": 0.8403, "step": 9540 }, { "epoch": 4.3211050724637685, "grad_norm": 0.0727246470317382, "learning_rate": 4.409249868972954e-06, "loss": 0.8376, "step": 9541 }, { "epoch": 4.321557971014493, "grad_norm": 0.07994901284657752, "learning_rate": 4.4034792673104974e-06, "loss": 0.8545, "step": 9542 }, { "epoch": 4.322010869565218, "grad_norm": 0.08727980921202892, "learning_rate": 4.397712224288624e-06, "loss": 0.8507, "step": 9543 }, { "epoch": 4.322463768115942, "grad_norm": 0.08223527099006353, "learning_rate": 4.39194874048388e-06, "loss": 0.8593, "step": 9544 }, { "epoch": 4.322916666666667, "grad_norm": 0.07381002559753765, "learning_rate": 4.386188816472441e-06, "loss": 0.8332, "step": 9545 }, { "epoch": 4.323369565217392, "grad_norm": 0.09679211995642298, "learning_rate": 4.380432452830148e-06, "loss": 0.8581, "step": 9546 }, { "epoch": 4.323822463768116, "grad_norm": 0.09222766397387226, "learning_rate": 4.374679650132461e-06, "loss": 0.8586, "step": 9547 }, { "epoch": 4.324275362318841, "grad_norm": 0.07774066491331105, "learning_rate": 4.36893040895451e-06, "loss": 0.8441, "step": 9548 }, { "epoch": 4.324728260869565, "grad_norm": 0.0977928929189919, "learning_rate": 4.363184729871046e-06, "loss": 0.8657, "step": 9549 }, { "epoch": 4.32518115942029, "grad_norm": 0.09795605407238986, "learning_rate": 4.357442613456475e-06, "loss": 0.8467, "step": 9550 }, { "epoch": 4.325634057971015, "grad_norm": 0.09703447951435865, "learning_rate": 4.35170406028485e-06, "loss": 0.8675, "step": 9551 }, { "epoch": 4.326086956521739, "grad_norm": 0.09691677996833462, "learning_rate": 4.345969070929848e-06, "loss": 0.8333, "step": 9552 }, { "epoch": 4.326539855072464, "grad_norm": 0.10231346983862867, "learning_rate": 4.340237645964829e-06, "loss": 0.863, "step": 9553 }, { "epoch": 4.3269927536231885, "grad_norm": 0.08809613587028746, "learning_rate": 4.334509785962749e-06, "loss": 0.846, "step": 9554 }, { "epoch": 4.327445652173913, "grad_norm": 0.08847577693390488, "learning_rate": 4.32878549149625e-06, "loss": 0.8234, "step": 9555 }, { "epoch": 4.327898550724638, "grad_norm": 0.08203571708138341, "learning_rate": 4.323064763137587e-06, "loss": 0.8414, "step": 9556 }, { "epoch": 4.328351449275362, "grad_norm": 0.09935982334548467, "learning_rate": 4.317347601458677e-06, "loss": 0.858, "step": 9557 }, { "epoch": 4.328804347826087, "grad_norm": 0.08583135281208583, "learning_rate": 4.311634007031078e-06, "loss": 0.8488, "step": 9558 }, { "epoch": 4.3292572463768115, "grad_norm": 0.08877834617758826, "learning_rate": 4.3059239804259744e-06, "loss": 0.8449, "step": 9559 }, { "epoch": 4.329710144927536, "grad_norm": 0.08269761386545058, "learning_rate": 4.300217522214208e-06, "loss": 0.8456, "step": 9560 }, { "epoch": 4.330163043478261, "grad_norm": 0.07638022658614191, "learning_rate": 4.2945146329662756e-06, "loss": 0.8326, "step": 9561 }, { "epoch": 4.330615942028985, "grad_norm": 0.1071451591131385, "learning_rate": 4.288815313252284e-06, "loss": 0.8474, "step": 9562 }, { "epoch": 4.33106884057971, "grad_norm": 0.09027198840060614, "learning_rate": 4.283119563642024e-06, "loss": 0.8492, "step": 9563 }, { "epoch": 4.331521739130435, "grad_norm": 0.08595674051951006, "learning_rate": 4.277427384704895e-06, "loss": 0.8634, "step": 9564 }, { "epoch": 4.331974637681159, "grad_norm": 0.09500973668371429, "learning_rate": 4.271738777009962e-06, "loss": 0.8361, "step": 9565 }, { "epoch": 4.332427536231884, "grad_norm": 0.10401462123334693, "learning_rate": 4.266053741125919e-06, "loss": 0.8538, "step": 9566 }, { "epoch": 4.332880434782608, "grad_norm": 0.09079912266454004, "learning_rate": 4.260372277621114e-06, "loss": 0.8482, "step": 9567 }, { "epoch": 4.333333333333333, "grad_norm": 0.0962627543957252, "learning_rate": 4.254694387063514e-06, "loss": 0.8533, "step": 9568 }, { "epoch": 4.333786231884058, "grad_norm": 0.1047625832224612, "learning_rate": 4.249020070020766e-06, "loss": 0.8518, "step": 9569 }, { "epoch": 4.334239130434782, "grad_norm": 0.08537172875280197, "learning_rate": 4.243349327060129e-06, "loss": 0.8433, "step": 9570 }, { "epoch": 4.334692028985507, "grad_norm": 0.07680466062802055, "learning_rate": 4.237682158748531e-06, "loss": 0.8349, "step": 9571 }, { "epoch": 4.3351449275362315, "grad_norm": 0.10882383178613737, "learning_rate": 4.232018565652505e-06, "loss": 0.8542, "step": 9572 }, { "epoch": 4.335597826086957, "grad_norm": 0.10354478197814437, "learning_rate": 4.226358548338274e-06, "loss": 0.87, "step": 9573 }, { "epoch": 4.336050724637682, "grad_norm": 0.07913079393430313, "learning_rate": 4.2207021073716656e-06, "loss": 0.8504, "step": 9574 }, { "epoch": 4.336503623188406, "grad_norm": 0.09252406260260876, "learning_rate": 4.215049243318165e-06, "loss": 0.8428, "step": 9575 }, { "epoch": 4.336956521739131, "grad_norm": 0.10607009232798718, "learning_rate": 4.20939995674289e-06, "loss": 0.8394, "step": 9576 }, { "epoch": 4.337409420289855, "grad_norm": 0.08633112746379762, "learning_rate": 4.203754248210623e-06, "loss": 0.8448, "step": 9577 }, { "epoch": 4.33786231884058, "grad_norm": 0.0897654698079039, "learning_rate": 4.1981121182857624e-06, "loss": 0.832, "step": 9578 }, { "epoch": 4.338315217391305, "grad_norm": 0.08642143458073182, "learning_rate": 4.192473567532372e-06, "loss": 0.8476, "step": 9579 }, { "epoch": 4.338768115942029, "grad_norm": 0.09164074027175756, "learning_rate": 4.186838596514133e-06, "loss": 0.8623, "step": 9580 }, { "epoch": 4.339221014492754, "grad_norm": 0.08981584260174222, "learning_rate": 4.181207205794397e-06, "loss": 0.8338, "step": 9581 }, { "epoch": 4.3396739130434785, "grad_norm": 0.08148709239436427, "learning_rate": 4.175579395936131e-06, "loss": 0.8534, "step": 9582 }, { "epoch": 4.340126811594203, "grad_norm": 0.08679905218147725, "learning_rate": 4.169955167501969e-06, "loss": 0.8649, "step": 9583 }, { "epoch": 4.340579710144928, "grad_norm": 0.10634631464956826, "learning_rate": 4.164334521054167e-06, "loss": 0.8464, "step": 9584 }, { "epoch": 4.341032608695652, "grad_norm": 0.09298515901200885, "learning_rate": 4.158717457154628e-06, "loss": 0.8504, "step": 9585 }, { "epoch": 4.341485507246377, "grad_norm": 0.09881146249165523, "learning_rate": 4.153103976364894e-06, "loss": 0.8496, "step": 9586 }, { "epoch": 4.3419384057971016, "grad_norm": 0.10187042747821368, "learning_rate": 4.147494079246164e-06, "loss": 0.8577, "step": 9587 }, { "epoch": 4.342391304347826, "grad_norm": 0.08436045721669792, "learning_rate": 4.1418877663592605e-06, "loss": 0.8485, "step": 9588 }, { "epoch": 4.342844202898551, "grad_norm": 0.08887717115829076, "learning_rate": 4.136285038264665e-06, "loss": 0.876, "step": 9589 }, { "epoch": 4.343297101449275, "grad_norm": 0.10060547119716676, "learning_rate": 4.130685895522479e-06, "loss": 0.8627, "step": 9590 }, { "epoch": 4.34375, "grad_norm": 0.09194314760977237, "learning_rate": 4.12509033869247e-06, "loss": 0.8435, "step": 9591 }, { "epoch": 4.344202898550725, "grad_norm": 0.08171682951801043, "learning_rate": 4.119498368334029e-06, "loss": 0.854, "step": 9592 }, { "epoch": 4.344655797101449, "grad_norm": 0.08535295077925467, "learning_rate": 4.11390998500619e-06, "loss": 0.8493, "step": 9593 }, { "epoch": 4.345108695652174, "grad_norm": 0.08854698842680205, "learning_rate": 4.108325189267634e-06, "loss": 0.8383, "step": 9594 }, { "epoch": 4.3455615942028984, "grad_norm": 0.09999910160658988, "learning_rate": 4.102743981676689e-06, "loss": 0.8591, "step": 9595 }, { "epoch": 4.346014492753623, "grad_norm": 0.08568495289635181, "learning_rate": 4.0971663627913025e-06, "loss": 0.8429, "step": 9596 }, { "epoch": 4.346467391304348, "grad_norm": 0.0844789976939759, "learning_rate": 4.091592333169101e-06, "loss": 0.8566, "step": 9597 }, { "epoch": 4.346920289855072, "grad_norm": 0.07484800135022489, "learning_rate": 4.086021893367305e-06, "loss": 0.8495, "step": 9598 }, { "epoch": 4.347373188405797, "grad_norm": 0.08408517885229716, "learning_rate": 4.0804550439428195e-06, "loss": 0.8565, "step": 9599 }, { "epoch": 4.3478260869565215, "grad_norm": 0.08102659842250283, "learning_rate": 4.07489178545216e-06, "loss": 0.8413, "step": 9600 }, { "epoch": 4.348278985507246, "grad_norm": 0.08109742347258517, "learning_rate": 4.069332118451499e-06, "loss": 0.8528, "step": 9601 }, { "epoch": 4.348731884057971, "grad_norm": 0.08472913610517399, "learning_rate": 4.063776043496632e-06, "loss": 0.8446, "step": 9602 }, { "epoch": 4.349184782608695, "grad_norm": 0.10632018345540949, "learning_rate": 4.058223561143031e-06, "loss": 0.8478, "step": 9603 }, { "epoch": 4.34963768115942, "grad_norm": 0.08136173481363425, "learning_rate": 4.052674671945767e-06, "loss": 0.8555, "step": 9604 }, { "epoch": 4.350090579710145, "grad_norm": 0.08077480787466065, "learning_rate": 4.047129376459586e-06, "loss": 0.8545, "step": 9605 }, { "epoch": 4.350543478260869, "grad_norm": 0.10827755897256514, "learning_rate": 4.041587675238847e-06, "loss": 0.8349, "step": 9606 }, { "epoch": 4.350996376811594, "grad_norm": 0.08787572987626217, "learning_rate": 4.036049568837577e-06, "loss": 0.8688, "step": 9607 }, { "epoch": 4.351449275362318, "grad_norm": 0.07771878074427073, "learning_rate": 4.030515057809421e-06, "loss": 0.859, "step": 9608 }, { "epoch": 4.351902173913043, "grad_norm": 0.11038931002692635, "learning_rate": 4.0249841427076756e-06, "loss": 0.8669, "step": 9609 }, { "epoch": 4.3523550724637685, "grad_norm": 0.07999909873723005, "learning_rate": 4.019456824085262e-06, "loss": 0.8507, "step": 9610 }, { "epoch": 4.352807971014493, "grad_norm": 0.08723955517825235, "learning_rate": 4.013933102494774e-06, "loss": 0.8403, "step": 9611 }, { "epoch": 4.353260869565218, "grad_norm": 0.09126820404480165, "learning_rate": 4.008412978488409e-06, "loss": 0.8606, "step": 9612 }, { "epoch": 4.353713768115942, "grad_norm": 0.11840623960696198, "learning_rate": 4.002896452618044e-06, "loss": 0.8505, "step": 9613 }, { "epoch": 4.354166666666667, "grad_norm": 0.0995693776758568, "learning_rate": 3.997383525435154e-06, "loss": 0.8432, "step": 9614 }, { "epoch": 4.354619565217392, "grad_norm": 0.10720009978099863, "learning_rate": 3.991874197490888e-06, "loss": 0.8462, "step": 9615 }, { "epoch": 4.355072463768116, "grad_norm": 0.1326532981845215, "learning_rate": 3.986368469336022e-06, "loss": 0.8486, "step": 9616 }, { "epoch": 4.355525362318841, "grad_norm": 0.07806961852514142, "learning_rate": 3.980866341520968e-06, "loss": 0.8695, "step": 9617 }, { "epoch": 4.355978260869565, "grad_norm": 0.09202389921819236, "learning_rate": 3.975367814595776e-06, "loss": 0.8489, "step": 9618 }, { "epoch": 4.35643115942029, "grad_norm": 0.1284781208840976, "learning_rate": 3.969872889110153e-06, "loss": 0.8416, "step": 9619 }, { "epoch": 4.356884057971015, "grad_norm": 0.11962599844044959, "learning_rate": 3.9643815656134285e-06, "loss": 0.8402, "step": 9620 }, { "epoch": 4.357336956521739, "grad_norm": 0.11871367785571416, "learning_rate": 3.95889384465459e-06, "loss": 0.8498, "step": 9621 }, { "epoch": 4.357789855072464, "grad_norm": 0.10705134297783547, "learning_rate": 3.953409726782238e-06, "loss": 0.8518, "step": 9622 }, { "epoch": 4.3582427536231885, "grad_norm": 0.12063276700373117, "learning_rate": 3.947929212544641e-06, "loss": 0.8325, "step": 9623 }, { "epoch": 4.358695652173913, "grad_norm": 0.12503656890954545, "learning_rate": 3.942452302489686e-06, "loss": 0.8369, "step": 9624 }, { "epoch": 4.359148550724638, "grad_norm": 0.12069725068453054, "learning_rate": 3.936978997164915e-06, "loss": 0.8541, "step": 9625 }, { "epoch": 4.359601449275362, "grad_norm": 0.13307123150439606, "learning_rate": 3.931509297117506e-06, "loss": 0.8331, "step": 9626 }, { "epoch": 4.360054347826087, "grad_norm": 0.11888046911056778, "learning_rate": 3.926043202894265e-06, "loss": 0.8557, "step": 9627 }, { "epoch": 4.3605072463768115, "grad_norm": 0.08646170669985087, "learning_rate": 3.920580715041644e-06, "loss": 0.8496, "step": 9628 }, { "epoch": 4.360960144927536, "grad_norm": 0.12857041573321104, "learning_rate": 3.915121834105748e-06, "loss": 0.8582, "step": 9629 }, { "epoch": 4.361413043478261, "grad_norm": 0.11849852350038444, "learning_rate": 3.909666560632292e-06, "loss": 0.8604, "step": 9630 }, { "epoch": 4.361865942028985, "grad_norm": 0.10187668921356166, "learning_rate": 3.904214895166672e-06, "loss": 0.869, "step": 9631 }, { "epoch": 4.36231884057971, "grad_norm": 0.1128165448705145, "learning_rate": 3.89876683825388e-06, "loss": 0.8409, "step": 9632 }, { "epoch": 4.362771739130435, "grad_norm": 0.11627487807709234, "learning_rate": 3.893322390438581e-06, "loss": 0.8524, "step": 9633 }, { "epoch": 4.363224637681159, "grad_norm": 0.08817126700175633, "learning_rate": 3.887881552265058e-06, "loss": 0.8624, "step": 9634 }, { "epoch": 4.363677536231884, "grad_norm": 0.12718147500256125, "learning_rate": 3.882444324277241e-06, "loss": 0.8658, "step": 9635 }, { "epoch": 4.364130434782608, "grad_norm": 0.10284322121947492, "learning_rate": 3.877010707018691e-06, "loss": 0.8559, "step": 9636 }, { "epoch": 4.364583333333333, "grad_norm": 0.10112382588336324, "learning_rate": 3.871580701032631e-06, "loss": 0.8646, "step": 9637 }, { "epoch": 4.365036231884058, "grad_norm": 0.14363849203264212, "learning_rate": 3.866154306861889e-06, "loss": 0.8624, "step": 9638 }, { "epoch": 4.365489130434782, "grad_norm": 0.09230594177172624, "learning_rate": 3.860731525048969e-06, "loss": 0.8467, "step": 9639 }, { "epoch": 4.365942028985507, "grad_norm": 0.11769595275657355, "learning_rate": 3.855312356135978e-06, "loss": 0.8225, "step": 9640 }, { "epoch": 4.3663949275362315, "grad_norm": 0.1053094639626374, "learning_rate": 3.849896800664694e-06, "loss": 0.8533, "step": 9641 }, { "epoch": 4.366847826086957, "grad_norm": 0.09226485776006949, "learning_rate": 3.844484859176514e-06, "loss": 0.8558, "step": 9642 }, { "epoch": 4.367300724637682, "grad_norm": 0.10301887481663304, "learning_rate": 3.839076532212476e-06, "loss": 0.8704, "step": 9643 }, { "epoch": 4.367753623188406, "grad_norm": 0.09357274177742268, "learning_rate": 3.83367182031325e-06, "loss": 0.8583, "step": 9644 }, { "epoch": 4.368206521739131, "grad_norm": 0.10831071342581011, "learning_rate": 3.828270724019167e-06, "loss": 0.8485, "step": 9645 }, { "epoch": 4.368659420289855, "grad_norm": 0.11000331052557838, "learning_rate": 3.8228732438701755e-06, "loss": 0.8531, "step": 9646 }, { "epoch": 4.36911231884058, "grad_norm": 0.08808342658058982, "learning_rate": 3.8174793804058775e-06, "loss": 0.8505, "step": 9647 }, { "epoch": 4.369565217391305, "grad_norm": 0.11549772132632605, "learning_rate": 3.8120891341654955e-06, "loss": 0.8545, "step": 9648 }, { "epoch": 4.370018115942029, "grad_norm": 0.10596544022890558, "learning_rate": 3.8067025056879182e-06, "loss": 0.8477, "step": 9649 }, { "epoch": 4.370471014492754, "grad_norm": 0.08270804145284565, "learning_rate": 3.8013194955116397e-06, "loss": 0.8494, "step": 9650 }, { "epoch": 4.3709239130434785, "grad_norm": 0.1296308640357696, "learning_rate": 3.7959401041748155e-06, "loss": 0.8327, "step": 9651 }, { "epoch": 4.371376811594203, "grad_norm": 0.11798268297640303, "learning_rate": 3.790564332215225e-06, "loss": 0.8508, "step": 9652 }, { "epoch": 4.371829710144928, "grad_norm": 0.09285192702824145, "learning_rate": 3.785192180170301e-06, "loss": 0.8517, "step": 9653 }, { "epoch": 4.372282608695652, "grad_norm": 0.11521238547190156, "learning_rate": 3.7798236485770925e-06, "loss": 0.8505, "step": 9654 }, { "epoch": 4.372735507246377, "grad_norm": 0.12523733311459911, "learning_rate": 3.774458737972322e-06, "loss": 0.864, "step": 9655 }, { "epoch": 4.3731884057971016, "grad_norm": 0.09530513243084507, "learning_rate": 3.7690974488923027e-06, "loss": 0.8553, "step": 9656 }, { "epoch": 4.373641304347826, "grad_norm": 0.10290712909851396, "learning_rate": 3.763739781873037e-06, "loss": 0.8428, "step": 9657 }, { "epoch": 4.374094202898551, "grad_norm": 0.09953261959962363, "learning_rate": 3.758385737450123e-06, "loss": 0.8496, "step": 9658 }, { "epoch": 4.374547101449275, "grad_norm": 0.11655125865614731, "learning_rate": 3.7530353161588175e-06, "loss": 0.8442, "step": 9659 }, { "epoch": 4.375, "grad_norm": 0.10556903590625517, "learning_rate": 3.747688518534003e-06, "loss": 0.8577, "step": 9660 }, { "epoch": 4.375452898550725, "grad_norm": 0.090689822002905, "learning_rate": 3.742345345110221e-06, "loss": 0.8646, "step": 9661 }, { "epoch": 4.375905797101449, "grad_norm": 0.09783637155768236, "learning_rate": 3.7370057964216223e-06, "loss": 0.836, "step": 9662 }, { "epoch": 4.376358695652174, "grad_norm": 0.09005676246460727, "learning_rate": 3.731669873002024e-06, "loss": 0.8508, "step": 9663 }, { "epoch": 4.3768115942028984, "grad_norm": 0.11909091131336326, "learning_rate": 3.7263375753848573e-06, "loss": 0.8471, "step": 9664 }, { "epoch": 4.377264492753623, "grad_norm": 0.09496388571338851, "learning_rate": 3.7210089041032073e-06, "loss": 0.8529, "step": 9665 }, { "epoch": 4.377717391304348, "grad_norm": 0.08789335482367762, "learning_rate": 3.7156838596897803e-06, "loss": 0.8511, "step": 9666 }, { "epoch": 4.378170289855072, "grad_norm": 0.0847340423743701, "learning_rate": 3.7103624426769424e-06, "loss": 0.8607, "step": 9667 }, { "epoch": 4.378623188405797, "grad_norm": 0.10425791571630777, "learning_rate": 3.7050446535966765e-06, "loss": 0.8644, "step": 9668 }, { "epoch": 4.3790760869565215, "grad_norm": 0.08124628603414923, "learning_rate": 3.699730492980611e-06, "loss": 0.8821, "step": 9669 }, { "epoch": 4.379528985507246, "grad_norm": 0.08257617090123336, "learning_rate": 3.694419961360005e-06, "loss": 0.8507, "step": 9670 }, { "epoch": 4.379981884057971, "grad_norm": 0.08197989934086017, "learning_rate": 3.689113059265772e-06, "loss": 0.8421, "step": 9671 }, { "epoch": 4.380434782608695, "grad_norm": 0.07305610709315534, "learning_rate": 3.6838097872284382e-06, "loss": 0.8255, "step": 9672 }, { "epoch": 4.38088768115942, "grad_norm": 0.10296055532498344, "learning_rate": 3.678510145778198e-06, "loss": 0.8631, "step": 9673 }, { "epoch": 4.381340579710145, "grad_norm": 0.08291335020195219, "learning_rate": 3.6732141354448536e-06, "loss": 0.8363, "step": 9674 }, { "epoch": 4.381793478260869, "grad_norm": 0.0881580535106841, "learning_rate": 3.6679217567578486e-06, "loss": 0.8457, "step": 9675 }, { "epoch": 4.382246376811594, "grad_norm": 0.07461915442846306, "learning_rate": 3.662633010246288e-06, "loss": 0.8428, "step": 9676 }, { "epoch": 4.382699275362318, "grad_norm": 0.08948240235579803, "learning_rate": 3.657347896438883e-06, "loss": 0.8439, "step": 9677 }, { "epoch": 4.383152173913043, "grad_norm": 0.08525704447690777, "learning_rate": 3.652066415864002e-06, "loss": 0.8435, "step": 9678 }, { "epoch": 4.3836050724637685, "grad_norm": 0.07823038280115877, "learning_rate": 3.646788569049631e-06, "loss": 0.8435, "step": 9679 }, { "epoch": 4.384057971014493, "grad_norm": 0.0982630756691353, "learning_rate": 3.641514356523419e-06, "loss": 0.8217, "step": 9680 }, { "epoch": 4.384510869565218, "grad_norm": 0.0818913744132946, "learning_rate": 3.6362437788126292e-06, "loss": 0.8618, "step": 9681 }, { "epoch": 4.384963768115942, "grad_norm": 0.07296205797993738, "learning_rate": 3.6309768364441776e-06, "loss": 0.8471, "step": 9682 }, { "epoch": 4.385416666666667, "grad_norm": 0.08938748782996346, "learning_rate": 3.6257135299445943e-06, "loss": 0.8391, "step": 9683 }, { "epoch": 4.385869565217392, "grad_norm": 0.0799657625273312, "learning_rate": 3.620453859840076e-06, "loss": 0.8423, "step": 9684 }, { "epoch": 4.386322463768116, "grad_norm": 0.07589963709792366, "learning_rate": 3.615197826656438e-06, "loss": 0.8526, "step": 9685 }, { "epoch": 4.386775362318841, "grad_norm": 0.09046567768707106, "learning_rate": 3.6099454309191262e-06, "loss": 0.8593, "step": 9686 }, { "epoch": 4.387228260869565, "grad_norm": 0.07893680877710108, "learning_rate": 3.6046966731532275e-06, "loss": 0.8466, "step": 9687 }, { "epoch": 4.38768115942029, "grad_norm": 0.0728454821237571, "learning_rate": 3.599451553883482e-06, "loss": 0.8608, "step": 9688 }, { "epoch": 4.388134057971015, "grad_norm": 0.08340188341153108, "learning_rate": 3.5942100736342433e-06, "loss": 0.8588, "step": 9689 }, { "epoch": 4.388586956521739, "grad_norm": 0.08366047616828691, "learning_rate": 3.5889722329295152e-06, "loss": 0.867, "step": 9690 }, { "epoch": 4.389039855072464, "grad_norm": 0.07875909246425745, "learning_rate": 3.5837380322929273e-06, "loss": 0.8529, "step": 9691 }, { "epoch": 4.3894927536231885, "grad_norm": 0.07105169385989138, "learning_rate": 3.578507472247763e-06, "loss": 0.8316, "step": 9692 }, { "epoch": 4.389945652173913, "grad_norm": 0.07404510251042763, "learning_rate": 3.57328055331692e-06, "loss": 0.8417, "step": 9693 }, { "epoch": 4.390398550724638, "grad_norm": 0.11266833972047195, "learning_rate": 3.5680572760229446e-06, "loss": 0.8521, "step": 9694 }, { "epoch": 4.390851449275362, "grad_norm": 0.07502681228282188, "learning_rate": 3.5628376408880062e-06, "loss": 0.8232, "step": 9695 }, { "epoch": 4.391304347826087, "grad_norm": 0.08229555903663845, "learning_rate": 3.5576216484339356e-06, "loss": 0.8371, "step": 9696 }, { "epoch": 4.3917572463768115, "grad_norm": 0.08113583418347341, "learning_rate": 3.552409299182169e-06, "loss": 0.8712, "step": 9697 }, { "epoch": 4.392210144927536, "grad_norm": 0.07955591887975538, "learning_rate": 3.54720059365381e-06, "loss": 0.8596, "step": 9698 }, { "epoch": 4.392663043478261, "grad_norm": 0.08387558200609412, "learning_rate": 3.541995532369562e-06, "loss": 0.8377, "step": 9699 }, { "epoch": 4.393115942028985, "grad_norm": 0.10792185486108148, "learning_rate": 3.536794115849804e-06, "loss": 0.8433, "step": 9700 }, { "epoch": 4.39356884057971, "grad_norm": 0.07574289394317506, "learning_rate": 3.531596344614516e-06, "loss": 0.8519, "step": 9701 }, { "epoch": 4.394021739130435, "grad_norm": 0.0773808466506208, "learning_rate": 3.5264022191833313e-06, "loss": 0.8477, "step": 9702 }, { "epoch": 4.394474637681159, "grad_norm": 0.09922493178791891, "learning_rate": 3.5212117400755097e-06, "loss": 0.8317, "step": 9703 }, { "epoch": 4.394927536231884, "grad_norm": 0.0909332761889818, "learning_rate": 3.5160249078099606e-06, "loss": 0.8409, "step": 9704 }, { "epoch": 4.395380434782608, "grad_norm": 0.09473603819949798, "learning_rate": 3.510841722905207e-06, "loss": 0.8663, "step": 9705 }, { "epoch": 4.395833333333333, "grad_norm": 0.08179508059409783, "learning_rate": 3.5056621858794393e-06, "loss": 0.8655, "step": 9706 }, { "epoch": 4.396286231884058, "grad_norm": 0.095693515425119, "learning_rate": 3.5004862972504426e-06, "loss": 0.8356, "step": 9707 }, { "epoch": 4.396739130434782, "grad_norm": 0.0873504418395937, "learning_rate": 3.495314057535679e-06, "loss": 0.8515, "step": 9708 }, { "epoch": 4.397192028985507, "grad_norm": 0.07981171363887953, "learning_rate": 3.4901454672522063e-06, "loss": 0.8571, "step": 9709 }, { "epoch": 4.3976449275362315, "grad_norm": 0.09149916447602485, "learning_rate": 3.4849805269167615e-06, "loss": 0.8549, "step": 9710 }, { "epoch": 4.398097826086957, "grad_norm": 0.10496347992922656, "learning_rate": 3.479819237045661e-06, "loss": 0.8459, "step": 9711 }, { "epoch": 4.398550724637682, "grad_norm": 0.09140642630667084, "learning_rate": 3.4746615981549094e-06, "loss": 0.8632, "step": 9712 }, { "epoch": 4.399003623188406, "grad_norm": 0.08825241344222445, "learning_rate": 3.4695076107601124e-06, "loss": 0.8428, "step": 9713 }, { "epoch": 4.399456521739131, "grad_norm": 0.07828493203443293, "learning_rate": 3.464357275376533e-06, "loss": 0.8525, "step": 9714 }, { "epoch": 4.399909420289855, "grad_norm": 0.08563091211959281, "learning_rate": 3.4592105925190446e-06, "loss": 0.8616, "step": 9715 }, { "epoch": 4.40036231884058, "grad_norm": 0.10497210783594871, "learning_rate": 3.4540675627021856e-06, "loss": 0.8257, "step": 9716 }, { "epoch": 4.400815217391305, "grad_norm": 0.08002596436420047, "learning_rate": 3.448928186440097e-06, "loss": 0.8734, "step": 9717 }, { "epoch": 4.401268115942029, "grad_norm": 0.07820770334805978, "learning_rate": 3.4437924642465846e-06, "loss": 0.8398, "step": 9718 }, { "epoch": 4.401721014492754, "grad_norm": 0.07695766687679637, "learning_rate": 3.438660396635065e-06, "loss": 0.8487, "step": 9719 }, { "epoch": 4.4021739130434785, "grad_norm": 0.07297939095846159, "learning_rate": 3.4335319841186077e-06, "loss": 0.8464, "step": 9720 }, { "epoch": 4.402626811594203, "grad_norm": 0.07786931683535117, "learning_rate": 3.428407227209891e-06, "loss": 0.8634, "step": 9721 }, { "epoch": 4.403079710144928, "grad_norm": 0.07600821118164831, "learning_rate": 3.4232861264212657e-06, "loss": 0.8477, "step": 9722 }, { "epoch": 4.403532608695652, "grad_norm": 0.08350835476014323, "learning_rate": 3.4181686822646777e-06, "loss": 0.871, "step": 9723 }, { "epoch": 4.403985507246377, "grad_norm": 0.08203659603869178, "learning_rate": 3.413054895251744e-06, "loss": 0.8679, "step": 9724 }, { "epoch": 4.4044384057971016, "grad_norm": 0.07421868144348398, "learning_rate": 3.4079447658936825e-06, "loss": 0.8513, "step": 9725 }, { "epoch": 4.404891304347826, "grad_norm": 0.07810405273699784, "learning_rate": 3.402838294701374e-06, "loss": 0.8268, "step": 9726 }, { "epoch": 4.405344202898551, "grad_norm": 0.08086626369115028, "learning_rate": 3.397735482185316e-06, "loss": 0.8514, "step": 9727 }, { "epoch": 4.405797101449275, "grad_norm": 0.07855641511946639, "learning_rate": 3.392636328855643e-06, "loss": 0.8621, "step": 9728 }, { "epoch": 4.40625, "grad_norm": 0.07296788795724729, "learning_rate": 3.3875408352221164e-06, "loss": 0.8455, "step": 9729 }, { "epoch": 4.406702898550725, "grad_norm": 0.08343430583756589, "learning_rate": 3.3824490017941592e-06, "loss": 0.8551, "step": 9730 }, { "epoch": 4.407155797101449, "grad_norm": 0.0835343447522183, "learning_rate": 3.3773608290807914e-06, "loss": 0.8426, "step": 9731 }, { "epoch": 4.407608695652174, "grad_norm": 0.07003989623499483, "learning_rate": 3.3722763175907037e-06, "loss": 0.8577, "step": 9732 }, { "epoch": 4.4080615942028984, "grad_norm": 0.07998917986542209, "learning_rate": 3.3671954678321873e-06, "loss": 0.8465, "step": 9733 }, { "epoch": 4.408514492753623, "grad_norm": 0.08279128876308951, "learning_rate": 3.3621182803131957e-06, "loss": 0.841, "step": 9734 }, { "epoch": 4.408967391304348, "grad_norm": 0.10512721299378987, "learning_rate": 3.3570447555412964e-06, "loss": 0.8595, "step": 9735 }, { "epoch": 4.409420289855072, "grad_norm": 0.0910715218310315, "learning_rate": 3.3519748940237017e-06, "loss": 0.8536, "step": 9736 }, { "epoch": 4.409873188405797, "grad_norm": 0.07238696395950443, "learning_rate": 3.3469086962672413e-06, "loss": 0.8532, "step": 9737 }, { "epoch": 4.4103260869565215, "grad_norm": 0.07976737924822547, "learning_rate": 3.3418461627784083e-06, "loss": 0.823, "step": 9738 }, { "epoch": 4.410778985507246, "grad_norm": 0.07490587034554771, "learning_rate": 3.3367872940632952e-06, "loss": 0.8442, "step": 9739 }, { "epoch": 4.411231884057971, "grad_norm": 0.0758652267397819, "learning_rate": 3.331732090627666e-06, "loss": 0.855, "step": 9740 }, { "epoch": 4.411684782608695, "grad_norm": 0.07538563470335513, "learning_rate": 3.326680552976873e-06, "loss": 0.8665, "step": 9741 }, { "epoch": 4.41213768115942, "grad_norm": 0.09316781003336239, "learning_rate": 3.3216326816159517e-06, "loss": 0.8604, "step": 9742 }, { "epoch": 4.412590579710145, "grad_norm": 0.07344115267237764, "learning_rate": 3.316588477049529e-06, "loss": 0.8489, "step": 9743 }, { "epoch": 4.413043478260869, "grad_norm": 0.07990602900788041, "learning_rate": 3.311547939781887e-06, "loss": 0.8561, "step": 9744 }, { "epoch": 4.413496376811594, "grad_norm": 0.08658685581146235, "learning_rate": 3.3065110703169247e-06, "loss": 0.8813, "step": 9745 }, { "epoch": 4.413949275362318, "grad_norm": 0.07679436066800403, "learning_rate": 3.3014778691582073e-06, "loss": 0.8444, "step": 9746 }, { "epoch": 4.414402173913043, "grad_norm": 0.08377228480831189, "learning_rate": 3.2964483368088973e-06, "loss": 0.8529, "step": 9747 }, { "epoch": 4.4148550724637685, "grad_norm": 0.08862988638392083, "learning_rate": 3.2914224737718105e-06, "loss": 0.8565, "step": 9748 }, { "epoch": 4.415307971014493, "grad_norm": 0.08352675123774532, "learning_rate": 3.2864002805493845e-06, "loss": 0.8488, "step": 9749 }, { "epoch": 4.415760869565218, "grad_norm": 0.08234675907623001, "learning_rate": 3.2813817576437068e-06, "loss": 0.8559, "step": 9750 }, { "epoch": 4.416213768115942, "grad_norm": 0.07469461571835405, "learning_rate": 3.2763669055564784e-06, "loss": 0.8678, "step": 9751 }, { "epoch": 4.416666666666667, "grad_norm": 0.0779638543040613, "learning_rate": 3.2713557247890447e-06, "loss": 0.8238, "step": 9752 }, { "epoch": 4.417119565217392, "grad_norm": 0.07289600695712403, "learning_rate": 3.2663482158423743e-06, "loss": 0.8499, "step": 9753 }, { "epoch": 4.417572463768116, "grad_norm": 0.0790813381729796, "learning_rate": 3.261344379217093e-06, "loss": 0.8461, "step": 9754 }, { "epoch": 4.418025362318841, "grad_norm": 0.08896139677353299, "learning_rate": 3.256344215413418e-06, "loss": 0.8503, "step": 9755 }, { "epoch": 4.418478260869565, "grad_norm": 0.0715405644428383, "learning_rate": 3.2513477249312488e-06, "loss": 0.8535, "step": 9756 }, { "epoch": 4.41893115942029, "grad_norm": 0.07295320930537555, "learning_rate": 3.246354908270073e-06, "loss": 0.8236, "step": 9757 }, { "epoch": 4.419384057971015, "grad_norm": 0.0778346114859424, "learning_rate": 3.2413657659290433e-06, "loss": 0.8434, "step": 9758 }, { "epoch": 4.419836956521739, "grad_norm": 0.07300424899394037, "learning_rate": 3.2363802984069205e-06, "loss": 0.8524, "step": 9759 }, { "epoch": 4.420289855072464, "grad_norm": 0.0809214915368336, "learning_rate": 3.2313985062021236e-06, "loss": 0.8552, "step": 9760 }, { "epoch": 4.4207427536231885, "grad_norm": 0.08732970470005845, "learning_rate": 3.226420389812681e-06, "loss": 0.835, "step": 9761 }, { "epoch": 4.421195652173913, "grad_norm": 0.08438065319825408, "learning_rate": 3.221445949736266e-06, "loss": 0.8244, "step": 9762 }, { "epoch": 4.421648550724638, "grad_norm": 0.07311636293025402, "learning_rate": 3.2164751864701736e-06, "loss": 0.8536, "step": 9763 }, { "epoch": 4.422101449275362, "grad_norm": 0.0951590998748143, "learning_rate": 3.2115081005113536e-06, "loss": 0.8495, "step": 9764 }, { "epoch": 4.422554347826087, "grad_norm": 0.08743490114256897, "learning_rate": 3.206544692356359e-06, "loss": 0.8449, "step": 9765 }, { "epoch": 4.4230072463768115, "grad_norm": 0.07911545599388534, "learning_rate": 3.2015849625013985e-06, "loss": 0.8644, "step": 9766 }, { "epoch": 4.423460144927536, "grad_norm": 0.07396096143954703, "learning_rate": 3.1966289114422964e-06, "loss": 0.8753, "step": 9767 }, { "epoch": 4.423913043478261, "grad_norm": 0.08883518453259577, "learning_rate": 3.1916765396745288e-06, "loss": 0.8491, "step": 9768 }, { "epoch": 4.424365942028985, "grad_norm": 0.09337217746151205, "learning_rate": 3.1867278476931874e-06, "loss": 0.8662, "step": 9769 }, { "epoch": 4.42481884057971, "grad_norm": 0.07696396740702167, "learning_rate": 3.181782835992997e-06, "loss": 0.8361, "step": 9770 }, { "epoch": 4.425271739130435, "grad_norm": 0.08817440453944873, "learning_rate": 3.176841505068313e-06, "loss": 0.8452, "step": 9771 }, { "epoch": 4.425724637681159, "grad_norm": 0.09139058011386995, "learning_rate": 3.171903855413141e-06, "loss": 0.8604, "step": 9772 }, { "epoch": 4.426177536231884, "grad_norm": 0.0796786186811441, "learning_rate": 3.1669698875210943e-06, "loss": 0.8687, "step": 9773 }, { "epoch": 4.426630434782608, "grad_norm": 0.0783076046191799, "learning_rate": 3.1620396018854403e-06, "loss": 0.8411, "step": 9774 }, { "epoch": 4.427083333333333, "grad_norm": 0.0860512456456999, "learning_rate": 3.157112998999057e-06, "loss": 0.8603, "step": 9775 }, { "epoch": 4.427536231884058, "grad_norm": 0.07210745661551594, "learning_rate": 3.152190079354478e-06, "loss": 0.8593, "step": 9776 }, { "epoch": 4.427989130434782, "grad_norm": 0.07688861116248512, "learning_rate": 3.1472708434438437e-06, "loss": 0.8473, "step": 9777 }, { "epoch": 4.428442028985507, "grad_norm": 0.07116864691242909, "learning_rate": 3.1423552917589427e-06, "loss": 0.8327, "step": 9778 }, { "epoch": 4.4288949275362315, "grad_norm": 0.07042437915191918, "learning_rate": 3.1374434247911823e-06, "loss": 0.861, "step": 9779 }, { "epoch": 4.429347826086957, "grad_norm": 0.07066344485464508, "learning_rate": 3.1325352430316224e-06, "loss": 0.8473, "step": 9780 }, { "epoch": 4.429800724637682, "grad_norm": 0.077967562493872, "learning_rate": 3.1276307469709287e-06, "loss": 0.854, "step": 9781 }, { "epoch": 4.430253623188406, "grad_norm": 0.07888742581734266, "learning_rate": 3.1227299370994246e-06, "loss": 0.8613, "step": 9782 }, { "epoch": 4.430706521739131, "grad_norm": 0.07468350173691417, "learning_rate": 3.1178328139070422e-06, "loss": 0.8559, "step": 9783 }, { "epoch": 4.431159420289855, "grad_norm": 0.06900815967478398, "learning_rate": 3.1129393778833638e-06, "loss": 0.8591, "step": 9784 }, { "epoch": 4.43161231884058, "grad_norm": 0.0732278140809402, "learning_rate": 3.108049629517589e-06, "loss": 0.8456, "step": 9785 }, { "epoch": 4.432065217391305, "grad_norm": 0.07004216989345424, "learning_rate": 3.103163569298553e-06, "loss": 0.8578, "step": 9786 }, { "epoch": 4.432518115942029, "grad_norm": 0.076422055265984, "learning_rate": 3.098281197714714e-06, "loss": 0.8646, "step": 9787 }, { "epoch": 4.432971014492754, "grad_norm": 0.07378933206357721, "learning_rate": 3.0934025152541847e-06, "loss": 0.8495, "step": 9788 }, { "epoch": 4.4334239130434785, "grad_norm": 0.07813673119227771, "learning_rate": 3.088527522404685e-06, "loss": 0.8619, "step": 9789 }, { "epoch": 4.433876811594203, "grad_norm": 0.09176762097975516, "learning_rate": 3.08365621965359e-06, "loss": 0.8566, "step": 9790 }, { "epoch": 4.434329710144928, "grad_norm": 0.0733090539879139, "learning_rate": 3.07878860748787e-06, "loss": 0.8284, "step": 9791 }, { "epoch": 4.434782608695652, "grad_norm": 0.07402661922210366, "learning_rate": 3.0739246863941717e-06, "loss": 0.8379, "step": 9792 }, { "epoch": 4.435235507246377, "grad_norm": 0.07240133814382521, "learning_rate": 3.0690644568587324e-06, "loss": 0.8555, "step": 9793 }, { "epoch": 4.4356884057971016, "grad_norm": 0.07991633249227209, "learning_rate": 3.0642079193674436e-06, "loss": 0.8692, "step": 9794 }, { "epoch": 4.436141304347826, "grad_norm": 0.08672351255879482, "learning_rate": 3.05935507440581e-06, "loss": 0.8568, "step": 9795 }, { "epoch": 4.436594202898551, "grad_norm": 0.07837240663166141, "learning_rate": 3.0545059224589947e-06, "loss": 0.8569, "step": 9796 }, { "epoch": 4.437047101449275, "grad_norm": 0.07172940419075964, "learning_rate": 3.049660464011761e-06, "loss": 0.8396, "step": 9797 }, { "epoch": 4.4375, "grad_norm": 0.08270410883293827, "learning_rate": 3.0448186995485307e-06, "loss": 0.8425, "step": 9798 }, { "epoch": 4.437952898550725, "grad_norm": 0.08122567606789244, "learning_rate": 3.039980629553334e-06, "loss": 0.8345, "step": 9799 }, { "epoch": 4.438405797101449, "grad_norm": 0.07853297366762474, "learning_rate": 3.0351462545098374e-06, "loss": 0.8571, "step": 9800 }, { "epoch": 4.438858695652174, "grad_norm": 0.07609349555165608, "learning_rate": 3.0303155749013523e-06, "loss": 0.848, "step": 9801 }, { "epoch": 4.4393115942028984, "grad_norm": 0.07138321736147733, "learning_rate": 3.0254885912107946e-06, "loss": 0.8481, "step": 9802 }, { "epoch": 4.439764492753623, "grad_norm": 0.07813951137965185, "learning_rate": 3.0206653039207424e-06, "loss": 0.8418, "step": 9803 }, { "epoch": 4.440217391304348, "grad_norm": 0.07350712393974056, "learning_rate": 3.0158457135133745e-06, "loss": 0.8495, "step": 9804 }, { "epoch": 4.440670289855072, "grad_norm": 0.07095027772235292, "learning_rate": 3.011029820470519e-06, "loss": 0.8575, "step": 9805 }, { "epoch": 4.441123188405797, "grad_norm": 0.08105546882857839, "learning_rate": 3.006217625273622e-06, "loss": 0.8574, "step": 9806 }, { "epoch": 4.4415760869565215, "grad_norm": 0.07545229575682232, "learning_rate": 3.0014091284037737e-06, "loss": 0.8396, "step": 9807 }, { "epoch": 4.442028985507246, "grad_norm": 0.07843736675114638, "learning_rate": 2.996604330341679e-06, "loss": 0.8509, "step": 9808 }, { "epoch": 4.442481884057971, "grad_norm": 0.07093676822146922, "learning_rate": 2.9918032315676957e-06, "loss": 0.8375, "step": 9809 }, { "epoch": 4.442934782608695, "grad_norm": 0.08340716852835929, "learning_rate": 2.987005832561778e-06, "loss": 0.8492, "step": 9810 }, { "epoch": 4.44338768115942, "grad_norm": 0.08643551749917336, "learning_rate": 2.9822121338035457e-06, "loss": 0.8302, "step": 9811 }, { "epoch": 4.443840579710145, "grad_norm": 0.07542660132954768, "learning_rate": 2.9774221357722298e-06, "loss": 0.8438, "step": 9812 }, { "epoch": 4.444293478260869, "grad_norm": 0.07729638166056112, "learning_rate": 2.972635838946687e-06, "loss": 0.8365, "step": 9813 }, { "epoch": 4.444746376811594, "grad_norm": 0.08512854060417918, "learning_rate": 2.9678532438054096e-06, "loss": 0.8376, "step": 9814 }, { "epoch": 4.445199275362318, "grad_norm": 0.07831462270722227, "learning_rate": 2.9630743508265313e-06, "loss": 0.8504, "step": 9815 }, { "epoch": 4.445652173913043, "grad_norm": 0.0799300656751256, "learning_rate": 2.95829916048779e-06, "loss": 0.8423, "step": 9816 }, { "epoch": 4.4461050724637685, "grad_norm": 0.0776464076271682, "learning_rate": 2.9535276732665897e-06, "loss": 0.8497, "step": 9817 }, { "epoch": 4.446557971014493, "grad_norm": 0.07819059808614816, "learning_rate": 2.948759889639927e-06, "loss": 0.8606, "step": 9818 }, { "epoch": 4.447010869565218, "grad_norm": 0.07218611886829092, "learning_rate": 2.9439958100844567e-06, "loss": 0.8619, "step": 9819 }, { "epoch": 4.447463768115942, "grad_norm": 0.09152415889774806, "learning_rate": 2.939235435076446e-06, "loss": 0.8629, "step": 9820 }, { "epoch": 4.447916666666667, "grad_norm": 0.08229277635176105, "learning_rate": 2.934478765091795e-06, "loss": 0.8562, "step": 9821 }, { "epoch": 4.448369565217392, "grad_norm": 0.08499479647180828, "learning_rate": 2.9297258006060335e-06, "loss": 0.8493, "step": 9822 }, { "epoch": 4.448822463768116, "grad_norm": 0.08292019598619635, "learning_rate": 2.9249765420943287e-06, "loss": 0.8666, "step": 9823 }, { "epoch": 4.449275362318841, "grad_norm": 0.07670414475595995, "learning_rate": 2.920230990031465e-06, "loss": 0.8259, "step": 9824 }, { "epoch": 4.449728260869565, "grad_norm": 0.07686670496657837, "learning_rate": 2.9154891448918764e-06, "loss": 0.8629, "step": 9825 }, { "epoch": 4.45018115942029, "grad_norm": 0.0782197565036777, "learning_rate": 2.9107510071495928e-06, "loss": 0.8509, "step": 9826 }, { "epoch": 4.450634057971015, "grad_norm": 0.07496156892277174, "learning_rate": 2.9060165772783144e-06, "loss": 0.865, "step": 9827 }, { "epoch": 4.451086956521739, "grad_norm": 0.08023354578823436, "learning_rate": 2.9012858557513344e-06, "loss": 0.8402, "step": 9828 }, { "epoch": 4.451539855072464, "grad_norm": 0.0769493115541479, "learning_rate": 2.8965588430415993e-06, "loss": 0.8496, "step": 9829 }, { "epoch": 4.4519927536231885, "grad_norm": 0.0795152066521359, "learning_rate": 2.8918355396216636e-06, "loss": 0.851, "step": 9830 }, { "epoch": 4.452445652173913, "grad_norm": 0.08726119339759321, "learning_rate": 2.887115945963741e-06, "loss": 0.8457, "step": 9831 }, { "epoch": 4.452898550724638, "grad_norm": 0.0765381514768441, "learning_rate": 2.882400062539641e-06, "loss": 0.8586, "step": 9832 }, { "epoch": 4.453351449275362, "grad_norm": 0.0721735844263542, "learning_rate": 2.877687889820826e-06, "loss": 0.8403, "step": 9833 }, { "epoch": 4.453804347826087, "grad_norm": 0.0849350327021626, "learning_rate": 2.8729794282783776e-06, "loss": 0.8549, "step": 9834 }, { "epoch": 4.4542572463768115, "grad_norm": 0.08601990358222163, "learning_rate": 2.8682746783830118e-06, "loss": 0.8399, "step": 9835 }, { "epoch": 4.454710144927536, "grad_norm": 0.07610608837150898, "learning_rate": 2.8635736406050683e-06, "loss": 0.8535, "step": 9836 }, { "epoch": 4.455163043478261, "grad_norm": 0.07655136373861361, "learning_rate": 2.858876315414514e-06, "loss": 0.8452, "step": 9837 }, { "epoch": 4.455615942028985, "grad_norm": 0.08083778147362584, "learning_rate": 2.8541827032809453e-06, "loss": 0.8654, "step": 9838 }, { "epoch": 4.45606884057971, "grad_norm": 0.08652621506340881, "learning_rate": 2.849492804673597e-06, "loss": 0.8622, "step": 9839 }, { "epoch": 4.456521739130435, "grad_norm": 0.08718072890582361, "learning_rate": 2.84480662006132e-06, "loss": 0.849, "step": 9840 }, { "epoch": 4.456974637681159, "grad_norm": 0.07313436077293795, "learning_rate": 2.8401241499126064e-06, "loss": 0.85, "step": 9841 }, { "epoch": 4.457427536231884, "grad_norm": 0.07225688380659676, "learning_rate": 2.835445394695562e-06, "loss": 0.8511, "step": 9842 }, { "epoch": 4.457880434782608, "grad_norm": 0.07008865185436361, "learning_rate": 2.8307703548779406e-06, "loss": 0.8563, "step": 9843 }, { "epoch": 4.458333333333333, "grad_norm": 0.08265678696401643, "learning_rate": 2.8260990309270987e-06, "loss": 0.8541, "step": 9844 }, { "epoch": 4.458786231884058, "grad_norm": 0.07657985201657032, "learning_rate": 2.8214314233100526e-06, "loss": 0.8291, "step": 9845 }, { "epoch": 4.459239130434782, "grad_norm": 0.07537172030788528, "learning_rate": 2.8167675324934207e-06, "loss": 0.8664, "step": 9846 }, { "epoch": 4.459692028985507, "grad_norm": 0.0772948390299464, "learning_rate": 2.8121073589434566e-06, "loss": 0.8391, "step": 9847 }, { "epoch": 4.4601449275362315, "grad_norm": 0.07420856839952085, "learning_rate": 2.8074509031260454e-06, "loss": 0.8486, "step": 9848 }, { "epoch": 4.460597826086957, "grad_norm": 0.08497681427689184, "learning_rate": 2.8027981655067126e-06, "loss": 0.8492, "step": 9849 }, { "epoch": 4.461050724637682, "grad_norm": 0.07518532605858619, "learning_rate": 2.798149146550584e-06, "loss": 0.8517, "step": 9850 }, { "epoch": 4.461503623188406, "grad_norm": 0.07376579876729944, "learning_rate": 2.793503846722443e-06, "loss": 0.8424, "step": 9851 }, { "epoch": 4.461956521739131, "grad_norm": 0.07476910555155467, "learning_rate": 2.7888622664866736e-06, "loss": 0.8627, "step": 9852 }, { "epoch": 4.462409420289855, "grad_norm": 0.08456972611198914, "learning_rate": 2.7842244063073143e-06, "loss": 0.8352, "step": 9853 }, { "epoch": 4.46286231884058, "grad_norm": 0.08035742017710402, "learning_rate": 2.779590266648016e-06, "loss": 0.8286, "step": 9854 }, { "epoch": 4.463315217391305, "grad_norm": 0.07501336742167515, "learning_rate": 2.7749598479720608e-06, "loss": 0.8543, "step": 9855 }, { "epoch": 4.463768115942029, "grad_norm": 0.0810756497729412, "learning_rate": 2.7703331507423504e-06, "loss": 0.8479, "step": 9856 }, { "epoch": 4.464221014492754, "grad_norm": 0.08908458317197668, "learning_rate": 2.7657101754214343e-06, "loss": 0.8344, "step": 9857 }, { "epoch": 4.4646739130434785, "grad_norm": 0.07707205921250475, "learning_rate": 2.761090922471472e-06, "loss": 0.8616, "step": 9858 }, { "epoch": 4.465126811594203, "grad_norm": 0.08694137715018405, "learning_rate": 2.7564753923542676e-06, "loss": 0.8534, "step": 9859 }, { "epoch": 4.465579710144928, "grad_norm": 0.07869557895608313, "learning_rate": 2.751863585531225e-06, "loss": 0.8528, "step": 9860 }, { "epoch": 4.466032608695652, "grad_norm": 0.07440464380356746, "learning_rate": 2.747255502463415e-06, "loss": 0.8422, "step": 9861 }, { "epoch": 4.466485507246377, "grad_norm": 0.07462911226979108, "learning_rate": 2.742651143611501e-06, "loss": 0.8441, "step": 9862 }, { "epoch": 4.4669384057971016, "grad_norm": 0.08140386282747676, "learning_rate": 2.7380505094357947e-06, "loss": 0.8359, "step": 9863 }, { "epoch": 4.467391304347826, "grad_norm": 0.08012517907446098, "learning_rate": 2.7334536003962166e-06, "loss": 0.8546, "step": 9864 }, { "epoch": 4.467844202898551, "grad_norm": 0.08657987485266905, "learning_rate": 2.7288604169523456e-06, "loss": 0.8283, "step": 9865 }, { "epoch": 4.468297101449275, "grad_norm": 0.07685290876430996, "learning_rate": 2.7242709595633486e-06, "loss": 0.8582, "step": 9866 }, { "epoch": 4.46875, "grad_norm": 0.07982725287604751, "learning_rate": 2.7196852286880624e-06, "loss": 0.8587, "step": 9867 }, { "epoch": 4.469202898550725, "grad_norm": 0.07114220209090438, "learning_rate": 2.715103224784912e-06, "loss": 0.8516, "step": 9868 }, { "epoch": 4.469655797101449, "grad_norm": 0.06995160626874698, "learning_rate": 2.7105249483119834e-06, "loss": 0.842, "step": 9869 }, { "epoch": 4.470108695652174, "grad_norm": 0.07876844283164026, "learning_rate": 2.705950399726964e-06, "loss": 0.8322, "step": 9870 }, { "epoch": 4.4705615942028984, "grad_norm": 0.0711496938607041, "learning_rate": 2.7013795794871776e-06, "loss": 0.8515, "step": 9871 }, { "epoch": 4.471014492753623, "grad_norm": 0.07657072270680028, "learning_rate": 2.6968124880495783e-06, "loss": 0.847, "step": 9872 }, { "epoch": 4.471467391304348, "grad_norm": 0.06783561805977897, "learning_rate": 2.6922491258707516e-06, "loss": 0.857, "step": 9873 }, { "epoch": 4.471920289855072, "grad_norm": 0.07807725425915121, "learning_rate": 2.687689493406889e-06, "loss": 0.847, "step": 9874 }, { "epoch": 4.472373188405797, "grad_norm": 0.07137145375228662, "learning_rate": 2.683133591113847e-06, "loss": 0.8298, "step": 9875 }, { "epoch": 4.4728260869565215, "grad_norm": 0.08289930048481599, "learning_rate": 2.6785814194470616e-06, "loss": 0.8614, "step": 9876 }, { "epoch": 4.473278985507246, "grad_norm": 0.07313626017913166, "learning_rate": 2.67403297886164e-06, "loss": 0.849, "step": 9877 }, { "epoch": 4.473731884057971, "grad_norm": 0.0701427208222515, "learning_rate": 2.6694882698122903e-06, "loss": 0.8633, "step": 9878 }, { "epoch": 4.474184782608695, "grad_norm": 0.07785631602795988, "learning_rate": 2.664947292753355e-06, "loss": 0.8661, "step": 9879 }, { "epoch": 4.47463768115942, "grad_norm": 0.07601166946627697, "learning_rate": 2.6604100481387952e-06, "loss": 0.8598, "step": 9880 }, { "epoch": 4.475090579710145, "grad_norm": 0.07888322740283414, "learning_rate": 2.6558765364222174e-06, "loss": 0.8502, "step": 9881 }, { "epoch": 4.475543478260869, "grad_norm": 0.07985073684017323, "learning_rate": 2.651346758056832e-06, "loss": 0.8551, "step": 9882 }, { "epoch": 4.475996376811594, "grad_norm": 0.07952267208993345, "learning_rate": 2.6468207134955036e-06, "loss": 0.848, "step": 9883 }, { "epoch": 4.476449275362318, "grad_norm": 0.07799117251567682, "learning_rate": 2.6422984031906886e-06, "loss": 0.8682, "step": 9884 }, { "epoch": 4.476902173913043, "grad_norm": 0.08614836591803664, "learning_rate": 2.637779827594513e-06, "loss": 0.8429, "step": 9885 }, { "epoch": 4.4773550724637685, "grad_norm": 0.08404532649741905, "learning_rate": 2.6332649871586836e-06, "loss": 0.8311, "step": 9886 }, { "epoch": 4.477807971014493, "grad_norm": 0.07638161956902602, "learning_rate": 2.6287538823345805e-06, "loss": 0.8681, "step": 9887 }, { "epoch": 4.478260869565218, "grad_norm": 0.08431123684596337, "learning_rate": 2.6242465135731587e-06, "loss": 0.8465, "step": 9888 }, { "epoch": 4.478713768115942, "grad_norm": 0.07791178964575984, "learning_rate": 2.6197428813250448e-06, "loss": 0.8345, "step": 9889 }, { "epoch": 4.479166666666667, "grad_norm": 0.09010004857961765, "learning_rate": 2.6152429860404647e-06, "loss": 0.8477, "step": 9890 }, { "epoch": 4.479619565217392, "grad_norm": 0.08008601008948317, "learning_rate": 2.6107468281692904e-06, "loss": 0.8532, "step": 9891 }, { "epoch": 4.480072463768116, "grad_norm": 0.07356479233674705, "learning_rate": 2.606254408160993e-06, "loss": 0.8474, "step": 9892 }, { "epoch": 4.480525362318841, "grad_norm": 0.08328905710601579, "learning_rate": 2.6017657264647113e-06, "loss": 0.8596, "step": 9893 }, { "epoch": 4.480978260869565, "grad_norm": 0.078577243423494, "learning_rate": 2.5972807835291613e-06, "loss": 0.8524, "step": 9894 }, { "epoch": 4.48143115942029, "grad_norm": 0.086408255320065, "learning_rate": 2.5927995798027317e-06, "loss": 0.8374, "step": 9895 }, { "epoch": 4.481884057971015, "grad_norm": 0.06576507408790572, "learning_rate": 2.5883221157334015e-06, "loss": 0.8447, "step": 9896 }, { "epoch": 4.482336956521739, "grad_norm": 0.0786497031896817, "learning_rate": 2.583848391768795e-06, "loss": 0.8527, "step": 9897 }, { "epoch": 4.482789855072464, "grad_norm": 0.09359508746537447, "learning_rate": 2.57937840835615e-06, "loss": 0.8499, "step": 9898 }, { "epoch": 4.4832427536231885, "grad_norm": 0.08380404955606321, "learning_rate": 2.574912165942345e-06, "loss": 0.8555, "step": 9899 }, { "epoch": 4.483695652173913, "grad_norm": 0.07872066461406882, "learning_rate": 2.5704496649738754e-06, "loss": 0.8477, "step": 9900 }, { "epoch": 4.484148550724638, "grad_norm": 0.08368559747799836, "learning_rate": 2.565990905896869e-06, "loss": 0.8505, "step": 9901 }, { "epoch": 4.484601449275362, "grad_norm": 0.0748267994021611, "learning_rate": 2.561535889157063e-06, "loss": 0.8264, "step": 9902 }, { "epoch": 4.485054347826087, "grad_norm": 0.08591277647941852, "learning_rate": 2.5570846151998517e-06, "loss": 0.849, "step": 9903 }, { "epoch": 4.4855072463768115, "grad_norm": 0.08264600021426342, "learning_rate": 2.552637084470222e-06, "loss": 0.8591, "step": 9904 }, { "epoch": 4.485960144927536, "grad_norm": 0.09251350319784021, "learning_rate": 2.5481932974128e-06, "loss": 0.8537, "step": 9905 }, { "epoch": 4.486413043478261, "grad_norm": 0.07900789695618687, "learning_rate": 2.543753254471839e-06, "loss": 0.8459, "step": 9906 }, { "epoch": 4.486865942028985, "grad_norm": 0.10495759327499568, "learning_rate": 2.5393169560912247e-06, "loss": 0.8458, "step": 9907 }, { "epoch": 4.48731884057971, "grad_norm": 0.07844065049617811, "learning_rate": 2.5348844027144503e-06, "loss": 0.8395, "step": 9908 }, { "epoch": 4.487771739130435, "grad_norm": 0.08162062317171218, "learning_rate": 2.53045559478466e-06, "loss": 0.8512, "step": 9909 }, { "epoch": 4.488224637681159, "grad_norm": 0.08282920275588046, "learning_rate": 2.526030532744588e-06, "loss": 0.8551, "step": 9910 }, { "epoch": 4.488677536231884, "grad_norm": 0.07767711811223334, "learning_rate": 2.521609217036636e-06, "loss": 0.8516, "step": 9911 }, { "epoch": 4.489130434782608, "grad_norm": 0.07560246169585894, "learning_rate": 2.5171916481027968e-06, "loss": 0.8701, "step": 9912 }, { "epoch": 4.489583333333333, "grad_norm": 0.06959851030216893, "learning_rate": 2.512777826384709e-06, "loss": 0.8419, "step": 9913 }, { "epoch": 4.490036231884058, "grad_norm": 0.07606706590967784, "learning_rate": 2.5083677523236194e-06, "loss": 0.8418, "step": 9914 }, { "epoch": 4.490489130434782, "grad_norm": 0.07736909129844406, "learning_rate": 2.5039614263604194e-06, "loss": 0.8306, "step": 9915 }, { "epoch": 4.490942028985507, "grad_norm": 0.07264956058297725, "learning_rate": 2.499558848935606e-06, "loss": 0.8645, "step": 9916 }, { "epoch": 4.4913949275362315, "grad_norm": 0.07357827115771802, "learning_rate": 2.495160020489329e-06, "loss": 0.8531, "step": 9917 }, { "epoch": 4.491847826086957, "grad_norm": 0.06949599494369503, "learning_rate": 2.4907649414613256e-06, "loss": 0.8646, "step": 9918 }, { "epoch": 4.492300724637682, "grad_norm": 0.07045777062274042, "learning_rate": 2.486373612291e-06, "loss": 0.8493, "step": 9919 }, { "epoch": 4.492753623188406, "grad_norm": 0.08371120827549768, "learning_rate": 2.4819860334173473e-06, "loss": 0.8452, "step": 9920 }, { "epoch": 4.493206521739131, "grad_norm": 0.06604614976065687, "learning_rate": 2.477602205279004e-06, "loss": 0.8519, "step": 9921 }, { "epoch": 4.493659420289855, "grad_norm": 0.07751019202831884, "learning_rate": 2.473222128314223e-06, "loss": 0.8459, "step": 9922 }, { "epoch": 4.49411231884058, "grad_norm": 0.07269026069584544, "learning_rate": 2.4688458029608952e-06, "loss": 0.8495, "step": 9923 }, { "epoch": 4.494565217391305, "grad_norm": 0.07514413033985023, "learning_rate": 2.4644732296565277e-06, "loss": 0.8646, "step": 9924 }, { "epoch": 4.495018115942029, "grad_norm": 0.07619879629903138, "learning_rate": 2.4601044088382465e-06, "loss": 0.8386, "step": 9925 }, { "epoch": 4.495471014492754, "grad_norm": 0.08012967566249413, "learning_rate": 2.4557393409428176e-06, "loss": 0.8322, "step": 9926 }, { "epoch": 4.4959239130434785, "grad_norm": 0.07866142231498696, "learning_rate": 2.4513780264066165e-06, "loss": 0.8533, "step": 9927 }, { "epoch": 4.496376811594203, "grad_norm": 0.09135420676114722, "learning_rate": 2.4470204656656638e-06, "loss": 0.8601, "step": 9928 }, { "epoch": 4.496829710144928, "grad_norm": 0.071112350028167, "learning_rate": 2.4426666591555794e-06, "loss": 0.8449, "step": 9929 }, { "epoch": 4.497282608695652, "grad_norm": 0.07673905835709492, "learning_rate": 2.4383166073116236e-06, "loss": 0.8536, "step": 9930 }, { "epoch": 4.497735507246377, "grad_norm": 0.08154217859866514, "learning_rate": 2.4339703105686717e-06, "loss": 0.8536, "step": 9931 }, { "epoch": 4.4981884057971016, "grad_norm": 0.07380594989164596, "learning_rate": 2.4296277693612424e-06, "loss": 0.8468, "step": 9932 }, { "epoch": 4.498641304347826, "grad_norm": 0.06694320790577218, "learning_rate": 2.4252889841234594e-06, "loss": 0.8539, "step": 9933 }, { "epoch": 4.499094202898551, "grad_norm": 0.0851438219849406, "learning_rate": 2.420953955289078e-06, "loss": 0.844, "step": 9934 }, { "epoch": 4.499547101449275, "grad_norm": 0.07740074514540515, "learning_rate": 2.4166226832914764e-06, "loss": 0.841, "step": 9935 }, { "epoch": 4.5, "grad_norm": 0.07534165009247014, "learning_rate": 2.4122951685636674e-06, "loss": 0.8487, "step": 9936 }, { "epoch": 4.500452898550725, "grad_norm": 0.07162302393973102, "learning_rate": 2.407971411538266e-06, "loss": 0.8554, "step": 9937 }, { "epoch": 4.500905797101449, "grad_norm": 0.07616392431957647, "learning_rate": 2.4036514126475384e-06, "loss": 0.866, "step": 9938 }, { "epoch": 4.501358695652174, "grad_norm": 0.07264621356297428, "learning_rate": 2.3993351723233584e-06, "loss": 0.8299, "step": 9939 }, { "epoch": 4.5018115942028984, "grad_norm": 0.07860484532022947, "learning_rate": 2.39502269099722e-06, "loss": 0.8516, "step": 9940 }, { "epoch": 4.502264492753623, "grad_norm": 0.08430231679580351, "learning_rate": 2.390713969100249e-06, "loss": 0.8375, "step": 9941 }, { "epoch": 4.502717391304348, "grad_norm": 0.07134286529226, "learning_rate": 2.386409007063204e-06, "loss": 0.8519, "step": 9942 }, { "epoch": 4.503170289855072, "grad_norm": 0.07365643850555992, "learning_rate": 2.3821078053164514e-06, "loss": 0.8649, "step": 9943 }, { "epoch": 4.503623188405797, "grad_norm": 0.08672609316189978, "learning_rate": 2.3778103642899985e-06, "loss": 0.8546, "step": 9944 }, { "epoch": 4.5040760869565215, "grad_norm": 0.073904788530189, "learning_rate": 2.3735166844134526e-06, "loss": 0.8379, "step": 9945 }, { "epoch": 4.504528985507246, "grad_norm": 0.07597369135613058, "learning_rate": 2.369226766116075e-06, "loss": 0.8643, "step": 9946 }, { "epoch": 4.504981884057971, "grad_norm": 0.0782194977252165, "learning_rate": 2.3649406098267315e-06, "loss": 0.8522, "step": 9947 }, { "epoch": 4.505434782608695, "grad_norm": 0.07990717140345337, "learning_rate": 2.36065821597391e-06, "loss": 0.848, "step": 9948 }, { "epoch": 4.50588768115942, "grad_norm": 0.08379365817861988, "learning_rate": 2.356379584985722e-06, "loss": 0.8642, "step": 9949 }, { "epoch": 4.506340579710145, "grad_norm": 0.08242534985959327, "learning_rate": 2.3521047172899313e-06, "loss": 0.8611, "step": 9950 }, { "epoch": 4.506793478260869, "grad_norm": 0.08144543263943205, "learning_rate": 2.3478336133138812e-06, "loss": 0.8591, "step": 9951 }, { "epoch": 4.507246376811594, "grad_norm": 0.08040022312462027, "learning_rate": 2.343566273484572e-06, "loss": 0.8384, "step": 9952 }, { "epoch": 4.507699275362318, "grad_norm": 0.07620880351053957, "learning_rate": 2.339302698228614e-06, "loss": 0.845, "step": 9953 }, { "epoch": 4.508152173913043, "grad_norm": 0.08396932468512891, "learning_rate": 2.3350428879722476e-06, "loss": 0.8274, "step": 9954 }, { "epoch": 4.508605072463768, "grad_norm": 0.0800469721469858, "learning_rate": 2.3307868431413283e-06, "loss": 0.8389, "step": 9955 }, { "epoch": 4.509057971014493, "grad_norm": 0.07188171203300188, "learning_rate": 2.326534564161338e-06, "loss": 0.8565, "step": 9956 }, { "epoch": 4.509510869565218, "grad_norm": 0.08427631498854737, "learning_rate": 2.322286051457385e-06, "loss": 0.8384, "step": 9957 }, { "epoch": 4.509963768115942, "grad_norm": 0.08224258287218138, "learning_rate": 2.318041305454202e-06, "loss": 0.8625, "step": 9958 }, { "epoch": 4.510416666666667, "grad_norm": 0.07448385087902383, "learning_rate": 2.313800326576141e-06, "loss": 0.8434, "step": 9959 }, { "epoch": 4.510869565217392, "grad_norm": 0.07365521074867745, "learning_rate": 2.309563115247184e-06, "loss": 0.8546, "step": 9960 }, { "epoch": 4.511322463768116, "grad_norm": 0.08271556803183522, "learning_rate": 2.3053296718909215e-06, "loss": 0.8516, "step": 9961 }, { "epoch": 4.511775362318841, "grad_norm": 0.08562649099760637, "learning_rate": 2.3010999969305913e-06, "loss": 0.8414, "step": 9962 }, { "epoch": 4.512228260869565, "grad_norm": 0.07215559774707207, "learning_rate": 2.2968740907890296e-06, "loss": 0.8505, "step": 9963 }, { "epoch": 4.51268115942029, "grad_norm": 0.06866154361696379, "learning_rate": 2.292651953888716e-06, "loss": 0.8439, "step": 9964 }, { "epoch": 4.513134057971015, "grad_norm": 0.07632336649940666, "learning_rate": 2.2884335866517305e-06, "loss": 0.8437, "step": 9965 }, { "epoch": 4.513586956521739, "grad_norm": 0.08150789337724983, "learning_rate": 2.284218989499802e-06, "loss": 0.846, "step": 9966 }, { "epoch": 4.514039855072464, "grad_norm": 0.0735033502336729, "learning_rate": 2.2800081628542657e-06, "loss": 0.864, "step": 9967 }, { "epoch": 4.5144927536231885, "grad_norm": 0.070461209643792, "learning_rate": 2.275801107136086e-06, "loss": 0.8489, "step": 9968 }, { "epoch": 4.514945652173913, "grad_norm": 0.07187876838070897, "learning_rate": 2.2715978227658473e-06, "loss": 0.8527, "step": 9969 }, { "epoch": 4.515398550724638, "grad_norm": 0.0736419752659102, "learning_rate": 2.267398310163764e-06, "loss": 0.8596, "step": 9970 }, { "epoch": 4.515851449275362, "grad_norm": 0.0729306664135104, "learning_rate": 2.263202569749665e-06, "loss": 0.8565, "step": 9971 }, { "epoch": 4.516304347826087, "grad_norm": 0.08373557606335767, "learning_rate": 2.2590106019430057e-06, "loss": 0.8651, "step": 9972 }, { "epoch": 4.5167572463768115, "grad_norm": 0.0809138262401395, "learning_rate": 2.2548224071628557e-06, "loss": 0.8531, "step": 9973 }, { "epoch": 4.517210144927536, "grad_norm": 0.07975431622542785, "learning_rate": 2.2506379858279283e-06, "loss": 0.8408, "step": 9974 }, { "epoch": 4.517663043478261, "grad_norm": 0.07466837553519697, "learning_rate": 2.246457338356529e-06, "loss": 0.8617, "step": 9975 }, { "epoch": 4.518115942028985, "grad_norm": 0.07351068953930522, "learning_rate": 2.2422804651666265e-06, "loss": 0.8507, "step": 9976 }, { "epoch": 4.51856884057971, "grad_norm": 0.08052753057376037, "learning_rate": 2.23810736667577e-06, "loss": 0.8436, "step": 9977 }, { "epoch": 4.519021739130435, "grad_norm": 0.07371563208251679, "learning_rate": 2.2339380433011693e-06, "loss": 0.8453, "step": 9978 }, { "epoch": 4.519474637681159, "grad_norm": 0.07351457741015674, "learning_rate": 2.229772495459619e-06, "loss": 0.8531, "step": 9979 }, { "epoch": 4.519927536231884, "grad_norm": 0.07217570848900161, "learning_rate": 2.2256107235675684e-06, "loss": 0.843, "step": 9980 }, { "epoch": 4.520380434782608, "grad_norm": 0.07773336035523465, "learning_rate": 2.2214527280410712e-06, "loss": 0.8586, "step": 9981 }, { "epoch": 4.520833333333333, "grad_norm": 0.07139193681957015, "learning_rate": 2.217298509295813e-06, "loss": 0.8566, "step": 9982 }, { "epoch": 4.521286231884058, "grad_norm": 0.07102247110255415, "learning_rate": 2.2131480677470886e-06, "loss": 0.8597, "step": 9983 }, { "epoch": 4.521739130434782, "grad_norm": 0.07242240148281315, "learning_rate": 2.2090014038098364e-06, "loss": 0.8521, "step": 9984 }, { "epoch": 4.522192028985507, "grad_norm": 0.0668200173911473, "learning_rate": 2.204858517898596e-06, "loss": 0.8394, "step": 9985 }, { "epoch": 4.522644927536232, "grad_norm": 0.07147497353672753, "learning_rate": 2.200719410427543e-06, "loss": 0.8473, "step": 9986 }, { "epoch": 4.523097826086957, "grad_norm": 0.08323334114819826, "learning_rate": 2.1965840818104623e-06, "loss": 0.8339, "step": 9987 }, { "epoch": 4.523550724637682, "grad_norm": 0.0913713113344302, "learning_rate": 2.1924525324607825e-06, "loss": 0.8571, "step": 9988 }, { "epoch": 4.524003623188406, "grad_norm": 0.07149669887299558, "learning_rate": 2.1883247627915338e-06, "loss": 0.8572, "step": 9989 }, { "epoch": 4.524456521739131, "grad_norm": 0.08172153032203426, "learning_rate": 2.184200773215377e-06, "loss": 0.8661, "step": 9990 }, { "epoch": 4.524909420289855, "grad_norm": 0.091253681012271, "learning_rate": 2.1800805641445867e-06, "loss": 0.8644, "step": 9991 }, { "epoch": 4.52536231884058, "grad_norm": 0.08923463302619097, "learning_rate": 2.1759641359910777e-06, "loss": 0.8614, "step": 9992 }, { "epoch": 4.525815217391305, "grad_norm": 0.07589188524715586, "learning_rate": 2.17185148916637e-06, "loss": 0.8545, "step": 9993 }, { "epoch": 4.526268115942029, "grad_norm": 0.07446657155866457, "learning_rate": 2.1677426240816146e-06, "loss": 0.8603, "step": 9994 }, { "epoch": 4.526721014492754, "grad_norm": 0.07965498160737872, "learning_rate": 2.1636375411475765e-06, "loss": 0.8414, "step": 9995 }, { "epoch": 4.5271739130434785, "grad_norm": 0.07940126056786137, "learning_rate": 2.159536240774651e-06, "loss": 0.8549, "step": 9996 }, { "epoch": 4.527626811594203, "grad_norm": 0.09787758604676251, "learning_rate": 2.155438723372858e-06, "loss": 0.8604, "step": 9997 }, { "epoch": 4.528079710144928, "grad_norm": 0.07829557256878936, "learning_rate": 2.1513449893518203e-06, "loss": 0.8592, "step": 9998 }, { "epoch": 4.528532608695652, "grad_norm": 0.07828008527969746, "learning_rate": 2.147255039120797e-06, "loss": 0.8465, "step": 9999 }, { "epoch": 4.528985507246377, "grad_norm": 0.0788309182934986, "learning_rate": 2.1431688730886747e-06, "loss": 0.8556, "step": 10000 }, { "epoch": 4.5294384057971016, "grad_norm": 0.07634497067980138, "learning_rate": 2.139086491663944e-06, "loss": 0.8553, "step": 10001 }, { "epoch": 4.529891304347826, "grad_norm": 0.07810959024326056, "learning_rate": 2.135007895254737e-06, "loss": 0.8463, "step": 10002 }, { "epoch": 4.530344202898551, "grad_norm": 0.07713621666077429, "learning_rate": 2.1309330842687893e-06, "loss": 0.8554, "step": 10003 }, { "epoch": 4.530797101449275, "grad_norm": 0.07230069474000735, "learning_rate": 2.126862059113477e-06, "loss": 0.8731, "step": 10004 }, { "epoch": 4.53125, "grad_norm": 0.06783686782939194, "learning_rate": 2.122794820195777e-06, "loss": 0.8396, "step": 10005 }, { "epoch": 4.531702898550725, "grad_norm": 0.07391718674969161, "learning_rate": 2.118731367922302e-06, "loss": 0.8508, "step": 10006 }, { "epoch": 4.532155797101449, "grad_norm": 0.08595712210016099, "learning_rate": 2.1146717026992735e-06, "loss": 0.8588, "step": 10007 }, { "epoch": 4.532608695652174, "grad_norm": 0.07447378140857913, "learning_rate": 2.1106158249325536e-06, "loss": 0.8569, "step": 10008 }, { "epoch": 4.5330615942028984, "grad_norm": 0.07815477864122246, "learning_rate": 2.1065637350276045e-06, "loss": 0.8614, "step": 10009 }, { "epoch": 4.533514492753623, "grad_norm": 0.08331897935496503, "learning_rate": 2.1025154333895337e-06, "loss": 0.8528, "step": 10010 }, { "epoch": 4.533967391304348, "grad_norm": 0.06934549962208486, "learning_rate": 2.0984709204230437e-06, "loss": 0.8466, "step": 10011 }, { "epoch": 4.534420289855072, "grad_norm": 0.07303736271751145, "learning_rate": 2.094430196532482e-06, "loss": 0.8479, "step": 10012 }, { "epoch": 4.534873188405797, "grad_norm": 0.06729353804862086, "learning_rate": 2.0903932621217972e-06, "loss": 0.85, "step": 10013 }, { "epoch": 4.5353260869565215, "grad_norm": 0.07723512576732901, "learning_rate": 2.0863601175945723e-06, "loss": 0.8614, "step": 10014 }, { "epoch": 4.535778985507246, "grad_norm": 0.06532239917417092, "learning_rate": 2.082330763354001e-06, "loss": 0.8433, "step": 10015 }, { "epoch": 4.536231884057971, "grad_norm": 0.06632760180024296, "learning_rate": 2.078305199802917e-06, "loss": 0.8677, "step": 10016 }, { "epoch": 4.536684782608695, "grad_norm": 0.08130507848661735, "learning_rate": 2.074283427343744e-06, "loss": 0.8393, "step": 10017 }, { "epoch": 4.53713768115942, "grad_norm": 0.06769691043036086, "learning_rate": 2.070265446378561e-06, "loss": 0.8569, "step": 10018 }, { "epoch": 4.537590579710145, "grad_norm": 0.07209495721841812, "learning_rate": 2.0662512573090422e-06, "loss": 0.8417, "step": 10019 }, { "epoch": 4.538043478260869, "grad_norm": 0.0677513384468818, "learning_rate": 2.0622408605365017e-06, "loss": 0.8218, "step": 10020 }, { "epoch": 4.538496376811594, "grad_norm": 0.06745845199865698, "learning_rate": 2.058234256461855e-06, "loss": 0.8534, "step": 10021 }, { "epoch": 4.538949275362318, "grad_norm": 0.06600828327635891, "learning_rate": 2.0542314454856614e-06, "loss": 0.8348, "step": 10022 }, { "epoch": 4.539402173913043, "grad_norm": 0.06667285929785734, "learning_rate": 2.0502324280080766e-06, "loss": 0.8509, "step": 10023 }, { "epoch": 4.539855072463768, "grad_norm": 0.07120020356279655, "learning_rate": 2.0462372044288916e-06, "loss": 0.8564, "step": 10024 }, { "epoch": 4.540307971014493, "grad_norm": 0.07160355154867275, "learning_rate": 2.0422457751475157e-06, "loss": 0.8799, "step": 10025 }, { "epoch": 4.540760869565218, "grad_norm": 0.06946474699737094, "learning_rate": 2.0382581405629807e-06, "loss": 0.8273, "step": 10026 }, { "epoch": 4.541213768115942, "grad_norm": 0.06668085686959001, "learning_rate": 2.0342743010739286e-06, "loss": 0.8201, "step": 10027 }, { "epoch": 4.541666666666667, "grad_norm": 0.06616125681459219, "learning_rate": 2.0302942570786446e-06, "loss": 0.8591, "step": 10028 }, { "epoch": 4.542119565217392, "grad_norm": 0.08557266382106019, "learning_rate": 2.0263180089750057e-06, "loss": 0.8627, "step": 10029 }, { "epoch": 4.542572463768116, "grad_norm": 0.07419322922353494, "learning_rate": 2.0223455571605344e-06, "loss": 0.8543, "step": 10030 }, { "epoch": 4.543025362318841, "grad_norm": 0.06972810461103675, "learning_rate": 2.0183769020323574e-06, "loss": 0.8514, "step": 10031 }, { "epoch": 4.543478260869565, "grad_norm": 0.06753210298999904, "learning_rate": 2.014412043987233e-06, "loss": 0.8668, "step": 10032 }, { "epoch": 4.54393115942029, "grad_norm": 0.07061630791813356, "learning_rate": 2.010450983421519e-06, "loss": 0.8456, "step": 10033 }, { "epoch": 4.544384057971015, "grad_norm": 0.07475088260663983, "learning_rate": 2.0064937207312286e-06, "loss": 0.8544, "step": 10034 }, { "epoch": 4.544836956521739, "grad_norm": 0.06843523059301602, "learning_rate": 2.00254025631196e-06, "loss": 0.8393, "step": 10035 }, { "epoch": 4.545289855072464, "grad_norm": 0.07389758315135934, "learning_rate": 1.998590590558962e-06, "loss": 0.8541, "step": 10036 }, { "epoch": 4.5457427536231885, "grad_norm": 0.06885519742275843, "learning_rate": 1.9946447238670742e-06, "loss": 0.8302, "step": 10037 }, { "epoch": 4.546195652173913, "grad_norm": 0.07383387642666113, "learning_rate": 1.990702656630781e-06, "loss": 0.8549, "step": 10038 }, { "epoch": 4.546648550724638, "grad_norm": 0.08791098197192553, "learning_rate": 1.9867643892441755e-06, "loss": 0.8618, "step": 10039 }, { "epoch": 4.547101449275362, "grad_norm": 0.07089686264317695, "learning_rate": 1.982829922100975e-06, "loss": 0.8643, "step": 10040 }, { "epoch": 4.547554347826087, "grad_norm": 0.0754975373119244, "learning_rate": 1.9788992555944996e-06, "loss": 0.871, "step": 10041 }, { "epoch": 4.5480072463768115, "grad_norm": 0.062299433758022396, "learning_rate": 1.974972390117724e-06, "loss": 0.8442, "step": 10042 }, { "epoch": 4.548460144927536, "grad_norm": 0.06773873633792765, "learning_rate": 1.971049326063206e-06, "loss": 0.8446, "step": 10043 }, { "epoch": 4.548913043478261, "grad_norm": 0.07281370373669811, "learning_rate": 1.9671300638231594e-06, "loss": 0.8715, "step": 10044 }, { "epoch": 4.549365942028985, "grad_norm": 0.07105905566282524, "learning_rate": 1.9632146037893786e-06, "loss": 0.8531, "step": 10045 }, { "epoch": 4.54981884057971, "grad_norm": 0.06932495566849088, "learning_rate": 1.9593029463533143e-06, "loss": 0.86, "step": 10046 }, { "epoch": 4.550271739130435, "grad_norm": 0.06699752063863865, "learning_rate": 1.9553950919060182e-06, "loss": 0.8473, "step": 10047 }, { "epoch": 4.550724637681159, "grad_norm": 0.07545285199741937, "learning_rate": 1.951491040838156e-06, "loss": 0.8536, "step": 10048 }, { "epoch": 4.551177536231884, "grad_norm": 0.0698571885636565, "learning_rate": 1.947590793540033e-06, "loss": 0.8552, "step": 10049 }, { "epoch": 4.551630434782608, "grad_norm": 0.06572944544196985, "learning_rate": 1.9436943504015505e-06, "loss": 0.8485, "step": 10050 }, { "epoch": 4.552083333333333, "grad_norm": 0.06982026757367714, "learning_rate": 1.9398017118122546e-06, "loss": 0.8185, "step": 10051 }, { "epoch": 4.552536231884058, "grad_norm": 0.06753929698696923, "learning_rate": 1.935912878161288e-06, "loss": 0.8395, "step": 10052 }, { "epoch": 4.552989130434782, "grad_norm": 0.06851290326692819, "learning_rate": 1.9320278498374323e-06, "loss": 0.8402, "step": 10053 }, { "epoch": 4.553442028985507, "grad_norm": 0.06756094150593392, "learning_rate": 1.9281466272290704e-06, "loss": 0.8585, "step": 10054 }, { "epoch": 4.553894927536232, "grad_norm": 0.06941619567775013, "learning_rate": 1.9242692107242257e-06, "loss": 0.8653, "step": 10055 }, { "epoch": 4.554347826086957, "grad_norm": 0.06640916118320211, "learning_rate": 1.9203956007105293e-06, "loss": 0.8301, "step": 10056 }, { "epoch": 4.554800724637682, "grad_norm": 0.0731747439612123, "learning_rate": 1.9165257975752193e-06, "loss": 0.8617, "step": 10057 }, { "epoch": 4.555253623188406, "grad_norm": 0.0701614739296728, "learning_rate": 1.9126598017051766e-06, "loss": 0.8157, "step": 10058 }, { "epoch": 4.555706521739131, "grad_norm": 0.06690345452598463, "learning_rate": 1.908797613486888e-06, "loss": 0.8473, "step": 10059 }, { "epoch": 4.556159420289855, "grad_norm": 0.07538834772431001, "learning_rate": 1.9049392333064577e-06, "loss": 0.8629, "step": 10060 }, { "epoch": 4.55661231884058, "grad_norm": 0.06843211706608472, "learning_rate": 1.901084661549626e-06, "loss": 0.8391, "step": 10061 }, { "epoch": 4.557065217391305, "grad_norm": 0.08494584175761997, "learning_rate": 1.8972338986017247e-06, "loss": 0.8449, "step": 10062 }, { "epoch": 4.557518115942029, "grad_norm": 0.07916167815563592, "learning_rate": 1.8933869448477394e-06, "loss": 0.8448, "step": 10063 }, { "epoch": 4.557971014492754, "grad_norm": 0.063627217845119, "learning_rate": 1.8895438006722377e-06, "loss": 0.8325, "step": 10064 }, { "epoch": 4.5584239130434785, "grad_norm": 0.08823221769848819, "learning_rate": 1.885704466459446e-06, "loss": 0.8518, "step": 10065 }, { "epoch": 4.558876811594203, "grad_norm": 0.06824000504864405, "learning_rate": 1.881868942593168e-06, "loss": 0.8485, "step": 10066 }, { "epoch": 4.559329710144928, "grad_norm": 0.07162491234302874, "learning_rate": 1.8780372294568573e-06, "loss": 0.8423, "step": 10067 }, { "epoch": 4.559782608695652, "grad_norm": 0.07191464370967794, "learning_rate": 1.874209327433567e-06, "loss": 0.8255, "step": 10068 }, { "epoch": 4.560235507246377, "grad_norm": 0.06372203312980743, "learning_rate": 1.8703852369059961e-06, "loss": 0.8279, "step": 10069 }, { "epoch": 4.5606884057971016, "grad_norm": 0.07352836408120023, "learning_rate": 1.8665649582564248e-06, "loss": 0.8212, "step": 10070 }, { "epoch": 4.561141304347826, "grad_norm": 0.06767024331623336, "learning_rate": 1.862748491866788e-06, "loss": 0.8494, "step": 10071 }, { "epoch": 4.561594202898551, "grad_norm": 0.06946949145067693, "learning_rate": 1.8589358381186163e-06, "loss": 0.8569, "step": 10072 }, { "epoch": 4.562047101449275, "grad_norm": 0.07903168730840271, "learning_rate": 1.8551269973930707e-06, "loss": 0.8742, "step": 10073 }, { "epoch": 4.5625, "grad_norm": 0.06481519091167717, "learning_rate": 1.8513219700709272e-06, "loss": 0.8523, "step": 10074 }, { "epoch": 4.562952898550725, "grad_norm": 0.0810049508420131, "learning_rate": 1.8475207565325747e-06, "loss": 0.8519, "step": 10075 }, { "epoch": 4.563405797101449, "grad_norm": 0.07390294139993708, "learning_rate": 1.8437233571580294e-06, "loss": 0.8527, "step": 10076 }, { "epoch": 4.563858695652174, "grad_norm": 0.06588610009701303, "learning_rate": 1.839929772326925e-06, "loss": 0.858, "step": 10077 }, { "epoch": 4.5643115942028984, "grad_norm": 0.07322338988849804, "learning_rate": 1.836140002418505e-06, "loss": 0.8403, "step": 10078 }, { "epoch": 4.564764492753623, "grad_norm": 0.07871216170584083, "learning_rate": 1.832354047811653e-06, "loss": 0.8525, "step": 10079 }, { "epoch": 4.565217391304348, "grad_norm": 0.06302271996742594, "learning_rate": 1.8285719088848442e-06, "loss": 0.8526, "step": 10080 }, { "epoch": 4.565670289855072, "grad_norm": 0.06758526150073609, "learning_rate": 1.8247935860161937e-06, "loss": 0.8667, "step": 10081 }, { "epoch": 4.566123188405797, "grad_norm": 0.07107895090081343, "learning_rate": 1.8210190795834216e-06, "loss": 0.8631, "step": 10082 }, { "epoch": 4.5665760869565215, "grad_norm": 0.06645748182638708, "learning_rate": 1.8172483899638704e-06, "loss": 0.8448, "step": 10083 }, { "epoch": 4.567028985507246, "grad_norm": 0.06813073491926286, "learning_rate": 1.8134815175345011e-06, "loss": 0.8379, "step": 10084 }, { "epoch": 4.567481884057971, "grad_norm": 0.06319613248721591, "learning_rate": 1.8097184626719011e-06, "loss": 0.8394, "step": 10085 }, { "epoch": 4.567934782608695, "grad_norm": 0.06581597918102035, "learning_rate": 1.8059592257522584e-06, "loss": 0.8475, "step": 10086 }, { "epoch": 4.56838768115942, "grad_norm": 0.06569287371518938, "learning_rate": 1.802203807151397e-06, "loss": 0.8265, "step": 10087 }, { "epoch": 4.568840579710145, "grad_norm": 0.07001064012707489, "learning_rate": 1.7984522072447497e-06, "loss": 0.8596, "step": 10088 }, { "epoch": 4.569293478260869, "grad_norm": 0.06484728312602761, "learning_rate": 1.7947044264073766e-06, "loss": 0.8408, "step": 10089 }, { "epoch": 4.569746376811594, "grad_norm": 0.06749359450056233, "learning_rate": 1.7909604650139423e-06, "loss": 0.8521, "step": 10090 }, { "epoch": 4.570199275362318, "grad_norm": 0.07280202257648742, "learning_rate": 1.7872203234387342e-06, "loss": 0.8783, "step": 10091 }, { "epoch": 4.570652173913043, "grad_norm": 0.06314124431045952, "learning_rate": 1.783484002055662e-06, "loss": 0.8391, "step": 10092 }, { "epoch": 4.571105072463768, "grad_norm": 0.06390283571947045, "learning_rate": 1.7797515012382583e-06, "loss": 0.8339, "step": 10093 }, { "epoch": 4.571557971014493, "grad_norm": 0.06595385518174682, "learning_rate": 1.7760228213596598e-06, "loss": 0.8616, "step": 10094 }, { "epoch": 4.572010869565218, "grad_norm": 0.0630051912971934, "learning_rate": 1.7722979627926307e-06, "loss": 0.8285, "step": 10095 }, { "epoch": 4.572463768115942, "grad_norm": 0.06765043037701508, "learning_rate": 1.7685769259095486e-06, "loss": 0.8641, "step": 10096 }, { "epoch": 4.572916666666667, "grad_norm": 0.07307370692612819, "learning_rate": 1.7648597110824183e-06, "loss": 0.8404, "step": 10097 }, { "epoch": 4.573369565217392, "grad_norm": 0.07263585528670821, "learning_rate": 1.7611463186828537e-06, "loss": 0.858, "step": 10098 }, { "epoch": 4.573822463768116, "grad_norm": 0.06566743912198766, "learning_rate": 1.7574367490820865e-06, "loss": 0.8189, "step": 10099 }, { "epoch": 4.574275362318841, "grad_norm": 0.07239703130239351, "learning_rate": 1.753731002650958e-06, "loss": 0.8475, "step": 10100 }, { "epoch": 4.574728260869565, "grad_norm": 0.06727811870607325, "learning_rate": 1.7500290797599585e-06, "loss": 0.8371, "step": 10101 }, { "epoch": 4.57518115942029, "grad_norm": 0.0759316099833996, "learning_rate": 1.7463309807791562e-06, "loss": 0.8626, "step": 10102 }, { "epoch": 4.575634057971015, "grad_norm": 0.0729447217583676, "learning_rate": 1.742636706078269e-06, "loss": 0.8578, "step": 10103 }, { "epoch": 4.576086956521739, "grad_norm": 0.07014309892594155, "learning_rate": 1.7389462560266102e-06, "loss": 0.8394, "step": 10104 }, { "epoch": 4.576539855072464, "grad_norm": 0.07064683717243928, "learning_rate": 1.7352596309931292e-06, "loss": 0.8582, "step": 10105 }, { "epoch": 4.5769927536231885, "grad_norm": 0.06693161839389276, "learning_rate": 1.7315768313463755e-06, "loss": 0.8498, "step": 10106 }, { "epoch": 4.577445652173913, "grad_norm": 0.06841875813577493, "learning_rate": 1.7278978574545301e-06, "loss": 0.8391, "step": 10107 }, { "epoch": 4.577898550724638, "grad_norm": 0.06447003759668483, "learning_rate": 1.724222709685379e-06, "loss": 0.8475, "step": 10108 }, { "epoch": 4.578351449275362, "grad_norm": 0.0673712391636628, "learning_rate": 1.7205513884063396e-06, "loss": 0.8505, "step": 10109 }, { "epoch": 4.578804347826087, "grad_norm": 0.071817397917317, "learning_rate": 1.7168838939844335e-06, "loss": 0.8496, "step": 10110 }, { "epoch": 4.5792572463768115, "grad_norm": 0.07292772856902935, "learning_rate": 1.713220226786314e-06, "loss": 0.8546, "step": 10111 }, { "epoch": 4.579710144927536, "grad_norm": 0.07356814837343627, "learning_rate": 1.709560387178235e-06, "loss": 0.8466, "step": 10112 }, { "epoch": 4.580163043478261, "grad_norm": 0.06970927236481148, "learning_rate": 1.7059043755260817e-06, "loss": 0.8501, "step": 10113 }, { "epoch": 4.580615942028985, "grad_norm": 0.06948571819956058, "learning_rate": 1.7022521921953482e-06, "loss": 0.864, "step": 10114 }, { "epoch": 4.58106884057971, "grad_norm": 0.07060336896584227, "learning_rate": 1.698603837551156e-06, "loss": 0.8376, "step": 10115 }, { "epoch": 4.581521739130435, "grad_norm": 0.07776167446274158, "learning_rate": 1.6949593119582352e-06, "loss": 0.8612, "step": 10116 }, { "epoch": 4.581974637681159, "grad_norm": 0.06713266008552282, "learning_rate": 1.6913186157809257e-06, "loss": 0.8625, "step": 10117 }, { "epoch": 4.582427536231884, "grad_norm": 0.06856899241724922, "learning_rate": 1.6876817493831987e-06, "loss": 0.8615, "step": 10118 }, { "epoch": 4.582880434782608, "grad_norm": 0.07038128033985735, "learning_rate": 1.6840487131286388e-06, "loss": 0.8542, "step": 10119 }, { "epoch": 4.583333333333333, "grad_norm": 0.0660614478874669, "learning_rate": 1.6804195073804442e-06, "loss": 0.8587, "step": 10120 }, { "epoch": 4.583786231884058, "grad_norm": 0.06821240918236886, "learning_rate": 1.6767941325014402e-06, "loss": 0.8392, "step": 10121 }, { "epoch": 4.584239130434782, "grad_norm": 0.0698834621807543, "learning_rate": 1.6731725888540485e-06, "loss": 0.8633, "step": 10122 }, { "epoch": 4.584692028985507, "grad_norm": 0.0637743798278594, "learning_rate": 1.66955487680033e-06, "loss": 0.841, "step": 10123 }, { "epoch": 4.585144927536232, "grad_norm": 0.07622393732171497, "learning_rate": 1.6659409967019513e-06, "loss": 0.8486, "step": 10124 }, { "epoch": 4.585597826086957, "grad_norm": 0.0693735847043641, "learning_rate": 1.6623309489201967e-06, "loss": 0.8476, "step": 10125 }, { "epoch": 4.586050724637682, "grad_norm": 0.06838297344340835, "learning_rate": 1.658724733815964e-06, "loss": 0.8508, "step": 10126 }, { "epoch": 4.586503623188406, "grad_norm": 0.0680439370514386, "learning_rate": 1.6551223517497827e-06, "loss": 0.8571, "step": 10127 }, { "epoch": 4.586956521739131, "grad_norm": 0.07427694849675469, "learning_rate": 1.651523803081778e-06, "loss": 0.8392, "step": 10128 }, { "epoch": 4.587409420289855, "grad_norm": 0.07151211828514258, "learning_rate": 1.6479290881717114e-06, "loss": 0.8452, "step": 10129 }, { "epoch": 4.58786231884058, "grad_norm": 0.07666986081843347, "learning_rate": 1.6443382073789393e-06, "loss": 0.8414, "step": 10130 }, { "epoch": 4.588315217391305, "grad_norm": 0.06254084512943023, "learning_rate": 1.6407511610624638e-06, "loss": 0.8483, "step": 10131 }, { "epoch": 4.588768115942029, "grad_norm": 0.08211010599512486, "learning_rate": 1.6371679495808823e-06, "loss": 0.8572, "step": 10132 }, { "epoch": 4.589221014492754, "grad_norm": 0.07730290646265546, "learning_rate": 1.6335885732924107e-06, "loss": 0.8394, "step": 10133 }, { "epoch": 4.5896739130434785, "grad_norm": 0.07146593021953875, "learning_rate": 1.6300130325548825e-06, "loss": 0.8453, "step": 10134 }, { "epoch": 4.590126811594203, "grad_norm": 0.07279239323724701, "learning_rate": 1.626441327725754e-06, "loss": 0.8402, "step": 10135 }, { "epoch": 4.590579710144928, "grad_norm": 0.09244432048763732, "learning_rate": 1.6228734591620953e-06, "loss": 0.8378, "step": 10136 }, { "epoch": 4.591032608695652, "grad_norm": 0.07863846879960665, "learning_rate": 1.61930942722059e-06, "loss": 0.843, "step": 10137 }, { "epoch": 4.591485507246377, "grad_norm": 0.07826697062835017, "learning_rate": 1.6157492322575396e-06, "loss": 0.8557, "step": 10138 }, { "epoch": 4.5919384057971016, "grad_norm": 0.07220155502575736, "learning_rate": 1.6121928746288684e-06, "loss": 0.85, "step": 10139 }, { "epoch": 4.592391304347826, "grad_norm": 0.07033123962641481, "learning_rate": 1.6086403546901053e-06, "loss": 0.8343, "step": 10140 }, { "epoch": 4.592844202898551, "grad_norm": 0.07103579035073336, "learning_rate": 1.6050916727964017e-06, "loss": 0.847, "step": 10141 }, { "epoch": 4.593297101449275, "grad_norm": 0.07314390807916712, "learning_rate": 1.6015468293025227e-06, "loss": 0.8515, "step": 10142 }, { "epoch": 4.59375, "grad_norm": 0.07274901119186225, "learning_rate": 1.598005824562856e-06, "loss": 0.8442, "step": 10143 }, { "epoch": 4.594202898550725, "grad_norm": 0.070085795963993, "learning_rate": 1.5944686589313984e-06, "loss": 0.8377, "step": 10144 }, { "epoch": 4.594655797101449, "grad_norm": 0.07321632279175883, "learning_rate": 1.5909353327617693e-06, "loss": 0.8455, "step": 10145 }, { "epoch": 4.595108695652174, "grad_norm": 0.06470274744518326, "learning_rate": 1.5874058464071974e-06, "loss": 0.8527, "step": 10146 }, { "epoch": 4.5955615942028984, "grad_norm": 0.07093267385989684, "learning_rate": 1.5838802002205377e-06, "loss": 0.8641, "step": 10147 }, { "epoch": 4.596014492753623, "grad_norm": 0.08044233092811022, "learning_rate": 1.5803583945542511e-06, "loss": 0.8543, "step": 10148 }, { "epoch": 4.596467391304348, "grad_norm": 0.07335044288872515, "learning_rate": 1.5768404297604112e-06, "loss": 0.8431, "step": 10149 }, { "epoch": 4.596920289855072, "grad_norm": 0.06316218294830452, "learning_rate": 1.5733263061907234e-06, "loss": 0.8583, "step": 10150 }, { "epoch": 4.597373188405797, "grad_norm": 0.07272217637580194, "learning_rate": 1.5698160241964978e-06, "loss": 0.8597, "step": 10151 }, { "epoch": 4.5978260869565215, "grad_norm": 0.06609319495578866, "learning_rate": 1.5663095841286535e-06, "loss": 0.8431, "step": 10152 }, { "epoch": 4.598278985507246, "grad_norm": 0.0683036105605981, "learning_rate": 1.5628069863377548e-06, "loss": 0.8617, "step": 10153 }, { "epoch": 4.598731884057971, "grad_norm": 0.06996987031567395, "learning_rate": 1.5593082311739439e-06, "loss": 0.8316, "step": 10154 }, { "epoch": 4.599184782608695, "grad_norm": 0.07333055639903648, "learning_rate": 1.5558133189870073e-06, "loss": 0.851, "step": 10155 }, { "epoch": 4.59963768115942, "grad_norm": 0.06802709715453356, "learning_rate": 1.552322250126328e-06, "loss": 0.8649, "step": 10156 }, { "epoch": 4.600090579710145, "grad_norm": 0.0798764043105079, "learning_rate": 1.5488350249409245e-06, "loss": 0.848, "step": 10157 }, { "epoch": 4.600543478260869, "grad_norm": 0.0761014718594131, "learning_rate": 1.5453516437794159e-06, "loss": 0.8459, "step": 10158 }, { "epoch": 4.600996376811594, "grad_norm": 0.07800102106300824, "learning_rate": 1.5418721069900388e-06, "loss": 0.8605, "step": 10159 }, { "epoch": 4.601449275362318, "grad_norm": 0.07666454079373419, "learning_rate": 1.5383964149206442e-06, "loss": 0.835, "step": 10160 }, { "epoch": 4.601902173913043, "grad_norm": 0.09221614224099006, "learning_rate": 1.5349245679187141e-06, "loss": 0.8383, "step": 10161 }, { "epoch": 4.602355072463768, "grad_norm": 0.07147123942552558, "learning_rate": 1.5314565663313219e-06, "loss": 0.8598, "step": 10162 }, { "epoch": 4.602807971014493, "grad_norm": 0.06198476112730373, "learning_rate": 1.5279924105051813e-06, "loss": 0.8707, "step": 10163 }, { "epoch": 4.603260869565218, "grad_norm": 0.07501332689737986, "learning_rate": 1.5245321007865977e-06, "loss": 0.8425, "step": 10164 }, { "epoch": 4.603713768115942, "grad_norm": 0.07003307728121086, "learning_rate": 1.5210756375215118e-06, "loss": 0.8543, "step": 10165 }, { "epoch": 4.604166666666667, "grad_norm": 0.07480254981115429, "learning_rate": 1.5176230210554744e-06, "loss": 0.8643, "step": 10166 }, { "epoch": 4.604619565217392, "grad_norm": 0.06727741230884453, "learning_rate": 1.51417425173364e-06, "loss": 0.8664, "step": 10167 }, { "epoch": 4.605072463768116, "grad_norm": 0.07220272989299095, "learning_rate": 1.5107293299007907e-06, "loss": 0.8629, "step": 10168 }, { "epoch": 4.605525362318841, "grad_norm": 0.06228087213646275, "learning_rate": 1.5072882559013225e-06, "loss": 0.8312, "step": 10169 }, { "epoch": 4.605978260869565, "grad_norm": 0.06615710923817816, "learning_rate": 1.503851030079244e-06, "loss": 0.8655, "step": 10170 }, { "epoch": 4.60643115942029, "grad_norm": 0.062008000003872486, "learning_rate": 1.5004176527781833e-06, "loss": 0.8442, "step": 10171 }, { "epoch": 4.606884057971015, "grad_norm": 0.06780649292075537, "learning_rate": 1.4969881243413764e-06, "loss": 0.8412, "step": 10172 }, { "epoch": 4.607336956521739, "grad_norm": 0.07091831743270002, "learning_rate": 1.4935624451116737e-06, "loss": 0.851, "step": 10173 }, { "epoch": 4.607789855072464, "grad_norm": 0.06941270113046273, "learning_rate": 1.4901406154315612e-06, "loss": 0.8469, "step": 10174 }, { "epoch": 4.6082427536231885, "grad_norm": 0.08175270709982635, "learning_rate": 1.4867226356431118e-06, "loss": 0.8339, "step": 10175 }, { "epoch": 4.608695652173913, "grad_norm": 0.07209943754535526, "learning_rate": 1.4833085060880349e-06, "loss": 0.8493, "step": 10176 }, { "epoch": 4.609148550724638, "grad_norm": 0.06409220772509679, "learning_rate": 1.4798982271076302e-06, "loss": 0.8465, "step": 10177 }, { "epoch": 4.609601449275362, "grad_norm": 0.08021476811178038, "learning_rate": 1.4764917990428518e-06, "loss": 0.8323, "step": 10178 }, { "epoch": 4.610054347826087, "grad_norm": 0.06911152058239377, "learning_rate": 1.473089222234223e-06, "loss": 0.8629, "step": 10179 }, { "epoch": 4.6105072463768115, "grad_norm": 0.07125335063418058, "learning_rate": 1.469690497021925e-06, "loss": 0.846, "step": 10180 }, { "epoch": 4.610960144927536, "grad_norm": 0.07627196582679817, "learning_rate": 1.4662956237457172e-06, "loss": 0.8573, "step": 10181 }, { "epoch": 4.611413043478261, "grad_norm": 0.07389675257149708, "learning_rate": 1.4629046027450034e-06, "loss": 0.8373, "step": 10182 }, { "epoch": 4.611865942028985, "grad_norm": 0.06945585394666262, "learning_rate": 1.4595174343587838e-06, "loss": 0.8538, "step": 10183 }, { "epoch": 4.61231884057971, "grad_norm": 0.07142852808090128, "learning_rate": 1.4561341189256762e-06, "loss": 0.8478, "step": 10184 }, { "epoch": 4.612771739130435, "grad_norm": 0.07405401887970196, "learning_rate": 1.4527546567839169e-06, "loss": 0.8483, "step": 10185 }, { "epoch": 4.613224637681159, "grad_norm": 0.073988872830089, "learning_rate": 1.4493790482713644e-06, "loss": 0.8526, "step": 10186 }, { "epoch": 4.613677536231884, "grad_norm": 0.07057092646837886, "learning_rate": 1.4460072937254688e-06, "loss": 0.8693, "step": 10187 }, { "epoch": 4.614130434782608, "grad_norm": 0.0642901168192915, "learning_rate": 1.4426393934833249e-06, "loss": 0.8652, "step": 10188 }, { "epoch": 4.614583333333333, "grad_norm": 0.07453001454664239, "learning_rate": 1.4392753478816145e-06, "loss": 0.8416, "step": 10189 }, { "epoch": 4.615036231884058, "grad_norm": 0.06667788154052637, "learning_rate": 1.4359151572566598e-06, "loss": 0.8495, "step": 10190 }, { "epoch": 4.615489130434782, "grad_norm": 0.07149890283949448, "learning_rate": 1.4325588219443743e-06, "loss": 0.8631, "step": 10191 }, { "epoch": 4.615942028985507, "grad_norm": 0.06353534174788002, "learning_rate": 1.4292063422802982e-06, "loss": 0.8325, "step": 10192 }, { "epoch": 4.616394927536232, "grad_norm": 0.06177837823472016, "learning_rate": 1.4258577185995815e-06, "loss": 0.8541, "step": 10193 }, { "epoch": 4.616847826086957, "grad_norm": 0.07601701882631379, "learning_rate": 1.4225129512369962e-06, "loss": 0.8482, "step": 10194 }, { "epoch": 4.617300724637682, "grad_norm": 0.06626892655615561, "learning_rate": 1.4191720405269194e-06, "loss": 0.8535, "step": 10195 }, { "epoch": 4.617753623188406, "grad_norm": 0.06159298224865895, "learning_rate": 1.4158349868033594e-06, "loss": 0.8411, "step": 10196 }, { "epoch": 4.618206521739131, "grad_norm": 0.06264701926534257, "learning_rate": 1.4125017903999073e-06, "loss": 0.8447, "step": 10197 }, { "epoch": 4.618659420289855, "grad_norm": 0.07151732200243624, "learning_rate": 1.409172451649803e-06, "loss": 0.8573, "step": 10198 }, { "epoch": 4.61911231884058, "grad_norm": 0.064500645540858, "learning_rate": 1.405846970885878e-06, "loss": 0.8508, "step": 10199 }, { "epoch": 4.619565217391305, "grad_norm": 0.06329560329951683, "learning_rate": 1.4025253484405954e-06, "loss": 0.8721, "step": 10200 }, { "epoch": 4.620018115942029, "grad_norm": 0.0644975537479953, "learning_rate": 1.399207584646014e-06, "loss": 0.8691, "step": 10201 }, { "epoch": 4.620471014492754, "grad_norm": 0.06776765903484494, "learning_rate": 1.395893679833824e-06, "loss": 0.8598, "step": 10202 }, { "epoch": 4.6209239130434785, "grad_norm": 0.07055101185054276, "learning_rate": 1.3925836343353073e-06, "loss": 0.8435, "step": 10203 }, { "epoch": 4.621376811594203, "grad_norm": 0.07055851863350221, "learning_rate": 1.3892774484813853e-06, "loss": 0.8565, "step": 10204 }, { "epoch": 4.621829710144928, "grad_norm": 0.06818005203944237, "learning_rate": 1.3859751226025808e-06, "loss": 0.8523, "step": 10205 }, { "epoch": 4.622282608695652, "grad_norm": 0.06609579349118112, "learning_rate": 1.382676657029034e-06, "loss": 0.8488, "step": 10206 }, { "epoch": 4.622735507246377, "grad_norm": 0.07529242812343766, "learning_rate": 1.3793820520904944e-06, "loss": 0.8574, "step": 10207 }, { "epoch": 4.6231884057971016, "grad_norm": 0.0737020214820034, "learning_rate": 1.3760913081163341e-06, "loss": 0.8648, "step": 10208 }, { "epoch": 4.623641304347826, "grad_norm": 0.07145679775901641, "learning_rate": 1.37280442543553e-06, "loss": 0.8507, "step": 10209 }, { "epoch": 4.624094202898551, "grad_norm": 0.060708488079382485, "learning_rate": 1.369521404376677e-06, "loss": 0.8487, "step": 10210 }, { "epoch": 4.624547101449275, "grad_norm": 0.06665330277670858, "learning_rate": 1.3662422452679792e-06, "loss": 0.8357, "step": 10211 }, { "epoch": 4.625, "grad_norm": 0.08349849565191475, "learning_rate": 1.3629669484372722e-06, "loss": 0.8416, "step": 10212 }, { "epoch": 4.625452898550725, "grad_norm": 0.06607090366008761, "learning_rate": 1.3596955142119784e-06, "loss": 0.8623, "step": 10213 }, { "epoch": 4.625905797101449, "grad_norm": 0.06404507662678953, "learning_rate": 1.3564279429191607e-06, "loss": 0.8676, "step": 10214 }, { "epoch": 4.626358695652174, "grad_norm": 0.07191700458887162, "learning_rate": 1.353164234885469e-06, "loss": 0.8492, "step": 10215 }, { "epoch": 4.6268115942028984, "grad_norm": 0.08603511154325007, "learning_rate": 1.3499043904371978e-06, "loss": 0.8762, "step": 10216 }, { "epoch": 4.627264492753623, "grad_norm": 0.07140089415438358, "learning_rate": 1.346648409900233e-06, "loss": 0.8446, "step": 10217 }, { "epoch": 4.627717391304348, "grad_norm": 0.0716301840649527, "learning_rate": 1.3433962936000745e-06, "loss": 0.8393, "step": 10218 }, { "epoch": 4.628170289855072, "grad_norm": 0.07781383795501996, "learning_rate": 1.340148041861844e-06, "loss": 0.8767, "step": 10219 }, { "epoch": 4.628623188405797, "grad_norm": 0.07379564628124526, "learning_rate": 1.3369036550102776e-06, "loss": 0.8428, "step": 10220 }, { "epoch": 4.6290760869565215, "grad_norm": 0.08123456253447003, "learning_rate": 1.333663133369716e-06, "loss": 0.8523, "step": 10221 }, { "epoch": 4.629528985507246, "grad_norm": 0.06872067205398502, "learning_rate": 1.3304264772641263e-06, "loss": 0.8493, "step": 10222 }, { "epoch": 4.629981884057971, "grad_norm": 0.07189435606365041, "learning_rate": 1.3271936870170764e-06, "loss": 0.8448, "step": 10223 }, { "epoch": 4.630434782608695, "grad_norm": 0.07092757089395343, "learning_rate": 1.3239647629517572e-06, "loss": 0.8491, "step": 10224 }, { "epoch": 4.63088768115942, "grad_norm": 0.07999734720158233, "learning_rate": 1.320739705390972e-06, "loss": 0.8401, "step": 10225 }, { "epoch": 4.631340579710145, "grad_norm": 0.07306325259911432, "learning_rate": 1.3175185146571302e-06, "loss": 0.8327, "step": 10226 }, { "epoch": 4.631793478260869, "grad_norm": 0.08406441946596986, "learning_rate": 1.3143011910722536e-06, "loss": 0.8467, "step": 10227 }, { "epoch": 4.632246376811594, "grad_norm": 0.0664662499755651, "learning_rate": 1.311087734957992e-06, "loss": 0.8437, "step": 10228 }, { "epoch": 4.632699275362318, "grad_norm": 0.06281253647515819, "learning_rate": 1.3078781466355995e-06, "loss": 0.8555, "step": 10229 }, { "epoch": 4.633152173913043, "grad_norm": 0.07438923558654748, "learning_rate": 1.3046724264259435e-06, "loss": 0.8567, "step": 10230 }, { "epoch": 4.633605072463768, "grad_norm": 0.06683060779791887, "learning_rate": 1.3014705746494972e-06, "loss": 0.8728, "step": 10231 }, { "epoch": 4.634057971014493, "grad_norm": 0.07724927495520992, "learning_rate": 1.2982725916263638e-06, "loss": 0.8537, "step": 10232 }, { "epoch": 4.634510869565218, "grad_norm": 0.06293685015774769, "learning_rate": 1.295078477676248e-06, "loss": 0.8301, "step": 10233 }, { "epoch": 4.634963768115942, "grad_norm": 0.06753446501804684, "learning_rate": 1.2918882331184724e-06, "loss": 0.8568, "step": 10234 }, { "epoch": 4.635416666666667, "grad_norm": 0.07381135114973134, "learning_rate": 1.2887018582719634e-06, "loss": 0.8519, "step": 10235 }, { "epoch": 4.635869565217392, "grad_norm": 0.06633289268581109, "learning_rate": 1.2855193534552756e-06, "loss": 0.8382, "step": 10236 }, { "epoch": 4.636322463768116, "grad_norm": 0.06484503422301241, "learning_rate": 1.2823407189865633e-06, "loss": 0.8604, "step": 10237 }, { "epoch": 4.636775362318841, "grad_norm": 0.07603279728331319, "learning_rate": 1.2791659551836078e-06, "loss": 0.8497, "step": 10238 }, { "epoch": 4.637228260869565, "grad_norm": 0.06860754825357077, "learning_rate": 1.2759950623637863e-06, "loss": 0.8363, "step": 10239 }, { "epoch": 4.63768115942029, "grad_norm": 0.07007466902156512, "learning_rate": 1.272828040844103e-06, "loss": 0.8593, "step": 10240 }, { "epoch": 4.638134057971015, "grad_norm": 0.0675598856374838, "learning_rate": 1.269664890941167e-06, "loss": 0.8617, "step": 10241 }, { "epoch": 4.638586956521739, "grad_norm": 0.06826143783584006, "learning_rate": 1.2665056129712138e-06, "loss": 0.8548, "step": 10242 }, { "epoch": 4.639039855072464, "grad_norm": 0.0671984906774052, "learning_rate": 1.263350207250067e-06, "loss": 0.88, "step": 10243 }, { "epoch": 4.6394927536231885, "grad_norm": 0.06595231263176576, "learning_rate": 1.260198674093185e-06, "loss": 0.8425, "step": 10244 }, { "epoch": 4.639945652173913, "grad_norm": 0.06803997575120257, "learning_rate": 1.2570510138156223e-06, "loss": 0.8587, "step": 10245 }, { "epoch": 4.640398550724638, "grad_norm": 0.07701264676471538, "learning_rate": 1.253907226732074e-06, "loss": 0.8463, "step": 10246 }, { "epoch": 4.640851449275362, "grad_norm": 0.06546348304603683, "learning_rate": 1.2507673131568087e-06, "loss": 0.8471, "step": 10247 }, { "epoch": 4.641304347826087, "grad_norm": 0.06929708495296805, "learning_rate": 1.2476312734037488e-06, "loss": 0.8456, "step": 10248 }, { "epoch": 4.6417572463768115, "grad_norm": 0.06525584332568661, "learning_rate": 1.24449910778639e-06, "loss": 0.8444, "step": 10249 }, { "epoch": 4.642210144927536, "grad_norm": 0.06564599705373818, "learning_rate": 1.241370816617873e-06, "loss": 0.8383, "step": 10250 }, { "epoch": 4.642663043478261, "grad_norm": 0.07066516009023509, "learning_rate": 1.238246400210934e-06, "loss": 0.8402, "step": 10251 }, { "epoch": 4.643115942028985, "grad_norm": 0.07083184752531696, "learning_rate": 1.235125858877928e-06, "loss": 0.8411, "step": 10252 }, { "epoch": 4.64356884057971, "grad_norm": 0.0747793261238447, "learning_rate": 1.2320091929308142e-06, "loss": 0.8499, "step": 10253 }, { "epoch": 4.644021739130435, "grad_norm": 0.06585576131389546, "learning_rate": 1.2288964026811744e-06, "loss": 0.8228, "step": 10254 }, { "epoch": 4.644474637681159, "grad_norm": 0.07800121890327728, "learning_rate": 1.2257874884402e-06, "loss": 0.8549, "step": 10255 }, { "epoch": 4.644927536231884, "grad_norm": 0.06313648069909678, "learning_rate": 1.222682450518695e-06, "loss": 0.8657, "step": 10256 }, { "epoch": 4.645380434782608, "grad_norm": 0.06693621571355264, "learning_rate": 1.2195812892270698e-06, "loss": 0.8541, "step": 10257 }, { "epoch": 4.645833333333333, "grad_norm": 0.06825444334083541, "learning_rate": 1.2164840048753602e-06, "loss": 0.8365, "step": 10258 }, { "epoch": 4.646286231884058, "grad_norm": 0.07092344305436993, "learning_rate": 1.2133905977731986e-06, "loss": 0.8556, "step": 10259 }, { "epoch": 4.646739130434782, "grad_norm": 0.07280428481490744, "learning_rate": 1.2103010682298489e-06, "loss": 0.8392, "step": 10260 }, { "epoch": 4.647192028985507, "grad_norm": 0.06997276680627128, "learning_rate": 1.2072154165541572e-06, "loss": 0.8513, "step": 10261 }, { "epoch": 4.647644927536232, "grad_norm": 0.06331929523521462, "learning_rate": 1.2041336430546235e-06, "loss": 0.8625, "step": 10262 }, { "epoch": 4.648097826086957, "grad_norm": 0.07002484523546613, "learning_rate": 1.201055748039317e-06, "loss": 0.8484, "step": 10263 }, { "epoch": 4.648550724637682, "grad_norm": 0.06668025437664642, "learning_rate": 1.1979817318159559e-06, "loss": 0.8417, "step": 10264 }, { "epoch": 4.649003623188406, "grad_norm": 0.06644720453986769, "learning_rate": 1.1949115946918454e-06, "loss": 0.8501, "step": 10265 }, { "epoch": 4.649456521739131, "grad_norm": 0.06298088040618684, "learning_rate": 1.191845336973918e-06, "loss": 0.8603, "step": 10266 }, { "epoch": 4.649909420289855, "grad_norm": 0.0714322901025598, "learning_rate": 1.1887829589687105e-06, "loss": 0.8361, "step": 10267 }, { "epoch": 4.65036231884058, "grad_norm": 0.07058103010963253, "learning_rate": 1.1857244609823693e-06, "loss": 0.8705, "step": 10268 }, { "epoch": 4.650815217391305, "grad_norm": 0.06984460529492542, "learning_rate": 1.1826698433206584e-06, "loss": 0.8686, "step": 10269 }, { "epoch": 4.651268115942029, "grad_norm": 0.06778591757464954, "learning_rate": 1.1796191062889606e-06, "loss": 0.8417, "step": 10270 }, { "epoch": 4.651721014492754, "grad_norm": 0.06372962172718448, "learning_rate": 1.1765722501922494e-06, "loss": 0.8376, "step": 10271 }, { "epoch": 4.6521739130434785, "grad_norm": 0.07160153858225565, "learning_rate": 1.1735292753351346e-06, "loss": 0.8656, "step": 10272 }, { "epoch": 4.652626811594203, "grad_norm": 0.061473789854343173, "learning_rate": 1.1704901820218217e-06, "loss": 0.8601, "step": 10273 }, { "epoch": 4.653079710144928, "grad_norm": 0.0719696032970881, "learning_rate": 1.167454970556139e-06, "loss": 0.8372, "step": 10274 }, { "epoch": 4.653532608695652, "grad_norm": 0.0754349399175678, "learning_rate": 1.1644236412415189e-06, "loss": 0.8513, "step": 10275 }, { "epoch": 4.653985507246377, "grad_norm": 0.0707327288903637, "learning_rate": 1.161396194381008e-06, "loss": 0.8349, "step": 10276 }, { "epoch": 4.6544384057971016, "grad_norm": 0.07166918905580007, "learning_rate": 1.1583726302772625e-06, "loss": 0.8354, "step": 10277 }, { "epoch": 4.654891304347826, "grad_norm": 0.06688344961439531, "learning_rate": 1.1553529492325555e-06, "loss": 0.8622, "step": 10278 }, { "epoch": 4.655344202898551, "grad_norm": 0.07180617980954286, "learning_rate": 1.1523371515487657e-06, "loss": 0.8276, "step": 10279 }, { "epoch": 4.655797101449275, "grad_norm": 0.06524204205927821, "learning_rate": 1.1493252375273944e-06, "loss": 0.8366, "step": 10280 }, { "epoch": 4.65625, "grad_norm": 0.07176509266950772, "learning_rate": 1.1463172074695428e-06, "loss": 0.8457, "step": 10281 }, { "epoch": 4.656702898550725, "grad_norm": 0.06599062955073721, "learning_rate": 1.1433130616759304e-06, "loss": 0.84, "step": 10282 }, { "epoch": 4.657155797101449, "grad_norm": 0.07143872711018374, "learning_rate": 1.140312800446881e-06, "loss": 0.821, "step": 10283 }, { "epoch": 4.657608695652174, "grad_norm": 0.07439918428114135, "learning_rate": 1.137316424082342e-06, "loss": 0.8388, "step": 10284 }, { "epoch": 4.6580615942028984, "grad_norm": 0.06768318457601942, "learning_rate": 1.13432393288186e-06, "loss": 0.8461, "step": 10285 }, { "epoch": 4.658514492753623, "grad_norm": 0.07752028993437692, "learning_rate": 1.1313353271446093e-06, "loss": 0.8514, "step": 10286 }, { "epoch": 4.658967391304348, "grad_norm": 0.06725336933573198, "learning_rate": 1.1283506071693506e-06, "loss": 0.8439, "step": 10287 }, { "epoch": 4.659420289855072, "grad_norm": 0.07346218803702675, "learning_rate": 1.1253697732544855e-06, "loss": 0.84, "step": 10288 }, { "epoch": 4.659873188405797, "grad_norm": 0.06980316468942126, "learning_rate": 1.1223928256979977e-06, "loss": 0.8533, "step": 10289 }, { "epoch": 4.6603260869565215, "grad_norm": 0.0669156596586051, "learning_rate": 1.1194197647975158e-06, "loss": 0.8506, "step": 10290 }, { "epoch": 4.660778985507246, "grad_norm": 0.11542117927708855, "learning_rate": 1.1164505908502465e-06, "loss": 0.8354, "step": 10291 }, { "epoch": 4.661231884057971, "grad_norm": 0.060122537513001124, "learning_rate": 1.113485304153028e-06, "loss": 0.8667, "step": 10292 }, { "epoch": 4.661684782608695, "grad_norm": 0.06552022370365732, "learning_rate": 1.1105239050023075e-06, "loss": 0.8559, "step": 10293 }, { "epoch": 4.66213768115942, "grad_norm": 0.0851862628279751, "learning_rate": 1.1075663936941373e-06, "loss": 0.8635, "step": 10294 }, { "epoch": 4.662590579710145, "grad_norm": 0.06132451584939715, "learning_rate": 1.1046127705241871e-06, "loss": 0.8656, "step": 10295 }, { "epoch": 4.663043478260869, "grad_norm": 0.06397095971443521, "learning_rate": 1.1016630357877278e-06, "loss": 0.8493, "step": 10296 }, { "epoch": 4.663496376811594, "grad_norm": 0.06415972145485516, "learning_rate": 1.0987171897796567e-06, "loss": 0.8477, "step": 10297 }, { "epoch": 4.663949275362318, "grad_norm": 0.06288763538106522, "learning_rate": 1.095775232794467e-06, "loss": 0.8605, "step": 10298 }, { "epoch": 4.664402173913043, "grad_norm": 0.07012752988971473, "learning_rate": 1.0928371651262837e-06, "loss": 0.8425, "step": 10299 }, { "epoch": 4.664855072463768, "grad_norm": 0.06335506686934275, "learning_rate": 1.0899029870688183e-06, "loss": 0.8419, "step": 10300 }, { "epoch": 4.665307971014493, "grad_norm": 0.07259404811344414, "learning_rate": 1.086972698915414e-06, "loss": 0.8154, "step": 10301 }, { "epoch": 4.665760869565218, "grad_norm": 0.06467085290870508, "learning_rate": 1.084046300959014e-06, "loss": 0.8383, "step": 10302 }, { "epoch": 4.666213768115942, "grad_norm": 0.06950622551727192, "learning_rate": 1.081123793492167e-06, "loss": 0.8537, "step": 10303 }, { "epoch": 4.666666666666667, "grad_norm": 0.06251895398637261, "learning_rate": 1.0782051768070477e-06, "loss": 0.8573, "step": 10304 }, { "epoch": 4.667119565217392, "grad_norm": 0.06338913485997916, "learning_rate": 1.0752904511954364e-06, "loss": 0.8567, "step": 10305 }, { "epoch": 4.667572463768116, "grad_norm": 0.06423453983408912, "learning_rate": 1.0723796169487178e-06, "loss": 0.8758, "step": 10306 }, { "epoch": 4.668025362318841, "grad_norm": 0.06073032145965118, "learning_rate": 1.0694726743578944e-06, "loss": 0.8387, "step": 10307 }, { "epoch": 4.668478260869565, "grad_norm": 0.059271429495924254, "learning_rate": 1.0665696237135782e-06, "loss": 0.849, "step": 10308 }, { "epoch": 4.66893115942029, "grad_norm": 0.06644652143238695, "learning_rate": 1.0636704653059949e-06, "loss": 0.8649, "step": 10309 }, { "epoch": 4.669384057971015, "grad_norm": 0.08350064441237777, "learning_rate": 1.0607751994249749e-06, "loss": 0.8506, "step": 10310 }, { "epoch": 4.669836956521739, "grad_norm": 0.06515454062604144, "learning_rate": 1.0578838263599622e-06, "loss": 0.8591, "step": 10311 }, { "epoch": 4.670289855072464, "grad_norm": 0.06539218303863603, "learning_rate": 1.0549963464000102e-06, "loss": 0.8529, "step": 10312 }, { "epoch": 4.6707427536231885, "grad_norm": 0.05860572905619627, "learning_rate": 1.0521127598337899e-06, "loss": 0.8585, "step": 10313 }, { "epoch": 4.671195652173913, "grad_norm": 0.06459665512353031, "learning_rate": 1.0492330669495688e-06, "loss": 0.8577, "step": 10314 }, { "epoch": 4.671648550724638, "grad_norm": 0.07021682390140144, "learning_rate": 1.0463572680352495e-06, "loss": 0.8533, "step": 10315 }, { "epoch": 4.672101449275362, "grad_norm": 0.06809085531039953, "learning_rate": 1.0434853633783137e-06, "loss": 0.8245, "step": 10316 }, { "epoch": 4.672554347826087, "grad_norm": 0.06392495444920773, "learning_rate": 1.040617353265887e-06, "loss": 0.846, "step": 10317 }, { "epoch": 4.6730072463768115, "grad_norm": 0.06012867891975018, "learning_rate": 1.0377532379846734e-06, "loss": 0.852, "step": 10318 }, { "epoch": 4.673460144927536, "grad_norm": 0.06810900094983019, "learning_rate": 1.0348930178210125e-06, "loss": 0.8372, "step": 10319 }, { "epoch": 4.673913043478261, "grad_norm": 0.0680996057636141, "learning_rate": 1.0320366930608405e-06, "loss": 0.8364, "step": 10320 }, { "epoch": 4.674365942028985, "grad_norm": 0.06406152874288396, "learning_rate": 1.029184263989711e-06, "loss": 0.8576, "step": 10321 }, { "epoch": 4.67481884057971, "grad_norm": 0.0701882820880485, "learning_rate": 1.0263357308927823e-06, "loss": 0.8427, "step": 10322 }, { "epoch": 4.675271739130435, "grad_norm": 0.06541224316707334, "learning_rate": 1.023491094054836e-06, "loss": 0.8395, "step": 10323 }, { "epoch": 4.675724637681159, "grad_norm": 0.07555861241433254, "learning_rate": 1.0206503537602396e-06, "loss": 0.8572, "step": 10324 }, { "epoch": 4.676177536231884, "grad_norm": 0.06811134778918594, "learning_rate": 1.017813510293002e-06, "loss": 0.8689, "step": 10325 }, { "epoch": 4.676630434782608, "grad_norm": 0.06760016607111918, "learning_rate": 1.0149805639367226e-06, "loss": 0.852, "step": 10326 }, { "epoch": 4.677083333333333, "grad_norm": 0.0605100991518275, "learning_rate": 1.0121515149746108e-06, "loss": 0.8449, "step": 10327 }, { "epoch": 4.677536231884058, "grad_norm": 0.06758463707899176, "learning_rate": 1.009326363689489e-06, "loss": 0.845, "step": 10328 }, { "epoch": 4.677989130434782, "grad_norm": 0.0728310656440874, "learning_rate": 1.0065051103638025e-06, "loss": 0.8524, "step": 10329 }, { "epoch": 4.678442028985507, "grad_norm": 0.06992472931346459, "learning_rate": 1.003687755279583e-06, "loss": 0.8326, "step": 10330 }, { "epoch": 4.678894927536232, "grad_norm": 0.06123560403670673, "learning_rate": 1.000874298718504e-06, "loss": 0.8695, "step": 10331 }, { "epoch": 4.679347826086957, "grad_norm": 0.06318423986303842, "learning_rate": 9.980647409618106e-07, "loss": 0.8624, "step": 10332 }, { "epoch": 4.679800724637682, "grad_norm": 0.06584102986352781, "learning_rate": 9.95259082290394e-07, "loss": 0.8554, "step": 10333 }, { "epoch": 4.680253623188406, "grad_norm": 0.06410844730084835, "learning_rate": 9.924573229847368e-07, "loss": 0.8715, "step": 10334 }, { "epoch": 4.680706521739131, "grad_norm": 0.07370053265172126, "learning_rate": 9.896594633249301e-07, "loss": 0.8448, "step": 10335 }, { "epoch": 4.681159420289855, "grad_norm": 0.0639434376901701, "learning_rate": 9.868655035906883e-07, "loss": 0.8366, "step": 10336 }, { "epoch": 4.68161231884058, "grad_norm": 0.07016170557292506, "learning_rate": 9.840754440613253e-07, "loss": 0.8552, "step": 10337 }, { "epoch": 4.682065217391305, "grad_norm": 0.06501644866991514, "learning_rate": 9.812892850157563e-07, "loss": 0.8423, "step": 10338 }, { "epoch": 4.682518115942029, "grad_norm": 0.06419605242394627, "learning_rate": 9.785070267325358e-07, "loss": 0.8554, "step": 10339 }, { "epoch": 4.682971014492754, "grad_norm": 0.07160985648645309, "learning_rate": 9.757286694898016e-07, "loss": 0.8396, "step": 10340 }, { "epoch": 4.6834239130434785, "grad_norm": 0.06785301540232999, "learning_rate": 9.729542135653092e-07, "loss": 0.8493, "step": 10341 }, { "epoch": 4.683876811594203, "grad_norm": 0.06454076456432849, "learning_rate": 9.701836592364278e-07, "loss": 0.833, "step": 10342 }, { "epoch": 4.684329710144928, "grad_norm": 0.06328517699362375, "learning_rate": 9.674170067801359e-07, "loss": 0.8515, "step": 10343 }, { "epoch": 4.684782608695652, "grad_norm": 0.06009494602837124, "learning_rate": 9.64654256473021e-07, "loss": 0.8436, "step": 10344 }, { "epoch": 4.685235507246377, "grad_norm": 0.06014695580449757, "learning_rate": 9.618954085912757e-07, "loss": 0.8685, "step": 10345 }, { "epoch": 4.6856884057971016, "grad_norm": 0.06519937496565495, "learning_rate": 9.591404634107015e-07, "loss": 0.8327, "step": 10346 }, { "epoch": 4.686141304347826, "grad_norm": 0.06396774521020829, "learning_rate": 9.56389421206727e-07, "loss": 0.8675, "step": 10347 }, { "epoch": 4.686594202898551, "grad_norm": 0.07541760396658098, "learning_rate": 9.536422822543678e-07, "loss": 0.8574, "step": 10348 }, { "epoch": 4.687047101449275, "grad_norm": 0.07255992993478969, "learning_rate": 9.508990468282664e-07, "loss": 0.8415, "step": 10349 }, { "epoch": 4.6875, "grad_norm": 0.06632547020074035, "learning_rate": 9.481597152026656e-07, "loss": 0.8353, "step": 10350 }, { "epoch": 4.687952898550725, "grad_norm": 0.0636198365462289, "learning_rate": 9.454242876514219e-07, "loss": 0.8607, "step": 10351 }, { "epoch": 4.688405797101449, "grad_norm": 0.0612931762148889, "learning_rate": 9.42692764448001e-07, "loss": 0.8488, "step": 10352 }, { "epoch": 4.688858695652174, "grad_norm": 0.0634536052991197, "learning_rate": 9.399651458654735e-07, "loss": 0.852, "step": 10353 }, { "epoch": 4.6893115942028984, "grad_norm": 0.06617947939649808, "learning_rate": 9.372414321765233e-07, "loss": 0.8789, "step": 10354 }, { "epoch": 4.689764492753623, "grad_norm": 0.059795135307353586, "learning_rate": 9.345216236534527e-07, "loss": 0.8522, "step": 10355 }, { "epoch": 4.690217391304348, "grad_norm": 0.07403314090595872, "learning_rate": 9.318057205681552e-07, "loss": 0.8597, "step": 10356 }, { "epoch": 4.690670289855072, "grad_norm": 0.06382558194319124, "learning_rate": 9.290937231921516e-07, "loss": 0.845, "step": 10357 }, { "epoch": 4.691123188405797, "grad_norm": 0.06445557695471062, "learning_rate": 9.26385631796558e-07, "loss": 0.8433, "step": 10358 }, { "epoch": 4.6915760869565215, "grad_norm": 0.0700464931045587, "learning_rate": 9.236814466521138e-07, "loss": 0.8531, "step": 10359 }, { "epoch": 4.692028985507246, "grad_norm": 0.06951128449655034, "learning_rate": 9.209811680291581e-07, "loss": 0.8601, "step": 10360 }, { "epoch": 4.692481884057971, "grad_norm": 0.07214535573258182, "learning_rate": 9.182847961976438e-07, "loss": 0.8507, "step": 10361 }, { "epoch": 4.692934782608695, "grad_norm": 0.06779286599925541, "learning_rate": 9.155923314271197e-07, "loss": 0.8575, "step": 10362 }, { "epoch": 4.69338768115942, "grad_norm": 0.059902808572710395, "learning_rate": 9.129037739867707e-07, "loss": 0.8378, "step": 10363 }, { "epoch": 4.693840579710145, "grad_norm": 0.060773582569257835, "learning_rate": 9.102191241453684e-07, "loss": 0.8587, "step": 10364 }, { "epoch": 4.694293478260869, "grad_norm": 0.06342459285396441, "learning_rate": 9.075383821713069e-07, "loss": 0.8509, "step": 10365 }, { "epoch": 4.694746376811594, "grad_norm": 0.07111804241024398, "learning_rate": 9.048615483325762e-07, "loss": 0.8453, "step": 10366 }, { "epoch": 4.695199275362318, "grad_norm": 0.06180222209399906, "learning_rate": 9.021886228967891e-07, "loss": 0.8534, "step": 10367 }, { "epoch": 4.695652173913043, "grad_norm": 0.06445327310310749, "learning_rate": 8.995196061311628e-07, "loss": 0.8476, "step": 10368 }, { "epoch": 4.696105072463768, "grad_norm": 0.06383201218523434, "learning_rate": 8.96854498302524e-07, "loss": 0.8627, "step": 10369 }, { "epoch": 4.696557971014493, "grad_norm": 0.05951922137330308, "learning_rate": 8.941932996772995e-07, "loss": 0.8438, "step": 10370 }, { "epoch": 4.697010869565218, "grad_norm": 0.07094568336452213, "learning_rate": 8.915360105215431e-07, "loss": 0.8552, "step": 10371 }, { "epoch": 4.697463768115942, "grad_norm": 0.06758140569148223, "learning_rate": 8.888826311009002e-07, "loss": 0.8377, "step": 10372 }, { "epoch": 4.697916666666667, "grad_norm": 0.06610116258326913, "learning_rate": 8.862331616806385e-07, "loss": 0.8642, "step": 10373 }, { "epoch": 4.698369565217392, "grad_norm": 0.0664098625689114, "learning_rate": 8.835876025256307e-07, "loss": 0.8751, "step": 10374 }, { "epoch": 4.698822463768116, "grad_norm": 0.06304703832710125, "learning_rate": 8.809459539003584e-07, "loss": 0.848, "step": 10375 }, { "epoch": 4.699275362318841, "grad_norm": 0.06714736011326165, "learning_rate": 8.783082160688994e-07, "loss": 0.8536, "step": 10376 }, { "epoch": 4.699728260869565, "grad_norm": 0.06780741120592866, "learning_rate": 8.756743892949715e-07, "loss": 0.8457, "step": 10377 }, { "epoch": 4.70018115942029, "grad_norm": 0.07520392460906346, "learning_rate": 8.730444738418708e-07, "loss": 0.8452, "step": 10378 }, { "epoch": 4.700634057971015, "grad_norm": 0.06481035304820038, "learning_rate": 8.70418469972516e-07, "loss": 0.8596, "step": 10379 }, { "epoch": 4.701086956521739, "grad_norm": 0.06144719983952692, "learning_rate": 8.677963779494259e-07, "loss": 0.8493, "step": 10380 }, { "epoch": 4.701539855072464, "grad_norm": 0.06692469114122172, "learning_rate": 8.651781980347507e-07, "loss": 0.8412, "step": 10381 }, { "epoch": 4.7019927536231885, "grad_norm": 0.05899826795486706, "learning_rate": 8.625639304902189e-07, "loss": 0.8471, "step": 10382 }, { "epoch": 4.702445652173913, "grad_norm": 0.06621550691089793, "learning_rate": 8.599535755771948e-07, "loss": 0.8351, "step": 10383 }, { "epoch": 4.702898550724638, "grad_norm": 0.06718374428280681, "learning_rate": 8.573471335566297e-07, "loss": 0.8518, "step": 10384 }, { "epoch": 4.703351449275362, "grad_norm": 0.06397351782408234, "learning_rate": 8.547446046891017e-07, "loss": 0.85, "step": 10385 }, { "epoch": 4.703804347826087, "grad_norm": 0.06755909489837902, "learning_rate": 8.521459892347894e-07, "loss": 0.8354, "step": 10386 }, { "epoch": 4.7042572463768115, "grad_norm": 0.0618549067214179, "learning_rate": 8.495512874534761e-07, "loss": 0.8422, "step": 10387 }, { "epoch": 4.704710144927536, "grad_norm": 0.060921475068504406, "learning_rate": 8.469604996045544e-07, "loss": 0.8627, "step": 10388 }, { "epoch": 4.705163043478261, "grad_norm": 0.06687077303025839, "learning_rate": 8.443736259470392e-07, "loss": 0.8616, "step": 10389 }, { "epoch": 4.705615942028985, "grad_norm": 0.05841043898719834, "learning_rate": 8.41790666739537e-07, "loss": 0.8544, "step": 10390 }, { "epoch": 4.70606884057971, "grad_norm": 0.06713490925423743, "learning_rate": 8.392116222402768e-07, "loss": 0.8427, "step": 10391 }, { "epoch": 4.706521739130435, "grad_norm": 0.07038353507142553, "learning_rate": 8.366364927070836e-07, "loss": 0.8271, "step": 10392 }, { "epoch": 4.706974637681159, "grad_norm": 0.06364530418841657, "learning_rate": 8.340652783974046e-07, "loss": 0.8675, "step": 10393 }, { "epoch": 4.707427536231884, "grad_norm": 0.0647781668707772, "learning_rate": 8.314979795682787e-07, "loss": 0.8457, "step": 10394 }, { "epoch": 4.707880434782608, "grad_norm": 0.0733194584790151, "learning_rate": 8.289345964763718e-07, "loss": 0.8745, "step": 10395 }, { "epoch": 4.708333333333333, "grad_norm": 0.06377817109414387, "learning_rate": 8.263751293779409e-07, "loss": 0.8619, "step": 10396 }, { "epoch": 4.708786231884058, "grad_norm": 0.06957723875509943, "learning_rate": 8.238195785288661e-07, "loss": 0.8546, "step": 10397 }, { "epoch": 4.709239130434782, "grad_norm": 0.06246477227728374, "learning_rate": 8.212679441846272e-07, "loss": 0.8605, "step": 10398 }, { "epoch": 4.709692028985507, "grad_norm": 0.06809740256605376, "learning_rate": 8.187202266003181e-07, "loss": 0.8476, "step": 10399 }, { "epoch": 4.710144927536232, "grad_norm": 0.07194965432289617, "learning_rate": 8.161764260306327e-07, "loss": 0.8495, "step": 10400 }, { "epoch": 4.710597826086957, "grad_norm": 0.06720572956642672, "learning_rate": 8.136365427298831e-07, "loss": 0.849, "step": 10401 }, { "epoch": 4.711050724637682, "grad_norm": 0.06611386597314158, "learning_rate": 8.111005769519864e-07, "loss": 0.8597, "step": 10402 }, { "epoch": 4.711503623188406, "grad_norm": 0.06582359073107312, "learning_rate": 8.085685289504641e-07, "loss": 0.8632, "step": 10403 }, { "epoch": 4.711956521739131, "grad_norm": 0.070761841171238, "learning_rate": 8.060403989784516e-07, "loss": 0.8515, "step": 10404 }, { "epoch": 4.712409420289855, "grad_norm": 0.06840524912239458, "learning_rate": 8.035161872886887e-07, "loss": 0.8573, "step": 10405 }, { "epoch": 4.71286231884058, "grad_norm": 0.06742651367579532, "learning_rate": 8.009958941335205e-07, "loss": 0.8799, "step": 10406 }, { "epoch": 4.713315217391305, "grad_norm": 0.06775488777317326, "learning_rate": 7.984795197649098e-07, "loss": 0.8649, "step": 10407 }, { "epoch": 4.713768115942029, "grad_norm": 0.07029444104468688, "learning_rate": 7.959670644344242e-07, "loss": 0.875, "step": 10408 }, { "epoch": 4.714221014492754, "grad_norm": 0.06725466331337568, "learning_rate": 7.934585283932317e-07, "loss": 0.8492, "step": 10409 }, { "epoch": 4.7146739130434785, "grad_norm": 0.06479453813464443, "learning_rate": 7.909539118921228e-07, "loss": 0.8445, "step": 10410 }, { "epoch": 4.715126811594203, "grad_norm": 0.06410101757293611, "learning_rate": 7.884532151814839e-07, "loss": 0.8296, "step": 10411 }, { "epoch": 4.715579710144928, "grad_norm": 0.06775727769087382, "learning_rate": 7.859564385113105e-07, "loss": 0.8539, "step": 10412 }, { "epoch": 4.716032608695652, "grad_norm": 0.060262239371706015, "learning_rate": 7.83463582131212e-07, "loss": 0.8541, "step": 10413 }, { "epoch": 4.716485507246377, "grad_norm": 0.06762499578080551, "learning_rate": 7.809746462904022e-07, "loss": 0.8565, "step": 10414 }, { "epoch": 4.7169384057971016, "grad_norm": 0.06649720622301153, "learning_rate": 7.784896312377044e-07, "loss": 0.8378, "step": 10415 }, { "epoch": 4.717391304347826, "grad_norm": 0.06834971858464113, "learning_rate": 7.760085372215509e-07, "loss": 0.858, "step": 10416 }, { "epoch": 4.717844202898551, "grad_norm": 0.07067993226709404, "learning_rate": 7.735313644899789e-07, "loss": 0.8616, "step": 10417 }, { "epoch": 4.718297101449275, "grad_norm": 0.06549941160697789, "learning_rate": 7.710581132906391e-07, "loss": 0.8371, "step": 10418 }, { "epoch": 4.71875, "grad_norm": 0.06397825693638286, "learning_rate": 7.685887838707828e-07, "loss": 0.8522, "step": 10419 }, { "epoch": 4.719202898550725, "grad_norm": 0.05729831876525508, "learning_rate": 7.661233764772746e-07, "loss": 0.85, "step": 10420 }, { "epoch": 4.719655797101449, "grad_norm": 0.06541790418768341, "learning_rate": 7.636618913565796e-07, "loss": 0.8402, "step": 10421 }, { "epoch": 4.720108695652174, "grad_norm": 0.06318100634920515, "learning_rate": 7.612043287547855e-07, "loss": 0.8359, "step": 10422 }, { "epoch": 4.7205615942028984, "grad_norm": 0.0630590264452115, "learning_rate": 7.587506889175666e-07, "loss": 0.8492, "step": 10423 }, { "epoch": 4.721014492753623, "grad_norm": 0.07081829346061305, "learning_rate": 7.563009720902292e-07, "loss": 0.858, "step": 10424 }, { "epoch": 4.721467391304348, "grad_norm": 0.06738615373026637, "learning_rate": 7.538551785176707e-07, "loss": 0.8468, "step": 10425 }, { "epoch": 4.721920289855072, "grad_norm": 0.06774183280980607, "learning_rate": 7.514133084443975e-07, "loss": 0.8585, "step": 10426 }, { "epoch": 4.722373188405797, "grad_norm": 0.06860439740163167, "learning_rate": 7.489753621145346e-07, "loss": 0.8313, "step": 10427 }, { "epoch": 4.7228260869565215, "grad_norm": 0.07765185536131702, "learning_rate": 7.465413397718024e-07, "loss": 0.8653, "step": 10428 }, { "epoch": 4.723278985507246, "grad_norm": 0.06264499725737481, "learning_rate": 7.441112416595353e-07, "loss": 0.8352, "step": 10429 }, { "epoch": 4.723731884057971, "grad_norm": 0.0634059244945265, "learning_rate": 7.416850680206722e-07, "loss": 0.8332, "step": 10430 }, { "epoch": 4.724184782608695, "grad_norm": 0.0652955309997085, "learning_rate": 7.392628190977613e-07, "loss": 0.8429, "step": 10431 }, { "epoch": 4.72463768115942, "grad_norm": 0.06643144844898581, "learning_rate": 7.368444951329645e-07, "loss": 0.8483, "step": 10432 }, { "epoch": 4.725090579710145, "grad_norm": 0.06214939900272799, "learning_rate": 7.344300963680395e-07, "loss": 0.852, "step": 10433 }, { "epoch": 4.725543478260869, "grad_norm": 0.06481312033947041, "learning_rate": 7.320196230443577e-07, "loss": 0.8589, "step": 10434 }, { "epoch": 4.725996376811594, "grad_norm": 0.06139509710758574, "learning_rate": 7.296130754028997e-07, "loss": 0.8373, "step": 10435 }, { "epoch": 4.726449275362318, "grad_norm": 0.06339668779698603, "learning_rate": 7.27210453684255e-07, "loss": 0.844, "step": 10436 }, { "epoch": 4.726902173913043, "grad_norm": 0.061891424615103, "learning_rate": 7.24811758128614e-07, "loss": 0.854, "step": 10437 }, { "epoch": 4.727355072463768, "grad_norm": 0.058922663078617626, "learning_rate": 7.224169889757804e-07, "loss": 0.8712, "step": 10438 }, { "epoch": 4.727807971014493, "grad_norm": 0.06554348727808822, "learning_rate": 7.200261464651537e-07, "loss": 0.8411, "step": 10439 }, { "epoch": 4.728260869565218, "grad_norm": 0.0682744434628946, "learning_rate": 7.176392308357649e-07, "loss": 0.8537, "step": 10440 }, { "epoch": 4.728713768115942, "grad_norm": 0.06562500341961508, "learning_rate": 7.152562423262232e-07, "loss": 0.8517, "step": 10441 }, { "epoch": 4.729166666666667, "grad_norm": 0.06495427267560003, "learning_rate": 7.128771811747737e-07, "loss": 0.8593, "step": 10442 }, { "epoch": 4.729619565217392, "grad_norm": 0.062389513418612555, "learning_rate": 7.105020476192437e-07, "loss": 0.8631, "step": 10443 }, { "epoch": 4.730072463768116, "grad_norm": 0.06047433337353177, "learning_rate": 7.081308418970878e-07, "loss": 0.8637, "step": 10444 }, { "epoch": 4.730525362318841, "grad_norm": 0.06120626401744506, "learning_rate": 7.057635642453519e-07, "loss": 0.8749, "step": 10445 }, { "epoch": 4.730978260869565, "grad_norm": 0.07171339020383172, "learning_rate": 7.034002149007002e-07, "loss": 0.8602, "step": 10446 }, { "epoch": 4.73143115942029, "grad_norm": 0.06976411833593858, "learning_rate": 7.010407940993969e-07, "loss": 0.8652, "step": 10447 }, { "epoch": 4.731884057971015, "grad_norm": 0.05937962032080448, "learning_rate": 6.986853020773243e-07, "loss": 0.8554, "step": 10448 }, { "epoch": 4.732336956521739, "grad_norm": 0.06770695723655368, "learning_rate": 6.963337390699565e-07, "loss": 0.8557, "step": 10449 }, { "epoch": 4.732789855072464, "grad_norm": 0.06119247414714818, "learning_rate": 6.939861053123897e-07, "loss": 0.8578, "step": 10450 }, { "epoch": 4.7332427536231885, "grad_norm": 0.06298903218352825, "learning_rate": 6.916424010393164e-07, "loss": 0.8562, "step": 10451 }, { "epoch": 4.733695652173913, "grad_norm": 0.06197945909178466, "learning_rate": 6.893026264850466e-07, "loss": 0.8541, "step": 10452 }, { "epoch": 4.734148550724638, "grad_norm": 0.06460877076275659, "learning_rate": 6.869667818834824e-07, "loss": 0.8422, "step": 10453 }, { "epoch": 4.734601449275362, "grad_norm": 0.06470826058728747, "learning_rate": 6.846348674681436e-07, "loss": 0.8415, "step": 10454 }, { "epoch": 4.735054347826087, "grad_norm": 0.061308915871043076, "learning_rate": 6.823068834721591e-07, "loss": 0.8537, "step": 10455 }, { "epoch": 4.7355072463768115, "grad_norm": 0.06931215741160711, "learning_rate": 6.799828301282629e-07, "loss": 0.8446, "step": 10456 }, { "epoch": 4.735960144927536, "grad_norm": 0.06048538731791855, "learning_rate": 6.776627076687847e-07, "loss": 0.8445, "step": 10457 }, { "epoch": 4.736413043478261, "grad_norm": 0.06635974138238787, "learning_rate": 6.753465163256812e-07, "loss": 0.8436, "step": 10458 }, { "epoch": 4.736865942028985, "grad_norm": 0.06199818274125426, "learning_rate": 6.730342563305003e-07, "loss": 0.8455, "step": 10459 }, { "epoch": 4.73731884057971, "grad_norm": 0.06463108178674065, "learning_rate": 6.707259279144085e-07, "loss": 0.8442, "step": 10460 }, { "epoch": 4.737771739130435, "grad_norm": 0.061650144164698426, "learning_rate": 6.684215313081632e-07, "loss": 0.8645, "step": 10461 }, { "epoch": 4.738224637681159, "grad_norm": 0.0710131577112456, "learning_rate": 6.661210667421491e-07, "loss": 0.8655, "step": 10462 }, { "epoch": 4.738677536231884, "grad_norm": 0.07041182832595405, "learning_rate": 6.638245344463335e-07, "loss": 0.8564, "step": 10463 }, { "epoch": 4.739130434782608, "grad_norm": 0.060290343796528784, "learning_rate": 6.61531934650319e-07, "loss": 0.8471, "step": 10464 }, { "epoch": 4.739583333333333, "grad_norm": 0.06412356486809818, "learning_rate": 6.592432675832916e-07, "loss": 0.8439, "step": 10465 }, { "epoch": 4.740036231884058, "grad_norm": 0.06634109814007856, "learning_rate": 6.569585334740546e-07, "loss": 0.8457, "step": 10466 }, { "epoch": 4.740489130434782, "grad_norm": 0.062967781264645, "learning_rate": 6.546777325510167e-07, "loss": 0.8484, "step": 10467 }, { "epoch": 4.740942028985507, "grad_norm": 0.05927574128455519, "learning_rate": 6.524008650421998e-07, "loss": 0.8691, "step": 10468 }, { "epoch": 4.741394927536232, "grad_norm": 0.06976641107213677, "learning_rate": 6.501279311752129e-07, "loss": 0.8582, "step": 10469 }, { "epoch": 4.741847826086957, "grad_norm": 0.06089432444483877, "learning_rate": 6.478589311772965e-07, "loss": 0.8456, "step": 10470 }, { "epoch": 4.742300724637682, "grad_norm": 0.06678705123787106, "learning_rate": 6.455938652752824e-07, "loss": 0.8407, "step": 10471 }, { "epoch": 4.742753623188406, "grad_norm": 0.061627969772110346, "learning_rate": 6.433327336956119e-07, "loss": 0.8646, "step": 10472 }, { "epoch": 4.743206521739131, "grad_norm": 0.07833000700074177, "learning_rate": 6.410755366643307e-07, "loss": 0.8548, "step": 10473 }, { "epoch": 4.743659420289855, "grad_norm": 0.06472154659923043, "learning_rate": 6.388222744070982e-07, "loss": 0.8605, "step": 10474 }, { "epoch": 4.74411231884058, "grad_norm": 0.06934647552707461, "learning_rate": 6.365729471491743e-07, "loss": 0.8527, "step": 10475 }, { "epoch": 4.744565217391305, "grad_norm": 0.05821886634522241, "learning_rate": 6.343275551154326e-07, "loss": 0.8525, "step": 10476 }, { "epoch": 4.745018115942029, "grad_norm": 0.06698675271414217, "learning_rate": 6.320860985303467e-07, "loss": 0.8371, "step": 10477 }, { "epoch": 4.745471014492754, "grad_norm": 0.06259260316972205, "learning_rate": 6.298485776179953e-07, "loss": 0.8557, "step": 10478 }, { "epoch": 4.7459239130434785, "grad_norm": 0.05861953890186562, "learning_rate": 6.276149926020703e-07, "loss": 0.8583, "step": 10479 }, { "epoch": 4.746376811594203, "grad_norm": 0.06379751491694141, "learning_rate": 6.253853437058644e-07, "loss": 0.8364, "step": 10480 }, { "epoch": 4.746829710144928, "grad_norm": 0.06318091654353462, "learning_rate": 6.231596311522792e-07, "loss": 0.8317, "step": 10481 }, { "epoch": 4.747282608695652, "grad_norm": 0.0725476238716438, "learning_rate": 6.209378551638256e-07, "loss": 0.8446, "step": 10482 }, { "epoch": 4.747735507246377, "grad_norm": 0.06799297592311175, "learning_rate": 6.18720015962615e-07, "loss": 0.8532, "step": 10483 }, { "epoch": 4.7481884057971016, "grad_norm": 0.05972455050472923, "learning_rate": 6.165061137703677e-07, "loss": 0.8459, "step": 10484 }, { "epoch": 4.748641304347826, "grad_norm": 0.06933672887204721, "learning_rate": 6.142961488084131e-07, "loss": 0.8499, "step": 10485 }, { "epoch": 4.749094202898551, "grad_norm": 0.06231096979207709, "learning_rate": 6.12090121297686e-07, "loss": 0.8575, "step": 10486 }, { "epoch": 4.749547101449275, "grad_norm": 0.06022586721905267, "learning_rate": 6.098880314587296e-07, "loss": 0.8487, "step": 10487 }, { "epoch": 4.75, "grad_norm": 0.07016091144745874, "learning_rate": 6.076898795116792e-07, "loss": 0.8331, "step": 10488 }, { "epoch": 4.750452898550725, "grad_norm": 0.06733753153906544, "learning_rate": 6.054956656762967e-07, "loss": 0.839, "step": 10489 }, { "epoch": 4.750905797101449, "grad_norm": 0.06970455480723273, "learning_rate": 6.033053901719355e-07, "loss": 0.8486, "step": 10490 }, { "epoch": 4.751358695652174, "grad_norm": 0.0685950727076548, "learning_rate": 6.01119053217567e-07, "loss": 0.8433, "step": 10491 }, { "epoch": 4.7518115942028984, "grad_norm": 0.06281596789279878, "learning_rate": 5.989366550317588e-07, "loss": 0.8699, "step": 10492 }, { "epoch": 4.752264492753623, "grad_norm": 0.06490847586774992, "learning_rate": 5.967581958326873e-07, "loss": 0.8626, "step": 10493 }, { "epoch": 4.752717391304348, "grad_norm": 0.07021272556214112, "learning_rate": 5.945836758381429e-07, "loss": 0.8596, "step": 10494 }, { "epoch": 4.753170289855072, "grad_norm": 0.06408280151471003, "learning_rate": 5.924130952655071e-07, "loss": 0.8534, "step": 10495 }, { "epoch": 4.753623188405797, "grad_norm": 0.06404721296142726, "learning_rate": 5.902464543317843e-07, "loss": 0.8572, "step": 10496 }, { "epoch": 4.7540760869565215, "grad_norm": 0.06935086595469152, "learning_rate": 5.8808375325357e-07, "loss": 0.8501, "step": 10497 }, { "epoch": 4.754528985507246, "grad_norm": 0.061183610905293964, "learning_rate": 5.859249922470823e-07, "loss": 0.847, "step": 10498 }, { "epoch": 4.754981884057971, "grad_norm": 0.061393583481559204, "learning_rate": 5.837701715281263e-07, "loss": 0.8529, "step": 10499 }, { "epoch": 4.755434782608695, "grad_norm": 0.06118023021876861, "learning_rate": 5.816192913121255e-07, "loss": 0.8325, "step": 10500 }, { "epoch": 4.75588768115942, "grad_norm": 0.061085369959791934, "learning_rate": 5.794723518141077e-07, "loss": 0.8446, "step": 10501 }, { "epoch": 4.756340579710145, "grad_norm": 0.062358883185242, "learning_rate": 5.773293532487101e-07, "loss": 0.8208, "step": 10502 }, { "epoch": 4.756793478260869, "grad_norm": 0.0648049212030908, "learning_rate": 5.751902958301658e-07, "loss": 0.838, "step": 10503 }, { "epoch": 4.757246376811594, "grad_norm": 0.06245472611027701, "learning_rate": 5.730551797723216e-07, "loss": 0.8698, "step": 10504 }, { "epoch": 4.757699275362318, "grad_norm": 0.0625286989497511, "learning_rate": 5.709240052886289e-07, "loss": 0.8464, "step": 10505 }, { "epoch": 4.758152173913043, "grad_norm": 0.057430548734116356, "learning_rate": 5.687967725921439e-07, "loss": 0.8257, "step": 10506 }, { "epoch": 4.758605072463768, "grad_norm": 0.05856624641697053, "learning_rate": 5.666734818955233e-07, "loss": 0.8368, "step": 10507 }, { "epoch": 4.759057971014493, "grad_norm": 0.06187868736185657, "learning_rate": 5.645541334110505e-07, "loss": 0.8564, "step": 10508 }, { "epoch": 4.759510869565218, "grad_norm": 0.06434730866414384, "learning_rate": 5.624387273505871e-07, "loss": 0.8491, "step": 10509 }, { "epoch": 4.759963768115942, "grad_norm": 0.062485123957204465, "learning_rate": 5.603272639256219e-07, "loss": 0.8669, "step": 10510 }, { "epoch": 4.760416666666667, "grad_norm": 0.0631083896471288, "learning_rate": 5.582197433472348e-07, "loss": 0.8436, "step": 10511 }, { "epoch": 4.760869565217392, "grad_norm": 0.05983533941662258, "learning_rate": 5.561161658261238e-07, "loss": 0.8505, "step": 10512 }, { "epoch": 4.761322463768116, "grad_norm": 0.06575416587632772, "learning_rate": 5.540165315725832e-07, "loss": 0.8465, "step": 10513 }, { "epoch": 4.761775362318841, "grad_norm": 0.06861218848707124, "learning_rate": 5.51920840796516e-07, "loss": 0.8605, "step": 10514 }, { "epoch": 4.762228260869565, "grad_norm": 0.06534617908653277, "learning_rate": 5.498290937074303e-07, "loss": 0.8721, "step": 10515 }, { "epoch": 4.76268115942029, "grad_norm": 0.05989639603262375, "learning_rate": 5.477412905144519e-07, "loss": 0.8627, "step": 10516 }, { "epoch": 4.763134057971015, "grad_norm": 0.06072800649890113, "learning_rate": 5.456574314262853e-07, "loss": 0.8649, "step": 10517 }, { "epoch": 4.763586956521739, "grad_norm": 0.07450412613355048, "learning_rate": 5.435775166512747e-07, "loss": 0.8429, "step": 10518 }, { "epoch": 4.764039855072464, "grad_norm": 0.07135998544399069, "learning_rate": 5.415015463973383e-07, "loss": 0.8345, "step": 10519 }, { "epoch": 4.7644927536231885, "grad_norm": 0.06832519657754885, "learning_rate": 5.394295208720213e-07, "loss": 0.8274, "step": 10520 }, { "epoch": 4.764945652173913, "grad_norm": 0.06570234604521556, "learning_rate": 5.373614402824645e-07, "loss": 0.8436, "step": 10521 }, { "epoch": 4.765398550724638, "grad_norm": 0.06143807978746117, "learning_rate": 5.352973048354226e-07, "loss": 0.8557, "step": 10522 }, { "epoch": 4.765851449275362, "grad_norm": 0.06611408090984396, "learning_rate": 5.332371147372417e-07, "loss": 0.8495, "step": 10523 }, { "epoch": 4.766304347826087, "grad_norm": 0.06803683079543985, "learning_rate": 5.311808701938859e-07, "loss": 0.8546, "step": 10524 }, { "epoch": 4.7667572463768115, "grad_norm": 0.06521284685755054, "learning_rate": 5.29128571410924e-07, "loss": 0.8656, "step": 10525 }, { "epoch": 4.767210144927536, "grad_norm": 0.06426526977795843, "learning_rate": 5.270802185935298e-07, "loss": 0.8505, "step": 10526 }, { "epoch": 4.767663043478261, "grad_norm": 0.06894523788477877, "learning_rate": 5.250358119464682e-07, "loss": 0.8448, "step": 10527 }, { "epoch": 4.768115942028985, "grad_norm": 0.06337812514571371, "learning_rate": 5.229953516741359e-07, "loss": 0.8188, "step": 10528 }, { "epoch": 4.76856884057971, "grad_norm": 0.057679036439403884, "learning_rate": 5.209588379805164e-07, "loss": 0.8417, "step": 10529 }, { "epoch": 4.769021739130435, "grad_norm": 0.07362247563089919, "learning_rate": 5.189262710691977e-07, "loss": 0.8323, "step": 10530 }, { "epoch": 4.769474637681159, "grad_norm": 0.07284321357409296, "learning_rate": 5.168976511433865e-07, "loss": 0.8553, "step": 10531 }, { "epoch": 4.769927536231884, "grad_norm": 0.06926174200116224, "learning_rate": 5.148729784058804e-07, "loss": 0.8549, "step": 10532 }, { "epoch": 4.770380434782608, "grad_norm": 0.06209275282948411, "learning_rate": 5.128522530590907e-07, "loss": 0.8549, "step": 10533 }, { "epoch": 4.770833333333333, "grad_norm": 0.06375407800788184, "learning_rate": 5.108354753050381e-07, "loss": 0.8397, "step": 10534 }, { "epoch": 4.771286231884058, "grad_norm": 0.05772066002256404, "learning_rate": 5.088226453453349e-07, "loss": 0.8425, "step": 10535 }, { "epoch": 4.771739130434782, "grad_norm": 0.06152698763723052, "learning_rate": 5.068137633812153e-07, "loss": 0.8579, "step": 10536 }, { "epoch": 4.772192028985507, "grad_norm": 0.06149682833340665, "learning_rate": 5.048088296135056e-07, "loss": 0.8484, "step": 10537 }, { "epoch": 4.772644927536232, "grad_norm": 0.06214310945556125, "learning_rate": 5.028078442426409e-07, "loss": 0.8552, "step": 10538 }, { "epoch": 4.773097826086957, "grad_norm": 0.06291804822554241, "learning_rate": 5.008108074686657e-07, "loss": 0.8513, "step": 10539 }, { "epoch": 4.773550724637682, "grad_norm": 0.05790186136310249, "learning_rate": 4.988177194912247e-07, "loss": 0.8329, "step": 10540 }, { "epoch": 4.774003623188406, "grad_norm": 0.057739940830498236, "learning_rate": 4.968285805095674e-07, "loss": 0.8454, "step": 10541 }, { "epoch": 4.774456521739131, "grad_norm": 0.060673745803839725, "learning_rate": 4.948433907225614e-07, "loss": 0.8546, "step": 10542 }, { "epoch": 4.774909420289855, "grad_norm": 0.05919931151653634, "learning_rate": 4.928621503286613e-07, "loss": 0.8481, "step": 10543 }, { "epoch": 4.77536231884058, "grad_norm": 0.05980986996029763, "learning_rate": 4.908848595259352e-07, "loss": 0.8603, "step": 10544 }, { "epoch": 4.775815217391305, "grad_norm": 0.05670189723612065, "learning_rate": 4.889115185120608e-07, "loss": 0.8496, "step": 10545 }, { "epoch": 4.776268115942029, "grad_norm": 0.06436786224357129, "learning_rate": 4.869421274843112e-07, "loss": 0.8656, "step": 10546 }, { "epoch": 4.776721014492754, "grad_norm": 0.05807036722922942, "learning_rate": 4.849766866395689e-07, "loss": 0.8421, "step": 10547 }, { "epoch": 4.7771739130434785, "grad_norm": 0.06237655838788795, "learning_rate": 4.830151961743257e-07, "loss": 0.8509, "step": 10548 }, { "epoch": 4.777626811594203, "grad_norm": 0.06307843215275658, "learning_rate": 4.810576562846737e-07, "loss": 0.8741, "step": 10549 }, { "epoch": 4.778079710144928, "grad_norm": 0.061654165329552986, "learning_rate": 4.79104067166305e-07, "loss": 0.8469, "step": 10550 }, { "epoch": 4.778532608695652, "grad_norm": 0.06049174707553811, "learning_rate": 4.771544290145347e-07, "loss": 0.8582, "step": 10551 }, { "epoch": 4.778985507246377, "grad_norm": 0.05801965180773041, "learning_rate": 4.752087420242646e-07, "loss": 0.8614, "step": 10552 }, { "epoch": 4.7794384057971016, "grad_norm": 0.06065665658528178, "learning_rate": 4.732670063900102e-07, "loss": 0.833, "step": 10553 }, { "epoch": 4.779891304347826, "grad_norm": 0.061543781511445385, "learning_rate": 4.71329222305883e-07, "loss": 0.8407, "step": 10554 }, { "epoch": 4.780344202898551, "grad_norm": 0.05913752490356741, "learning_rate": 4.693953899656167e-07, "loss": 0.8594, "step": 10555 }, { "epoch": 4.780797101449275, "grad_norm": 0.061464199125626104, "learning_rate": 4.6746550956253687e-07, "loss": 0.8452, "step": 10556 }, { "epoch": 4.78125, "grad_norm": 0.06140111416235055, "learning_rate": 4.6553958128957355e-07, "loss": 0.8602, "step": 10557 }, { "epoch": 4.781702898550725, "grad_norm": 0.06565779878545654, "learning_rate": 4.636176053392616e-07, "loss": 0.8671, "step": 10558 }, { "epoch": 4.782155797101449, "grad_norm": 0.06344328102025651, "learning_rate": 4.6169958190375395e-07, "loss": 0.8396, "step": 10559 }, { "epoch": 4.782608695652174, "grad_norm": 0.06864698986156645, "learning_rate": 4.5978551117479063e-07, "loss": 0.8402, "step": 10560 }, { "epoch": 4.7830615942028984, "grad_norm": 0.0656479250137962, "learning_rate": 4.578753933437252e-07, "loss": 0.8453, "step": 10561 }, { "epoch": 4.783514492753623, "grad_norm": 0.06769179648506046, "learning_rate": 4.5596922860151603e-07, "loss": 0.8467, "step": 10562 }, { "epoch": 4.783967391304348, "grad_norm": 0.06487883075662287, "learning_rate": 4.540670171387307e-07, "loss": 0.8642, "step": 10563 }, { "epoch": 4.784420289855072, "grad_norm": 0.06166950443985821, "learning_rate": 4.5216875914552815e-07, "loss": 0.8383, "step": 10564 }, { "epoch": 4.784873188405797, "grad_norm": 0.058050187531221605, "learning_rate": 4.502744548116855e-07, "loss": 0.84, "step": 10565 }, { "epoch": 4.7853260869565215, "grad_norm": 0.06988800686183831, "learning_rate": 4.4838410432657573e-07, "loss": 0.8307, "step": 10566 }, { "epoch": 4.785778985507246, "grad_norm": 0.0611391867056405, "learning_rate": 4.46497707879181e-07, "loss": 0.8517, "step": 10567 }, { "epoch": 4.786231884057971, "grad_norm": 0.0619939185075916, "learning_rate": 4.446152656580882e-07, "loss": 0.8257, "step": 10568 }, { "epoch": 4.786684782608695, "grad_norm": 0.0622679351672466, "learning_rate": 4.427367778514935e-07, "loss": 0.8333, "step": 10569 }, { "epoch": 4.78713768115942, "grad_norm": 0.0624991133247191, "learning_rate": 4.4086224464717995e-07, "loss": 0.8688, "step": 10570 }, { "epoch": 4.787590579710145, "grad_norm": 0.06197360884267676, "learning_rate": 4.389916662325577e-07, "loss": 0.8469, "step": 10571 }, { "epoch": 4.788043478260869, "grad_norm": 0.1053997327069018, "learning_rate": 4.3712504279463275e-07, "loss": 0.8529, "step": 10572 }, { "epoch": 4.788496376811594, "grad_norm": 0.06294706959689135, "learning_rate": 4.3526237452000684e-07, "loss": 0.8618, "step": 10573 }, { "epoch": 4.788949275362318, "grad_norm": 0.059781211280165704, "learning_rate": 4.3340366159489553e-07, "loss": 0.8645, "step": 10574 }, { "epoch": 4.789402173913043, "grad_norm": 0.060632012924725996, "learning_rate": 4.315489042051191e-07, "loss": 0.8586, "step": 10575 }, { "epoch": 4.789855072463768, "grad_norm": 0.06144719586505624, "learning_rate": 4.29698102536098e-07, "loss": 0.8532, "step": 10576 }, { "epoch": 4.790307971014493, "grad_norm": 0.06417250772659423, "learning_rate": 4.2785125677286654e-07, "loss": 0.8505, "step": 10577 }, { "epoch": 4.790760869565218, "grad_norm": 0.0639033585460222, "learning_rate": 4.260083671000459e-07, "loss": 0.8413, "step": 10578 }, { "epoch": 4.791213768115942, "grad_norm": 0.061679728435228905, "learning_rate": 4.241694337018887e-07, "loss": 0.8452, "step": 10579 }, { "epoch": 4.791666666666667, "grad_norm": 0.05695317448418829, "learning_rate": 4.223344567622212e-07, "loss": 0.8491, "step": 10580 }, { "epoch": 4.792119565217392, "grad_norm": 0.07841718820299137, "learning_rate": 4.205034364644966e-07, "loss": 0.8553, "step": 10581 }, { "epoch": 4.792572463768116, "grad_norm": 0.07528817143117486, "learning_rate": 4.1867637299175536e-07, "loss": 0.8514, "step": 10582 }, { "epoch": 4.793025362318841, "grad_norm": 0.06407534892747725, "learning_rate": 4.168532665266689e-07, "loss": 0.8556, "step": 10583 }, { "epoch": 4.793478260869565, "grad_norm": 0.060059296541390104, "learning_rate": 4.1503411725147824e-07, "loss": 0.8377, "step": 10584 }, { "epoch": 4.79393115942029, "grad_norm": 0.06292338444605453, "learning_rate": 4.132189253480601e-07, "loss": 0.8543, "step": 10585 }, { "epoch": 4.794384057971015, "grad_norm": 0.05729770699248009, "learning_rate": 4.114076909978737e-07, "loss": 0.8334, "step": 10586 }, { "epoch": 4.794836956521739, "grad_norm": 0.0682802580209639, "learning_rate": 4.096004143819965e-07, "loss": 0.8391, "step": 10587 }, { "epoch": 4.795289855072464, "grad_norm": 0.06245586690104328, "learning_rate": 4.0779709568109727e-07, "loss": 0.8486, "step": 10588 }, { "epoch": 4.7957427536231885, "grad_norm": 0.06589916723320546, "learning_rate": 4.0599773507546735e-07, "loss": 0.8579, "step": 10589 }, { "epoch": 4.796195652173913, "grad_norm": 0.06692789800916829, "learning_rate": 4.0420233274498064e-07, "loss": 0.8427, "step": 10590 }, { "epoch": 4.796648550724638, "grad_norm": 0.0599916195096832, "learning_rate": 4.02410888869138e-07, "loss": 0.8437, "step": 10591 }, { "epoch": 4.797101449275362, "grad_norm": 0.06491578391599939, "learning_rate": 4.006234036270229e-07, "loss": 0.8486, "step": 10592 }, { "epoch": 4.797554347826087, "grad_norm": 0.061025487490040276, "learning_rate": 3.9883987719733673e-07, "loss": 0.8412, "step": 10593 }, { "epoch": 4.7980072463768115, "grad_norm": 0.06626206397883662, "learning_rate": 3.9706030975837697e-07, "loss": 0.8631, "step": 10594 }, { "epoch": 4.798460144927536, "grad_norm": 0.060551051800383775, "learning_rate": 3.9528470148806344e-07, "loss": 0.837, "step": 10595 }, { "epoch": 4.798913043478261, "grad_norm": 0.06376368739758732, "learning_rate": 3.935130525638897e-07, "loss": 0.8376, "step": 10596 }, { "epoch": 4.799365942028985, "grad_norm": 0.06512609280435763, "learning_rate": 3.917453631629853e-07, "loss": 0.8503, "step": 10597 }, { "epoch": 4.79981884057971, "grad_norm": 0.06887335651474712, "learning_rate": 3.899816334620532e-07, "loss": 0.8446, "step": 10598 }, { "epoch": 4.800271739130435, "grad_norm": 0.05957136354468094, "learning_rate": 3.882218636374324e-07, "loss": 0.8379, "step": 10599 }, { "epoch": 4.800724637681159, "grad_norm": 0.058992940190410986, "learning_rate": 3.8646605386503554e-07, "loss": 0.8393, "step": 10600 }, { "epoch": 4.801177536231884, "grad_norm": 0.058458790853633355, "learning_rate": 3.8471420432040664e-07, "loss": 0.8501, "step": 10601 }, { "epoch": 4.801630434782608, "grad_norm": 0.061935226731423536, "learning_rate": 3.829663151786678e-07, "loss": 0.8596, "step": 10602 }, { "epoch": 4.802083333333333, "grad_norm": 0.05859717678807966, "learning_rate": 3.8122238661456814e-07, "loss": 0.8398, "step": 10603 }, { "epoch": 4.802536231884058, "grad_norm": 0.057543859756385624, "learning_rate": 3.794824188024482e-07, "loss": 0.8554, "step": 10604 }, { "epoch": 4.802989130434782, "grad_norm": 0.06409957780745452, "learning_rate": 3.7774641191625773e-07, "loss": 0.8586, "step": 10605 }, { "epoch": 4.803442028985507, "grad_norm": 0.061035207201336567, "learning_rate": 3.7601436612954236e-07, "loss": 0.8586, "step": 10606 }, { "epoch": 4.803894927536232, "grad_norm": 0.06154978181623301, "learning_rate": 3.7428628161546135e-07, "loss": 0.8517, "step": 10607 }, { "epoch": 4.804347826086957, "grad_norm": 0.06071166801586499, "learning_rate": 3.725621585467698e-07, "loss": 0.8487, "step": 10608 }, { "epoch": 4.804800724637682, "grad_norm": 0.06565874932722936, "learning_rate": 3.708419970958366e-07, "loss": 0.8452, "step": 10609 }, { "epoch": 4.805253623188406, "grad_norm": 0.06023190646415399, "learning_rate": 3.691257974346263e-07, "loss": 0.8466, "step": 10610 }, { "epoch": 4.805706521739131, "grad_norm": 0.05792433898046567, "learning_rate": 3.674135597347128e-07, "loss": 0.868, "step": 10611 }, { "epoch": 4.806159420289855, "grad_norm": 0.05890846364850198, "learning_rate": 3.6570528416726593e-07, "loss": 0.8666, "step": 10612 }, { "epoch": 4.80661231884058, "grad_norm": 0.06183610081566497, "learning_rate": 3.6400097090306894e-07, "loss": 0.8524, "step": 10613 }, { "epoch": 4.807065217391305, "grad_norm": 0.062216121348600446, "learning_rate": 3.6230062011250565e-07, "loss": 0.8409, "step": 10614 }, { "epoch": 4.807518115942029, "grad_norm": 0.06068031771214214, "learning_rate": 3.6060423196556005e-07, "loss": 0.8515, "step": 10615 }, { "epoch": 4.807971014492754, "grad_norm": 0.064107172312354, "learning_rate": 3.589118066318209e-07, "loss": 0.8558, "step": 10616 }, { "epoch": 4.8084239130434785, "grad_norm": 0.0561129591665299, "learning_rate": 3.572233442804862e-07, "loss": 0.8427, "step": 10617 }, { "epoch": 4.808876811594203, "grad_norm": 0.05904300234879874, "learning_rate": 3.555388450803499e-07, "loss": 0.8556, "step": 10618 }, { "epoch": 4.809329710144928, "grad_norm": 0.06291117052963391, "learning_rate": 3.538583091998238e-07, "loss": 0.8608, "step": 10619 }, { "epoch": 4.809782608695652, "grad_norm": 0.06113637678920443, "learning_rate": 3.521817368069025e-07, "loss": 0.8587, "step": 10620 }, { "epoch": 4.810235507246377, "grad_norm": 0.0579489674964958, "learning_rate": 3.50509128069203e-07, "loss": 0.8451, "step": 10621 }, { "epoch": 4.8106884057971016, "grad_norm": 0.058392418109650006, "learning_rate": 3.4884048315393384e-07, "loss": 0.8411, "step": 10622 }, { "epoch": 4.811141304347826, "grad_norm": 0.060304442231868115, "learning_rate": 3.47175802227917e-07, "loss": 0.8219, "step": 10623 }, { "epoch": 4.811594202898551, "grad_norm": 0.06689599737150269, "learning_rate": 3.455150854575662e-07, "loss": 0.8393, "step": 10624 }, { "epoch": 4.812047101449275, "grad_norm": 0.06310159350201423, "learning_rate": 3.4385833300891735e-07, "loss": 0.8542, "step": 10625 }, { "epoch": 4.8125, "grad_norm": 0.07798885382668709, "learning_rate": 3.4220554504758475e-07, "loss": 0.8565, "step": 10626 }, { "epoch": 4.812952898550725, "grad_norm": 0.05805575097765378, "learning_rate": 3.405567217388095e-07, "loss": 0.8352, "step": 10627 }, { "epoch": 4.813405797101449, "grad_norm": 0.061224881296842185, "learning_rate": 3.3891186324741974e-07, "loss": 0.8539, "step": 10628 }, { "epoch": 4.813858695652174, "grad_norm": 0.06463750970622582, "learning_rate": 3.372709697378662e-07, "loss": 0.8421, "step": 10629 }, { "epoch": 4.8143115942028984, "grad_norm": 0.0593155859655167, "learning_rate": 3.356340413741821e-07, "loss": 0.8509, "step": 10630 }, { "epoch": 4.814764492753623, "grad_norm": 0.062082344917160616, "learning_rate": 3.3400107832001873e-07, "loss": 0.8356, "step": 10631 }, { "epoch": 4.815217391304348, "grad_norm": 0.06632451510379578, "learning_rate": 3.3237208073862324e-07, "loss": 0.8545, "step": 10632 }, { "epoch": 4.815670289855072, "grad_norm": 0.06099383959429983, "learning_rate": 3.3074704879284767e-07, "loss": 0.8504, "step": 10633 }, { "epoch": 4.816123188405797, "grad_norm": 0.06285652619779188, "learning_rate": 3.291259826451487e-07, "loss": 0.837, "step": 10634 }, { "epoch": 4.8165760869565215, "grad_norm": 0.06516949931086988, "learning_rate": 3.2750888245759226e-07, "loss": 0.8588, "step": 10635 }, { "epoch": 4.817028985507246, "grad_norm": 0.060534619695734385, "learning_rate": 3.2589574839184015e-07, "loss": 0.8419, "step": 10636 }, { "epoch": 4.817481884057971, "grad_norm": 0.05868110308273312, "learning_rate": 3.2428658060915887e-07, "loss": 0.8311, "step": 10637 }, { "epoch": 4.817934782608695, "grad_norm": 0.057029009942631825, "learning_rate": 3.2268137927041977e-07, "loss": 0.8399, "step": 10638 }, { "epoch": 4.81838768115942, "grad_norm": 0.061923623076186735, "learning_rate": 3.2108014453609446e-07, "loss": 0.8544, "step": 10639 }, { "epoch": 4.818840579710145, "grad_norm": 0.06452809810355234, "learning_rate": 3.1948287656626387e-07, "loss": 0.8712, "step": 10640 }, { "epoch": 4.819293478260869, "grad_norm": 0.06116192712798751, "learning_rate": 3.178895755206135e-07, "loss": 0.8681, "step": 10641 }, { "epoch": 4.819746376811594, "grad_norm": 0.060504646272017196, "learning_rate": 3.163002415584204e-07, "loss": 0.8524, "step": 10642 }, { "epoch": 4.820199275362318, "grad_norm": 0.061346524194731834, "learning_rate": 3.147148748385753e-07, "loss": 0.8518, "step": 10643 }, { "epoch": 4.820652173913043, "grad_norm": 0.0598214383115195, "learning_rate": 3.131334755195736e-07, "loss": 0.8589, "step": 10644 }, { "epoch": 4.821105072463768, "grad_norm": 0.05988025166408644, "learning_rate": 3.1155604375950666e-07, "loss": 0.8604, "step": 10645 }, { "epoch": 4.821557971014493, "grad_norm": 0.062475538994177314, "learning_rate": 3.0998257971607494e-07, "loss": 0.872, "step": 10646 }, { "epoch": 4.822010869565218, "grad_norm": 0.06483252029333884, "learning_rate": 3.0841308354657486e-07, "loss": 0.8497, "step": 10647 }, { "epoch": 4.822463768115942, "grad_norm": 0.059422273848769934, "learning_rate": 3.068475554079209e-07, "loss": 0.8506, "step": 10648 }, { "epoch": 4.822916666666667, "grad_norm": 0.06012057757518074, "learning_rate": 3.0528599545661453e-07, "loss": 0.846, "step": 10649 }, { "epoch": 4.823369565217392, "grad_norm": 0.05739776216911038, "learning_rate": 3.0372840384876646e-07, "loss": 0.8325, "step": 10650 }, { "epoch": 4.823822463768116, "grad_norm": 0.06032915989208731, "learning_rate": 3.0217478074009656e-07, "loss": 0.8632, "step": 10651 }, { "epoch": 4.824275362318841, "grad_norm": 0.060227130792865305, "learning_rate": 3.0062512628592054e-07, "loss": 0.8697, "step": 10652 }, { "epoch": 4.824728260869565, "grad_norm": 0.05960037830812795, "learning_rate": 2.9907944064115903e-07, "loss": 0.8391, "step": 10653 }, { "epoch": 4.82518115942029, "grad_norm": 0.06901564714535997, "learning_rate": 2.9753772396033276e-07, "loss": 0.8498, "step": 10654 }, { "epoch": 4.825634057971015, "grad_norm": 0.057649101680123806, "learning_rate": 2.959999763975807e-07, "loss": 0.8443, "step": 10655 }, { "epoch": 4.826086956521739, "grad_norm": 0.05520806881776836, "learning_rate": 2.9446619810662434e-07, "loss": 0.8426, "step": 10656 }, { "epoch": 4.826539855072464, "grad_norm": 0.0618934922920496, "learning_rate": 2.9293638924080327e-07, "loss": 0.8574, "step": 10657 }, { "epoch": 4.8269927536231885, "grad_norm": 0.05864503360673752, "learning_rate": 2.9141054995305285e-07, "loss": 0.8445, "step": 10658 }, { "epoch": 4.827445652173913, "grad_norm": 0.06013371349927166, "learning_rate": 2.898886803959089e-07, "loss": 0.86, "step": 10659 }, { "epoch": 4.827898550724638, "grad_norm": 0.059763392462909855, "learning_rate": 2.8837078072152083e-07, "loss": 0.8665, "step": 10660 }, { "epoch": 4.828351449275362, "grad_norm": 0.05955888956642423, "learning_rate": 2.868568510816383e-07, "loss": 0.8412, "step": 10661 }, { "epoch": 4.828804347826087, "grad_norm": 0.060801425173928406, "learning_rate": 2.8534689162760255e-07, "loss": 0.852, "step": 10662 }, { "epoch": 4.8292572463768115, "grad_norm": 0.05646096996036014, "learning_rate": 2.8384090251037724e-07, "loss": 0.8376, "step": 10663 }, { "epoch": 4.829710144927536, "grad_norm": 0.06301488860948154, "learning_rate": 2.823388838805086e-07, "loss": 0.8442, "step": 10664 }, { "epoch": 4.830163043478261, "grad_norm": 0.05556188039310898, "learning_rate": 2.8084083588816093e-07, "loss": 0.8637, "step": 10665 }, { "epoch": 4.830615942028985, "grad_norm": 0.057566915075530245, "learning_rate": 2.7934675868309e-07, "loss": 0.8478, "step": 10666 }, { "epoch": 4.83106884057971, "grad_norm": 0.059764787212324276, "learning_rate": 2.778566524146742e-07, "loss": 0.8475, "step": 10667 }, { "epoch": 4.831521739130435, "grad_norm": 0.06239200655440578, "learning_rate": 2.763705172318698e-07, "loss": 0.843, "step": 10668 }, { "epoch": 4.831974637681159, "grad_norm": 0.06223929669051953, "learning_rate": 2.7488835328325137e-07, "loss": 0.8458, "step": 10669 }, { "epoch": 4.832427536231884, "grad_norm": 0.0655646797119681, "learning_rate": 2.734101607169981e-07, "loss": 0.8656, "step": 10670 }, { "epoch": 4.832880434782608, "grad_norm": 0.05961567561914831, "learning_rate": 2.7193593968088074e-07, "loss": 0.8685, "step": 10671 }, { "epoch": 4.833333333333333, "grad_norm": 0.06266375832630366, "learning_rate": 2.704656903222791e-07, "loss": 0.8492, "step": 10672 }, { "epoch": 4.833786231884058, "grad_norm": 0.05537704124428962, "learning_rate": 2.6899941278818673e-07, "loss": 0.8638, "step": 10673 }, { "epoch": 4.834239130434782, "grad_norm": 0.061753994406998206, "learning_rate": 2.6753710722517536e-07, "loss": 0.8366, "step": 10674 }, { "epoch": 4.834692028985507, "grad_norm": 0.062226000266017316, "learning_rate": 2.660787737794435e-07, "loss": 0.8338, "step": 10675 }, { "epoch": 4.835144927536232, "grad_norm": 0.06488277454129754, "learning_rate": 2.6462441259678115e-07, "loss": 0.869, "step": 10676 }, { "epoch": 4.835597826086957, "grad_norm": 0.062299525344428584, "learning_rate": 2.631740238225788e-07, "loss": 0.8453, "step": 10677 }, { "epoch": 4.836050724637682, "grad_norm": 0.06076228059390151, "learning_rate": 2.6172760760184047e-07, "loss": 0.8488, "step": 10678 }, { "epoch": 4.836503623188406, "grad_norm": 0.06341009373944295, "learning_rate": 2.602851640791659e-07, "loss": 0.8451, "step": 10679 }, { "epoch": 4.836956521739131, "grad_norm": 0.056551093824593224, "learning_rate": 2.5884669339875546e-07, "loss": 0.8666, "step": 10680 }, { "epoch": 4.837409420289855, "grad_norm": 0.05667255959408948, "learning_rate": 2.57412195704414e-07, "loss": 0.8409, "step": 10681 }, { "epoch": 4.83786231884058, "grad_norm": 0.06264345693070399, "learning_rate": 2.5598167113955574e-07, "loss": 0.8582, "step": 10682 }, { "epoch": 4.838315217391305, "grad_norm": 0.06614240591944769, "learning_rate": 2.5455511984718626e-07, "loss": 0.8731, "step": 10683 }, { "epoch": 4.838768115942029, "grad_norm": 0.061034018857529064, "learning_rate": 2.531325419699249e-07, "loss": 0.8423, "step": 10684 }, { "epoch": 4.839221014492754, "grad_norm": 0.05530930932983766, "learning_rate": 2.517139376499866e-07, "loss": 0.8483, "step": 10685 }, { "epoch": 4.8396739130434785, "grad_norm": 0.06543029309169993, "learning_rate": 2.5029930702919594e-07, "loss": 0.8555, "step": 10686 }, { "epoch": 4.840126811594203, "grad_norm": 0.06263777453819867, "learning_rate": 2.488886502489685e-07, "loss": 0.8519, "step": 10687 }, { "epoch": 4.840579710144928, "grad_norm": 0.06167740152522837, "learning_rate": 2.474819674503381e-07, "loss": 0.8566, "step": 10688 }, { "epoch": 4.841032608695652, "grad_norm": 0.06218835143947199, "learning_rate": 2.4607925877392135e-07, "loss": 0.8623, "step": 10689 }, { "epoch": 4.841485507246377, "grad_norm": 0.05962561005605198, "learning_rate": 2.4468052435996146e-07, "loss": 0.8443, "step": 10690 }, { "epoch": 4.8419384057971016, "grad_norm": 0.056254731775190615, "learning_rate": 2.4328576434828887e-07, "loss": 0.8637, "step": 10691 }, { "epoch": 4.842391304347826, "grad_norm": 0.06589972764932038, "learning_rate": 2.4189497887833866e-07, "loss": 0.8428, "step": 10692 }, { "epoch": 4.842844202898551, "grad_norm": 0.05958080162269494, "learning_rate": 2.4050816808914635e-07, "loss": 0.8425, "step": 10693 }, { "epoch": 4.843297101449275, "grad_norm": 0.10097419864585465, "learning_rate": 2.3912533211935653e-07, "loss": 0.8659, "step": 10694 }, { "epoch": 4.84375, "grad_norm": 0.057132289051155215, "learning_rate": 2.3774647110721415e-07, "loss": 0.8272, "step": 10695 }, { "epoch": 4.844202898550725, "grad_norm": 0.06080571242609175, "learning_rate": 2.3637158519056902e-07, "loss": 0.8368, "step": 10696 }, { "epoch": 4.844655797101449, "grad_norm": 0.058414631062296246, "learning_rate": 2.3500067450686225e-07, "loss": 0.8546, "step": 10697 }, { "epoch": 4.845108695652174, "grad_norm": 0.06166246346056158, "learning_rate": 2.3363373919315757e-07, "loss": 0.8461, "step": 10698 }, { "epoch": 4.8455615942028984, "grad_norm": 0.06249879905668596, "learning_rate": 2.3227077938609678e-07, "loss": 0.8541, "step": 10699 }, { "epoch": 4.846014492753623, "grad_norm": 0.0623365851955473, "learning_rate": 2.3091179522194862e-07, "loss": 0.8436, "step": 10700 }, { "epoch": 4.846467391304348, "grad_norm": 0.05856595717787584, "learning_rate": 2.2955678683656445e-07, "loss": 0.8358, "step": 10701 }, { "epoch": 4.846920289855072, "grad_norm": 0.05740350763097656, "learning_rate": 2.282057543654137e-07, "loss": 0.8531, "step": 10702 }, { "epoch": 4.847373188405797, "grad_norm": 0.06347713253239472, "learning_rate": 2.2685869794355274e-07, "loss": 0.865, "step": 10703 }, { "epoch": 4.8478260869565215, "grad_norm": 0.056166649103599524, "learning_rate": 2.2551561770565611e-07, "loss": 0.8493, "step": 10704 }, { "epoch": 4.848278985507246, "grad_norm": 0.056619342444330915, "learning_rate": 2.241765137859897e-07, "loss": 0.8665, "step": 10705 }, { "epoch": 4.848731884057971, "grad_norm": 0.06088252124186459, "learning_rate": 2.2284138631842867e-07, "loss": 0.8658, "step": 10706 }, { "epoch": 4.849184782608695, "grad_norm": 0.05842624244220646, "learning_rate": 2.2151023543644845e-07, "loss": 0.8455, "step": 10707 }, { "epoch": 4.84963768115942, "grad_norm": 0.057492501226954075, "learning_rate": 2.201830612731204e-07, "loss": 0.8616, "step": 10708 }, { "epoch": 4.850090579710145, "grad_norm": 0.058645818415676385, "learning_rate": 2.1885986396112947e-07, "loss": 0.8481, "step": 10709 }, { "epoch": 4.850543478260869, "grad_norm": 0.05832043633039579, "learning_rate": 2.1754064363276096e-07, "loss": 0.8475, "step": 10710 }, { "epoch": 4.850996376811594, "grad_norm": 0.06505698512806452, "learning_rate": 2.1622540041988716e-07, "loss": 0.8456, "step": 10711 }, { "epoch": 4.851449275362318, "grad_norm": 0.05987459163065896, "learning_rate": 2.1491413445401177e-07, "loss": 0.8604, "step": 10712 }, { "epoch": 4.851902173913043, "grad_norm": 0.06120519840298144, "learning_rate": 2.1360684586620772e-07, "loss": 0.8452, "step": 10713 }, { "epoch": 4.852355072463768, "grad_norm": 0.062624331173814, "learning_rate": 2.1230353478717935e-07, "loss": 0.8599, "step": 10714 }, { "epoch": 4.852807971014493, "grad_norm": 0.06098848118950813, "learning_rate": 2.1100420134721356e-07, "loss": 0.8279, "step": 10715 }, { "epoch": 4.853260869565218, "grad_norm": 0.05663414999902751, "learning_rate": 2.0970884567620642e-07, "loss": 0.8363, "step": 10716 }, { "epoch": 4.853713768115942, "grad_norm": 0.06076062699527474, "learning_rate": 2.0841746790366325e-07, "loss": 0.8546, "step": 10717 }, { "epoch": 4.854166666666667, "grad_norm": 0.060744434554713095, "learning_rate": 2.0713006815868075e-07, "loss": 0.8463, "step": 10718 }, { "epoch": 4.854619565217392, "grad_norm": 0.05682241990677342, "learning_rate": 2.0584664656995602e-07, "loss": 0.8427, "step": 10719 }, { "epoch": 4.855072463768116, "grad_norm": 0.05870109311246752, "learning_rate": 2.0456720326580858e-07, "loss": 0.8353, "step": 10720 }, { "epoch": 4.855525362318841, "grad_norm": 0.061998516579369015, "learning_rate": 2.032917383741362e-07, "loss": 0.8602, "step": 10721 }, { "epoch": 4.855978260869565, "grad_norm": 0.06073958801099608, "learning_rate": 2.0202025202245457e-07, "loss": 0.876, "step": 10722 }, { "epoch": 4.85643115942029, "grad_norm": 0.05533008702336329, "learning_rate": 2.0075274433787096e-07, "loss": 0.845, "step": 10723 }, { "epoch": 4.856884057971015, "grad_norm": 0.055628824165188544, "learning_rate": 1.9948921544710177e-07, "loss": 0.8393, "step": 10724 }, { "epoch": 4.857336956521739, "grad_norm": 0.0561102763024394, "learning_rate": 1.9822966547646815e-07, "loss": 0.8694, "step": 10725 }, { "epoch": 4.857789855072464, "grad_norm": 0.05644110184150861, "learning_rate": 1.9697409455188278e-07, "loss": 0.8251, "step": 10726 }, { "epoch": 4.8582427536231885, "grad_norm": 0.05764044973803662, "learning_rate": 1.9572250279887185e-07, "loss": 0.8474, "step": 10727 }, { "epoch": 4.858695652173913, "grad_norm": 0.06495294783123143, "learning_rate": 1.9447489034256196e-07, "loss": 0.851, "step": 10728 }, { "epoch": 4.859148550724638, "grad_norm": 0.062282387851113424, "learning_rate": 1.9323125730766666e-07, "loss": 0.8505, "step": 10729 }, { "epoch": 4.859601449275362, "grad_norm": 0.05976034909829673, "learning_rate": 1.9199160381852656e-07, "loss": 0.8461, "step": 10730 }, { "epoch": 4.860054347826087, "grad_norm": 0.05811849673898074, "learning_rate": 1.907559299990691e-07, "loss": 0.8218, "step": 10731 }, { "epoch": 4.8605072463768115, "grad_norm": 0.05940639648119267, "learning_rate": 1.895242359728222e-07, "loss": 0.846, "step": 10732 }, { "epoch": 4.860960144927536, "grad_norm": 0.06279774703114137, "learning_rate": 1.882965218629229e-07, "loss": 0.8401, "step": 10733 }, { "epoch": 4.861413043478261, "grad_norm": 0.06112493430196695, "learning_rate": 1.8707278779210415e-07, "loss": 0.8518, "step": 10734 }, { "epoch": 4.861865942028985, "grad_norm": 0.06706837168603463, "learning_rate": 1.8585303388271247e-07, "loss": 0.8383, "step": 10735 }, { "epoch": 4.86231884057971, "grad_norm": 0.05909921948854443, "learning_rate": 1.846372602566815e-07, "loss": 0.8562, "step": 10736 }, { "epoch": 4.862771739130435, "grad_norm": 0.05600630753619483, "learning_rate": 1.8342546703555396e-07, "loss": 0.8303, "step": 10737 }, { "epoch": 4.863224637681159, "grad_norm": 0.06421504683245308, "learning_rate": 1.822176543404819e-07, "loss": 0.8469, "step": 10738 }, { "epoch": 4.863677536231884, "grad_norm": 0.06332772167956387, "learning_rate": 1.810138222922042e-07, "loss": 0.8578, "step": 10739 }, { "epoch": 4.864130434782608, "grad_norm": 0.05843118583626075, "learning_rate": 1.7981397101107355e-07, "loss": 0.8383, "step": 10740 }, { "epoch": 4.864583333333333, "grad_norm": 0.05869132517022124, "learning_rate": 1.7861810061704287e-07, "loss": 0.8503, "step": 10741 }, { "epoch": 4.865036231884058, "grad_norm": 0.05870155143678865, "learning_rate": 1.774262112296654e-07, "loss": 0.8478, "step": 10742 }, { "epoch": 4.865489130434782, "grad_norm": 0.06137481730461772, "learning_rate": 1.7623830296808587e-07, "loss": 0.8602, "step": 10743 }, { "epoch": 4.865942028985507, "grad_norm": 0.0617756604444056, "learning_rate": 1.750543759510759e-07, "loss": 0.8482, "step": 10744 }, { "epoch": 4.866394927536232, "grad_norm": 0.06203217248902586, "learning_rate": 1.7387443029698525e-07, "loss": 0.8338, "step": 10745 }, { "epoch": 4.866847826086957, "grad_norm": 0.0581170660416843, "learning_rate": 1.7269846612378182e-07, "loss": 0.8357, "step": 10746 }, { "epoch": 4.867300724637682, "grad_norm": 0.057289468549134644, "learning_rate": 1.7152648354902047e-07, "loss": 0.8508, "step": 10747 }, { "epoch": 4.867753623188406, "grad_norm": 0.06091937117423543, "learning_rate": 1.7035848268986966e-07, "loss": 0.8474, "step": 10748 }, { "epoch": 4.868206521739131, "grad_norm": 0.05880290732374456, "learning_rate": 1.6919446366309822e-07, "loss": 0.8479, "step": 10749 }, { "epoch": 4.868659420289855, "grad_norm": 0.06735330079993482, "learning_rate": 1.680344265850753e-07, "loss": 0.8352, "step": 10750 }, { "epoch": 4.86911231884058, "grad_norm": 0.05927257722163186, "learning_rate": 1.668783715717659e-07, "loss": 0.8478, "step": 10751 }, { "epoch": 4.869565217391305, "grad_norm": 0.05927862539769891, "learning_rate": 1.6572629873874869e-07, "loss": 0.8254, "step": 10752 }, { "epoch": 4.870018115942029, "grad_norm": 0.05847805089510684, "learning_rate": 1.645782082011982e-07, "loss": 0.8151, "step": 10753 }, { "epoch": 4.870471014492754, "grad_norm": 0.05691274367596002, "learning_rate": 1.6343410007388483e-07, "loss": 0.847, "step": 10754 }, { "epoch": 4.8709239130434785, "grad_norm": 0.061152652422609505, "learning_rate": 1.6229397447119267e-07, "loss": 0.8618, "step": 10755 }, { "epoch": 4.871376811594203, "grad_norm": 0.06486149104984446, "learning_rate": 1.6115783150710162e-07, "loss": 0.8437, "step": 10756 }, { "epoch": 4.871829710144928, "grad_norm": 0.05929999667455538, "learning_rate": 1.6002567129519197e-07, "loss": 0.8512, "step": 10757 }, { "epoch": 4.872282608695652, "grad_norm": 0.05651955094282065, "learning_rate": 1.5889749394864872e-07, "loss": 0.8365, "step": 10758 }, { "epoch": 4.872735507246377, "grad_norm": 0.06087376413167484, "learning_rate": 1.577732995802528e-07, "loss": 0.8515, "step": 10759 }, { "epoch": 4.8731884057971016, "grad_norm": 0.061761861345117956, "learning_rate": 1.5665308830239867e-07, "loss": 0.8433, "step": 10760 }, { "epoch": 4.873641304347826, "grad_norm": 0.062316504218957916, "learning_rate": 1.5553686022707237e-07, "loss": 0.8363, "step": 10761 }, { "epoch": 4.874094202898551, "grad_norm": 0.06441711925464433, "learning_rate": 1.5442461546586461e-07, "loss": 0.8471, "step": 10762 }, { "epoch": 4.874547101449275, "grad_norm": 0.05941327846974608, "learning_rate": 1.533163541299665e-07, "loss": 0.8734, "step": 10763 }, { "epoch": 4.875, "grad_norm": 0.05978150043633088, "learning_rate": 1.522120763301782e-07, "loss": 0.8531, "step": 10764 }, { "epoch": 4.875452898550725, "grad_norm": 0.05624766232641056, "learning_rate": 1.511117821768915e-07, "loss": 0.856, "step": 10765 }, { "epoch": 4.875905797101449, "grad_norm": 0.05738054607169611, "learning_rate": 1.5001547178010723e-07, "loss": 0.8489, "step": 10766 }, { "epoch": 4.876358695652174, "grad_norm": 0.0624601763051747, "learning_rate": 1.489231452494222e-07, "loss": 0.8296, "step": 10767 }, { "epoch": 4.8768115942028984, "grad_norm": 0.06005515863979519, "learning_rate": 1.4783480269404238e-07, "loss": 0.8392, "step": 10768 }, { "epoch": 4.877264492753623, "grad_norm": 0.06027426649600592, "learning_rate": 1.4675044422276517e-07, "loss": 0.8406, "step": 10769 }, { "epoch": 4.877717391304348, "grad_norm": 0.06538966879910015, "learning_rate": 1.4567006994400167e-07, "loss": 0.8468, "step": 10770 }, { "epoch": 4.878170289855072, "grad_norm": 0.06641817173621373, "learning_rate": 1.4459367996575436e-07, "loss": 0.8476, "step": 10771 }, { "epoch": 4.878623188405797, "grad_norm": 0.0642838657648493, "learning_rate": 1.4352127439563045e-07, "loss": 0.8391, "step": 10772 }, { "epoch": 4.8790760869565215, "grad_norm": 0.06211571021257246, "learning_rate": 1.4245285334084646e-07, "loss": 0.8361, "step": 10773 }, { "epoch": 4.879528985507246, "grad_norm": 0.0594504494498744, "learning_rate": 1.4138841690821027e-07, "loss": 0.8407, "step": 10774 }, { "epoch": 4.879981884057971, "grad_norm": 0.06211889652020368, "learning_rate": 1.403279652041345e-07, "loss": 0.8452, "step": 10775 }, { "epoch": 4.880434782608695, "grad_norm": 0.059864159029128, "learning_rate": 1.392714983346366e-07, "loss": 0.8526, "step": 10776 }, { "epoch": 4.88088768115942, "grad_norm": 0.05901114950831972, "learning_rate": 1.3821901640532986e-07, "loss": 0.8494, "step": 10777 }, { "epoch": 4.881340579710145, "grad_norm": 0.054027811294131055, "learning_rate": 1.371705195214368e-07, "loss": 0.8439, "step": 10778 }, { "epoch": 4.881793478260869, "grad_norm": 0.05875446127570483, "learning_rate": 1.361260077877713e-07, "loss": 0.8265, "step": 10779 }, { "epoch": 4.882246376811594, "grad_norm": 0.05819511412552701, "learning_rate": 1.3508548130876097e-07, "loss": 0.8575, "step": 10780 }, { "epoch": 4.882699275362318, "grad_norm": 0.061664036622356616, "learning_rate": 1.3404894018842484e-07, "loss": 0.8448, "step": 10781 }, { "epoch": 4.883152173913043, "grad_norm": 0.057420653522607334, "learning_rate": 1.3301638453039112e-07, "loss": 0.8577, "step": 10782 }, { "epoch": 4.883605072463768, "grad_norm": 0.05471324082899356, "learning_rate": 1.319878144378839e-07, "loss": 0.8677, "step": 10783 }, { "epoch": 4.884057971014493, "grad_norm": 0.05927308945788297, "learning_rate": 1.30963230013732e-07, "loss": 0.8712, "step": 10784 }, { "epoch": 4.884510869565218, "grad_norm": 0.05792783383282973, "learning_rate": 1.2994263136036024e-07, "loss": 0.8405, "step": 10785 }, { "epoch": 4.884963768115942, "grad_norm": 0.06418981324494598, "learning_rate": 1.289260185798069e-07, "loss": 0.8352, "step": 10786 }, { "epoch": 4.885416666666667, "grad_norm": 0.06264417646112078, "learning_rate": 1.2791339177369745e-07, "loss": 0.8741, "step": 10787 }, { "epoch": 4.885869565217392, "grad_norm": 0.05778375306955302, "learning_rate": 1.2690475104327082e-07, "loss": 0.8501, "step": 10788 }, { "epoch": 4.886322463768116, "grad_norm": 0.0601274456144687, "learning_rate": 1.2590009648935752e-07, "loss": 0.84, "step": 10789 }, { "epoch": 4.886775362318841, "grad_norm": 0.06164916975910647, "learning_rate": 1.2489942821240164e-07, "loss": 0.8575, "step": 10790 }, { "epoch": 4.887228260869565, "grad_norm": 0.060021528932032595, "learning_rate": 1.2390274631243872e-07, "loss": 0.8442, "step": 10791 }, { "epoch": 4.88768115942029, "grad_norm": 0.05875405455236428, "learning_rate": 1.2291005088910457e-07, "loss": 0.8462, "step": 10792 }, { "epoch": 4.888134057971015, "grad_norm": 0.05918067849223825, "learning_rate": 1.219213420416443e-07, "loss": 0.8527, "step": 10793 }, { "epoch": 4.888586956521739, "grad_norm": 0.06125829554330302, "learning_rate": 1.2093661986889883e-07, "loss": 0.8638, "step": 10794 }, { "epoch": 4.889039855072464, "grad_norm": 0.06283713730690296, "learning_rate": 1.1995588446931382e-07, "loss": 0.8448, "step": 10795 }, { "epoch": 4.8894927536231885, "grad_norm": 0.05883206041709408, "learning_rate": 1.1897913594093091e-07, "loss": 0.8554, "step": 10796 }, { "epoch": 4.889945652173913, "grad_norm": 0.06080452287763924, "learning_rate": 1.1800637438140528e-07, "loss": 0.8621, "step": 10797 }, { "epoch": 4.890398550724638, "grad_norm": 0.0566142076219197, "learning_rate": 1.1703759988797914e-07, "loss": 0.8442, "step": 10798 }, { "epoch": 4.890851449275362, "grad_norm": 0.05764163170405229, "learning_rate": 1.1607281255750391e-07, "loss": 0.8345, "step": 10799 }, { "epoch": 4.891304347826087, "grad_norm": 0.058097306729302765, "learning_rate": 1.1511201248643134e-07, "loss": 0.8406, "step": 10800 }, { "epoch": 4.8917572463768115, "grad_norm": 0.05767320884340104, "learning_rate": 1.1415519977081347e-07, "loss": 0.8511, "step": 10801 }, { "epoch": 4.892210144927536, "grad_norm": 0.05995252036857827, "learning_rate": 1.132023745063071e-07, "loss": 0.8609, "step": 10802 }, { "epoch": 4.892663043478261, "grad_norm": 0.0626840304678074, "learning_rate": 1.1225353678816497e-07, "loss": 0.8587, "step": 10803 }, { "epoch": 4.893115942028985, "grad_norm": 0.06145209423688036, "learning_rate": 1.1130868671124006e-07, "loss": 0.8593, "step": 10804 }, { "epoch": 4.89356884057971, "grad_norm": 0.06224493941832574, "learning_rate": 1.1036782436999905e-07, "loss": 0.8507, "step": 10805 }, { "epoch": 4.894021739130435, "grad_norm": 0.05632186752226071, "learning_rate": 1.0943094985849556e-07, "loss": 0.8501, "step": 10806 }, { "epoch": 4.894474637681159, "grad_norm": 0.05938817258255206, "learning_rate": 1.0849806327039248e-07, "loss": 0.8563, "step": 10807 }, { "epoch": 4.894927536231884, "grad_norm": 0.059019359359475845, "learning_rate": 1.0756916469894851e-07, "loss": 0.8686, "step": 10808 }, { "epoch": 4.895380434782608, "grad_norm": 0.05807101648649475, "learning_rate": 1.0664425423703606e-07, "loss": 0.8511, "step": 10809 }, { "epoch": 4.895833333333333, "grad_norm": 0.057836699786648064, "learning_rate": 1.0572333197711005e-07, "loss": 0.848, "step": 10810 }, { "epoch": 4.896286231884058, "grad_norm": 0.06285326965612817, "learning_rate": 1.0480639801124348e-07, "loss": 0.8644, "step": 10811 }, { "epoch": 4.896739130434782, "grad_norm": 0.05705357277905981, "learning_rate": 1.0389345243109639e-07, "loss": 0.8624, "step": 10812 }, { "epoch": 4.897192028985507, "grad_norm": 0.05840751735308607, "learning_rate": 1.0298449532794241e-07, "loss": 0.8477, "step": 10813 }, { "epoch": 4.897644927536232, "grad_norm": 0.05405598939084173, "learning_rate": 1.020795267926511e-07, "loss": 0.8456, "step": 10814 }, { "epoch": 4.898097826086957, "grad_norm": 0.05831861468282176, "learning_rate": 1.0117854691569672e-07, "loss": 0.8615, "step": 10815 }, { "epoch": 4.898550724637682, "grad_norm": 0.06009427805189653, "learning_rate": 1.0028155578714504e-07, "loss": 0.8629, "step": 10816 }, { "epoch": 4.899003623188406, "grad_norm": 0.08385410984062451, "learning_rate": 9.938855349667542e-08, "loss": 0.8529, "step": 10817 }, { "epoch": 4.899456521739131, "grad_norm": 0.061781927143271215, "learning_rate": 9.849954013355423e-08, "loss": 0.8374, "step": 10818 }, { "epoch": 4.899909420289855, "grad_norm": 0.05426186873184908, "learning_rate": 9.761451578666592e-08, "loss": 0.8455, "step": 10819 }, { "epoch": 4.90036231884058, "grad_norm": 0.061797298063407155, "learning_rate": 9.673348054448639e-08, "loss": 0.8505, "step": 10820 }, { "epoch": 4.900815217391305, "grad_norm": 0.06300277138688823, "learning_rate": 9.58564344950963e-08, "loss": 0.8509, "step": 10821 }, { "epoch": 4.901268115942029, "grad_norm": 0.054998360442994075, "learning_rate": 9.498337772616773e-08, "loss": 0.8425, "step": 10822 }, { "epoch": 4.901721014492754, "grad_norm": 0.05877848360313202, "learning_rate": 9.411431032498642e-08, "loss": 0.8497, "step": 10823 }, { "epoch": 4.9021739130434785, "grad_norm": 0.06755409063047084, "learning_rate": 9.324923237842953e-08, "loss": 0.8389, "step": 10824 }, { "epoch": 4.902626811594203, "grad_norm": 0.06196245577350782, "learning_rate": 9.238814397299234e-08, "loss": 0.8561, "step": 10825 }, { "epoch": 4.903079710144928, "grad_norm": 0.062432207933879096, "learning_rate": 9.15310451947482e-08, "loss": 0.8547, "step": 10826 }, { "epoch": 4.903532608695652, "grad_norm": 0.0579326614711128, "learning_rate": 9.067793612938414e-08, "loss": 0.8628, "step": 10827 }, { "epoch": 4.903985507246377, "grad_norm": 0.06028773747794205, "learning_rate": 8.982881686219191e-08, "loss": 0.8544, "step": 10828 }, { "epoch": 4.9044384057971016, "grad_norm": 0.056312742784981766, "learning_rate": 8.898368747805474e-08, "loss": 0.8185, "step": 10829 }, { "epoch": 4.904891304347826, "grad_norm": 0.05896840449056783, "learning_rate": 8.81425480614606e-08, "loss": 0.871, "step": 10830 }, { "epoch": 4.905344202898551, "grad_norm": 0.05946854247349216, "learning_rate": 8.730539869650224e-08, "loss": 0.8591, "step": 10831 }, { "epoch": 4.905797101449275, "grad_norm": 0.05715551794020551, "learning_rate": 8.647223946687266e-08, "loss": 0.8387, "step": 10832 }, { "epoch": 4.90625, "grad_norm": 0.06470801567999307, "learning_rate": 8.564307045586085e-08, "loss": 0.8711, "step": 10833 }, { "epoch": 4.906702898550725, "grad_norm": 0.05903302251614535, "learning_rate": 8.481789174636046e-08, "loss": 0.8561, "step": 10834 }, { "epoch": 4.907155797101449, "grad_norm": 0.056386536764020265, "learning_rate": 8.399670342086553e-08, "loss": 0.8507, "step": 10835 }, { "epoch": 4.907608695652174, "grad_norm": 0.05872166661898688, "learning_rate": 8.31795055614748e-08, "loss": 0.846, "step": 10836 }, { "epoch": 4.9080615942028984, "grad_norm": 0.059666415218869014, "learning_rate": 8.236629824987851e-08, "loss": 0.8703, "step": 10837 }, { "epoch": 4.908514492753623, "grad_norm": 0.060395416789500535, "learning_rate": 8.155708156738052e-08, "loss": 0.8561, "step": 10838 }, { "epoch": 4.908967391304348, "grad_norm": 0.06390865751168567, "learning_rate": 8.075185559488053e-08, "loss": 0.85, "step": 10839 }, { "epoch": 4.909420289855072, "grad_norm": 0.056190037742813996, "learning_rate": 7.995062041286972e-08, "loss": 0.8391, "step": 10840 }, { "epoch": 4.909873188405797, "grad_norm": 0.059991957922332605, "learning_rate": 7.915337610145735e-08, "loss": 0.8451, "step": 10841 }, { "epoch": 4.9103260869565215, "grad_norm": 0.057628202238242734, "learning_rate": 7.83601227403441e-08, "loss": 0.8438, "step": 10842 }, { "epoch": 4.910778985507246, "grad_norm": 0.056242964852772706, "learning_rate": 7.757086040882655e-08, "loss": 0.8484, "step": 10843 }, { "epoch": 4.911231884057971, "grad_norm": 0.05621623310376714, "learning_rate": 7.678558918581492e-08, "loss": 0.8498, "step": 10844 }, { "epoch": 4.911684782608695, "grad_norm": 0.05903632916267992, "learning_rate": 7.600430914981083e-08, "loss": 0.8295, "step": 10845 }, { "epoch": 4.91213768115942, "grad_norm": 0.05358031130293572, "learning_rate": 7.522702037892071e-08, "loss": 0.82, "step": 10846 }, { "epoch": 4.912590579710145, "grad_norm": 0.0627970913031624, "learning_rate": 7.44537229508513e-08, "loss": 0.8481, "step": 10847 }, { "epoch": 4.913043478260869, "grad_norm": 0.055478961201739535, "learning_rate": 7.368441694290963e-08, "loss": 0.8362, "step": 10848 }, { "epoch": 4.913496376811594, "grad_norm": 0.05820186766817191, "learning_rate": 7.291910243200751e-08, "loss": 0.8357, "step": 10849 }, { "epoch": 4.913949275362318, "grad_norm": 0.05751841350035765, "learning_rate": 7.215777949465264e-08, "loss": 0.8414, "step": 10850 }, { "epoch": 4.914402173913043, "grad_norm": 0.05650327115971179, "learning_rate": 7.140044820695302e-08, "loss": 0.8335, "step": 10851 }, { "epoch": 4.914855072463768, "grad_norm": 0.06016347397581373, "learning_rate": 7.064710864462587e-08, "loss": 0.8655, "step": 10852 }, { "epoch": 4.915307971014493, "grad_norm": 0.0576373061061278, "learning_rate": 6.989776088297983e-08, "loss": 0.8483, "step": 10853 }, { "epoch": 4.915760869565218, "grad_norm": 0.060888578579698795, "learning_rate": 6.915240499692832e-08, "loss": 0.8526, "step": 10854 }, { "epoch": 4.916213768115942, "grad_norm": 0.06115989913170766, "learning_rate": 6.84110410609895e-08, "loss": 0.8357, "step": 10855 }, { "epoch": 4.916666666666667, "grad_norm": 0.058705844196509406, "learning_rate": 6.767366914927298e-08, "loss": 0.8486, "step": 10856 }, { "epoch": 4.917119565217392, "grad_norm": 0.06225176177618016, "learning_rate": 6.694028933549757e-08, "loss": 0.8502, "step": 10857 }, { "epoch": 4.917572463768116, "grad_norm": 0.05657554389611485, "learning_rate": 6.621090169298238e-08, "loss": 0.873, "step": 10858 }, { "epoch": 4.918025362318841, "grad_norm": 0.06964984267837127, "learning_rate": 6.548550629464245e-08, "loss": 0.8538, "step": 10859 }, { "epoch": 4.918478260869565, "grad_norm": 0.05586553959926002, "learning_rate": 6.476410321300197e-08, "loss": 0.8511, "step": 10860 }, { "epoch": 4.91893115942029, "grad_norm": 0.06065507417949602, "learning_rate": 6.404669252017214e-08, "loss": 0.8379, "step": 10861 }, { "epoch": 4.919384057971015, "grad_norm": 0.059377914234197485, "learning_rate": 6.333327428788227e-08, "loss": 0.8523, "step": 10862 }, { "epoch": 4.919836956521739, "grad_norm": 0.05748527729685333, "learning_rate": 6.262384858744863e-08, "loss": 0.8578, "step": 10863 }, { "epoch": 4.920289855072464, "grad_norm": 0.05941974755720278, "learning_rate": 6.191841548979227e-08, "loss": 0.8493, "step": 10864 }, { "epoch": 4.9207427536231885, "grad_norm": 0.05868168857277074, "learning_rate": 6.121697506544345e-08, "loss": 0.858, "step": 10865 }, { "epoch": 4.921195652173913, "grad_norm": 0.05817000164441224, "learning_rate": 6.051952738451939e-08, "loss": 0.837, "step": 10866 }, { "epoch": 4.921648550724638, "grad_norm": 0.05881001073420884, "learning_rate": 5.982607251675098e-08, "loss": 0.8626, "step": 10867 }, { "epoch": 4.922101449275362, "grad_norm": 0.05613074317558411, "learning_rate": 5.913661053145614e-08, "loss": 0.8455, "step": 10868 }, { "epoch": 4.922554347826087, "grad_norm": 0.0603887824005639, "learning_rate": 5.8451141497570804e-08, "loss": 0.8446, "step": 10869 }, { "epoch": 4.9230072463768115, "grad_norm": 0.05717159926423047, "learning_rate": 5.776966548361351e-08, "loss": 0.8661, "step": 10870 }, { "epoch": 4.923460144927536, "grad_norm": 0.05627585244573711, "learning_rate": 5.709218255771643e-08, "loss": 0.8453, "step": 10871 }, { "epoch": 4.923913043478261, "grad_norm": 0.05773340506103724, "learning_rate": 5.641869278761203e-08, "loss": 0.8586, "step": 10872 }, { "epoch": 4.924365942028985, "grad_norm": 0.061154180664565586, "learning_rate": 5.574919624062425e-08, "loss": 0.8326, "step": 10873 }, { "epoch": 4.92481884057971, "grad_norm": 0.05575501586836949, "learning_rate": 5.508369298369065e-08, "loss": 0.8215, "step": 10874 }, { "epoch": 4.925271739130435, "grad_norm": 0.05827875152140528, "learning_rate": 5.44221830833358e-08, "loss": 0.8402, "step": 10875 }, { "epoch": 4.925724637681159, "grad_norm": 0.057493743976470064, "learning_rate": 5.376466660569346e-08, "loss": 0.8544, "step": 10876 }, { "epoch": 4.926177536231884, "grad_norm": 0.05877661153321956, "learning_rate": 5.311114361649772e-08, "loss": 0.8486, "step": 10877 }, { "epoch": 4.926630434782608, "grad_norm": 0.0576592063286108, "learning_rate": 5.246161418108298e-08, "loss": 0.8459, "step": 10878 }, { "epoch": 4.927083333333333, "grad_norm": 0.056353048873034675, "learning_rate": 5.1816078364383956e-08, "loss": 0.8637, "step": 10879 }, { "epoch": 4.927536231884058, "grad_norm": 0.05642277385967805, "learning_rate": 5.117453623093127e-08, "loss": 0.8599, "step": 10880 }, { "epoch": 4.927989130434782, "grad_norm": 0.057383914068562566, "learning_rate": 5.053698784486916e-08, "loss": 0.8476, "step": 10881 }, { "epoch": 4.928442028985507, "grad_norm": 0.05732765352213133, "learning_rate": 4.9903433269928856e-08, "loss": 0.8644, "step": 10882 }, { "epoch": 4.928894927536232, "grad_norm": 0.05759178040265987, "learning_rate": 4.927387256944638e-08, "loss": 0.8536, "step": 10883 }, { "epoch": 4.929347826086957, "grad_norm": 0.05991709679158084, "learning_rate": 4.864830580636248e-08, "loss": 0.8473, "step": 10884 }, { "epoch": 4.929800724637682, "grad_norm": 0.05352012151143873, "learning_rate": 4.8026733043218256e-08, "loss": 0.8738, "step": 10885 }, { "epoch": 4.930253623188406, "grad_norm": 0.05558143579483712, "learning_rate": 4.7409154342146215e-08, "loss": 0.8612, "step": 10886 }, { "epoch": 4.930706521739131, "grad_norm": 0.054611421603241984, "learning_rate": 4.6795569764896966e-08, "loss": 0.85, "step": 10887 }, { "epoch": 4.931159420289855, "grad_norm": 0.058362781490170414, "learning_rate": 4.618597937280367e-08, "loss": 0.8588, "step": 10888 }, { "epoch": 4.93161231884058, "grad_norm": 0.05576807342407621, "learning_rate": 4.558038322680869e-08, "loss": 0.8682, "step": 10889 }, { "epoch": 4.932065217391305, "grad_norm": 0.05709938393417205, "learning_rate": 4.497878138745915e-08, "loss": 0.8385, "step": 10890 }, { "epoch": 4.932518115942029, "grad_norm": 0.055642419943933735, "learning_rate": 4.4381173914893605e-08, "loss": 0.8363, "step": 10891 }, { "epoch": 4.932971014492754, "grad_norm": 0.06258500701075749, "learning_rate": 4.378756086885538e-08, "loss": 0.8609, "step": 10892 }, { "epoch": 4.9334239130434785, "grad_norm": 0.054645432516463886, "learning_rate": 4.3197942308692564e-08, "loss": 0.8567, "step": 10893 }, { "epoch": 4.933876811594203, "grad_norm": 0.057677285697378664, "learning_rate": 4.261231829334467e-08, "loss": 0.8686, "step": 10894 }, { "epoch": 4.934329710144928, "grad_norm": 0.05904146181168144, "learning_rate": 4.2030688881364854e-08, "loss": 0.842, "step": 10895 }, { "epoch": 4.934782608695652, "grad_norm": 0.056993511494958314, "learning_rate": 4.145305413089773e-08, "loss": 0.8136, "step": 10896 }, { "epoch": 4.935235507246377, "grad_norm": 0.05570628926196265, "learning_rate": 4.0879414099688205e-08, "loss": 0.846, "step": 10897 }, { "epoch": 4.9356884057971016, "grad_norm": 0.05803724205685816, "learning_rate": 4.030976884508153e-08, "loss": 0.8373, "step": 10898 }, { "epoch": 4.936141304347826, "grad_norm": 0.05597703913376401, "learning_rate": 3.97441184240277e-08, "loss": 0.8405, "step": 10899 }, { "epoch": 4.936594202898551, "grad_norm": 0.05722351116364244, "learning_rate": 3.918246289308147e-08, "loss": 0.8483, "step": 10900 }, { "epoch": 4.937047101449275, "grad_norm": 0.05536720298503385, "learning_rate": 3.862480230838461e-08, "loss": 0.8393, "step": 10901 }, { "epoch": 4.9375, "grad_norm": 0.0556001094547862, "learning_rate": 3.8071136725688074e-08, "loss": 0.8623, "step": 10902 }, { "epoch": 4.937952898550725, "grad_norm": 0.06221518431714879, "learning_rate": 3.752146620035202e-08, "loss": 0.8664, "step": 10903 }, { "epoch": 4.938405797101449, "grad_norm": 0.05903046926447769, "learning_rate": 3.697579078731473e-08, "loss": 0.8505, "step": 10904 }, { "epoch": 4.938858695652174, "grad_norm": 0.05918097560115979, "learning_rate": 3.643411054113699e-08, "loss": 0.8416, "step": 10905 }, { "epoch": 4.9393115942028984, "grad_norm": 0.059157575947062425, "learning_rate": 3.589642551596662e-08, "loss": 0.8456, "step": 10906 }, { "epoch": 4.939764492753623, "grad_norm": 0.05609940611305396, "learning_rate": 3.5362735765565036e-08, "loss": 0.8573, "step": 10907 }, { "epoch": 4.940217391304348, "grad_norm": 0.05688278017513296, "learning_rate": 3.483304134327181e-08, "loss": 0.8547, "step": 10908 }, { "epoch": 4.940670289855072, "grad_norm": 0.0575717163847831, "learning_rate": 3.43073423020579e-08, "loss": 0.88, "step": 10909 }, { "epoch": 4.941123188405797, "grad_norm": 0.06354327370968384, "learning_rate": 3.37856386944635e-08, "loss": 0.8685, "step": 10910 }, { "epoch": 4.9415760869565215, "grad_norm": 0.06496229802799669, "learning_rate": 3.326793057265576e-08, "loss": 0.8589, "step": 10911 }, { "epoch": 4.942028985507246, "grad_norm": 0.05387027088351333, "learning_rate": 3.2754217988384406e-08, "loss": 0.8635, "step": 10912 }, { "epoch": 4.942481884057971, "grad_norm": 0.055114036747268086, "learning_rate": 3.2244500993003916e-08, "loss": 0.851, "step": 10913 }, { "epoch": 4.942934782608695, "grad_norm": 0.06220596431028797, "learning_rate": 3.173877963747796e-08, "loss": 0.8404, "step": 10914 }, { "epoch": 4.94338768115942, "grad_norm": 0.05965773911762804, "learning_rate": 3.1237053972361653e-08, "loss": 0.8548, "step": 10915 }, { "epoch": 4.943840579710145, "grad_norm": 0.060355445052841135, "learning_rate": 3.0739324047810436e-08, "loss": 0.8419, "step": 10916 }, { "epoch": 4.944293478260869, "grad_norm": 0.06175102571699631, "learning_rate": 3.024558991358895e-08, "loss": 0.865, "step": 10917 }, { "epoch": 4.944746376811594, "grad_norm": 0.059827830656293414, "learning_rate": 2.9755851619053256e-08, "loss": 0.8586, "step": 10918 }, { "epoch": 4.945199275362318, "grad_norm": 0.06162595538412127, "learning_rate": 2.9270109213159757e-08, "loss": 0.8458, "step": 10919 }, { "epoch": 4.945652173913043, "grad_norm": 0.05628944663875044, "learning_rate": 2.8788362744478492e-08, "loss": 0.8374, "step": 10920 }, { "epoch": 4.946105072463768, "grad_norm": 0.0571563770720865, "learning_rate": 2.8310612261157612e-08, "loss": 0.8401, "step": 10921 }, { "epoch": 4.946557971014493, "grad_norm": 0.055985777011719855, "learning_rate": 2.7836857810972227e-08, "loss": 0.8435, "step": 10922 }, { "epoch": 4.947010869565218, "grad_norm": 0.056761967809840594, "learning_rate": 2.736709944127114e-08, "loss": 0.8583, "step": 10923 }, { "epoch": 4.947463768115942, "grad_norm": 0.058868681954311546, "learning_rate": 2.6901337199025656e-08, "loss": 0.8447, "step": 10924 }, { "epoch": 4.947916666666667, "grad_norm": 0.058121269083933176, "learning_rate": 2.6439571130798536e-08, "loss": 0.8501, "step": 10925 }, { "epoch": 4.948369565217392, "grad_norm": 0.059961696587013925, "learning_rate": 2.598180128274841e-08, "loss": 0.8303, "step": 10926 }, { "epoch": 4.948822463768116, "grad_norm": 0.06162718020089699, "learning_rate": 2.552802770064311e-08, "loss": 0.8363, "step": 10927 }, { "epoch": 4.949275362318841, "grad_norm": 0.055970317073322645, "learning_rate": 2.5078250429846352e-08, "loss": 0.8297, "step": 10928 }, { "epoch": 4.949728260869565, "grad_norm": 0.05729808799863792, "learning_rate": 2.4632469515317725e-08, "loss": 0.8616, "step": 10929 }, { "epoch": 4.95018115942029, "grad_norm": 0.057606641280472874, "learning_rate": 2.4190685001634906e-08, "loss": 0.8447, "step": 10930 }, { "epoch": 4.950634057971015, "grad_norm": 0.05786755194771917, "learning_rate": 2.3752896932949244e-08, "loss": 0.851, "step": 10931 }, { "epoch": 4.951086956521739, "grad_norm": 0.056953299792932385, "learning_rate": 2.331910535303905e-08, "loss": 0.84, "step": 10932 }, { "epoch": 4.951539855072464, "grad_norm": 0.055538731820496685, "learning_rate": 2.28893103052652e-08, "loss": 0.8348, "step": 10933 }, { "epoch": 4.9519927536231885, "grad_norm": 0.05833943209915267, "learning_rate": 2.246351183259776e-08, "loss": 0.8396, "step": 10934 }, { "epoch": 4.952445652173913, "grad_norm": 0.056361134975285665, "learning_rate": 2.2041709977598248e-08, "loss": 0.8488, "step": 10935 }, { "epoch": 4.952898550724638, "grad_norm": 0.06033003126690857, "learning_rate": 2.1623904782446246e-08, "loss": 0.852, "step": 10936 }, { "epoch": 4.953351449275362, "grad_norm": 0.05728591881107417, "learning_rate": 2.1210096288899473e-08, "loss": 0.8259, "step": 10937 }, { "epoch": 4.953804347826087, "grad_norm": 0.06076304848931477, "learning_rate": 2.0800284538329274e-08, "loss": 0.857, "step": 10938 }, { "epoch": 4.9542572463768115, "grad_norm": 0.061266762478723376, "learning_rate": 2.039446957170732e-08, "loss": 0.8671, "step": 10939 }, { "epoch": 4.954710144927536, "grad_norm": 0.05592591292824048, "learning_rate": 1.9992651429610042e-08, "loss": 0.8714, "step": 10940 }, { "epoch": 4.955163043478261, "grad_norm": 0.05760541689862919, "learning_rate": 1.959483015219199e-08, "loss": 0.8272, "step": 10941 }, { "epoch": 4.955615942028985, "grad_norm": 0.05946109249197253, "learning_rate": 1.9201005779239113e-08, "loss": 0.876, "step": 10942 }, { "epoch": 4.95606884057971, "grad_norm": 0.055184823183343105, "learning_rate": 1.8811178350115476e-08, "loss": 0.8452, "step": 10943 }, { "epoch": 4.956521739130435, "grad_norm": 0.059237200925342764, "learning_rate": 1.8425347903794355e-08, "loss": 0.8378, "step": 10944 }, { "epoch": 4.956974637681159, "grad_norm": 0.06181313836244047, "learning_rate": 1.8043514478849334e-08, "loss": 0.86, "step": 10945 }, { "epoch": 4.957427536231884, "grad_norm": 0.05387720851144327, "learning_rate": 1.766567811344544e-08, "loss": 0.8311, "step": 10946 }, { "epoch": 4.957880434782608, "grad_norm": 0.059078498461259024, "learning_rate": 1.7291838845365783e-08, "loss": 0.8361, "step": 10947 }, { "epoch": 4.958333333333333, "grad_norm": 0.05481884082360749, "learning_rate": 1.6921996711976028e-08, "loss": 0.8668, "step": 10948 }, { "epoch": 4.958786231884058, "grad_norm": 0.05720793827392879, "learning_rate": 1.655615175025549e-08, "loss": 0.8381, "step": 10949 }, { "epoch": 4.959239130434782, "grad_norm": 0.05866478816926048, "learning_rate": 1.6194303996774906e-08, "loss": 0.8539, "step": 10950 }, { "epoch": 4.959692028985507, "grad_norm": 0.06226124346727618, "learning_rate": 1.583645348770979e-08, "loss": 0.8595, "step": 10951 }, { "epoch": 4.960144927536232, "grad_norm": 0.06090493459878218, "learning_rate": 1.5482600258831526e-08, "loss": 0.8438, "step": 10952 }, { "epoch": 4.960597826086957, "grad_norm": 0.05911885270130727, "learning_rate": 1.5132744345525142e-08, "loss": 0.8428, "step": 10953 }, { "epoch": 4.961050724637682, "grad_norm": 0.06687405351318526, "learning_rate": 1.478688578275378e-08, "loss": 0.8344, "step": 10954 }, { "epoch": 4.961503623188406, "grad_norm": 0.054286495938987885, "learning_rate": 1.4445024605103108e-08, "loss": 0.8613, "step": 10955 }, { "epoch": 4.961956521739131, "grad_norm": 0.054532130202252074, "learning_rate": 1.4107160846741353e-08, "loss": 0.8461, "step": 10956 }, { "epoch": 4.962409420289855, "grad_norm": 0.05389698540261016, "learning_rate": 1.3773294541454818e-08, "loss": 0.8364, "step": 10957 }, { "epoch": 4.96286231884058, "grad_norm": 0.054308192284451086, "learning_rate": 1.3443425722612368e-08, "loss": 0.8363, "step": 10958 }, { "epoch": 4.963315217391305, "grad_norm": 0.0552325387468835, "learning_rate": 1.311755442319651e-08, "loss": 0.846, "step": 10959 }, { "epoch": 4.963768115942029, "grad_norm": 0.058178903610793184, "learning_rate": 1.2795680675781186e-08, "loss": 0.8514, "step": 10960 }, { "epoch": 4.964221014492754, "grad_norm": 0.054439812681498355, "learning_rate": 1.247780451254954e-08, "loss": 0.8265, "step": 10961 }, { "epoch": 4.9646739130434785, "grad_norm": 0.05673678840535832, "learning_rate": 1.2163925965276157e-08, "loss": 0.8445, "step": 10962 }, { "epoch": 4.965126811594203, "grad_norm": 0.060169232000873665, "learning_rate": 1.185404506534038e-08, "loss": 0.8548, "step": 10963 }, { "epoch": 4.965579710144928, "grad_norm": 0.06424049136632287, "learning_rate": 1.1548161843726313e-08, "loss": 0.8425, "step": 10964 }, { "epoch": 4.966032608695652, "grad_norm": 0.05868290986540072, "learning_rate": 1.1246276331005058e-08, "loss": 0.863, "step": 10965 }, { "epoch": 4.966485507246377, "grad_norm": 0.0598280461221794, "learning_rate": 1.0948388557365797e-08, "loss": 0.8633, "step": 10966 }, { "epoch": 4.9669384057971016, "grad_norm": 0.05594208231875176, "learning_rate": 1.0654498552580272e-08, "loss": 0.8263, "step": 10967 }, { "epoch": 4.967391304347826, "grad_norm": 0.056382445835436217, "learning_rate": 1.0364606346033867e-08, "loss": 0.8539, "step": 10968 }, { "epoch": 4.967844202898551, "grad_norm": 0.05637777051972172, "learning_rate": 1.0078711966707844e-08, "loss": 0.8596, "step": 10969 }, { "epoch": 4.968297101449275, "grad_norm": 0.0557194373292664, "learning_rate": 9.796815443179342e-09, "loss": 0.8523, "step": 10970 }, { "epoch": 4.96875, "grad_norm": 0.05708430118136589, "learning_rate": 9.518916803634703e-09, "loss": 0.8366, "step": 10971 }, { "epoch": 4.969202898550725, "grad_norm": 0.060652766979461356, "learning_rate": 9.245016075856151e-09, "loss": 0.8617, "step": 10972 }, { "epoch": 4.969655797101449, "grad_norm": 0.05658378210260057, "learning_rate": 8.975113287221782e-09, "loss": 0.8517, "step": 10973 }, { "epoch": 4.970108695652174, "grad_norm": 0.055842187990463414, "learning_rate": 8.7092084647189e-09, "loss": 0.8627, "step": 10974 }, { "epoch": 4.9705615942028984, "grad_norm": 0.05664645355321002, "learning_rate": 8.447301634926242e-09, "loss": 0.8538, "step": 10975 }, { "epoch": 4.971014492753623, "grad_norm": 0.05482333597744627, "learning_rate": 8.18939282403175e-09, "loss": 0.8244, "step": 10976 }, { "epoch": 4.971467391304348, "grad_norm": 0.05710989876931891, "learning_rate": 7.935482057814802e-09, "loss": 0.8408, "step": 10977 }, { "epoch": 4.971920289855072, "grad_norm": 0.056494584595173016, "learning_rate": 7.68556936165954e-09, "loss": 0.8371, "step": 10978 }, { "epoch": 4.972373188405797, "grad_norm": 0.05932353989499919, "learning_rate": 7.439654760550419e-09, "loss": 0.8389, "step": 10979 }, { "epoch": 4.9728260869565215, "grad_norm": 0.06157322566778094, "learning_rate": 7.197738279072219e-09, "loss": 0.8418, "step": 10980 }, { "epoch": 4.973278985507246, "grad_norm": 0.058506985845246356, "learning_rate": 6.95981994141448e-09, "loss": 0.8349, "step": 10981 }, { "epoch": 4.973731884057971, "grad_norm": 0.059988164712350855, "learning_rate": 6.725899771353739e-09, "loss": 0.8589, "step": 10982 }, { "epoch": 4.974184782608695, "grad_norm": 0.06015015355743011, "learning_rate": 6.495977792280173e-09, "loss": 0.8342, "step": 10983 }, { "epoch": 4.97463768115942, "grad_norm": 0.056674376445893467, "learning_rate": 6.27005402717984e-09, "loss": 0.8344, "step": 10984 }, { "epoch": 4.975090579710145, "grad_norm": 0.055137667527352496, "learning_rate": 6.048128498639116e-09, "loss": 0.8454, "step": 10985 }, { "epoch": 4.975543478260869, "grad_norm": 0.05720980087182989, "learning_rate": 5.8302012288402596e-09, "loss": 0.8377, "step": 10986 }, { "epoch": 4.975996376811594, "grad_norm": 0.05852872155495406, "learning_rate": 5.616272239570286e-09, "loss": 0.8572, "step": 10987 }, { "epoch": 4.976449275362318, "grad_norm": 0.05764915572610578, "learning_rate": 5.406341552225414e-09, "loss": 0.8463, "step": 10988 }, { "epoch": 4.976902173913043, "grad_norm": 0.056703935313798526, "learning_rate": 5.200409187779976e-09, "loss": 0.8494, "step": 10989 }, { "epoch": 4.977355072463768, "grad_norm": 0.061445502418713376, "learning_rate": 4.9984751668263885e-09, "loss": 0.8389, "step": 10990 }, { "epoch": 4.977807971014493, "grad_norm": 0.05560313847120384, "learning_rate": 4.800539509552948e-09, "loss": 0.8483, "step": 10991 }, { "epoch": 4.978260869565218, "grad_norm": 0.05498957442628572, "learning_rate": 4.606602235752711e-09, "loss": 0.8557, "step": 10992 }, { "epoch": 4.978713768115942, "grad_norm": 0.05984598931717806, "learning_rate": 4.416663364801288e-09, "loss": 0.8568, "step": 10993 }, { "epoch": 4.979166666666667, "grad_norm": 0.06023097082161434, "learning_rate": 4.230722915701257e-09, "loss": 0.8639, "step": 10994 }, { "epoch": 4.979619565217392, "grad_norm": 0.05929317490965049, "learning_rate": 4.0487809070288705e-09, "loss": 0.8467, "step": 10995 }, { "epoch": 4.980072463768116, "grad_norm": 0.05780067212805184, "learning_rate": 3.870837356982904e-09, "loss": 0.8207, "step": 10996 }, { "epoch": 4.980525362318841, "grad_norm": 0.05817288074775307, "learning_rate": 3.6968922833446885e-09, "loss": 0.8386, "step": 10997 }, { "epoch": 4.980978260869565, "grad_norm": 0.0629777972690487, "learning_rate": 3.526945703509199e-09, "loss": 0.8467, "step": 10998 }, { "epoch": 4.98143115942029, "grad_norm": 0.05633389221029704, "learning_rate": 3.360997634467289e-09, "loss": 0.8635, "step": 10999 }, { "epoch": 4.981884057971015, "grad_norm": 0.05669238854451952, "learning_rate": 3.19904809280569e-09, "loss": 0.8571, "step": 11000 }, { "epoch": 4.982336956521739, "grad_norm": 0.054236080292266474, "learning_rate": 3.041097094711454e-09, "loss": 0.8559, "step": 11001 }, { "epoch": 4.982789855072464, "grad_norm": 0.05904611206522206, "learning_rate": 2.8871446559852744e-09, "loss": 0.8373, "step": 11002 }, { "epoch": 4.9832427536231885, "grad_norm": 0.05483481453676402, "learning_rate": 2.7371907920104022e-09, "loss": 0.8625, "step": 11003 }, { "epoch": 4.983695652173913, "grad_norm": 0.05716058310789251, "learning_rate": 2.591235517779289e-09, "loss": 0.8694, "step": 11004 }, { "epoch": 4.984148550724638, "grad_norm": 0.05889029667906278, "learning_rate": 2.4492788478802654e-09, "loss": 0.8341, "step": 11005 }, { "epoch": 4.984601449275362, "grad_norm": 0.0534147745875404, "learning_rate": 2.311320796510863e-09, "loss": 0.8357, "step": 11006 }, { "epoch": 4.985054347826087, "grad_norm": 0.05992741467170223, "learning_rate": 2.177361377460052e-09, "loss": 0.8491, "step": 11007 }, { "epoch": 4.9855072463768115, "grad_norm": 0.05577307693015363, "learning_rate": 2.0474006041215634e-09, "loss": 0.8533, "step": 11008 }, { "epoch": 4.985960144927536, "grad_norm": 0.06073036695282761, "learning_rate": 1.9214384894894468e-09, "loss": 0.842, "step": 11009 }, { "epoch": 4.986413043478261, "grad_norm": 0.056278150772319756, "learning_rate": 1.7994750461491905e-09, "loss": 0.8504, "step": 11010 }, { "epoch": 4.986865942028985, "grad_norm": 0.05985311255362281, "learning_rate": 1.6815102862999255e-09, "loss": 0.8525, "step": 11011 }, { "epoch": 4.98731884057971, "grad_norm": 0.05500494275398649, "learning_rate": 1.5675442217322201e-09, "loss": 0.8455, "step": 11012 }, { "epoch": 4.987771739130435, "grad_norm": 0.05968400768368137, "learning_rate": 1.4575768638369626e-09, "loss": 0.8611, "step": 11013 }, { "epoch": 4.988224637681159, "grad_norm": 0.05579550053430792, "learning_rate": 1.3516082236142424e-09, "loss": 0.851, "step": 11014 }, { "epoch": 4.988677536231884, "grad_norm": 0.06061480994188329, "learning_rate": 1.2496383116555877e-09, "loss": 0.8526, "step": 11015 }, { "epoch": 4.989130434782608, "grad_norm": 0.05782726368997686, "learning_rate": 1.1516671381528454e-09, "loss": 0.8445, "step": 11016 }, { "epoch": 4.989583333333333, "grad_norm": 0.057758568120654226, "learning_rate": 1.057694712902624e-09, "loss": 0.85, "step": 11017 }, { "epoch": 4.990036231884058, "grad_norm": 0.054784107107156255, "learning_rate": 9.677210452974095e-10, "loss": 0.8478, "step": 11018 }, { "epoch": 4.990489130434782, "grad_norm": 0.06348931694626836, "learning_rate": 8.81746144330009e-10, "loss": 0.8464, "step": 11019 }, { "epoch": 4.990942028985507, "grad_norm": 0.05798592142505559, "learning_rate": 7.997700186024304e-10, "loss": 0.8505, "step": 11020 }, { "epoch": 4.991394927536232, "grad_norm": 0.0596311334955053, "learning_rate": 7.217926763036787e-10, "loss": 0.8664, "step": 11021 }, { "epoch": 4.991847826086957, "grad_norm": 0.0595769286797333, "learning_rate": 6.478141252319603e-10, "loss": 0.8598, "step": 11022 }, { "epoch": 4.992300724637682, "grad_norm": 0.057477967987800825, "learning_rate": 5.778343727858016e-10, "loss": 0.8482, "step": 11023 }, { "epoch": 4.992753623188406, "grad_norm": 0.06787914299584286, "learning_rate": 5.118534259551667e-10, "loss": 0.8524, "step": 11024 }, { "epoch": 4.993206521739131, "grad_norm": 0.05462523682563666, "learning_rate": 4.4987129133922116e-10, "loss": 0.8597, "step": 11025 }, { "epoch": 4.993659420289855, "grad_norm": 0.05859364997143069, "learning_rate": 3.9188797513300956e-10, "loss": 0.8636, "step": 11026 }, { "epoch": 4.99411231884058, "grad_norm": 0.0589028216526945, "learning_rate": 3.37903483136337e-10, "loss": 0.8372, "step": 11027 }, { "epoch": 4.994565217391305, "grad_norm": 0.06283945608305516, "learning_rate": 2.879178207448874e-10, "loss": 0.8454, "step": 11028 }, { "epoch": 4.995018115942029, "grad_norm": 0.05534400171423572, "learning_rate": 2.4193099295466426e-10, "loss": 0.8492, "step": 11029 }, { "epoch": 4.995471014492754, "grad_norm": 0.05865515960670916, "learning_rate": 1.9994300436643187e-10, "loss": 0.8475, "step": 11030 }, { "epoch": 4.9959239130434785, "grad_norm": 0.05768981528547375, "learning_rate": 1.6195385917239238e-10, "loss": 0.8446, "step": 11031 }, { "epoch": 4.996376811594203, "grad_norm": 0.0546638209604319, "learning_rate": 1.2796356117394936e-10, "loss": 0.832, "step": 11032 }, { "epoch": 4.996829710144928, "grad_norm": 0.057135233117212324, "learning_rate": 9.797211376394445e-11, "loss": 0.8433, "step": 11033 }, { "epoch": 4.997282608695652, "grad_norm": 0.060714771601945244, "learning_rate": 7.197951994886154e-11, "loss": 0.8464, "step": 11034 }, { "epoch": 4.997735507246377, "grad_norm": 0.05940965512078654, "learning_rate": 4.9985782322181654e-11, "loss": 0.8462, "step": 11035 }, { "epoch": 4.9981884057971016, "grad_norm": 0.055890811751821534, "learning_rate": 3.199090308658725e-11, "loss": 0.8387, "step": 11036 }, { "epoch": 4.998641304347826, "grad_norm": 0.0588219124919754, "learning_rate": 1.7994884036198757e-11, "loss": 0.844, "step": 11037 }, { "epoch": 4.999094202898551, "grad_norm": 0.05745954998968238, "learning_rate": 7.997726569897168e-12, "loss": 0.8443, "step": 11038 }, { "epoch": 4.999547101449275, "grad_norm": 0.05599988301611617, "learning_rate": 1.999431691324105e-12, "loss": 0.8389, "step": 11039 }, { "epoch": 5.0, "grad_norm": 0.06124825430748854, "learning_rate": 0.0, "loss": 0.8515, "step": 11040 }, { "epoch": 5.0, "step": 11040, "total_flos": 1.8520861053026304e+17, "train_loss": 0.1699829785126275, "train_runtime": 52220.1974, "train_samples_per_second": 108.211, "train_steps_per_second": 0.211 } ], "logging_steps": 1, "max_steps": 11040, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.8520861053026304e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }