{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.64, "eval_steps": 500, "global_step": 400, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0016, "grad_norm": 18.498022079467773, "learning_rate": 0.0, "loss": 11.9034, "step": 1 }, { "epoch": 0.0032, "grad_norm": 19.59610366821289, "learning_rate": 7.142857142857143e-07, "loss": 12.361, "step": 2 }, { "epoch": 0.0048, "grad_norm": 19.05824089050293, "learning_rate": 1.4285714285714286e-06, "loss": 12.3027, "step": 3 }, { "epoch": 0.0064, "grad_norm": 18.656831741333008, "learning_rate": 2.1428571428571427e-06, "loss": 11.9369, "step": 4 }, { "epoch": 0.008, "grad_norm": 19.260093688964844, "learning_rate": 2.8571428571428573e-06, "loss": 12.0124, "step": 5 }, { "epoch": 0.0096, "grad_norm": 18.27457046508789, "learning_rate": 3.5714285714285718e-06, "loss": 11.8957, "step": 6 }, { "epoch": 0.0112, "grad_norm": 18.090967178344727, "learning_rate": 4.2857142857142855e-06, "loss": 12.0549, "step": 7 }, { "epoch": 0.0128, "grad_norm": 18.144412994384766, "learning_rate": 5e-06, "loss": 12.0455, "step": 8 }, { "epoch": 0.0144, "grad_norm": 17.77107048034668, "learning_rate": 4.999967697817016e-06, "loss": 12.0381, "step": 9 }, { "epoch": 0.016, "grad_norm": 19.326251983642578, "learning_rate": 4.9998707921028104e-06, "loss": 12.2511, "step": 10 }, { "epoch": 0.0176, "grad_norm": 20.007396697998047, "learning_rate": 4.999709285361594e-06, "loss": 12.5214, "step": 11 }, { "epoch": 0.0192, "grad_norm": 18.9156551361084, "learning_rate": 4.999483181766986e-06, "loss": 12.226, "step": 12 }, { "epoch": 0.0208, "grad_norm": 16.71934700012207, "learning_rate": 4.999192487161895e-06, "loss": 11.63, "step": 13 }, { "epoch": 0.0224, "grad_norm": 16.819419860839844, "learning_rate": 4.998837209058379e-06, "loss": 11.8643, "step": 14 }, { "epoch": 0.024, "grad_norm": 17.38993263244629, "learning_rate": 4.998417356637445e-06, "loss": 12.0876, "step": 15 }, { "epoch": 0.0256, "grad_norm": 17.411874771118164, "learning_rate": 4.997932940748811e-06, "loss": 11.9954, "step": 16 }, { "epoch": 0.0272, "grad_norm": 17.002464294433594, "learning_rate": 4.997383973910631e-06, "loss": 12.0872, "step": 17 }, { "epoch": 0.0288, "grad_norm": 16.964523315429688, "learning_rate": 4.996770470309167e-06, "loss": 11.9659, "step": 18 }, { "epoch": 0.0304, "grad_norm": 17.12218475341797, "learning_rate": 4.9960924457984225e-06, "loss": 11.9041, "step": 19 }, { "epoch": 0.032, "grad_norm": 15.89167308807373, "learning_rate": 4.995349917899735e-06, "loss": 11.5677, "step": 20 }, { "epoch": 0.0336, "grad_norm": 17.404312133789062, "learning_rate": 4.994542905801323e-06, "loss": 12.0405, "step": 21 }, { "epoch": 0.0352, "grad_norm": 17.553056716918945, "learning_rate": 4.993671430357788e-06, "loss": 11.9337, "step": 22 }, { "epoch": 0.0368, "grad_norm": 17.339885711669922, "learning_rate": 4.9927355140895775e-06, "loss": 11.9247, "step": 23 }, { "epoch": 0.0384, "grad_norm": 16.14508056640625, "learning_rate": 4.991735181182401e-06, "loss": 11.5418, "step": 24 }, { "epoch": 0.04, "grad_norm": 14.626124382019043, "learning_rate": 4.9906704574866105e-06, "loss": 11.4515, "step": 25 }, { "epoch": 0.0416, "grad_norm": 17.168630599975586, "learning_rate": 4.989541370516523e-06, "loss": 11.9094, "step": 26 }, { "epoch": 0.0432, "grad_norm": 15.818598747253418, "learning_rate": 4.98834794944972e-06, "loss": 11.7249, "step": 27 }, { "epoch": 0.0448, "grad_norm": 15.550960540771484, "learning_rate": 4.987090225126285e-06, "loss": 11.3948, "step": 28 }, { "epoch": 0.0464, "grad_norm": 15.328129768371582, "learning_rate": 4.985768230048011e-06, "loss": 11.3551, "step": 29 }, { "epoch": 0.048, "grad_norm": 13.792865753173828, "learning_rate": 4.9843819983775575e-06, "loss": 11.1959, "step": 30 }, { "epoch": 0.0496, "grad_norm": 15.154494285583496, "learning_rate": 4.982931565937575e-06, "loss": 11.7186, "step": 31 }, { "epoch": 0.0512, "grad_norm": 14.210274696350098, "learning_rate": 4.98141697020977e-06, "loss": 11.3324, "step": 32 }, { "epoch": 0.0528, "grad_norm": 14.419412612915039, "learning_rate": 4.979838250333941e-06, "loss": 11.0964, "step": 33 }, { "epoch": 0.0544, "grad_norm": 14.476765632629395, "learning_rate": 4.978195447106965e-06, "loss": 11.4052, "step": 34 }, { "epoch": 0.056, "grad_norm": 13.166458129882812, "learning_rate": 4.976488602981748e-06, "loss": 11.1172, "step": 35 }, { "epoch": 0.0576, "grad_norm": 13.77320384979248, "learning_rate": 4.974717762066123e-06, "loss": 11.2174, "step": 36 }, { "epoch": 0.0592, "grad_norm": 13.610458374023438, "learning_rate": 4.972882970121711e-06, "loss": 11.0993, "step": 37 }, { "epoch": 0.0608, "grad_norm": 13.38510513305664, "learning_rate": 4.970984274562741e-06, "loss": 11.0244, "step": 38 }, { "epoch": 0.0624, "grad_norm": 13.513715744018555, "learning_rate": 4.969021724454821e-06, "loss": 11.0893, "step": 39 }, { "epoch": 0.064, "grad_norm": 12.476004600524902, "learning_rate": 4.966995370513675e-06, "loss": 10.9256, "step": 40 }, { "epoch": 0.0656, "grad_norm": 13.007659912109375, "learning_rate": 4.9649052651038255e-06, "loss": 10.897, "step": 41 }, { "epoch": 0.0672, "grad_norm": 12.74197006225586, "learning_rate": 4.962751462237248e-06, "loss": 10.9702, "step": 42 }, { "epoch": 0.0688, "grad_norm": 11.668892860412598, "learning_rate": 4.9605340175719686e-06, "loss": 10.6871, "step": 43 }, { "epoch": 0.0704, "grad_norm": 12.722009658813477, "learning_rate": 4.958252988410631e-06, "loss": 10.7933, "step": 44 }, { "epoch": 0.072, "grad_norm": 10.594651222229004, "learning_rate": 4.955908433699013e-06, "loss": 10.2903, "step": 45 }, { "epoch": 0.0736, "grad_norm": 12.946322441101074, "learning_rate": 4.9535004140245005e-06, "loss": 11.141, "step": 46 }, { "epoch": 0.0752, "grad_norm": 12.783811569213867, "learning_rate": 4.9510289916145295e-06, "loss": 10.779, "step": 47 }, { "epoch": 0.0768, "grad_norm": 11.354564666748047, "learning_rate": 4.94849423033497e-06, "loss": 10.5837, "step": 48 }, { "epoch": 0.0784, "grad_norm": 11.848569869995117, "learning_rate": 4.94589619568848e-06, "loss": 10.5706, "step": 49 }, { "epoch": 0.08, "grad_norm": 9.897051811218262, "learning_rate": 4.943234954812812e-06, "loss": 10.2364, "step": 50 }, { "epoch": 0.0816, "grad_norm": 11.316352844238281, "learning_rate": 4.940510576479079e-06, "loss": 10.6149, "step": 51 }, { "epoch": 0.0832, "grad_norm": 11.829602241516113, "learning_rate": 4.937723131089974e-06, "loss": 10.5496, "step": 52 }, { "epoch": 0.0848, "grad_norm": 10.069260597229004, "learning_rate": 4.934872690677953e-06, "loss": 10.0925, "step": 53 }, { "epoch": 0.0864, "grad_norm": 11.005655288696289, "learning_rate": 4.931959328903376e-06, "loss": 10.3874, "step": 54 }, { "epoch": 0.088, "grad_norm": 10.913771629333496, "learning_rate": 4.928983121052597e-06, "loss": 10.3699, "step": 55 }, { "epoch": 0.0896, "grad_norm": 10.943140983581543, "learning_rate": 4.925944144036027e-06, "loss": 10.3604, "step": 56 }, { "epoch": 0.0912, "grad_norm": 11.520155906677246, "learning_rate": 4.922842476386137e-06, "loss": 10.45, "step": 57 }, { "epoch": 0.0928, "grad_norm": 11.597639083862305, "learning_rate": 4.919678198255438e-06, "loss": 10.6551, "step": 58 }, { "epoch": 0.0944, "grad_norm": 11.26528549194336, "learning_rate": 4.9164513914144005e-06, "loss": 10.3729, "step": 59 }, { "epoch": 0.096, "grad_norm": 11.043292045593262, "learning_rate": 4.91316213924935e-06, "loss": 10.2723, "step": 60 }, { "epoch": 0.0976, "grad_norm": 10.670892715454102, "learning_rate": 4.909810526760307e-06, "loss": 10.3842, "step": 61 }, { "epoch": 0.0992, "grad_norm": 12.231328010559082, "learning_rate": 4.90639664055879e-06, "loss": 10.4897, "step": 62 }, { "epoch": 0.1008, "grad_norm": 10.968878746032715, "learning_rate": 4.902920568865582e-06, "loss": 10.225, "step": 63 }, { "epoch": 0.1024, "grad_norm": 10.761625289916992, "learning_rate": 4.899382401508446e-06, "loss": 10.0817, "step": 64 }, { "epoch": 0.104, "grad_norm": 9.868699073791504, "learning_rate": 4.8957822299198045e-06, "loss": 9.9514, "step": 65 }, { "epoch": 0.1056, "grad_norm": 10.76175308227539, "learning_rate": 4.892120147134378e-06, "loss": 10.2281, "step": 66 }, { "epoch": 0.1072, "grad_norm": 10.531647682189941, "learning_rate": 4.888396247786784e-06, "loss": 10.0424, "step": 67 }, { "epoch": 0.1088, "grad_norm": 9.720768928527832, "learning_rate": 4.884610628109082e-06, "loss": 9.8735, "step": 68 }, { "epoch": 0.1104, "grad_norm": 10.765035629272461, "learning_rate": 4.880763385928298e-06, "loss": 10.1977, "step": 69 }, { "epoch": 0.112, "grad_norm": 10.57542896270752, "learning_rate": 4.876854620663887e-06, "loss": 9.9432, "step": 70 }, { "epoch": 0.1136, "grad_norm": 10.321698188781738, "learning_rate": 4.872884433325169e-06, "loss": 9.9563, "step": 71 }, { "epoch": 0.1152, "grad_norm": 10.044979095458984, "learning_rate": 4.868852926508721e-06, "loss": 9.9412, "step": 72 }, { "epoch": 0.1168, "grad_norm": 10.604913711547852, "learning_rate": 4.864760204395717e-06, "loss": 10.0144, "step": 73 }, { "epoch": 0.1184, "grad_norm": 9.378023147583008, "learning_rate": 4.860606372749247e-06, "loss": 9.8055, "step": 74 }, { "epoch": 0.12, "grad_norm": 9.08212947845459, "learning_rate": 4.856391538911572e-06, "loss": 9.7179, "step": 75 }, { "epoch": 0.1216, "grad_norm": 10.120437622070312, "learning_rate": 4.8521158118013605e-06, "loss": 9.7055, "step": 76 }, { "epoch": 0.1232, "grad_norm": 9.448174476623535, "learning_rate": 4.847779301910868e-06, "loss": 9.7753, "step": 77 }, { "epoch": 0.1248, "grad_norm": 9.994352340698242, "learning_rate": 4.843382121303082e-06, "loss": 9.8005, "step": 78 }, { "epoch": 0.1264, "grad_norm": 10.175590515136719, "learning_rate": 4.83892438360883e-06, "loss": 9.7149, "step": 79 }, { "epoch": 0.128, "grad_norm": 9.187248229980469, "learning_rate": 4.83440620402384e-06, "loss": 9.5077, "step": 80 }, { "epoch": 0.1296, "grad_norm": 8.287508964538574, "learning_rate": 4.829827699305759e-06, "loss": 9.2745, "step": 81 }, { "epoch": 0.1312, "grad_norm": 9.502992630004883, "learning_rate": 4.825188987771149e-06, "loss": 9.7212, "step": 82 }, { "epoch": 0.1328, "grad_norm": 9.398665428161621, "learning_rate": 4.820490189292415e-06, "loss": 9.4711, "step": 83 }, { "epoch": 0.1344, "grad_norm": 9.67027473449707, "learning_rate": 4.815731425294716e-06, "loss": 9.6444, "step": 84 }, { "epoch": 0.136, "grad_norm": 9.999480247497559, "learning_rate": 4.810912818752825e-06, "loss": 9.6473, "step": 85 }, { "epoch": 0.1376, "grad_norm": 10.192490577697754, "learning_rate": 4.806034494187949e-06, "loss": 9.5, "step": 86 }, { "epoch": 0.1392, "grad_norm": 10.132280349731445, "learning_rate": 4.8010965776645145e-06, "loss": 9.5879, "step": 87 }, { "epoch": 0.1408, "grad_norm": 10.07225227355957, "learning_rate": 4.796099196786908e-06, "loss": 9.4842, "step": 88 }, { "epoch": 0.1424, "grad_norm": 9.903701782226562, "learning_rate": 4.791042480696179e-06, "loss": 9.5136, "step": 89 }, { "epoch": 0.144, "grad_norm": 9.270772933959961, "learning_rate": 4.785926560066703e-06, "loss": 9.367, "step": 90 }, { "epoch": 0.1456, "grad_norm": 8.604142189025879, "learning_rate": 4.780751567102801e-06, "loss": 9.1618, "step": 91 }, { "epoch": 0.1472, "grad_norm": 9.69423770904541, "learning_rate": 4.775517635535332e-06, "loss": 9.2597, "step": 92 }, { "epoch": 0.1488, "grad_norm": 10.679227828979492, "learning_rate": 4.770224900618226e-06, "loss": 9.3722, "step": 93 }, { "epoch": 0.1504, "grad_norm": 9.498805046081543, "learning_rate": 4.764873499124997e-06, "loss": 9.2149, "step": 94 }, { "epoch": 0.152, "grad_norm": 10.133721351623535, "learning_rate": 4.759463569345205e-06, "loss": 9.211, "step": 95 }, { "epoch": 0.1536, "grad_norm": 10.703787803649902, "learning_rate": 4.753995251080884e-06, "loss": 9.2456, "step": 96 }, { "epoch": 0.1552, "grad_norm": 9.839086532592773, "learning_rate": 4.748468685642926e-06, "loss": 9.0973, "step": 97 }, { "epoch": 0.1568, "grad_norm": 8.186079978942871, "learning_rate": 4.742884015847436e-06, "loss": 8.9378, "step": 98 }, { "epoch": 0.1584, "grad_norm": 9.491665840148926, "learning_rate": 4.737241386012032e-06, "loss": 9.0209, "step": 99 }, { "epoch": 0.16, "grad_norm": 9.313272476196289, "learning_rate": 4.731540941952126e-06, "loss": 9.0591, "step": 100 }, { "epoch": 0.1616, "grad_norm": 9.365504264831543, "learning_rate": 4.725782830977145e-06, "loss": 9.0187, "step": 101 }, { "epoch": 0.1632, "grad_norm": 9.761168479919434, "learning_rate": 4.719967201886734e-06, "loss": 8.9663, "step": 102 }, { "epoch": 0.1648, "grad_norm": 10.886737823486328, "learning_rate": 4.714094204966903e-06, "loss": 9.0523, "step": 103 }, { "epoch": 0.1664, "grad_norm": 11.529026985168457, "learning_rate": 4.708163991986152e-06, "loss": 8.9862, "step": 104 }, { "epoch": 0.168, "grad_norm": 11.003949165344238, "learning_rate": 4.702176716191539e-06, "loss": 8.9944, "step": 105 }, { "epoch": 0.1696, "grad_norm": 10.637957572937012, "learning_rate": 4.696132532304727e-06, "loss": 8.7661, "step": 106 }, { "epoch": 0.1712, "grad_norm": 10.600850105285645, "learning_rate": 4.690031596517984e-06, "loss": 8.6939, "step": 107 }, { "epoch": 0.1728, "grad_norm": 10.38036823272705, "learning_rate": 4.683874066490143e-06, "loss": 8.8789, "step": 108 }, { "epoch": 0.1744, "grad_norm": 10.32800006866455, "learning_rate": 4.677660101342536e-06, "loss": 8.7128, "step": 109 }, { "epoch": 0.176, "grad_norm": 9.01140022277832, "learning_rate": 4.671389861654873e-06, "loss": 8.5539, "step": 110 }, { "epoch": 0.1776, "grad_norm": 10.66911792755127, "learning_rate": 4.665063509461098e-06, "loss": 8.6381, "step": 111 }, { "epoch": 0.1792, "grad_norm": 12.453883171081543, "learning_rate": 4.658681208245198e-06, "loss": 8.744, "step": 112 }, { "epoch": 0.1808, "grad_norm": 10.25079345703125, "learning_rate": 4.652243122936987e-06, "loss": 8.4376, "step": 113 }, { "epoch": 0.1824, "grad_norm": 13.820728302001953, "learning_rate": 4.645749419907829e-06, "loss": 8.7434, "step": 114 }, { "epoch": 0.184, "grad_norm": 9.763056755065918, "learning_rate": 4.639200266966351e-06, "loss": 8.5522, "step": 115 }, { "epoch": 0.1856, "grad_norm": 12.80479907989502, "learning_rate": 4.632595833354105e-06, "loss": 8.5777, "step": 116 }, { "epoch": 0.1872, "grad_norm": 11.518301963806152, "learning_rate": 4.625936289741187e-06, "loss": 8.2751, "step": 117 }, { "epoch": 0.1888, "grad_norm": 13.24626350402832, "learning_rate": 4.619221808221833e-06, "loss": 8.4951, "step": 118 }, { "epoch": 0.1904, "grad_norm": 14.117487907409668, "learning_rate": 4.612452562309975e-06, "loss": 8.3492, "step": 119 }, { "epoch": 0.192, "grad_norm": 12.98033618927002, "learning_rate": 4.605628726934747e-06, "loss": 8.2238, "step": 120 }, { "epoch": 0.1936, "grad_norm": 11.61721420288086, "learning_rate": 4.598750478435973e-06, "loss": 8.2197, "step": 121 }, { "epoch": 0.1952, "grad_norm": 12.619826316833496, "learning_rate": 4.5918179945596055e-06, "loss": 8.2005, "step": 122 }, { "epoch": 0.1968, "grad_norm": 11.586774826049805, "learning_rate": 4.584831454453135e-06, "loss": 8.0181, "step": 123 }, { "epoch": 0.1984, "grad_norm": 11.792032241821289, "learning_rate": 4.577791038660959e-06, "loss": 8.1414, "step": 124 }, { "epoch": 0.2, "grad_norm": 14.116670608520508, "learning_rate": 4.570696929119717e-06, "loss": 7.9288, "step": 125 }, { "epoch": 0.2016, "grad_norm": 13.249540328979492, "learning_rate": 4.563549309153589e-06, "loss": 7.8991, "step": 126 }, { "epoch": 0.2032, "grad_norm": 13.10855770111084, "learning_rate": 4.556348363469556e-06, "loss": 7.7997, "step": 127 }, { "epoch": 0.2048, "grad_norm": 13.13003921508789, "learning_rate": 4.549094278152631e-06, "loss": 7.7856, "step": 128 }, { "epoch": 0.2064, "grad_norm": 11.22130012512207, "learning_rate": 4.541787240661049e-06, "loss": 7.8092, "step": 129 }, { "epoch": 0.208, "grad_norm": 12.641518592834473, "learning_rate": 4.534427439821416e-06, "loss": 7.6641, "step": 130 }, { "epoch": 0.2096, "grad_norm": 12.413702964782715, "learning_rate": 4.527015065823841e-06, "loss": 7.6125, "step": 131 }, { "epoch": 0.2112, "grad_norm": 12.156352043151855, "learning_rate": 4.519550310217013e-06, "loss": 7.5519, "step": 132 }, { "epoch": 0.2128, "grad_norm": 11.665326118469238, "learning_rate": 4.512033365903251e-06, "loss": 7.5902, "step": 133 }, { "epoch": 0.2144, "grad_norm": 12.924529075622559, "learning_rate": 4.504464427133527e-06, "loss": 7.5706, "step": 134 }, { "epoch": 0.216, "grad_norm": 15.648908615112305, "learning_rate": 4.496843689502435e-06, "loss": 7.3488, "step": 135 }, { "epoch": 0.2176, "grad_norm": 13.633544921875, "learning_rate": 4.489171349943144e-06, "loss": 7.2841, "step": 136 }, { "epoch": 0.2192, "grad_norm": 14.129197120666504, "learning_rate": 4.481447606722309e-06, "loss": 7.2605, "step": 137 }, { "epoch": 0.2208, "grad_norm": 15.414569854736328, "learning_rate": 4.473672659434941e-06, "loss": 7.1375, "step": 138 }, { "epoch": 0.2224, "grad_norm": 15.15605640411377, "learning_rate": 4.465846708999258e-06, "loss": 7.3157, "step": 139 }, { "epoch": 0.224, "grad_norm": 13.083776473999023, "learning_rate": 4.457969957651485e-06, "loss": 7.1391, "step": 140 }, { "epoch": 0.2256, "grad_norm": 14.658546447753906, "learning_rate": 4.450042608940632e-06, "loss": 6.8354, "step": 141 }, { "epoch": 0.2272, "grad_norm": 13.068192481994629, "learning_rate": 4.442064867723236e-06, "loss": 6.9534, "step": 142 }, { "epoch": 0.2288, "grad_norm": 13.683319091796875, "learning_rate": 4.434036940158062e-06, "loss": 6.7263, "step": 143 }, { "epoch": 0.2304, "grad_norm": 12.352782249450684, "learning_rate": 4.425959033700776e-06, "loss": 6.9508, "step": 144 }, { "epoch": 0.232, "grad_norm": 12.934060096740723, "learning_rate": 4.417831357098591e-06, "loss": 6.7608, "step": 145 }, { "epoch": 0.2336, "grad_norm": 10.694926261901855, "learning_rate": 4.409654120384863e-06, "loss": 6.9133, "step": 146 }, { "epoch": 0.2352, "grad_norm": 13.42801570892334, "learning_rate": 4.401427534873669e-06, "loss": 6.2184, "step": 147 }, { "epoch": 0.2368, "grad_norm": 12.629581451416016, "learning_rate": 4.393151813154345e-06, "loss": 6.1405, "step": 148 }, { "epoch": 0.2384, "grad_norm": 9.909472465515137, "learning_rate": 4.384827169085993e-06, "loss": 6.376, "step": 149 }, { "epoch": 0.24, "grad_norm": 9.083218574523926, "learning_rate": 4.3764538177919555e-06, "loss": 6.387, "step": 150 }, { "epoch": 0.2416, "grad_norm": 8.383381843566895, "learning_rate": 4.3680319756542525e-06, "loss": 6.7282, "step": 151 }, { "epoch": 0.2432, "grad_norm": 8.961660385131836, "learning_rate": 4.35956186030799e-06, "loss": 6.2623, "step": 152 }, { "epoch": 0.2448, "grad_norm": 9.286286354064941, "learning_rate": 4.351043690635744e-06, "loss": 6.0499, "step": 153 }, { "epoch": 0.2464, "grad_norm": 9.719377517700195, "learning_rate": 4.3424776867618935e-06, "loss": 6.4369, "step": 154 }, { "epoch": 0.248, "grad_norm": 7.7872700691223145, "learning_rate": 4.333864070046938e-06, "loss": 6.0176, "step": 155 }, { "epoch": 0.2496, "grad_norm": 8.78665828704834, "learning_rate": 4.325203063081776e-06, "loss": 6.0201, "step": 156 }, { "epoch": 0.2512, "grad_norm": 7.898557186126709, "learning_rate": 4.3164948896819535e-06, "loss": 6.1595, "step": 157 }, { "epoch": 0.2528, "grad_norm": 9.21320629119873, "learning_rate": 4.307739774881878e-06, "loss": 6.3917, "step": 158 }, { "epoch": 0.2544, "grad_norm": 6.031705379486084, "learning_rate": 4.298937944929007e-06, "loss": 5.6134, "step": 159 }, { "epoch": 0.256, "grad_norm": 7.0851006507873535, "learning_rate": 4.290089627277998e-06, "loss": 5.4909, "step": 160 }, { "epoch": 0.2576, "grad_norm": 7.172662258148193, "learning_rate": 4.28119505058483e-06, "loss": 6.1213, "step": 161 }, { "epoch": 0.2592, "grad_norm": 7.241125583648682, "learning_rate": 4.2722544447008995e-06, "loss": 5.9029, "step": 162 }, { "epoch": 0.2608, "grad_norm": 6.174571514129639, "learning_rate": 4.263268040667075e-06, "loss": 5.9515, "step": 163 }, { "epoch": 0.2624, "grad_norm": 5.6988935470581055, "learning_rate": 4.254236070707734e-06, "loss": 5.8603, "step": 164 }, { "epoch": 0.264, "grad_norm": 5.242663383483887, "learning_rate": 4.245158768224748e-06, "loss": 5.6214, "step": 165 }, { "epoch": 0.2656, "grad_norm": 5.694828987121582, "learning_rate": 4.236036367791471e-06, "loss": 6.3081, "step": 166 }, { "epoch": 0.2672, "grad_norm": 7.1838788986206055, "learning_rate": 4.226869105146658e-06, "loss": 5.4577, "step": 167 }, { "epoch": 0.2688, "grad_norm": 8.755837440490723, "learning_rate": 4.2176572171883865e-06, "loss": 5.5286, "step": 168 }, { "epoch": 0.2704, "grad_norm": 4.821361064910889, "learning_rate": 4.208400941967928e-06, "loss": 5.8776, "step": 169 }, { "epoch": 0.272, "grad_norm": 5.043092250823975, "learning_rate": 4.199100518683601e-06, "loss": 5.6913, "step": 170 }, { "epoch": 0.2736, "grad_norm": 5.266634941101074, "learning_rate": 4.189756187674584e-06, "loss": 5.8228, "step": 171 }, { "epoch": 0.2752, "grad_norm": 5.798868179321289, "learning_rate": 4.18036819041471e-06, "loss": 5.6908, "step": 172 }, { "epoch": 0.2768, "grad_norm": 5.337583065032959, "learning_rate": 4.170936769506222e-06, "loss": 6.0907, "step": 173 }, { "epoch": 0.2784, "grad_norm": 4.3411102294921875, "learning_rate": 4.161462168673508e-06, "loss": 5.8087, "step": 174 }, { "epoch": 0.28, "grad_norm": 6.649768829345703, "learning_rate": 4.1519446327567995e-06, "loss": 5.2901, "step": 175 }, { "epoch": 0.2816, "grad_norm": 6.4602556228637695, "learning_rate": 4.142384407705846e-06, "loss": 5.274, "step": 176 }, { "epoch": 0.2832, "grad_norm": 5.464128017425537, "learning_rate": 4.132781740573559e-06, "loss": 6.0373, "step": 177 }, { "epoch": 0.2848, "grad_norm": 6.308886528015137, "learning_rate": 4.123136879509626e-06, "loss": 5.2471, "step": 178 }, { "epoch": 0.2864, "grad_norm": 4.4718828201293945, "learning_rate": 4.1134500737541026e-06, "loss": 5.3702, "step": 179 }, { "epoch": 0.288, "grad_norm": 3.863276958465576, "learning_rate": 4.103721573630965e-06, "loss": 5.3922, "step": 180 }, { "epoch": 0.2896, "grad_norm": 4.226170063018799, "learning_rate": 4.093951630541646e-06, "loss": 5.4302, "step": 181 }, { "epoch": 0.2912, "grad_norm": 3.5398764610290527, "learning_rate": 4.084140496958539e-06, "loss": 5.3444, "step": 182 }, { "epoch": 0.2928, "grad_norm": 3.5163183212280273, "learning_rate": 4.074288426418467e-06, "loss": 5.3247, "step": 183 }, { "epoch": 0.2944, "grad_norm": 5.308356761932373, "learning_rate": 4.06439567351614e-06, "loss": 5.6837, "step": 184 }, { "epoch": 0.296, "grad_norm": 3.6820993423461914, "learning_rate": 4.054462493897569e-06, "loss": 5.1638, "step": 185 }, { "epoch": 0.2976, "grad_norm": 3.9401638507843018, "learning_rate": 4.0444891442534615e-06, "loss": 5.498, "step": 186 }, { "epoch": 0.2992, "grad_norm": 5.141748428344727, "learning_rate": 4.034475882312593e-06, "loss": 5.1034, "step": 187 }, { "epoch": 0.3008, "grad_norm": 3.4670636653900146, "learning_rate": 4.024422966835137e-06, "loss": 5.2071, "step": 188 }, { "epoch": 0.3024, "grad_norm": 3.7121994495391846, "learning_rate": 4.014330657605984e-06, "loss": 5.3885, "step": 189 }, { "epoch": 0.304, "grad_norm": 5.863914489746094, "learning_rate": 4.004199215428032e-06, "loss": 5.5361, "step": 190 }, { "epoch": 0.3056, "grad_norm": 3.343780517578125, "learning_rate": 3.994028902115439e-06, "loss": 5.4439, "step": 191 }, { "epoch": 0.3072, "grad_norm": 4.041346073150635, "learning_rate": 3.9838199804868635e-06, "loss": 5.6031, "step": 192 }, { "epoch": 0.3088, "grad_norm": 2.9033005237579346, "learning_rate": 3.973572714358668e-06, "loss": 5.3807, "step": 193 }, { "epoch": 0.3104, "grad_norm": 3.026655435562134, "learning_rate": 3.963287368538105e-06, "loss": 5.6681, "step": 194 }, { "epoch": 0.312, "grad_norm": 5.9403862953186035, "learning_rate": 3.9529642088164736e-06, "loss": 5.1567, "step": 195 }, { "epoch": 0.3136, "grad_norm": 4.5110883712768555, "learning_rate": 3.942603501962249e-06, "loss": 5.3055, "step": 196 }, { "epoch": 0.3152, "grad_norm": 4.24874210357666, "learning_rate": 3.932205515714189e-06, "loss": 5.3806, "step": 197 }, { "epoch": 0.3168, "grad_norm": 4.06620979309082, "learning_rate": 3.92177051877442e-06, "loss": 5.5074, "step": 198 }, { "epoch": 0.3184, "grad_norm": 3.4436845779418945, "learning_rate": 3.9112987808014824e-06, "loss": 5.3936, "step": 199 }, { "epoch": 0.32, "grad_norm": 3.5736324787139893, "learning_rate": 3.900790572403376e-06, "loss": 5.3905, "step": 200 }, { "epoch": 0.3216, "grad_norm": 2.719327926635742, "learning_rate": 3.890246165130556e-06, "loss": 5.3799, "step": 201 }, { "epoch": 0.3232, "grad_norm": 2.68566632270813, "learning_rate": 3.8796658314689205e-06, "loss": 5.1272, "step": 202 }, { "epoch": 0.3248, "grad_norm": 3.111459493637085, "learning_rate": 3.86904984483277e-06, "loss": 5.1238, "step": 203 }, { "epoch": 0.3264, "grad_norm": 3.006056785583496, "learning_rate": 3.858398479557739e-06, "loss": 5.0996, "step": 204 }, { "epoch": 0.328, "grad_norm": 6.713052749633789, "learning_rate": 3.847712010893706e-06, "loss": 5.1972, "step": 205 }, { "epoch": 0.3296, "grad_norm": 4.517950534820557, "learning_rate": 3.836990714997686e-06, "loss": 5.0826, "step": 206 }, { "epoch": 0.3312, "grad_norm": 3.7129902839660645, "learning_rate": 3.826234868926686e-06, "loss": 5.1656, "step": 207 }, { "epoch": 0.3328, "grad_norm": 4.543206691741943, "learning_rate": 3.815444750630555e-06, "loss": 5.0371, "step": 208 }, { "epoch": 0.3344, "grad_norm": 7.416889667510986, "learning_rate": 3.8046206389447916e-06, "loss": 5.46, "step": 209 }, { "epoch": 0.336, "grad_norm": 5.088094711303711, "learning_rate": 3.7937628135833453e-06, "loss": 5.313, "step": 210 }, { "epoch": 0.3376, "grad_norm": 2.3088865280151367, "learning_rate": 3.782871555131386e-06, "loss": 5.1569, "step": 211 }, { "epoch": 0.3392, "grad_norm": 4.071847915649414, "learning_rate": 3.7719471450380518e-06, "loss": 5.2869, "step": 212 }, { "epoch": 0.3408, "grad_norm": 5.414875030517578, "learning_rate": 3.760989865609178e-06, "loss": 4.9842, "step": 213 }, { "epoch": 0.3424, "grad_norm": 5.795034885406494, "learning_rate": 3.7500000000000005e-06, "loss": 5.0864, "step": 214 }, { "epoch": 0.344, "grad_norm": 3.6953725814819336, "learning_rate": 3.738977832207839e-06, "loss": 5.1704, "step": 215 }, { "epoch": 0.3456, "grad_norm": 2.568744659423828, "learning_rate": 3.7279236470647593e-06, "loss": 5.4024, "step": 216 }, { "epoch": 0.3472, "grad_norm": 2.6171984672546387, "learning_rate": 3.7168377302302096e-06, "loss": 5.4781, "step": 217 }, { "epoch": 0.3488, "grad_norm": 4.188714981079102, "learning_rate": 3.7057203681836407e-06, "loss": 5.0629, "step": 218 }, { "epoch": 0.3504, "grad_norm": 5.346961975097656, "learning_rate": 3.6945718482171044e-06, "loss": 4.7197, "step": 219 }, { "epoch": 0.352, "grad_norm": 3.4702541828155518, "learning_rate": 3.683392458427825e-06, "loss": 5.3812, "step": 220 }, { "epoch": 0.3536, "grad_norm": 2.4442174434661865, "learning_rate": 3.6721824877107588e-06, "loss": 5.0477, "step": 221 }, { "epoch": 0.3552, "grad_norm": 3.310760259628296, "learning_rate": 3.660942225751126e-06, "loss": 4.8426, "step": 222 }, { "epoch": 0.3568, "grad_norm": 2.951408624649048, "learning_rate": 3.6496719630169254e-06, "loss": 5.3388, "step": 223 }, { "epoch": 0.3584, "grad_norm": 4.146093368530273, "learning_rate": 3.638371990751428e-06, "loss": 4.9851, "step": 224 }, { "epoch": 0.36, "grad_norm": 2.433192729949951, "learning_rate": 3.6270426009656524e-06, "loss": 4.9225, "step": 225 }, { "epoch": 0.3616, "grad_norm": 5.279270648956299, "learning_rate": 3.615684086430815e-06, "loss": 5.291, "step": 226 }, { "epoch": 0.3632, "grad_norm": 2.825227737426758, "learning_rate": 3.604296740670768e-06, "loss": 5.0946, "step": 227 }, { "epoch": 0.3648, "grad_norm": 2.6808815002441406, "learning_rate": 3.592880857954413e-06, "loss": 4.8898, "step": 228 }, { "epoch": 0.3664, "grad_norm": 3.7098076343536377, "learning_rate": 3.5814367332880953e-06, "loss": 5.1529, "step": 229 }, { "epoch": 0.368, "grad_norm": 4.077409267425537, "learning_rate": 3.5699646624079824e-06, "loss": 5.3046, "step": 230 }, { "epoch": 0.3696, "grad_norm": 2.28322434425354, "learning_rate": 3.5584649417724217e-06, "loss": 5.1489, "step": 231 }, { "epoch": 0.3712, "grad_norm": 2.7317070960998535, "learning_rate": 3.5469378685542742e-06, "loss": 5.1974, "step": 232 }, { "epoch": 0.3728, "grad_norm": 3.298208475112915, "learning_rate": 3.5353837406332464e-06, "loss": 4.9805, "step": 233 }, { "epoch": 0.3744, "grad_norm": 3.0187830924987793, "learning_rate": 3.52380285658818e-06, "loss": 5.1739, "step": 234 }, { "epoch": 0.376, "grad_norm": 2.1207525730133057, "learning_rate": 3.512195515689343e-06, "loss": 5.0907, "step": 235 }, { "epoch": 0.3776, "grad_norm": 2.5246315002441406, "learning_rate": 3.500562017890695e-06, "loss": 4.9551, "step": 236 }, { "epoch": 0.3792, "grad_norm": 2.149975538253784, "learning_rate": 3.4889026638221376e-06, "loss": 5.0675, "step": 237 }, { "epoch": 0.3808, "grad_norm": 2.6374351978302, "learning_rate": 3.4772177547817387e-06, "loss": 5.2944, "step": 238 }, { "epoch": 0.3824, "grad_norm": 2.4661672115325928, "learning_rate": 3.4655075927279576e-06, "loss": 4.9949, "step": 239 }, { "epoch": 0.384, "grad_norm": 2.271843910217285, "learning_rate": 3.4537724802718294e-06, "loss": 5.044, "step": 240 }, { "epoch": 0.3856, "grad_norm": 2.4052512645721436, "learning_rate": 3.442012720669155e-06, "loss": 5.2165, "step": 241 }, { "epoch": 0.3872, "grad_norm": 2.379415988922119, "learning_rate": 3.430228617812661e-06, "loss": 5.0385, "step": 242 }, { "epoch": 0.3888, "grad_norm": 1.9934539794921875, "learning_rate": 3.418420476224143e-06, "loss": 4.865, "step": 243 }, { "epoch": 0.3904, "grad_norm": 2.073207378387451, "learning_rate": 3.4065886010466014e-06, "loss": 4.816, "step": 244 }, { "epoch": 0.392, "grad_norm": 3.5318527221679688, "learning_rate": 3.3947332980363552e-06, "loss": 4.7399, "step": 245 }, { "epoch": 0.3936, "grad_norm": 3.798776626586914, "learning_rate": 3.382854873555137e-06, "loss": 5.1423, "step": 246 }, { "epoch": 0.3952, "grad_norm": 3.180250644683838, "learning_rate": 3.370953634562179e-06, "loss": 4.8133, "step": 247 }, { "epoch": 0.3968, "grad_norm": 4.3103156089782715, "learning_rate": 3.3590298886062833e-06, "loss": 5.2605, "step": 248 }, { "epoch": 0.3984, "grad_norm": 2.771699905395508, "learning_rate": 3.347083943817867e-06, "loss": 4.6618, "step": 249 }, { "epoch": 0.4, "grad_norm": 2.18292236328125, "learning_rate": 3.3351161089010055e-06, "loss": 4.9174, "step": 250 }, { "epoch": 0.4016, "grad_norm": 4.050281047821045, "learning_rate": 3.3231266931254546e-06, "loss": 4.6317, "step": 251 }, { "epoch": 0.4032, "grad_norm": 2.1576311588287354, "learning_rate": 3.3111160063186553e-06, "loss": 4.6405, "step": 252 }, { "epoch": 0.4048, "grad_norm": 2.3556125164031982, "learning_rate": 3.299084358857731e-06, "loss": 4.8415, "step": 253 }, { "epoch": 0.4064, "grad_norm": 3.075002908706665, "learning_rate": 3.2870320616614626e-06, "loss": 5.0271, "step": 254 }, { "epoch": 0.408, "grad_norm": 2.8005783557891846, "learning_rate": 3.2749594261822586e-06, "loss": 4.9054, "step": 255 }, { "epoch": 0.4096, "grad_norm": 2.2018191814422607, "learning_rate": 3.2628667643981036e-06, "loss": 4.7503, "step": 256 }, { "epoch": 0.4112, "grad_norm": 3.0102760791778564, "learning_rate": 3.250754388804495e-06, "loss": 5.0094, "step": 257 }, { "epoch": 0.4128, "grad_norm": 2.608316659927368, "learning_rate": 3.238622612406373e-06, "loss": 4.8735, "step": 258 }, { "epoch": 0.4144, "grad_norm": 2.749432325363159, "learning_rate": 3.226471748710025e-06, "loss": 5.0013, "step": 259 }, { "epoch": 0.416, "grad_norm": 2.818932056427002, "learning_rate": 3.21430211171499e-06, "loss": 5.4434, "step": 260 }, { "epoch": 0.4176, "grad_norm": 5.755081653594971, "learning_rate": 3.2021140159059398e-06, "loss": 4.5975, "step": 261 }, { "epoch": 0.4192, "grad_norm": 2.349569320678711, "learning_rate": 3.189907776244556e-06, "loss": 5.1069, "step": 262 }, { "epoch": 0.4208, "grad_norm": 2.527462959289551, "learning_rate": 3.1776837081613893e-06, "loss": 4.875, "step": 263 }, { "epoch": 0.4224, "grad_norm": 2.4022037982940674, "learning_rate": 3.1654421275477045e-06, "loss": 5.1679, "step": 264 }, { "epoch": 0.424, "grad_norm": 5.29791784286499, "learning_rate": 3.1531833507473246e-06, "loss": 4.6665, "step": 265 }, { "epoch": 0.4256, "grad_norm": 2.1718266010284424, "learning_rate": 3.1409076945484513e-06, "loss": 5.0353, "step": 266 }, { "epoch": 0.4272, "grad_norm": 2.4552576541900635, "learning_rate": 3.128615476175477e-06, "loss": 5.2233, "step": 267 }, { "epoch": 0.4288, "grad_norm": 6.326014995574951, "learning_rate": 3.116307013280793e-06, "loss": 4.5923, "step": 268 }, { "epoch": 0.4304, "grad_norm": 2.048156261444092, "learning_rate": 3.1039826239365754e-06, "loss": 4.908, "step": 269 }, { "epoch": 0.432, "grad_norm": 3.1258327960968018, "learning_rate": 3.0916426266265676e-06, "loss": 4.6048, "step": 270 }, { "epoch": 0.4336, "grad_norm": 2.174611806869507, "learning_rate": 3.0792873402378516e-06, "loss": 4.8559, "step": 271 }, { "epoch": 0.4352, "grad_norm": 1.987095594406128, "learning_rate": 3.066917084052603e-06, "loss": 4.7892, "step": 272 }, { "epoch": 0.4368, "grad_norm": 5.872653484344482, "learning_rate": 3.0545321777398463e-06, "loss": 5.3549, "step": 273 }, { "epoch": 0.4384, "grad_norm": 3.6800520420074463, "learning_rate": 3.042132941347189e-06, "loss": 4.9779, "step": 274 }, { "epoch": 0.44, "grad_norm": 2.9514310359954834, "learning_rate": 3.0297196952925533e-06, "loss": 4.8666, "step": 275 }, { "epoch": 0.4416, "grad_norm": 2.6596765518188477, "learning_rate": 3.017292760355896e-06, "loss": 4.7164, "step": 276 }, { "epoch": 0.4432, "grad_norm": 2.4458515644073486, "learning_rate": 3.0048524576709175e-06, "loss": 4.6459, "step": 277 }, { "epoch": 0.4448, "grad_norm": 2.106074571609497, "learning_rate": 2.9923991087167657e-06, "loss": 4.7868, "step": 278 }, { "epoch": 0.4464, "grad_norm": 1.8992120027542114, "learning_rate": 2.9799330353097245e-06, "loss": 4.9395, "step": 279 }, { "epoch": 0.448, "grad_norm": 4.042559623718262, "learning_rate": 2.967454559594903e-06, "loss": 5.0813, "step": 280 }, { "epoch": 0.4496, "grad_norm": 4.273650646209717, "learning_rate": 2.9549640040379043e-06, "loss": 5.0635, "step": 281 }, { "epoch": 0.4512, "grad_norm": 2.039167881011963, "learning_rate": 2.9424616914164982e-06, "loss": 4.9402, "step": 282 }, { "epoch": 0.4528, "grad_norm": 3.257084369659424, "learning_rate": 2.929947944812277e-06, "loss": 5.2913, "step": 283 }, { "epoch": 0.4544, "grad_norm": 2.763265371322632, "learning_rate": 2.917423087602306e-06, "loss": 4.8663, "step": 284 }, { "epoch": 0.456, "grad_norm": 2.449984550476074, "learning_rate": 2.904887443450769e-06, "loss": 4.8823, "step": 285 }, { "epoch": 0.4576, "grad_norm": 5.364999294281006, "learning_rate": 2.8923413363006038e-06, "loss": 4.6396, "step": 286 }, { "epoch": 0.4592, "grad_norm": 2.5084285736083984, "learning_rate": 2.8797850903651274e-06, "loss": 4.8121, "step": 287 }, { "epoch": 0.4608, "grad_norm": 1.8236719369888306, "learning_rate": 2.8672190301196655e-06, "loss": 5.1095, "step": 288 }, { "epoch": 0.4624, "grad_norm": 5.353603839874268, "learning_rate": 2.854643480293159e-06, "loss": 4.6599, "step": 289 }, { "epoch": 0.464, "grad_norm": 2.1391165256500244, "learning_rate": 2.842058765859776e-06, "loss": 4.8299, "step": 290 }, { "epoch": 0.4656, "grad_norm": 2.8433682918548584, "learning_rate": 2.8294652120305167e-06, "loss": 4.7608, "step": 291 }, { "epoch": 0.4672, "grad_norm": 3.439664602279663, "learning_rate": 2.8168631442448046e-06, "loss": 4.6247, "step": 292 }, { "epoch": 0.4688, "grad_norm": 2.0286402702331543, "learning_rate": 2.804252888162079e-06, "loss": 5.0117, "step": 293 }, { "epoch": 0.4704, "grad_norm": 1.8686045408248901, "learning_rate": 2.791634769653381e-06, "loss": 4.7583, "step": 294 }, { "epoch": 0.472, "grad_norm": 2.822620391845703, "learning_rate": 2.779009114792928e-06, "loss": 4.8104, "step": 295 }, { "epoch": 0.4736, "grad_norm": 2.264028787612915, "learning_rate": 2.7663762498496905e-06, "loss": 4.6987, "step": 296 }, { "epoch": 0.4752, "grad_norm": 2.261293411254883, "learning_rate": 2.753736501278961e-06, "loss": 4.6871, "step": 297 }, { "epoch": 0.4768, "grad_norm": 3.6320760250091553, "learning_rate": 2.741090195713917e-06, "loss": 4.7803, "step": 298 }, { "epoch": 0.4784, "grad_norm": 2.178699254989624, "learning_rate": 2.7284376599571776e-06, "loss": 4.7506, "step": 299 }, { "epoch": 0.48, "grad_norm": 2.638765811920166, "learning_rate": 2.7157792209723654e-06, "loss": 4.4671, "step": 300 }, { "epoch": 0.4816, "grad_norm": 2.165921211242676, "learning_rate": 2.7031152058756495e-06, "loss": 4.6798, "step": 301 }, { "epoch": 0.4832, "grad_norm": 2.7200875282287598, "learning_rate": 2.6904459419272955e-06, "loss": 4.7178, "step": 302 }, { "epoch": 0.4848, "grad_norm": 3.2318854331970215, "learning_rate": 2.6777717565232098e-06, "loss": 4.8959, "step": 303 }, { "epoch": 0.4864, "grad_norm": 2.586148738861084, "learning_rate": 2.6650929771864776e-06, "loss": 4.8196, "step": 304 }, { "epoch": 0.488, "grad_norm": 1.7376669645309448, "learning_rate": 2.652409931558898e-06, "loss": 4.7363, "step": 305 }, { "epoch": 0.4896, "grad_norm": 1.8759428262710571, "learning_rate": 2.639722947392521e-06, "loss": 4.8715, "step": 306 }, { "epoch": 0.4912, "grad_norm": 1.9195789098739624, "learning_rate": 2.6270323525411724e-06, "loss": 4.9562, "step": 307 }, { "epoch": 0.4928, "grad_norm": 2.1402273178100586, "learning_rate": 2.614338474951987e-06, "loss": 4.7405, "step": 308 }, { "epoch": 0.4944, "grad_norm": 1.8862427473068237, "learning_rate": 2.6016416426569285e-06, "loss": 4.7351, "step": 309 }, { "epoch": 0.496, "grad_norm": 4.3101677894592285, "learning_rate": 2.5889421837643186e-06, "loss": 5.1979, "step": 310 }, { "epoch": 0.4976, "grad_norm": 2.146365165710449, "learning_rate": 2.5762404264503538e-06, "loss": 4.86, "step": 311 }, { "epoch": 0.4992, "grad_norm": 2.066309928894043, "learning_rate": 2.563536698950624e-06, "loss": 4.9702, "step": 312 }, { "epoch": 0.5008, "grad_norm": 1.9804738759994507, "learning_rate": 2.5508313295516326e-06, "loss": 4.8096, "step": 313 }, { "epoch": 0.5024, "grad_norm": 2.8938088417053223, "learning_rate": 2.538124646582315e-06, "loss": 5.2568, "step": 314 }, { "epoch": 0.504, "grad_norm": 1.8512638807296753, "learning_rate": 2.52541697840555e-06, "loss": 5.0433, "step": 315 }, { "epoch": 0.5056, "grad_norm": 4.223806858062744, "learning_rate": 2.512708653409674e-06, "loss": 4.7329, "step": 316 }, { "epoch": 0.5072, "grad_norm": 1.8683325052261353, "learning_rate": 2.5e-06, "loss": 5.0335, "step": 317 }, { "epoch": 0.5088, "grad_norm": 5.908133506774902, "learning_rate": 2.487291346590326e-06, "loss": 4.6557, "step": 318 }, { "epoch": 0.5104, "grad_norm": 5.461273670196533, "learning_rate": 2.4745830215944512e-06, "loss": 4.6582, "step": 319 }, { "epoch": 0.512, "grad_norm": 3.9137203693389893, "learning_rate": 2.4618753534176854e-06, "loss": 5.0977, "step": 320 }, { "epoch": 0.5136, "grad_norm": 2.8667490482330322, "learning_rate": 2.449168670448368e-06, "loss": 4.8414, "step": 321 }, { "epoch": 0.5152, "grad_norm": 2.354121685028076, "learning_rate": 2.436463301049378e-06, "loss": 4.95, "step": 322 }, { "epoch": 0.5168, "grad_norm": 3.92159366607666, "learning_rate": 2.423759573549647e-06, "loss": 4.6653, "step": 323 }, { "epoch": 0.5184, "grad_norm": 4.691086292266846, "learning_rate": 2.4110578162356814e-06, "loss": 5.3596, "step": 324 }, { "epoch": 0.52, "grad_norm": 3.4949967861175537, "learning_rate": 2.3983583573430723e-06, "loss": 4.9576, "step": 325 }, { "epoch": 0.5216, "grad_norm": 2.329535484313965, "learning_rate": 2.385661525048014e-06, "loss": 4.9078, "step": 326 }, { "epoch": 0.5232, "grad_norm": 1.949039340019226, "learning_rate": 2.372967647458829e-06, "loss": 4.8943, "step": 327 }, { "epoch": 0.5248, "grad_norm": 1.8755134344100952, "learning_rate": 2.3602770526074804e-06, "loss": 4.661, "step": 328 }, { "epoch": 0.5264, "grad_norm": 3.7387921810150146, "learning_rate": 2.3475900684411027e-06, "loss": 4.942, "step": 329 }, { "epoch": 0.528, "grad_norm": 2.0674028396606445, "learning_rate": 2.334907022813523e-06, "loss": 4.9509, "step": 330 }, { "epoch": 0.5296, "grad_norm": 3.522639036178589, "learning_rate": 2.3222282434767915e-06, "loss": 4.9442, "step": 331 }, { "epoch": 0.5312, "grad_norm": 3.0063114166259766, "learning_rate": 2.3095540580727054e-06, "loss": 4.9701, "step": 332 }, { "epoch": 0.5328, "grad_norm": 2.9866836071014404, "learning_rate": 2.296884794124351e-06, "loss": 4.8279, "step": 333 }, { "epoch": 0.5344, "grad_norm": 2.5366246700286865, "learning_rate": 2.2842207790276355e-06, "loss": 4.9832, "step": 334 }, { "epoch": 0.536, "grad_norm": 3.206937074661255, "learning_rate": 2.2715623400428228e-06, "loss": 4.6835, "step": 335 }, { "epoch": 0.5376, "grad_norm": 4.108307361602783, "learning_rate": 2.2589098042860838e-06, "loss": 4.6455, "step": 336 }, { "epoch": 0.5392, "grad_norm": 2.709421157836914, "learning_rate": 2.24626349872104e-06, "loss": 4.6987, "step": 337 }, { "epoch": 0.5408, "grad_norm": 2.0390799045562744, "learning_rate": 2.2336237501503103e-06, "loss": 4.8034, "step": 338 }, { "epoch": 0.5424, "grad_norm": 2.461397647857666, "learning_rate": 2.2209908852070723e-06, "loss": 4.6968, "step": 339 }, { "epoch": 0.544, "grad_norm": 3.7129275798797607, "learning_rate": 2.2083652303466196e-06, "loss": 4.5782, "step": 340 }, { "epoch": 0.5456, "grad_norm": 2.8389174938201904, "learning_rate": 2.1957471118379213e-06, "loss": 4.6995, "step": 341 }, { "epoch": 0.5472, "grad_norm": 1.771793246269226, "learning_rate": 2.1831368557551962e-06, "loss": 4.7748, "step": 342 }, { "epoch": 0.5488, "grad_norm": 4.029572486877441, "learning_rate": 2.170534787969484e-06, "loss": 5.0112, "step": 343 }, { "epoch": 0.5504, "grad_norm": 5.098694801330566, "learning_rate": 2.157941234140225e-06, "loss": 5.0988, "step": 344 }, { "epoch": 0.552, "grad_norm": 3.2921664714813232, "learning_rate": 2.145356519706842e-06, "loss": 4.7199, "step": 345 }, { "epoch": 0.5536, "grad_norm": 3.225116729736328, "learning_rate": 2.1327809698803354e-06, "loss": 5.062, "step": 346 }, { "epoch": 0.5552, "grad_norm": 2.9309229850769043, "learning_rate": 2.120214909634873e-06, "loss": 4.9291, "step": 347 }, { "epoch": 0.5568, "grad_norm": 2.9291505813598633, "learning_rate": 2.1076586636993975e-06, "loss": 4.9831, "step": 348 }, { "epoch": 0.5584, "grad_norm": 2.420208692550659, "learning_rate": 2.0951125565492313e-06, "loss": 4.9086, "step": 349 }, { "epoch": 0.56, "grad_norm": 1.8131273984909058, "learning_rate": 2.0825769123976954e-06, "loss": 4.6759, "step": 350 }, { "epoch": 0.5616, "grad_norm": 1.9805721044540405, "learning_rate": 2.070052055187724e-06, "loss": 4.7528, "step": 351 }, { "epoch": 0.5632, "grad_norm": 2.3578004837036133, "learning_rate": 2.057538308583502e-06, "loss": 4.8435, "step": 352 }, { "epoch": 0.5648, "grad_norm": 5.772895812988281, "learning_rate": 2.045035995962097e-06, "loss": 5.4256, "step": 353 }, { "epoch": 0.5664, "grad_norm": 2.147087574005127, "learning_rate": 2.0325454404050983e-06, "loss": 5.0171, "step": 354 }, { "epoch": 0.568, "grad_norm": 2.1338014602661133, "learning_rate": 2.0200669646902755e-06, "loss": 5.0236, "step": 355 }, { "epoch": 0.5696, "grad_norm": 1.909696102142334, "learning_rate": 2.0076008912832355e-06, "loss": 4.8829, "step": 356 }, { "epoch": 0.5712, "grad_norm": 4.729516983032227, "learning_rate": 1.995147542329083e-06, "loss": 4.891, "step": 357 }, { "epoch": 0.5728, "grad_norm": 5.204637050628662, "learning_rate": 1.9827072396441044e-06, "loss": 4.8071, "step": 358 }, { "epoch": 0.5744, "grad_norm": 9.13206672668457, "learning_rate": 1.970280304707447e-06, "loss": 4.3763, "step": 359 }, { "epoch": 0.576, "grad_norm": 3.977590560913086, "learning_rate": 1.957867058652812e-06, "loss": 4.9224, "step": 360 }, { "epoch": 0.5776, "grad_norm": 4.341916084289551, "learning_rate": 1.9454678222601545e-06, "loss": 4.727, "step": 361 }, { "epoch": 0.5792, "grad_norm": 2.2049553394317627, "learning_rate": 1.933082915947398e-06, "loss": 4.9447, "step": 362 }, { "epoch": 0.5808, "grad_norm": 5.574019908905029, "learning_rate": 1.9207126597621496e-06, "loss": 4.637, "step": 363 }, { "epoch": 0.5824, "grad_norm": 4.348245143890381, "learning_rate": 1.9083573733734328e-06, "loss": 5.3625, "step": 364 }, { "epoch": 0.584, "grad_norm": 1.8652396202087402, "learning_rate": 1.8960173760634257e-06, "loss": 4.9229, "step": 365 }, { "epoch": 0.5856, "grad_norm": 2.5181169509887695, "learning_rate": 1.8836929867192077e-06, "loss": 4.9309, "step": 366 }, { "epoch": 0.5872, "grad_norm": 1.853685975074768, "learning_rate": 1.8713845238245231e-06, "loss": 4.838, "step": 367 }, { "epoch": 0.5888, "grad_norm": 3.2928309440612793, "learning_rate": 1.8590923054515504e-06, "loss": 4.6349, "step": 368 }, { "epoch": 0.5904, "grad_norm": 2.0537586212158203, "learning_rate": 1.846816649252676e-06, "loss": 5.1173, "step": 369 }, { "epoch": 0.592, "grad_norm": 3.0493924617767334, "learning_rate": 1.8345578724522957e-06, "loss": 5.1656, "step": 370 }, { "epoch": 0.5936, "grad_norm": 1.8361153602600098, "learning_rate": 1.8223162918386122e-06, "loss": 4.7804, "step": 371 }, { "epoch": 0.5952, "grad_norm": 2.6936750411987305, "learning_rate": 1.8100922237554442e-06, "loss": 4.7275, "step": 372 }, { "epoch": 0.5968, "grad_norm": 3.0393989086151123, "learning_rate": 1.7978859840940604e-06, "loss": 4.6155, "step": 373 }, { "epoch": 0.5984, "grad_norm": 2.4623372554779053, "learning_rate": 1.7856978882850112e-06, "loss": 4.7711, "step": 374 }, { "epoch": 0.6, "grad_norm": 1.9320837259292603, "learning_rate": 1.7735282512899757e-06, "loss": 4.898, "step": 375 }, { "epoch": 0.6016, "grad_norm": 3.3518543243408203, "learning_rate": 1.7613773875936274e-06, "loss": 4.5005, "step": 376 }, { "epoch": 0.6032, "grad_norm": 5.2378458976745605, "learning_rate": 1.7492456111955052e-06, "loss": 5.2084, "step": 377 }, { "epoch": 0.6048, "grad_norm": 3.670921802520752, "learning_rate": 1.7371332356018972e-06, "loss": 5.069, "step": 378 }, { "epoch": 0.6064, "grad_norm": 2.2386555671691895, "learning_rate": 1.725040573817742e-06, "loss": 4.6441, "step": 379 }, { "epoch": 0.608, "grad_norm": 1.9244714975357056, "learning_rate": 1.7129679383385384e-06, "loss": 4.7077, "step": 380 }, { "epoch": 0.6096, "grad_norm": 4.245272159576416, "learning_rate": 1.70091564114227e-06, "loss": 5.0572, "step": 381 }, { "epoch": 0.6112, "grad_norm": 1.7141362428665161, "learning_rate": 1.688883993681345e-06, "loss": 4.5579, "step": 382 }, { "epoch": 0.6128, "grad_norm": 1.722749948501587, "learning_rate": 1.6768733068745468e-06, "loss": 4.619, "step": 383 }, { "epoch": 0.6144, "grad_norm": 2.210688352584839, "learning_rate": 1.6648838910989955e-06, "loss": 4.7759, "step": 384 }, { "epoch": 0.616, "grad_norm": 1.5690199136734009, "learning_rate": 1.652916056182134e-06, "loss": 4.6606, "step": 385 }, { "epoch": 0.6176, "grad_norm": 1.748518466949463, "learning_rate": 1.6409701113937182e-06, "loss": 4.7198, "step": 386 }, { "epoch": 0.6192, "grad_norm": 2.298208475112915, "learning_rate": 1.6290463654378213e-06, "loss": 4.9387, "step": 387 }, { "epoch": 0.6208, "grad_norm": 3.15093994140625, "learning_rate": 1.617145126444864e-06, "loss": 4.7299, "step": 388 }, { "epoch": 0.6224, "grad_norm": 2.0222041606903076, "learning_rate": 1.6052667019636462e-06, "loss": 4.6768, "step": 389 }, { "epoch": 0.624, "grad_norm": 3.3168511390686035, "learning_rate": 1.5934113989533992e-06, "loss": 4.9, "step": 390 }, { "epoch": 0.6256, "grad_norm": 2.9965100288391113, "learning_rate": 1.581579523775858e-06, "loss": 4.5604, "step": 391 }, { "epoch": 0.6272, "grad_norm": 1.9449670314788818, "learning_rate": 1.5697713821873401e-06, "loss": 4.7794, "step": 392 }, { "epoch": 0.6288, "grad_norm": 2.7728371620178223, "learning_rate": 1.5579872793308456e-06, "loss": 4.6898, "step": 393 }, { "epoch": 0.6304, "grad_norm": 1.9621694087982178, "learning_rate": 1.5462275197281717e-06, "loss": 4.7706, "step": 394 }, { "epoch": 0.632, "grad_norm": 3.6688954830169678, "learning_rate": 1.5344924072720434e-06, "loss": 5.0143, "step": 395 }, { "epoch": 0.6336, "grad_norm": 1.637315034866333, "learning_rate": 1.5227822452182617e-06, "loss": 4.6788, "step": 396 }, { "epoch": 0.6352, "grad_norm": 2.2460734844207764, "learning_rate": 1.5110973361778637e-06, "loss": 4.8856, "step": 397 }, { "epoch": 0.6368, "grad_norm": 2.410999298095703, "learning_rate": 1.499437982109305e-06, "loss": 4.5615, "step": 398 }, { "epoch": 0.6384, "grad_norm": 3.9339914321899414, "learning_rate": 1.4878044843106582e-06, "loss": 4.2681, "step": 399 }, { "epoch": 0.64, "grad_norm": 2.127648115158081, "learning_rate": 1.4761971434118207e-06, "loss": 4.8065, "step": 400 } ], "logging_steps": 1, "max_steps": 625, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 200, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.80528674619392e+16, "train_batch_size": 16, "trial_name": null, "trial_params": null }