| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 25.0, | |
| "eval_steps": 500, | |
| "global_step": 725, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.1724137931034483, | |
| "grad_norm": 24.532774084242813, | |
| "learning_rate": 1.999849788616454e-05, | |
| "loss": 1.4207, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.3448275862068966, | |
| "grad_norm": 15.039785697497418, | |
| "learning_rate": 1.9992396322115213e-05, | |
| "loss": 0.4531, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.5172413793103449, | |
| "grad_norm": 11.068602275701998, | |
| "learning_rate": 1.9981604287632104e-05, | |
| "loss": 0.2539, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.6896551724137931, | |
| "grad_norm": 10.064753194401916, | |
| "learning_rate": 1.996612684853896e-05, | |
| "loss": 0.2192, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.8620689655172413, | |
| "grad_norm": 11.621703211076142, | |
| "learning_rate": 1.994597127000704e-05, | |
| "loss": 0.2058, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 1.0344827586206897, | |
| "grad_norm": 11.053647982714269, | |
| "learning_rate": 1.9921147013144782e-05, | |
| "loss": 0.2045, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 1.206896551724138, | |
| "grad_norm": 12.029451451469665, | |
| "learning_rate": 1.9891665730556727e-05, | |
| "loss": 0.2097, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 1.3793103448275863, | |
| "grad_norm": 10.75340602132627, | |
| "learning_rate": 1.9857541260873764e-05, | |
| "loss": 0.1976, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 1.5517241379310345, | |
| "grad_norm": 12.305173501446632, | |
| "learning_rate": 1.9818789622257197e-05, | |
| "loss": 0.1954, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 1.7241379310344827, | |
| "grad_norm": 11.967313030536257, | |
| "learning_rate": 1.977542900487977e-05, | |
| "loss": 0.1867, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 1.896551724137931, | |
| "grad_norm": 12.226764504643585, | |
| "learning_rate": 1.9727479762387115e-05, | |
| "loss": 0.1967, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 2.0689655172413794, | |
| "grad_norm": 13.346169442214025, | |
| "learning_rate": 1.9674964402343684e-05, | |
| "loss": 0.1961, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 2.2413793103448274, | |
| "grad_norm": 11.702611356777243, | |
| "learning_rate": 1.9617907575667602e-05, | |
| "loss": 0.2044, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 2.413793103448276, | |
| "grad_norm": 11.798162323499401, | |
| "learning_rate": 1.955633606505943e-05, | |
| "loss": 0.2012, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 2.586206896551724, | |
| "grad_norm": 11.673995798827626, | |
| "learning_rate": 1.9490278772430255e-05, | |
| "loss": 0.1936, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 2.7586206896551726, | |
| "grad_norm": 10.686734043411159, | |
| "learning_rate": 1.9419766705335025e-05, | |
| "loss": 0.1927, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 2.9310344827586206, | |
| "grad_norm": 11.160586116670075, | |
| "learning_rate": 1.9344832962417475e-05, | |
| "loss": 0.1889, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 3.103448275862069, | |
| "grad_norm": 10.769874091338295, | |
| "learning_rate": 1.92655127178735e-05, | |
| "loss": 0.1858, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 3.2758620689655173, | |
| "grad_norm": 11.311883329593053, | |
| "learning_rate": 1.9181843204940232e-05, | |
| "loss": 0.2001, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 3.4482758620689653, | |
| "grad_norm": 12.707306428263719, | |
| "learning_rate": 1.9093863698418627e-05, | |
| "loss": 0.1937, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 3.6206896551724137, | |
| "grad_norm": 13.066551084589346, | |
| "learning_rate": 1.9001615496237714e-05, | |
| "loss": 0.2021, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 3.793103448275862, | |
| "grad_norm": 11.526972333003544, | |
| "learning_rate": 1.890514190006918e-05, | |
| "loss": 0.1961, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 3.9655172413793105, | |
| "grad_norm": 10.492067731277176, | |
| "learning_rate": 1.8804488195001394e-05, | |
| "loss": 0.2002, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 4.137931034482759, | |
| "grad_norm": 11.0280237578772, | |
| "learning_rate": 1.869970162828241e-05, | |
| "loss": 0.201, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 4.310344827586207, | |
| "grad_norm": 13.09901469391884, | |
| "learning_rate": 1.859083138714191e-05, | |
| "loss": 0.1916, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 4.482758620689655, | |
| "grad_norm": 9.77218748564033, | |
| "learning_rate": 1.847792857570255e-05, | |
| "loss": 0.1819, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 4.655172413793103, | |
| "grad_norm": 10.397235406767715, | |
| "learning_rate": 1.8361046190991457e-05, | |
| "loss": 0.1924, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 4.827586206896552, | |
| "grad_norm": 9.046262037311052, | |
| "learning_rate": 1.824023909806322e-05, | |
| "loss": 0.1813, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "grad_norm": 9.373790580903188, | |
| "learning_rate": 1.8115564004246025e-05, | |
| "loss": 0.1814, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 5.172413793103448, | |
| "grad_norm": 10.125444036531317, | |
| "learning_rate": 1.7987079432522997e-05, | |
| "loss": 0.1841, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 5.344827586206897, | |
| "grad_norm": 12.214021854754943, | |
| "learning_rate": 1.7854845694061294e-05, | |
| "loss": 0.1827, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 5.517241379310345, | |
| "grad_norm": 11.231831307080368, | |
| "learning_rate": 1.7718924859901793e-05, | |
| "loss": 0.1905, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 5.689655172413794, | |
| "grad_norm": 10.970143913905503, | |
| "learning_rate": 1.7579380731822712e-05, | |
| "loss": 0.1877, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 5.862068965517241, | |
| "grad_norm": 10.553953229075919, | |
| "learning_rate": 1.7436278812390788e-05, | |
| "loss": 0.1973, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 6.0344827586206895, | |
| "grad_norm": 9.844282254513766, | |
| "learning_rate": 1.7289686274214116e-05, | |
| "loss": 0.1912, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 6.206896551724138, | |
| "grad_norm": 10.48046238068855, | |
| "learning_rate": 1.7139671928411074e-05, | |
| "loss": 0.1895, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 6.379310344827586, | |
| "grad_norm": 10.884308915720293, | |
| "learning_rate": 1.6986306192310086e-05, | |
| "loss": 0.1931, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 6.551724137931035, | |
| "grad_norm": 10.530508782873323, | |
| "learning_rate": 1.6829661056395473e-05, | |
| "loss": 0.1845, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 6.724137931034483, | |
| "grad_norm": 8.923866337916758, | |
| "learning_rate": 1.6669810050514827e-05, | |
| "loss": 0.1805, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 6.896551724137931, | |
| "grad_norm": 10.099322440099796, | |
| "learning_rate": 1.6506828209363796e-05, | |
| "loss": 0.1772, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 7.068965517241379, | |
| "grad_norm": 10.458521429491142, | |
| "learning_rate": 1.634079203726453e-05, | |
| "loss": 0.1743, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 7.241379310344827, | |
| "grad_norm": 10.463312384855938, | |
| "learning_rate": 1.6171779472254206e-05, | |
| "loss": 0.1929, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 7.413793103448276, | |
| "grad_norm": 10.118912810787648, | |
| "learning_rate": 1.599986984950065e-05, | |
| "loss": 0.1805, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 7.586206896551724, | |
| "grad_norm": 9.327155921513027, | |
| "learning_rate": 1.5825143864062063e-05, | |
| "loss": 0.1727, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 7.758620689655173, | |
| "grad_norm": 8.24932430986195, | |
| "learning_rate": 1.5647683533008455e-05, | |
| "loss": 0.1864, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 7.931034482758621, | |
| "grad_norm": 10.172474933867472, | |
| "learning_rate": 1.5467572156922504e-05, | |
| "loss": 0.1764, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 8.10344827586207, | |
| "grad_norm": 9.969389715106933, | |
| "learning_rate": 1.528489428079793e-05, | |
| "loss": 0.1748, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 8.275862068965518, | |
| "grad_norm": 10.00608688194569, | |
| "learning_rate": 1.509973565435375e-05, | |
| "loss": 0.1841, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 8.448275862068966, | |
| "grad_norm": 9.994311879836584, | |
| "learning_rate": 1.4912183191782995e-05, | |
| "loss": 0.185, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 8.620689655172415, | |
| "grad_norm": 8.967701674753974, | |
| "learning_rate": 1.4722324930954885e-05, | |
| "loss": 0.1873, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 8.793103448275861, | |
| "grad_norm": 8.61339221417413, | |
| "learning_rate": 1.453024999208946e-05, | |
| "loss": 0.172, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 8.96551724137931, | |
| "grad_norm": 9.113590523548789, | |
| "learning_rate": 1.4336048535924223e-05, | |
| "loss": 0.177, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 9.137931034482758, | |
| "grad_norm": 6.366907049562298, | |
| "learning_rate": 1.4139811721392325e-05, | |
| "loss": 0.1766, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 9.310344827586206, | |
| "grad_norm": 8.257166469089526, | |
| "learning_rate": 1.3941631662832201e-05, | |
| "loss": 0.1738, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 9.482758620689655, | |
| "grad_norm": 9.95233344460982, | |
| "learning_rate": 1.3741601386748728e-05, | |
| "loss": 0.1681, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 9.655172413793103, | |
| "grad_norm": 9.655306228806518, | |
| "learning_rate": 1.3539814788146235e-05, | |
| "loss": 0.1866, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 9.827586206896552, | |
| "grad_norm": 8.289931367954654, | |
| "learning_rate": 1.3336366586453783e-05, | |
| "loss": 0.1913, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "grad_norm": 8.252904158335944, | |
| "learning_rate": 1.313135228106353e-05, | |
| "loss": 0.1829, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 10.172413793103448, | |
| "grad_norm": 10.550290488102121, | |
| "learning_rate": 1.292486810650289e-05, | |
| "loss": 0.178, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 10.344827586206897, | |
| "grad_norm": 8.345722020395192, | |
| "learning_rate": 1.2717010987261716e-05, | |
| "loss": 0.1738, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 10.517241379310345, | |
| "grad_norm": 8.134962325536728, | |
| "learning_rate": 1.250787849229552e-05, | |
| "loss": 0.1754, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 10.689655172413794, | |
| "grad_norm": 8.815840186959745, | |
| "learning_rate": 1.229756878922624e-05, | |
| "loss": 0.1761, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 10.862068965517242, | |
| "grad_norm": 8.690893129618297, | |
| "learning_rate": 1.2086180598261956e-05, | |
| "loss": 0.1723, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 11.03448275862069, | |
| "grad_norm": 8.231717795365554, | |
| "learning_rate": 1.187381314585725e-05, | |
| "loss": 0.1683, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 11.206896551724139, | |
| "grad_norm": 8.935068264024606, | |
| "learning_rate": 1.1660566118135894e-05, | |
| "loss": 0.1777, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 11.379310344827585, | |
| "grad_norm": 8.235723206312889, | |
| "learning_rate": 1.1446539614097814e-05, | |
| "loss": 0.1744, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 11.551724137931034, | |
| "grad_norm": 9.472032012397955, | |
| "learning_rate": 1.123183409863219e-05, | |
| "loss": 0.1786, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 11.724137931034482, | |
| "grad_norm": 7.458462811267308, | |
| "learning_rate": 1.1016550355358872e-05, | |
| "loss": 0.1782, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 11.89655172413793, | |
| "grad_norm": 8.818348138750482, | |
| "learning_rate": 1.0800789439320128e-05, | |
| "loss": 0.1699, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 12.068965517241379, | |
| "grad_norm": 7.046157313918523, | |
| "learning_rate": 1.0584652629545011e-05, | |
| "loss": 0.167, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 12.241379310344827, | |
| "grad_norm": 7.5773668951999555, | |
| "learning_rate": 1.036824138150859e-05, | |
| "loss": 0.1724, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 12.413793103448276, | |
| "grad_norm": 8.439620308050772, | |
| "learning_rate": 1.0151657279508335e-05, | |
| "loss": 0.1679, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 12.586206896551724, | |
| "grad_norm": 7.182234665957953, | |
| "learning_rate": 9.93500198898006e-06, | |
| "loss": 0.1683, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 12.758620689655173, | |
| "grad_norm": 7.45841562952522, | |
| "learning_rate": 9.718377208775744e-06, | |
| "loss": 0.1679, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 12.931034482758621, | |
| "grad_norm": 9.05065663213446, | |
| "learning_rate": 9.50188462342571e-06, | |
| "loss": 0.1679, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 13.10344827586207, | |
| "grad_norm": 8.609074622939076, | |
| "learning_rate": 9.285625855407485e-06, | |
| "loss": 0.1742, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 13.275862068965518, | |
| "grad_norm": 8.2061328299682, | |
| "learning_rate": 9.069702417443821e-06, | |
| "loss": 0.1732, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 13.448275862068966, | |
| "grad_norm": 7.900839577431847, | |
| "learning_rate": 8.854215664852207e-06, | |
| "loss": 0.171, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 13.620689655172415, | |
| "grad_norm": 9.118846376646786, | |
| "learning_rate": 8.63926674796829e-06, | |
| "loss": 0.1655, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 13.793103448275861, | |
| "grad_norm": 7.4554871018559945, | |
| "learning_rate": 8.424956564665508e-06, | |
| "loss": 0.1676, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 13.96551724137931, | |
| "grad_norm": 7.834220281291942, | |
| "learning_rate": 8.211385712993219e-06, | |
| "loss": 0.1683, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 14.137931034482758, | |
| "grad_norm": 7.057089973676238, | |
| "learning_rate": 7.998654443955586e-06, | |
| "loss": 0.1721, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 14.310344827586206, | |
| "grad_norm": 7.010954237310117, | |
| "learning_rate": 7.786862614453356e-06, | |
| "loss": 0.164, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 14.482758620689655, | |
| "grad_norm": 8.54121625744651, | |
| "learning_rate": 7.5761096404106335e-06, | |
| "loss": 0.1632, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 14.655172413793103, | |
| "grad_norm": 7.32836259078595, | |
| "learning_rate": 7.366494450108659e-06, | |
| "loss": 0.1653, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 14.827586206896552, | |
| "grad_norm": 6.161874927822607, | |
| "learning_rate": 7.158115437748467e-06, | |
| "loss": 0.1718, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "grad_norm": 5.960687988714388, | |
| "learning_rate": 6.951070417264278e-06, | |
| "loss": 0.1706, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 15.172413793103448, | |
| "grad_norm": 6.367501795811549, | |
| "learning_rate": 6.745456576409227e-06, | |
| "loss": 0.1611, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 15.344827586206897, | |
| "grad_norm": 5.06309133432644, | |
| "learning_rate": 6.541370431135073e-06, | |
| "loss": 0.1734, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 15.517241379310345, | |
| "grad_norm": 6.239307590864818, | |
| "learning_rate": 6.338907780287198e-06, | |
| "loss": 0.1753, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 15.689655172413794, | |
| "grad_norm": 7.09367607101717, | |
| "learning_rate": 6.138163660636285e-06, | |
| "loss": 0.1718, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 15.862068965517242, | |
| "grad_norm": 7.637051833850522, | |
| "learning_rate": 5.939232302267646e-06, | |
| "loss": 0.1613, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 16.03448275862069, | |
| "grad_norm": 6.919836610970123, | |
| "learning_rate": 5.742207084349274e-06, | |
| "loss": 0.1667, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 16.20689655172414, | |
| "grad_norm": 6.615501006073225, | |
| "learning_rate": 5.547180491299278e-06, | |
| "loss": 0.1653, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 16.379310344827587, | |
| "grad_norm": 6.3443598347944725, | |
| "learning_rate": 5.35424406937333e-06, | |
| "loss": 0.1691, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 16.551724137931036, | |
| "grad_norm": 6.545596807021073, | |
| "learning_rate": 5.163488383692499e-06, | |
| "loss": 0.167, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 16.724137931034484, | |
| "grad_norm": 5.794359907918518, | |
| "learning_rate": 4.975002975731613e-06, | |
| "loss": 0.1636, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 16.896551724137932, | |
| "grad_norm": 5.30655880304234, | |
| "learning_rate": 4.78887632128814e-06, | |
| "loss": 0.1732, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 17.06896551724138, | |
| "grad_norm": 5.564470023655747, | |
| "learning_rate": 4.6051957889513e-06, | |
| "loss": 0.1705, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 17.24137931034483, | |
| "grad_norm": 5.79246441022243, | |
| "learning_rate": 4.42404759909091e-06, | |
| "loss": 0.1739, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 17.413793103448278, | |
| "grad_norm": 4.968884353779036, | |
| "learning_rate": 4.2455167833851804e-06, | |
| "loss": 0.1684, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 17.586206896551722, | |
| "grad_norm": 5.820327270915885, | |
| "learning_rate": 4.069687144906532e-06, | |
| "loss": 0.1679, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 17.75862068965517, | |
| "grad_norm": 4.874594537233877, | |
| "learning_rate": 3.896641218784081e-06, | |
| "loss": 0.1671, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 17.93103448275862, | |
| "grad_norm": 4.322080959732308, | |
| "learning_rate": 3.7264602334613385e-06, | |
| "loss": 0.1615, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 18.103448275862068, | |
| "grad_norm": 3.394741423143991, | |
| "learning_rate": 3.5592240725672476e-06, | |
| "loss": 0.163, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 18.275862068965516, | |
| "grad_norm": 2.8899715660579917, | |
| "learning_rate": 3.395011237418494e-06, | |
| "loss": 0.1643, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 18.448275862068964, | |
| "grad_norm": 2.66333363991394, | |
| "learning_rate": 3.2338988101706727e-06, | |
| "loss": 0.1598, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 18.620689655172413, | |
| "grad_norm": 3.0652989937592787, | |
| "learning_rate": 3.075962417635634e-06, | |
| "loss": 0.1616, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 18.79310344827586, | |
| "grad_norm": 3.8408972623600373, | |
| "learning_rate": 2.9212761957819347e-06, | |
| "loss": 0.1673, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 18.96551724137931, | |
| "grad_norm": 3.031323684366623, | |
| "learning_rate": 2.769912754935146e-06, | |
| "loss": 0.1588, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 19.137931034482758, | |
| "grad_norm": 2.4766730582592027, | |
| "learning_rate": 2.6219431456942536e-06, | |
| "loss": 0.1626, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 19.310344827586206, | |
| "grad_norm": 3.347536737141708, | |
| "learning_rate": 2.4774368255802483e-06, | |
| "loss": 0.1607, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 19.482758620689655, | |
| "grad_norm": 2.75609785484982, | |
| "learning_rate": 2.3364616264324725e-06, | |
| "loss": 0.1664, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 19.655172413793103, | |
| "grad_norm": 3.7333685782452335, | |
| "learning_rate": 2.199083722568095e-06, | |
| "loss": 0.1634, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 19.82758620689655, | |
| "grad_norm": 3.008810658711964, | |
| "learning_rate": 2.065367599719621e-06, | |
| "loss": 0.1686, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "grad_norm": 3.4023386470750996, | |
| "learning_rate": 1.93537602476504e-06, | |
| "loss": 0.1692, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 20.17241379310345, | |
| "grad_norm": 2.9763118671554314, | |
| "learning_rate": 1.809170016264794e-06, | |
| "loss": 0.1768, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 20.344827586206897, | |
| "grad_norm": 2.7215403943473957, | |
| "learning_rate": 1.6868088158194351e-06, | |
| "loss": 0.1677, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 20.517241379310345, | |
| "grad_norm": 2.48039880730862, | |
| "learning_rate": 1.5683498602613689e-06, | |
| "loss": 0.1638, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 20.689655172413794, | |
| "grad_norm": 1.8624071329024212, | |
| "learning_rate": 1.4538487546937951e-06, | |
| "loss": 0.1687, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 20.862068965517242, | |
| "grad_norm": 2.871924413048312, | |
| "learning_rate": 1.3433592463894375e-06, | |
| "loss": 0.1617, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 21.03448275862069, | |
| "grad_norm": 2.7255286515937667, | |
| "learning_rate": 1.2369331995613664e-06, | |
| "loss": 0.1622, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 21.20689655172414, | |
| "grad_norm": 2.642540001372744, | |
| "learning_rate": 1.1346205710177304e-06, | |
| "loss": 0.1608, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 21.379310344827587, | |
| "grad_norm": 2.975135375218524, | |
| "learning_rate": 1.0364693867118425e-06, | |
| "loss": 0.1644, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 21.551724137931036, | |
| "grad_norm": 2.0876137084099784, | |
| "learning_rate": 9.425257191985859e-07, | |
| "loss": 0.1639, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 21.724137931034484, | |
| "grad_norm": 1.387475835397412, | |
| "learning_rate": 8.528336660077974e-07, | |
| "loss": 0.1677, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 21.896551724137932, | |
| "grad_norm": 2.869659930472805, | |
| "learning_rate": 7.674353289446946e-07, | |
| "loss": 0.1621, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 22.06896551724138, | |
| "grad_norm": 3.0026307802656196, | |
| "learning_rate": 6.863707943271325e-07, | |
| "loss": 0.1655, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 22.24137931034483, | |
| "grad_norm": 1.8278436522298838, | |
| "learning_rate": 6.096781141689223e-07, | |
| "loss": 0.163, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 22.413793103448278, | |
| "grad_norm": 2.0229424538116145, | |
| "learning_rate": 5.373932883180655e-07, | |
| "loss": 0.1665, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 22.586206896551722, | |
| "grad_norm": 2.1750833091387043, | |
| "learning_rate": 4.695502475582814e-07, | |
| "loss": 0.1611, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 22.75862068965517, | |
| "grad_norm": 1.948795256385187, | |
| "learning_rate": 4.0618083768176996e-07, | |
| "loss": 0.1619, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 22.93103448275862, | |
| "grad_norm": 1.5695625649630254, | |
| "learning_rate": 3.4731480454065823e-07, | |
| "loss": 0.1621, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 23.103448275862068, | |
| "grad_norm": 1.5429369658213046, | |
| "learning_rate": 2.9297978008419604e-07, | |
| "loss": 0.1687, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 23.275862068965516, | |
| "grad_norm": 2.0587233544995325, | |
| "learning_rate": 2.4320126938819023e-07, | |
| "loss": 0.1573, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 23.448275862068964, | |
| "grad_norm": 1.6851058427704997, | |
| "learning_rate": 1.9800263868283708e-07, | |
| "loss": 0.1651, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 23.620689655172413, | |
| "grad_norm": 1.2703207998322799, | |
| "learning_rate": 1.5740510438451374e-07, | |
| "loss": 0.1646, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 23.79310344827586, | |
| "grad_norm": 1.6546893474906568, | |
| "learning_rate": 1.214277231367078e-07, | |
| "loss": 0.1672, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 23.96551724137931, | |
| "grad_norm": 1.4962982730142882, | |
| "learning_rate": 9.008738286475748e-08, | |
| "loss": 0.1621, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 24.137931034482758, | |
| "grad_norm": 1.5704585386814611, | |
| "learning_rate": 6.339879484858924e-08, | |
| "loss": 0.1675, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 24.310344827586206, | |
| "grad_norm": 1.367562782086171, | |
| "learning_rate": 4.137448681718392e-08, | |
| "loss": 0.1637, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 24.482758620689655, | |
| "grad_norm": 1.3662424883362485, | |
| "learning_rate": 2.4024797068017414e-08, | |
| "loss": 0.1625, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 24.655172413793103, | |
| "grad_norm": 1.7154032971232247, | |
| "learning_rate": 1.135786961421248e-08, | |
| "loss": 0.1692, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 24.82758620689655, | |
| "grad_norm": 0.9411353813301526, | |
| "learning_rate": 3.3796503617167244e-09, | |
| "loss": 0.1681, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "grad_norm": 1.2181390829757093, | |
| "learning_rate": 9.388431826629074e-11, | |
| "loss": 0.1725, | |
| "step": 725 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 725, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 25, | |
| "save_steps": 200, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 50378664312832.0, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |