blockincup_oft / trainer_state.json
ymdou's picture
Upload folder using huggingface_hub
6896f12 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 25.0,
"eval_steps": 500,
"global_step": 725,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.1724137931034483,
"grad_norm": 24.532774084242813,
"learning_rate": 1.999849788616454e-05,
"loss": 1.4207,
"step": 5
},
{
"epoch": 0.3448275862068966,
"grad_norm": 15.039785697497418,
"learning_rate": 1.9992396322115213e-05,
"loss": 0.4531,
"step": 10
},
{
"epoch": 0.5172413793103449,
"grad_norm": 11.068602275701998,
"learning_rate": 1.9981604287632104e-05,
"loss": 0.2539,
"step": 15
},
{
"epoch": 0.6896551724137931,
"grad_norm": 10.064753194401916,
"learning_rate": 1.996612684853896e-05,
"loss": 0.2192,
"step": 20
},
{
"epoch": 0.8620689655172413,
"grad_norm": 11.621703211076142,
"learning_rate": 1.994597127000704e-05,
"loss": 0.2058,
"step": 25
},
{
"epoch": 1.0344827586206897,
"grad_norm": 11.053647982714269,
"learning_rate": 1.9921147013144782e-05,
"loss": 0.2045,
"step": 30
},
{
"epoch": 1.206896551724138,
"grad_norm": 12.029451451469665,
"learning_rate": 1.9891665730556727e-05,
"loss": 0.2097,
"step": 35
},
{
"epoch": 1.3793103448275863,
"grad_norm": 10.75340602132627,
"learning_rate": 1.9857541260873764e-05,
"loss": 0.1976,
"step": 40
},
{
"epoch": 1.5517241379310345,
"grad_norm": 12.305173501446632,
"learning_rate": 1.9818789622257197e-05,
"loss": 0.1954,
"step": 45
},
{
"epoch": 1.7241379310344827,
"grad_norm": 11.967313030536257,
"learning_rate": 1.977542900487977e-05,
"loss": 0.1867,
"step": 50
},
{
"epoch": 1.896551724137931,
"grad_norm": 12.226764504643585,
"learning_rate": 1.9727479762387115e-05,
"loss": 0.1967,
"step": 55
},
{
"epoch": 2.0689655172413794,
"grad_norm": 13.346169442214025,
"learning_rate": 1.9674964402343684e-05,
"loss": 0.1961,
"step": 60
},
{
"epoch": 2.2413793103448274,
"grad_norm": 11.702611356777243,
"learning_rate": 1.9617907575667602e-05,
"loss": 0.2044,
"step": 65
},
{
"epoch": 2.413793103448276,
"grad_norm": 11.798162323499401,
"learning_rate": 1.955633606505943e-05,
"loss": 0.2012,
"step": 70
},
{
"epoch": 2.586206896551724,
"grad_norm": 11.673995798827626,
"learning_rate": 1.9490278772430255e-05,
"loss": 0.1936,
"step": 75
},
{
"epoch": 2.7586206896551726,
"grad_norm": 10.686734043411159,
"learning_rate": 1.9419766705335025e-05,
"loss": 0.1927,
"step": 80
},
{
"epoch": 2.9310344827586206,
"grad_norm": 11.160586116670075,
"learning_rate": 1.9344832962417475e-05,
"loss": 0.1889,
"step": 85
},
{
"epoch": 3.103448275862069,
"grad_norm": 10.769874091338295,
"learning_rate": 1.92655127178735e-05,
"loss": 0.1858,
"step": 90
},
{
"epoch": 3.2758620689655173,
"grad_norm": 11.311883329593053,
"learning_rate": 1.9181843204940232e-05,
"loss": 0.2001,
"step": 95
},
{
"epoch": 3.4482758620689653,
"grad_norm": 12.707306428263719,
"learning_rate": 1.9093863698418627e-05,
"loss": 0.1937,
"step": 100
},
{
"epoch": 3.6206896551724137,
"grad_norm": 13.066551084589346,
"learning_rate": 1.9001615496237714e-05,
"loss": 0.2021,
"step": 105
},
{
"epoch": 3.793103448275862,
"grad_norm": 11.526972333003544,
"learning_rate": 1.890514190006918e-05,
"loss": 0.1961,
"step": 110
},
{
"epoch": 3.9655172413793105,
"grad_norm": 10.492067731277176,
"learning_rate": 1.8804488195001394e-05,
"loss": 0.2002,
"step": 115
},
{
"epoch": 4.137931034482759,
"grad_norm": 11.0280237578772,
"learning_rate": 1.869970162828241e-05,
"loss": 0.201,
"step": 120
},
{
"epoch": 4.310344827586207,
"grad_norm": 13.09901469391884,
"learning_rate": 1.859083138714191e-05,
"loss": 0.1916,
"step": 125
},
{
"epoch": 4.482758620689655,
"grad_norm": 9.77218748564033,
"learning_rate": 1.847792857570255e-05,
"loss": 0.1819,
"step": 130
},
{
"epoch": 4.655172413793103,
"grad_norm": 10.397235406767715,
"learning_rate": 1.8361046190991457e-05,
"loss": 0.1924,
"step": 135
},
{
"epoch": 4.827586206896552,
"grad_norm": 9.046262037311052,
"learning_rate": 1.824023909806322e-05,
"loss": 0.1813,
"step": 140
},
{
"epoch": 5.0,
"grad_norm": 9.373790580903188,
"learning_rate": 1.8115564004246025e-05,
"loss": 0.1814,
"step": 145
},
{
"epoch": 5.172413793103448,
"grad_norm": 10.125444036531317,
"learning_rate": 1.7987079432522997e-05,
"loss": 0.1841,
"step": 150
},
{
"epoch": 5.344827586206897,
"grad_norm": 12.214021854754943,
"learning_rate": 1.7854845694061294e-05,
"loss": 0.1827,
"step": 155
},
{
"epoch": 5.517241379310345,
"grad_norm": 11.231831307080368,
"learning_rate": 1.7718924859901793e-05,
"loss": 0.1905,
"step": 160
},
{
"epoch": 5.689655172413794,
"grad_norm": 10.970143913905503,
"learning_rate": 1.7579380731822712e-05,
"loss": 0.1877,
"step": 165
},
{
"epoch": 5.862068965517241,
"grad_norm": 10.553953229075919,
"learning_rate": 1.7436278812390788e-05,
"loss": 0.1973,
"step": 170
},
{
"epoch": 6.0344827586206895,
"grad_norm": 9.844282254513766,
"learning_rate": 1.7289686274214116e-05,
"loss": 0.1912,
"step": 175
},
{
"epoch": 6.206896551724138,
"grad_norm": 10.48046238068855,
"learning_rate": 1.7139671928411074e-05,
"loss": 0.1895,
"step": 180
},
{
"epoch": 6.379310344827586,
"grad_norm": 10.884308915720293,
"learning_rate": 1.6986306192310086e-05,
"loss": 0.1931,
"step": 185
},
{
"epoch": 6.551724137931035,
"grad_norm": 10.530508782873323,
"learning_rate": 1.6829661056395473e-05,
"loss": 0.1845,
"step": 190
},
{
"epoch": 6.724137931034483,
"grad_norm": 8.923866337916758,
"learning_rate": 1.6669810050514827e-05,
"loss": 0.1805,
"step": 195
},
{
"epoch": 6.896551724137931,
"grad_norm": 10.099322440099796,
"learning_rate": 1.6506828209363796e-05,
"loss": 0.1772,
"step": 200
},
{
"epoch": 7.068965517241379,
"grad_norm": 10.458521429491142,
"learning_rate": 1.634079203726453e-05,
"loss": 0.1743,
"step": 205
},
{
"epoch": 7.241379310344827,
"grad_norm": 10.463312384855938,
"learning_rate": 1.6171779472254206e-05,
"loss": 0.1929,
"step": 210
},
{
"epoch": 7.413793103448276,
"grad_norm": 10.118912810787648,
"learning_rate": 1.599986984950065e-05,
"loss": 0.1805,
"step": 215
},
{
"epoch": 7.586206896551724,
"grad_norm": 9.327155921513027,
"learning_rate": 1.5825143864062063e-05,
"loss": 0.1727,
"step": 220
},
{
"epoch": 7.758620689655173,
"grad_norm": 8.24932430986195,
"learning_rate": 1.5647683533008455e-05,
"loss": 0.1864,
"step": 225
},
{
"epoch": 7.931034482758621,
"grad_norm": 10.172474933867472,
"learning_rate": 1.5467572156922504e-05,
"loss": 0.1764,
"step": 230
},
{
"epoch": 8.10344827586207,
"grad_norm": 9.969389715106933,
"learning_rate": 1.528489428079793e-05,
"loss": 0.1748,
"step": 235
},
{
"epoch": 8.275862068965518,
"grad_norm": 10.00608688194569,
"learning_rate": 1.509973565435375e-05,
"loss": 0.1841,
"step": 240
},
{
"epoch": 8.448275862068966,
"grad_norm": 9.994311879836584,
"learning_rate": 1.4912183191782995e-05,
"loss": 0.185,
"step": 245
},
{
"epoch": 8.620689655172415,
"grad_norm": 8.967701674753974,
"learning_rate": 1.4722324930954885e-05,
"loss": 0.1873,
"step": 250
},
{
"epoch": 8.793103448275861,
"grad_norm": 8.61339221417413,
"learning_rate": 1.453024999208946e-05,
"loss": 0.172,
"step": 255
},
{
"epoch": 8.96551724137931,
"grad_norm": 9.113590523548789,
"learning_rate": 1.4336048535924223e-05,
"loss": 0.177,
"step": 260
},
{
"epoch": 9.137931034482758,
"grad_norm": 6.366907049562298,
"learning_rate": 1.4139811721392325e-05,
"loss": 0.1766,
"step": 265
},
{
"epoch": 9.310344827586206,
"grad_norm": 8.257166469089526,
"learning_rate": 1.3941631662832201e-05,
"loss": 0.1738,
"step": 270
},
{
"epoch": 9.482758620689655,
"grad_norm": 9.95233344460982,
"learning_rate": 1.3741601386748728e-05,
"loss": 0.1681,
"step": 275
},
{
"epoch": 9.655172413793103,
"grad_norm": 9.655306228806518,
"learning_rate": 1.3539814788146235e-05,
"loss": 0.1866,
"step": 280
},
{
"epoch": 9.827586206896552,
"grad_norm": 8.289931367954654,
"learning_rate": 1.3336366586453783e-05,
"loss": 0.1913,
"step": 285
},
{
"epoch": 10.0,
"grad_norm": 8.252904158335944,
"learning_rate": 1.313135228106353e-05,
"loss": 0.1829,
"step": 290
},
{
"epoch": 10.172413793103448,
"grad_norm": 10.550290488102121,
"learning_rate": 1.292486810650289e-05,
"loss": 0.178,
"step": 295
},
{
"epoch": 10.344827586206897,
"grad_norm": 8.345722020395192,
"learning_rate": 1.2717010987261716e-05,
"loss": 0.1738,
"step": 300
},
{
"epoch": 10.517241379310345,
"grad_norm": 8.134962325536728,
"learning_rate": 1.250787849229552e-05,
"loss": 0.1754,
"step": 305
},
{
"epoch": 10.689655172413794,
"grad_norm": 8.815840186959745,
"learning_rate": 1.229756878922624e-05,
"loss": 0.1761,
"step": 310
},
{
"epoch": 10.862068965517242,
"grad_norm": 8.690893129618297,
"learning_rate": 1.2086180598261956e-05,
"loss": 0.1723,
"step": 315
},
{
"epoch": 11.03448275862069,
"grad_norm": 8.231717795365554,
"learning_rate": 1.187381314585725e-05,
"loss": 0.1683,
"step": 320
},
{
"epoch": 11.206896551724139,
"grad_norm": 8.935068264024606,
"learning_rate": 1.1660566118135894e-05,
"loss": 0.1777,
"step": 325
},
{
"epoch": 11.379310344827585,
"grad_norm": 8.235723206312889,
"learning_rate": 1.1446539614097814e-05,
"loss": 0.1744,
"step": 330
},
{
"epoch": 11.551724137931034,
"grad_norm": 9.472032012397955,
"learning_rate": 1.123183409863219e-05,
"loss": 0.1786,
"step": 335
},
{
"epoch": 11.724137931034482,
"grad_norm": 7.458462811267308,
"learning_rate": 1.1016550355358872e-05,
"loss": 0.1782,
"step": 340
},
{
"epoch": 11.89655172413793,
"grad_norm": 8.818348138750482,
"learning_rate": 1.0800789439320128e-05,
"loss": 0.1699,
"step": 345
},
{
"epoch": 12.068965517241379,
"grad_norm": 7.046157313918523,
"learning_rate": 1.0584652629545011e-05,
"loss": 0.167,
"step": 350
},
{
"epoch": 12.241379310344827,
"grad_norm": 7.5773668951999555,
"learning_rate": 1.036824138150859e-05,
"loss": 0.1724,
"step": 355
},
{
"epoch": 12.413793103448276,
"grad_norm": 8.439620308050772,
"learning_rate": 1.0151657279508335e-05,
"loss": 0.1679,
"step": 360
},
{
"epoch": 12.586206896551724,
"grad_norm": 7.182234665957953,
"learning_rate": 9.93500198898006e-06,
"loss": 0.1683,
"step": 365
},
{
"epoch": 12.758620689655173,
"grad_norm": 7.45841562952522,
"learning_rate": 9.718377208775744e-06,
"loss": 0.1679,
"step": 370
},
{
"epoch": 12.931034482758621,
"grad_norm": 9.05065663213446,
"learning_rate": 9.50188462342571e-06,
"loss": 0.1679,
"step": 375
},
{
"epoch": 13.10344827586207,
"grad_norm": 8.609074622939076,
"learning_rate": 9.285625855407485e-06,
"loss": 0.1742,
"step": 380
},
{
"epoch": 13.275862068965518,
"grad_norm": 8.2061328299682,
"learning_rate": 9.069702417443821e-06,
"loss": 0.1732,
"step": 385
},
{
"epoch": 13.448275862068966,
"grad_norm": 7.900839577431847,
"learning_rate": 8.854215664852207e-06,
"loss": 0.171,
"step": 390
},
{
"epoch": 13.620689655172415,
"grad_norm": 9.118846376646786,
"learning_rate": 8.63926674796829e-06,
"loss": 0.1655,
"step": 395
},
{
"epoch": 13.793103448275861,
"grad_norm": 7.4554871018559945,
"learning_rate": 8.424956564665508e-06,
"loss": 0.1676,
"step": 400
},
{
"epoch": 13.96551724137931,
"grad_norm": 7.834220281291942,
"learning_rate": 8.211385712993219e-06,
"loss": 0.1683,
"step": 405
},
{
"epoch": 14.137931034482758,
"grad_norm": 7.057089973676238,
"learning_rate": 7.998654443955586e-06,
"loss": 0.1721,
"step": 410
},
{
"epoch": 14.310344827586206,
"grad_norm": 7.010954237310117,
"learning_rate": 7.786862614453356e-06,
"loss": 0.164,
"step": 415
},
{
"epoch": 14.482758620689655,
"grad_norm": 8.54121625744651,
"learning_rate": 7.5761096404106335e-06,
"loss": 0.1632,
"step": 420
},
{
"epoch": 14.655172413793103,
"grad_norm": 7.32836259078595,
"learning_rate": 7.366494450108659e-06,
"loss": 0.1653,
"step": 425
},
{
"epoch": 14.827586206896552,
"grad_norm": 6.161874927822607,
"learning_rate": 7.158115437748467e-06,
"loss": 0.1718,
"step": 430
},
{
"epoch": 15.0,
"grad_norm": 5.960687988714388,
"learning_rate": 6.951070417264278e-06,
"loss": 0.1706,
"step": 435
},
{
"epoch": 15.172413793103448,
"grad_norm": 6.367501795811549,
"learning_rate": 6.745456576409227e-06,
"loss": 0.1611,
"step": 440
},
{
"epoch": 15.344827586206897,
"grad_norm": 5.06309133432644,
"learning_rate": 6.541370431135073e-06,
"loss": 0.1734,
"step": 445
},
{
"epoch": 15.517241379310345,
"grad_norm": 6.239307590864818,
"learning_rate": 6.338907780287198e-06,
"loss": 0.1753,
"step": 450
},
{
"epoch": 15.689655172413794,
"grad_norm": 7.09367607101717,
"learning_rate": 6.138163660636285e-06,
"loss": 0.1718,
"step": 455
},
{
"epoch": 15.862068965517242,
"grad_norm": 7.637051833850522,
"learning_rate": 5.939232302267646e-06,
"loss": 0.1613,
"step": 460
},
{
"epoch": 16.03448275862069,
"grad_norm": 6.919836610970123,
"learning_rate": 5.742207084349274e-06,
"loss": 0.1667,
"step": 465
},
{
"epoch": 16.20689655172414,
"grad_norm": 6.615501006073225,
"learning_rate": 5.547180491299278e-06,
"loss": 0.1653,
"step": 470
},
{
"epoch": 16.379310344827587,
"grad_norm": 6.3443598347944725,
"learning_rate": 5.35424406937333e-06,
"loss": 0.1691,
"step": 475
},
{
"epoch": 16.551724137931036,
"grad_norm": 6.545596807021073,
"learning_rate": 5.163488383692499e-06,
"loss": 0.167,
"step": 480
},
{
"epoch": 16.724137931034484,
"grad_norm": 5.794359907918518,
"learning_rate": 4.975002975731613e-06,
"loss": 0.1636,
"step": 485
},
{
"epoch": 16.896551724137932,
"grad_norm": 5.30655880304234,
"learning_rate": 4.78887632128814e-06,
"loss": 0.1732,
"step": 490
},
{
"epoch": 17.06896551724138,
"grad_norm": 5.564470023655747,
"learning_rate": 4.6051957889513e-06,
"loss": 0.1705,
"step": 495
},
{
"epoch": 17.24137931034483,
"grad_norm": 5.79246441022243,
"learning_rate": 4.42404759909091e-06,
"loss": 0.1739,
"step": 500
},
{
"epoch": 17.413793103448278,
"grad_norm": 4.968884353779036,
"learning_rate": 4.2455167833851804e-06,
"loss": 0.1684,
"step": 505
},
{
"epoch": 17.586206896551722,
"grad_norm": 5.820327270915885,
"learning_rate": 4.069687144906532e-06,
"loss": 0.1679,
"step": 510
},
{
"epoch": 17.75862068965517,
"grad_norm": 4.874594537233877,
"learning_rate": 3.896641218784081e-06,
"loss": 0.1671,
"step": 515
},
{
"epoch": 17.93103448275862,
"grad_norm": 4.322080959732308,
"learning_rate": 3.7264602334613385e-06,
"loss": 0.1615,
"step": 520
},
{
"epoch": 18.103448275862068,
"grad_norm": 3.394741423143991,
"learning_rate": 3.5592240725672476e-06,
"loss": 0.163,
"step": 525
},
{
"epoch": 18.275862068965516,
"grad_norm": 2.8899715660579917,
"learning_rate": 3.395011237418494e-06,
"loss": 0.1643,
"step": 530
},
{
"epoch": 18.448275862068964,
"grad_norm": 2.66333363991394,
"learning_rate": 3.2338988101706727e-06,
"loss": 0.1598,
"step": 535
},
{
"epoch": 18.620689655172413,
"grad_norm": 3.0652989937592787,
"learning_rate": 3.075962417635634e-06,
"loss": 0.1616,
"step": 540
},
{
"epoch": 18.79310344827586,
"grad_norm": 3.8408972623600373,
"learning_rate": 2.9212761957819347e-06,
"loss": 0.1673,
"step": 545
},
{
"epoch": 18.96551724137931,
"grad_norm": 3.031323684366623,
"learning_rate": 2.769912754935146e-06,
"loss": 0.1588,
"step": 550
},
{
"epoch": 19.137931034482758,
"grad_norm": 2.4766730582592027,
"learning_rate": 2.6219431456942536e-06,
"loss": 0.1626,
"step": 555
},
{
"epoch": 19.310344827586206,
"grad_norm": 3.347536737141708,
"learning_rate": 2.4774368255802483e-06,
"loss": 0.1607,
"step": 560
},
{
"epoch": 19.482758620689655,
"grad_norm": 2.75609785484982,
"learning_rate": 2.3364616264324725e-06,
"loss": 0.1664,
"step": 565
},
{
"epoch": 19.655172413793103,
"grad_norm": 3.7333685782452335,
"learning_rate": 2.199083722568095e-06,
"loss": 0.1634,
"step": 570
},
{
"epoch": 19.82758620689655,
"grad_norm": 3.008810658711964,
"learning_rate": 2.065367599719621e-06,
"loss": 0.1686,
"step": 575
},
{
"epoch": 20.0,
"grad_norm": 3.4023386470750996,
"learning_rate": 1.93537602476504e-06,
"loss": 0.1692,
"step": 580
},
{
"epoch": 20.17241379310345,
"grad_norm": 2.9763118671554314,
"learning_rate": 1.809170016264794e-06,
"loss": 0.1768,
"step": 585
},
{
"epoch": 20.344827586206897,
"grad_norm": 2.7215403943473957,
"learning_rate": 1.6868088158194351e-06,
"loss": 0.1677,
"step": 590
},
{
"epoch": 20.517241379310345,
"grad_norm": 2.48039880730862,
"learning_rate": 1.5683498602613689e-06,
"loss": 0.1638,
"step": 595
},
{
"epoch": 20.689655172413794,
"grad_norm": 1.8624071329024212,
"learning_rate": 1.4538487546937951e-06,
"loss": 0.1687,
"step": 600
},
{
"epoch": 20.862068965517242,
"grad_norm": 2.871924413048312,
"learning_rate": 1.3433592463894375e-06,
"loss": 0.1617,
"step": 605
},
{
"epoch": 21.03448275862069,
"grad_norm": 2.7255286515937667,
"learning_rate": 1.2369331995613664e-06,
"loss": 0.1622,
"step": 610
},
{
"epoch": 21.20689655172414,
"grad_norm": 2.642540001372744,
"learning_rate": 1.1346205710177304e-06,
"loss": 0.1608,
"step": 615
},
{
"epoch": 21.379310344827587,
"grad_norm": 2.975135375218524,
"learning_rate": 1.0364693867118425e-06,
"loss": 0.1644,
"step": 620
},
{
"epoch": 21.551724137931036,
"grad_norm": 2.0876137084099784,
"learning_rate": 9.425257191985859e-07,
"loss": 0.1639,
"step": 625
},
{
"epoch": 21.724137931034484,
"grad_norm": 1.387475835397412,
"learning_rate": 8.528336660077974e-07,
"loss": 0.1677,
"step": 630
},
{
"epoch": 21.896551724137932,
"grad_norm": 2.869659930472805,
"learning_rate": 7.674353289446946e-07,
"loss": 0.1621,
"step": 635
},
{
"epoch": 22.06896551724138,
"grad_norm": 3.0026307802656196,
"learning_rate": 6.863707943271325e-07,
"loss": 0.1655,
"step": 640
},
{
"epoch": 22.24137931034483,
"grad_norm": 1.8278436522298838,
"learning_rate": 6.096781141689223e-07,
"loss": 0.163,
"step": 645
},
{
"epoch": 22.413793103448278,
"grad_norm": 2.0229424538116145,
"learning_rate": 5.373932883180655e-07,
"loss": 0.1665,
"step": 650
},
{
"epoch": 22.586206896551722,
"grad_norm": 2.1750833091387043,
"learning_rate": 4.695502475582814e-07,
"loss": 0.1611,
"step": 655
},
{
"epoch": 22.75862068965517,
"grad_norm": 1.948795256385187,
"learning_rate": 4.0618083768176996e-07,
"loss": 0.1619,
"step": 660
},
{
"epoch": 22.93103448275862,
"grad_norm": 1.5695625649630254,
"learning_rate": 3.4731480454065823e-07,
"loss": 0.1621,
"step": 665
},
{
"epoch": 23.103448275862068,
"grad_norm": 1.5429369658213046,
"learning_rate": 2.9297978008419604e-07,
"loss": 0.1687,
"step": 670
},
{
"epoch": 23.275862068965516,
"grad_norm": 2.0587233544995325,
"learning_rate": 2.4320126938819023e-07,
"loss": 0.1573,
"step": 675
},
{
"epoch": 23.448275862068964,
"grad_norm": 1.6851058427704997,
"learning_rate": 1.9800263868283708e-07,
"loss": 0.1651,
"step": 680
},
{
"epoch": 23.620689655172413,
"grad_norm": 1.2703207998322799,
"learning_rate": 1.5740510438451374e-07,
"loss": 0.1646,
"step": 685
},
{
"epoch": 23.79310344827586,
"grad_norm": 1.6546893474906568,
"learning_rate": 1.214277231367078e-07,
"loss": 0.1672,
"step": 690
},
{
"epoch": 23.96551724137931,
"grad_norm": 1.4962982730142882,
"learning_rate": 9.008738286475748e-08,
"loss": 0.1621,
"step": 695
},
{
"epoch": 24.137931034482758,
"grad_norm": 1.5704585386814611,
"learning_rate": 6.339879484858924e-08,
"loss": 0.1675,
"step": 700
},
{
"epoch": 24.310344827586206,
"grad_norm": 1.367562782086171,
"learning_rate": 4.137448681718392e-08,
"loss": 0.1637,
"step": 705
},
{
"epoch": 24.482758620689655,
"grad_norm": 1.3662424883362485,
"learning_rate": 2.4024797068017414e-08,
"loss": 0.1625,
"step": 710
},
{
"epoch": 24.655172413793103,
"grad_norm": 1.7154032971232247,
"learning_rate": 1.135786961421248e-08,
"loss": 0.1692,
"step": 715
},
{
"epoch": 24.82758620689655,
"grad_norm": 0.9411353813301526,
"learning_rate": 3.3796503617167244e-09,
"loss": 0.1681,
"step": 720
},
{
"epoch": 25.0,
"grad_norm": 1.2181390829757093,
"learning_rate": 9.388431826629074e-11,
"loss": 0.1725,
"step": 725
}
],
"logging_steps": 5,
"max_steps": 725,
"num_input_tokens_seen": 0,
"num_train_epochs": 25,
"save_steps": 200,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 50378664312832.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}