bulla_7b / trainer_state.json
billxbf's picture
Upload 14 files
ba605ff verified
Invalid JSON: Unexpected token 'I', ..."ad_norm": Infinity, "... is not valid JSON
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 4.0,
"eval_steps": 146,
"global_step": 1456,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.005494505494505495,
"grad_norm": Infinity,
"learning_rate": 0.0,
"loss": 2.0983,
"step": 2
},
{
"epoch": 0.01098901098901099,
"grad_norm": 5.516454696655273,
"learning_rate": 1.36986301369863e-07,
"loss": 2.0688,
"step": 4
},
{
"epoch": 0.016483516483516484,
"grad_norm": 5.619696617126465,
"learning_rate": 2.73972602739726e-07,
"loss": 2.0901,
"step": 6
},
{
"epoch": 0.02197802197802198,
"grad_norm": 5.722362995147705,
"learning_rate": 4.1095890410958903e-07,
"loss": 2.1,
"step": 8
},
{
"epoch": 0.027472527472527472,
"grad_norm": 5.279537200927734,
"learning_rate": 5.47945205479452e-07,
"loss": 2.0519,
"step": 10
},
{
"epoch": 0.03296703296703297,
"grad_norm": 4.9255170822143555,
"learning_rate": 6.849315068493151e-07,
"loss": 2.0328,
"step": 12
},
{
"epoch": 0.038461538461538464,
"grad_norm": 4.8295369148254395,
"learning_rate": 8.219178082191781e-07,
"loss": 1.9894,
"step": 14
},
{
"epoch": 0.04395604395604396,
"grad_norm": 6.103366374969482,
"learning_rate": 9.589041095890411e-07,
"loss": 1.9078,
"step": 16
},
{
"epoch": 0.04945054945054945,
"grad_norm": 5.239104747772217,
"learning_rate": 1.095890410958904e-06,
"loss": 1.8896,
"step": 18
},
{
"epoch": 0.054945054945054944,
"grad_norm": 4.494051933288574,
"learning_rate": 1.2328767123287673e-06,
"loss": 1.6875,
"step": 20
},
{
"epoch": 0.06043956043956044,
"grad_norm": 4.799252510070801,
"learning_rate": 1.3698630136986302e-06,
"loss": 1.6452,
"step": 22
},
{
"epoch": 0.06593406593406594,
"grad_norm": 4.425687789916992,
"learning_rate": 1.5068493150684932e-06,
"loss": 1.5281,
"step": 24
},
{
"epoch": 0.07142857142857142,
"grad_norm": 2.157963514328003,
"learning_rate": 1.6438356164383561e-06,
"loss": 1.3483,
"step": 26
},
{
"epoch": 0.07692307692307693,
"grad_norm": 1.2025972604751587,
"learning_rate": 1.7808219178082193e-06,
"loss": 1.3064,
"step": 28
},
{
"epoch": 0.08241758241758242,
"grad_norm": 0.9520919322967529,
"learning_rate": 1.9178082191780823e-06,
"loss": 1.2681,
"step": 30
},
{
"epoch": 0.08791208791208792,
"grad_norm": 0.6353013515472412,
"learning_rate": 2.0547945205479454e-06,
"loss": 1.2484,
"step": 32
},
{
"epoch": 0.09340659340659341,
"grad_norm": 0.46617600321769714,
"learning_rate": 2.191780821917808e-06,
"loss": 1.24,
"step": 34
},
{
"epoch": 0.0989010989010989,
"grad_norm": 0.44788283109664917,
"learning_rate": 2.3287671232876713e-06,
"loss": 1.2342,
"step": 36
},
{
"epoch": 0.1043956043956044,
"grad_norm": 0.3921670615673065,
"learning_rate": 2.4657534246575345e-06,
"loss": 1.217,
"step": 38
},
{
"epoch": 0.10989010989010989,
"grad_norm": 0.3610630929470062,
"learning_rate": 2.6027397260273973e-06,
"loss": 1.225,
"step": 40
},
{
"epoch": 0.11538461538461539,
"grad_norm": 0.3526926040649414,
"learning_rate": 2.7397260273972604e-06,
"loss": 1.2088,
"step": 42
},
{
"epoch": 0.12087912087912088,
"grad_norm": 0.3451712727546692,
"learning_rate": 2.876712328767123e-06,
"loss": 1.2198,
"step": 44
},
{
"epoch": 0.12637362637362637,
"grad_norm": 0.33328160643577576,
"learning_rate": 3.0136986301369864e-06,
"loss": 1.2173,
"step": 46
},
{
"epoch": 0.13186813186813187,
"grad_norm": 0.3273652195930481,
"learning_rate": 3.1506849315068495e-06,
"loss": 1.2088,
"step": 48
},
{
"epoch": 0.13736263736263737,
"grad_norm": 0.3195970952510834,
"learning_rate": 3.2876712328767123e-06,
"loss": 1.2054,
"step": 50
},
{
"epoch": 0.14285714285714285,
"grad_norm": 0.2997814416885376,
"learning_rate": 3.4246575342465754e-06,
"loss": 1.2046,
"step": 52
},
{
"epoch": 0.14835164835164835,
"grad_norm": 0.2609250247478485,
"learning_rate": 3.5616438356164386e-06,
"loss": 1.2058,
"step": 54
},
{
"epoch": 0.15384615384615385,
"grad_norm": 0.21140357851982117,
"learning_rate": 3.6986301369863014e-06,
"loss": 1.2001,
"step": 56
},
{
"epoch": 0.15934065934065933,
"grad_norm": 0.16791734099388123,
"learning_rate": 3.8356164383561645e-06,
"loss": 1.2017,
"step": 58
},
{
"epoch": 0.16483516483516483,
"grad_norm": 0.13620080053806305,
"learning_rate": 3.972602739726027e-06,
"loss": 1.1951,
"step": 60
},
{
"epoch": 0.17032967032967034,
"grad_norm": 0.12290852516889572,
"learning_rate": 4.109589041095891e-06,
"loss": 1.2006,
"step": 62
},
{
"epoch": 0.17582417582417584,
"grad_norm": 0.1265764832496643,
"learning_rate": 4.246575342465754e-06,
"loss": 1.2009,
"step": 64
},
{
"epoch": 0.1813186813186813,
"grad_norm": 0.10449671745300293,
"learning_rate": 4.383561643835616e-06,
"loss": 1.1948,
"step": 66
},
{
"epoch": 0.18681318681318682,
"grad_norm": 0.09835775941610336,
"learning_rate": 4.52054794520548e-06,
"loss": 1.186,
"step": 68
},
{
"epoch": 0.19230769230769232,
"grad_norm": 0.10505539923906326,
"learning_rate": 4.657534246575343e-06,
"loss": 1.1913,
"step": 70
},
{
"epoch": 0.1978021978021978,
"grad_norm": 0.13944971561431885,
"learning_rate": 4.7945205479452054e-06,
"loss": 1.1911,
"step": 72
},
{
"epoch": 0.2032967032967033,
"grad_norm": 0.1453215479850769,
"learning_rate": 4.931506849315069e-06,
"loss": 1.1969,
"step": 74
},
{
"epoch": 0.2087912087912088,
"grad_norm": 0.12393910437822342,
"learning_rate": 5.068493150684932e-06,
"loss": 1.199,
"step": 76
},
{
"epoch": 0.21428571428571427,
"grad_norm": 0.10112921893596649,
"learning_rate": 5.2054794520547945e-06,
"loss": 1.1914,
"step": 78
},
{
"epoch": 0.21978021978021978,
"grad_norm": 0.1048625037074089,
"learning_rate": 5.342465753424658e-06,
"loss": 1.1937,
"step": 80
},
{
"epoch": 0.22527472527472528,
"grad_norm": 0.1040705144405365,
"learning_rate": 5.479452054794521e-06,
"loss": 1.1889,
"step": 82
},
{
"epoch": 0.23076923076923078,
"grad_norm": 0.14360405504703522,
"learning_rate": 5.6164383561643845e-06,
"loss": 1.1946,
"step": 84
},
{
"epoch": 0.23626373626373626,
"grad_norm": 0.12936367094516754,
"learning_rate": 5.753424657534246e-06,
"loss": 1.1882,
"step": 86
},
{
"epoch": 0.24175824175824176,
"grad_norm": 0.08769886940717697,
"learning_rate": 5.89041095890411e-06,
"loss": 1.1873,
"step": 88
},
{
"epoch": 0.24725274725274726,
"grad_norm": 0.1197541207075119,
"learning_rate": 6.027397260273973e-06,
"loss": 1.1863,
"step": 90
},
{
"epoch": 0.25274725274725274,
"grad_norm": 0.12938760221004486,
"learning_rate": 6.164383561643836e-06,
"loss": 1.188,
"step": 92
},
{
"epoch": 0.25824175824175827,
"grad_norm": 0.14707419276237488,
"learning_rate": 6.301369863013699e-06,
"loss": 1.1886,
"step": 94
},
{
"epoch": 0.26373626373626374,
"grad_norm": 0.12229153513908386,
"learning_rate": 6.438356164383563e-06,
"loss": 1.1908,
"step": 96
},
{
"epoch": 0.2692307692307692,
"grad_norm": 0.1133919209241867,
"learning_rate": 6.5753424657534245e-06,
"loss": 1.1825,
"step": 98
},
{
"epoch": 0.27472527472527475,
"grad_norm": 0.12165709584951401,
"learning_rate": 6.712328767123288e-06,
"loss": 1.1807,
"step": 100
},
{
"epoch": 0.2802197802197802,
"grad_norm": 0.12554234266281128,
"learning_rate": 6.849315068493151e-06,
"loss": 1.1831,
"step": 102
},
{
"epoch": 0.2857142857142857,
"grad_norm": 0.11460216343402863,
"learning_rate": 6.9863013698630145e-06,
"loss": 1.1865,
"step": 104
},
{
"epoch": 0.29120879120879123,
"grad_norm": 0.1516668051481247,
"learning_rate": 7.123287671232877e-06,
"loss": 1.1871,
"step": 106
},
{
"epoch": 0.2967032967032967,
"grad_norm": 0.15025198459625244,
"learning_rate": 7.260273972602741e-06,
"loss": 1.1844,
"step": 108
},
{
"epoch": 0.3021978021978022,
"grad_norm": 0.13930697739124298,
"learning_rate": 7.397260273972603e-06,
"loss": 1.1803,
"step": 110
},
{
"epoch": 0.3076923076923077,
"grad_norm": 0.14465640485286713,
"learning_rate": 7.534246575342466e-06,
"loss": 1.1895,
"step": 112
},
{
"epoch": 0.3131868131868132,
"grad_norm": 0.12900669872760773,
"learning_rate": 7.671232876712329e-06,
"loss": 1.1874,
"step": 114
},
{
"epoch": 0.31868131868131866,
"grad_norm": 0.12059827893972397,
"learning_rate": 7.808219178082192e-06,
"loss": 1.184,
"step": 116
},
{
"epoch": 0.3241758241758242,
"grad_norm": 0.11196708679199219,
"learning_rate": 7.945205479452055e-06,
"loss": 1.1876,
"step": 118
},
{
"epoch": 0.32967032967032966,
"grad_norm": 0.11420655995607376,
"learning_rate": 8.082191780821919e-06,
"loss": 1.1799,
"step": 120
},
{
"epoch": 0.33516483516483514,
"grad_norm": 0.10233011841773987,
"learning_rate": 8.219178082191782e-06,
"loss": 1.1836,
"step": 122
},
{
"epoch": 0.34065934065934067,
"grad_norm": 0.13560251891613007,
"learning_rate": 8.356164383561644e-06,
"loss": 1.1825,
"step": 124
},
{
"epoch": 0.34615384615384615,
"grad_norm": 0.1417393684387207,
"learning_rate": 8.493150684931507e-06,
"loss": 1.181,
"step": 126
},
{
"epoch": 0.3516483516483517,
"grad_norm": 0.12368721514940262,
"learning_rate": 8.63013698630137e-06,
"loss": 1.1857,
"step": 128
},
{
"epoch": 0.35714285714285715,
"grad_norm": 0.14341862499713898,
"learning_rate": 8.767123287671233e-06,
"loss": 1.1839,
"step": 130
},
{
"epoch": 0.3626373626373626,
"grad_norm": 0.14925454556941986,
"learning_rate": 8.904109589041097e-06,
"loss": 1.1792,
"step": 132
},
{
"epoch": 0.36813186813186816,
"grad_norm": 0.1684715300798416,
"learning_rate": 9.04109589041096e-06,
"loss": 1.1806,
"step": 134
},
{
"epoch": 0.37362637362637363,
"grad_norm": 0.10095636546611786,
"learning_rate": 9.178082191780823e-06,
"loss": 1.176,
"step": 136
},
{
"epoch": 0.3791208791208791,
"grad_norm": 0.11856956034898758,
"learning_rate": 9.315068493150685e-06,
"loss": 1.1849,
"step": 138
},
{
"epoch": 0.38461538461538464,
"grad_norm": 0.19565819203853607,
"learning_rate": 9.452054794520548e-06,
"loss": 1.1803,
"step": 140
},
{
"epoch": 0.3901098901098901,
"grad_norm": 0.15888364613056183,
"learning_rate": 9.589041095890411e-06,
"loss": 1.1756,
"step": 142
},
{
"epoch": 0.3956043956043956,
"grad_norm": 0.16350317001342773,
"learning_rate": 9.726027397260275e-06,
"loss": 1.1777,
"step": 144
},
{
"epoch": 0.4010989010989011,
"grad_norm": 0.13230757415294647,
"learning_rate": 9.863013698630138e-06,
"loss": 1.1843,
"step": 146
},
{
"epoch": 0.4010989010989011,
"eval_loss": 1.1792316436767578,
"eval_runtime": 299.2609,
"eval_samples_per_second": 17.296,
"eval_steps_per_second": 0.271,
"step": 146
},
{
"epoch": 0.4065934065934066,
"grad_norm": 0.1627519279718399,
"learning_rate": 1e-05,
"loss": 1.174,
"step": 148
},
{
"epoch": 0.41208791208791207,
"grad_norm": 0.13225838541984558,
"learning_rate": 9.999942488284598e-06,
"loss": 1.1779,
"step": 150
},
{
"epoch": 0.4175824175824176,
"grad_norm": 0.15863117575645447,
"learning_rate": 9.999769954461425e-06,
"loss": 1.1767,
"step": 152
},
{
"epoch": 0.4230769230769231,
"grad_norm": 0.15866652131080627,
"learning_rate": 9.999482402499569e-06,
"loss": 1.1758,
"step": 154
},
{
"epoch": 0.42857142857142855,
"grad_norm": 0.1270761787891388,
"learning_rate": 9.999079839014074e-06,
"loss": 1.1756,
"step": 156
},
{
"epoch": 0.4340659340659341,
"grad_norm": 0.10430735349655151,
"learning_rate": 9.998562273265786e-06,
"loss": 1.1728,
"step": 158
},
{
"epoch": 0.43956043956043955,
"grad_norm": 0.13678255677223206,
"learning_rate": 9.997929717161142e-06,
"loss": 1.1774,
"step": 160
},
{
"epoch": 0.44505494505494503,
"grad_norm": 0.13817106187343597,
"learning_rate": 9.997182185251896e-06,
"loss": 1.1804,
"step": 162
},
{
"epoch": 0.45054945054945056,
"grad_norm": 0.15956689417362213,
"learning_rate": 9.996319694734787e-06,
"loss": 1.1802,
"step": 164
},
{
"epoch": 0.45604395604395603,
"grad_norm": 0.2157265841960907,
"learning_rate": 9.995342265451138e-06,
"loss": 1.1803,
"step": 166
},
{
"epoch": 0.46153846153846156,
"grad_norm": 0.17677265405654907,
"learning_rate": 9.994249919886402e-06,
"loss": 1.1764,
"step": 168
},
{
"epoch": 0.46703296703296704,
"grad_norm": 0.16796888411045074,
"learning_rate": 9.993042683169647e-06,
"loss": 1.1702,
"step": 170
},
{
"epoch": 0.4725274725274725,
"grad_norm": 0.16481636464595795,
"learning_rate": 9.991720583072975e-06,
"loss": 1.1844,
"step": 172
},
{
"epoch": 0.47802197802197804,
"grad_norm": 0.19735674560070038,
"learning_rate": 9.990283650010883e-06,
"loss": 1.1791,
"step": 174
},
{
"epoch": 0.4835164835164835,
"grad_norm": 0.15785232186317444,
"learning_rate": 9.988731917039564e-06,
"loss": 1.1783,
"step": 176
},
{
"epoch": 0.489010989010989,
"grad_norm": 0.17257508635520935,
"learning_rate": 9.98706541985615e-06,
"loss": 1.1769,
"step": 178
},
{
"epoch": 0.4945054945054945,
"grad_norm": 0.1660294085741043,
"learning_rate": 9.985284196797884e-06,
"loss": 1.181,
"step": 180
},
{
"epoch": 0.5,
"grad_norm": 0.1156650260090828,
"learning_rate": 9.983388288841246e-06,
"loss": 1.1865,
"step": 182
},
{
"epoch": 0.5054945054945055,
"grad_norm": 0.0954098179936409,
"learning_rate": 9.981377739601002e-06,
"loss": 1.1787,
"step": 184
},
{
"epoch": 0.510989010989011,
"grad_norm": 0.0957237258553505,
"learning_rate": 9.979252595329204e-06,
"loss": 1.1731,
"step": 186
},
{
"epoch": 0.5164835164835165,
"grad_norm": 0.11179114878177643,
"learning_rate": 9.977012904914133e-06,
"loss": 1.166,
"step": 188
},
{
"epoch": 0.521978021978022,
"grad_norm": 0.12088882923126221,
"learning_rate": 9.974658719879163e-06,
"loss": 1.1706,
"step": 190
},
{
"epoch": 0.5274725274725275,
"grad_norm": 0.1067051887512207,
"learning_rate": 9.972190094381578e-06,
"loss": 1.178,
"step": 192
},
{
"epoch": 0.532967032967033,
"grad_norm": 0.09633362293243408,
"learning_rate": 9.96960708521134e-06,
"loss": 1.1763,
"step": 194
},
{
"epoch": 0.5384615384615384,
"grad_norm": 0.11866485327482224,
"learning_rate": 9.966909751789758e-06,
"loss": 1.1777,
"step": 196
},
{
"epoch": 0.5439560439560439,
"grad_norm": 0.11669634282588959,
"learning_rate": 9.964098156168143e-06,
"loss": 1.1753,
"step": 198
},
{
"epoch": 0.5494505494505495,
"grad_norm": 0.1488696038722992,
"learning_rate": 9.96117236302637e-06,
"loss": 1.1656,
"step": 200
},
{
"epoch": 0.554945054945055,
"grad_norm": 0.1773654669523239,
"learning_rate": 9.958132439671392e-06,
"loss": 1.1809,
"step": 202
},
{
"epoch": 0.5604395604395604,
"grad_norm": 0.17689326405525208,
"learning_rate": 9.954978456035695e-06,
"loss": 1.1722,
"step": 204
},
{
"epoch": 0.5659340659340659,
"grad_norm": 0.13418439030647278,
"learning_rate": 9.951710484675677e-06,
"loss": 1.167,
"step": 206
},
{
"epoch": 0.5714285714285714,
"grad_norm": 0.14982537925243378,
"learning_rate": 9.948328600769996e-06,
"loss": 1.1713,
"step": 208
},
{
"epoch": 0.5769230769230769,
"grad_norm": 0.12595325708389282,
"learning_rate": 9.94483288211783e-06,
"loss": 1.1677,
"step": 210
},
{
"epoch": 0.5824175824175825,
"grad_norm": 0.10920244455337524,
"learning_rate": 9.941223409137088e-06,
"loss": 1.1727,
"step": 212
},
{
"epoch": 0.5879120879120879,
"grad_norm": 0.10856951773166656,
"learning_rate": 9.937500264862567e-06,
"loss": 1.1739,
"step": 214
},
{
"epoch": 0.5934065934065934,
"grad_norm": 0.13141585886478424,
"learning_rate": 9.933663534944029e-06,
"loss": 1.176,
"step": 216
},
{
"epoch": 0.5989010989010989,
"grad_norm": 0.13502047955989838,
"learning_rate": 9.929713307644245e-06,
"loss": 1.178,
"step": 218
},
{
"epoch": 0.6043956043956044,
"grad_norm": 0.1468881517648697,
"learning_rate": 9.925649673836949e-06,
"loss": 1.1819,
"step": 220
},
{
"epoch": 0.6098901098901099,
"grad_norm": 0.13739052414894104,
"learning_rate": 9.921472727004765e-06,
"loss": 1.1784,
"step": 222
},
{
"epoch": 0.6153846153846154,
"grad_norm": 0.13080830872058868,
"learning_rate": 9.917182563237045e-06,
"loss": 1.1762,
"step": 224
},
{
"epoch": 0.6208791208791209,
"grad_norm": 0.11544028669595718,
"learning_rate": 9.912779281227656e-06,
"loss": 1.1733,
"step": 226
},
{
"epoch": 0.6263736263736264,
"grad_norm": 0.14270079135894775,
"learning_rate": 9.908262982272724e-06,
"loss": 1.1812,
"step": 228
},
{
"epoch": 0.6318681318681318,
"grad_norm": 0.11610660701990128,
"learning_rate": 9.903633770268286e-06,
"loss": 1.1803,
"step": 230
},
{
"epoch": 0.6373626373626373,
"grad_norm": 0.10949360579252243,
"learning_rate": 9.89889175170791e-06,
"loss": 1.1776,
"step": 232
},
{
"epoch": 0.6428571428571429,
"grad_norm": 0.11754804104566574,
"learning_rate": 9.894037035680246e-06,
"loss": 1.169,
"step": 234
},
{
"epoch": 0.6483516483516484,
"grad_norm": 0.1328520029783249,
"learning_rate": 9.889069733866515e-06,
"loss": 1.1663,
"step": 236
},
{
"epoch": 0.6538461538461539,
"grad_norm": 0.12216369807720184,
"learning_rate": 9.883989960537934e-06,
"loss": 1.1779,
"step": 238
},
{
"epoch": 0.6593406593406593,
"grad_norm": 0.11219753324985504,
"learning_rate": 9.878797832553093e-06,
"loss": 1.1718,
"step": 240
},
{
"epoch": 0.6648351648351648,
"grad_norm": 0.11396931111812592,
"learning_rate": 9.873493469355271e-06,
"loss": 1.175,
"step": 242
},
{
"epoch": 0.6703296703296703,
"grad_norm": 0.104461669921875,
"learning_rate": 9.868076992969672e-06,
"loss": 1.1761,
"step": 244
},
{
"epoch": 0.6758241758241759,
"grad_norm": 0.11892971396446228,
"learning_rate": 9.862548528000644e-06,
"loss": 1.1713,
"step": 246
},
{
"epoch": 0.6813186813186813,
"grad_norm": 0.10305490344762802,
"learning_rate": 9.85690820162878e-06,
"loss": 1.1751,
"step": 248
},
{
"epoch": 0.6868131868131868,
"grad_norm": 0.11825086921453476,
"learning_rate": 9.851156143608025e-06,
"loss": 1.1749,
"step": 250
},
{
"epoch": 0.6923076923076923,
"grad_norm": 0.09655775129795074,
"learning_rate": 9.845292486262664e-06,
"loss": 1.1733,
"step": 252
},
{
"epoch": 0.6978021978021978,
"grad_norm": 0.09013015031814575,
"learning_rate": 9.839317364484295e-06,
"loss": 1.1736,
"step": 254
},
{
"epoch": 0.7032967032967034,
"grad_norm": 0.12731756269931793,
"learning_rate": 9.83323091572872e-06,
"loss": 1.1735,
"step": 256
},
{
"epoch": 0.7087912087912088,
"grad_norm": 0.11326012015342712,
"learning_rate": 9.827033280012783e-06,
"loss": 1.1678,
"step": 258
},
{
"epoch": 0.7142857142857143,
"grad_norm": 0.09636470675468445,
"learning_rate": 9.820724599911147e-06,
"loss": 1.1735,
"step": 260
},
{
"epoch": 0.7197802197802198,
"grad_norm": 0.12517008185386658,
"learning_rate": 9.81430502055302e-06,
"loss": 1.1839,
"step": 262
},
{
"epoch": 0.7252747252747253,
"grad_norm": 0.14088405668735504,
"learning_rate": 9.807774689618806e-06,
"loss": 1.1734,
"step": 264
},
{
"epoch": 0.7307692307692307,
"grad_norm": 0.09721764922142029,
"learning_rate": 9.801133757336726e-06,
"loss": 1.1681,
"step": 266
},
{
"epoch": 0.7362637362637363,
"grad_norm": 0.08106404542922974,
"learning_rate": 9.794382376479334e-06,
"loss": 1.1684,
"step": 268
},
{
"epoch": 0.7417582417582418,
"grad_norm": 0.1210760697722435,
"learning_rate": 9.787520702360035e-06,
"loss": 1.1656,
"step": 270
},
{
"epoch": 0.7472527472527473,
"grad_norm": 0.12102904915809631,
"learning_rate": 9.780548892829486e-06,
"loss": 1.1743,
"step": 272
},
{
"epoch": 0.7527472527472527,
"grad_norm": 0.11343714594841003,
"learning_rate": 9.773467108271978e-06,
"loss": 1.1756,
"step": 274
},
{
"epoch": 0.7582417582417582,
"grad_norm": 0.11999528110027313,
"learning_rate": 9.766275511601742e-06,
"loss": 1.1782,
"step": 276
},
{
"epoch": 0.7637362637362637,
"grad_norm": 0.148450568318367,
"learning_rate": 9.7589742682592e-06,
"loss": 1.1796,
"step": 278
},
{
"epoch": 0.7692307692307693,
"grad_norm": 0.09977131336927414,
"learning_rate": 9.751563546207167e-06,
"loss": 1.1686,
"step": 280
},
{
"epoch": 0.7747252747252747,
"grad_norm": 0.12250227481126785,
"learning_rate": 9.744043515926975e-06,
"loss": 1.1673,
"step": 282
},
{
"epoch": 0.7802197802197802,
"grad_norm": 0.12362032383680344,
"learning_rate": 9.736414350414564e-06,
"loss": 1.1707,
"step": 284
},
{
"epoch": 0.7857142857142857,
"grad_norm": 0.09825081378221512,
"learning_rate": 9.72867622517649e-06,
"loss": 1.1719,
"step": 286
},
{
"epoch": 0.7912087912087912,
"grad_norm": 0.13811303675174713,
"learning_rate": 9.720829318225897e-06,
"loss": 1.1711,
"step": 288
},
{
"epoch": 0.7967032967032966,
"grad_norm": 0.14676132798194885,
"learning_rate": 9.712873810078415e-06,
"loss": 1.1728,
"step": 290
},
{
"epoch": 0.8021978021978022,
"grad_norm": 0.09035024791955948,
"learning_rate": 9.704809883748012e-06,
"loss": 1.1776,
"step": 292
},
{
"epoch": 0.8021978021978022,
"eval_loss": 1.1732887029647827,
"eval_runtime": 298.7065,
"eval_samples_per_second": 17.328,
"eval_steps_per_second": 0.271,
"step": 292
},
{
"epoch": 0.8076923076923077,
"grad_norm": 0.10824787616729736,
"learning_rate": 9.696637724742785e-06,
"loss": 1.1735,
"step": 294
},
{
"epoch": 0.8131868131868132,
"grad_norm": 0.10512899607419968,
"learning_rate": 9.688357521060685e-06,
"loss": 1.1702,
"step": 296
},
{
"epoch": 0.8186813186813187,
"grad_norm": 0.12007107585668564,
"learning_rate": 9.6799694631852e-06,
"loss": 1.1687,
"step": 298
},
{
"epoch": 0.8241758241758241,
"grad_norm": 0.1161542534828186,
"learning_rate": 9.67147374408097e-06,
"loss": 1.1726,
"step": 300
},
{
"epoch": 0.8296703296703297,
"grad_norm": 0.16156896948814392,
"learning_rate": 9.662870559189344e-06,
"loss": 1.1639,
"step": 302
},
{
"epoch": 0.8351648351648352,
"grad_norm": 0.12350093573331833,
"learning_rate": 9.654160106423891e-06,
"loss": 1.1711,
"step": 304
},
{
"epoch": 0.8406593406593407,
"grad_norm": 0.1292959749698639,
"learning_rate": 9.645342586165845e-06,
"loss": 1.1677,
"step": 306
},
{
"epoch": 0.8461538461538461,
"grad_norm": 0.09656315296888351,
"learning_rate": 9.63641820125949e-06,
"loss": 1.1773,
"step": 308
},
{
"epoch": 0.8516483516483516,
"grad_norm": 0.13036847114562988,
"learning_rate": 9.627387157007502e-06,
"loss": 1.1819,
"step": 310
},
{
"epoch": 0.8571428571428571,
"grad_norm": 0.12335841357707977,
"learning_rate": 9.618249661166218e-06,
"loss": 1.1688,
"step": 312
},
{
"epoch": 0.8626373626373627,
"grad_norm": 0.13838279247283936,
"learning_rate": 9.609005923940865e-06,
"loss": 1.1665,
"step": 314
},
{
"epoch": 0.8681318681318682,
"grad_norm": 0.13689468801021576,
"learning_rate": 9.599656157980715e-06,
"loss": 1.1696,
"step": 316
},
{
"epoch": 0.8736263736263736,
"grad_norm": 0.13239113986492157,
"learning_rate": 9.590200578374198e-06,
"loss": 1.1686,
"step": 318
},
{
"epoch": 0.8791208791208791,
"grad_norm": 0.09575112909078598,
"learning_rate": 9.580639402643957e-06,
"loss": 1.1673,
"step": 320
},
{
"epoch": 0.8846153846153846,
"grad_norm": 0.11522715538740158,
"learning_rate": 9.570972850741839e-06,
"loss": 1.1685,
"step": 322
},
{
"epoch": 0.8901098901098901,
"grad_norm": 0.14126405119895935,
"learning_rate": 9.561201145043835e-06,
"loss": 1.1788,
"step": 324
},
{
"epoch": 0.8956043956043956,
"grad_norm": 0.17362023890018463,
"learning_rate": 9.551324510344972e-06,
"loss": 1.1746,
"step": 326
},
{
"epoch": 0.9010989010989011,
"grad_norm": 0.1431788206100464,
"learning_rate": 9.541343173854128e-06,
"loss": 1.1719,
"step": 328
},
{
"epoch": 0.9065934065934066,
"grad_norm": 0.11747945845127106,
"learning_rate": 9.531257365188818e-06,
"loss": 1.1763,
"step": 330
},
{
"epoch": 0.9120879120879121,
"grad_norm": 0.10873839259147644,
"learning_rate": 9.521067316369903e-06,
"loss": 1.173,
"step": 332
},
{
"epoch": 0.9175824175824175,
"grad_norm": 0.10153867304325104,
"learning_rate": 9.510773261816261e-06,
"loss": 1.1681,
"step": 334
},
{
"epoch": 0.9230769230769231,
"grad_norm": 0.07133303582668304,
"learning_rate": 9.500375438339384e-06,
"loss": 1.1643,
"step": 336
},
{
"epoch": 0.9285714285714286,
"grad_norm": 0.09183746576309204,
"learning_rate": 9.48987408513794e-06,
"loss": 1.1788,
"step": 338
},
{
"epoch": 0.9340659340659341,
"grad_norm": 0.10790558159351349,
"learning_rate": 9.47926944379226e-06,
"loss": 1.1817,
"step": 340
},
{
"epoch": 0.9395604395604396,
"grad_norm": 0.09112295508384705,
"learning_rate": 9.468561758258795e-06,
"loss": 1.1674,
"step": 342
},
{
"epoch": 0.945054945054945,
"grad_norm": 0.08475514501333237,
"learning_rate": 9.457751274864486e-06,
"loss": 1.1706,
"step": 344
},
{
"epoch": 0.9505494505494505,
"grad_norm": 0.10173246264457703,
"learning_rate": 9.446838242301113e-06,
"loss": 1.1644,
"step": 346
},
{
"epoch": 0.9560439560439561,
"grad_norm": 0.10165045410394669,
"learning_rate": 9.435822911619564e-06,
"loss": 1.1694,
"step": 348
},
{
"epoch": 0.9615384615384616,
"grad_norm": 0.10506884753704071,
"learning_rate": 9.424705536224065e-06,
"loss": 1.1708,
"step": 350
},
{
"epoch": 0.967032967032967,
"grad_norm": 0.10648410022258759,
"learning_rate": 9.41348637186635e-06,
"loss": 1.1712,
"step": 352
},
{
"epoch": 0.9725274725274725,
"grad_norm": 0.11728779226541519,
"learning_rate": 9.40216567663977e-06,
"loss": 1.1838,
"step": 354
},
{
"epoch": 0.978021978021978,
"grad_norm": 0.12955018877983093,
"learning_rate": 9.390743710973366e-06,
"loss": 1.1773,
"step": 356
},
{
"epoch": 0.9835164835164835,
"grad_norm": 0.14313377439975739,
"learning_rate": 9.379220737625877e-06,
"loss": 1.1693,
"step": 358
},
{
"epoch": 0.989010989010989,
"grad_norm": 0.17154935002326965,
"learning_rate": 9.367597021679686e-06,
"loss": 1.1674,
"step": 360
},
{
"epoch": 0.9945054945054945,
"grad_norm": 0.13390299677848816,
"learning_rate": 9.35587283053473e-06,
"loss": 1.1636,
"step": 362
},
{
"epoch": 1.0,
"grad_norm": 0.13049638271331787,
"learning_rate": 9.344048433902351e-06,
"loss": 1.1678,
"step": 364
},
{
"epoch": 1.0054945054945055,
"grad_norm": 0.11013516038656235,
"learning_rate": 9.332124103799075e-06,
"loss": 1.1627,
"step": 366
},
{
"epoch": 1.010989010989011,
"grad_norm": 0.11757113039493561,
"learning_rate": 9.320100114540382e-06,
"loss": 1.1679,
"step": 368
},
{
"epoch": 1.0164835164835164,
"grad_norm": 0.11947564780712128,
"learning_rate": 9.307976742734366e-06,
"loss": 1.1767,
"step": 370
},
{
"epoch": 1.021978021978022,
"grad_norm": 0.1045905202627182,
"learning_rate": 9.295754267275393e-06,
"loss": 1.1694,
"step": 372
},
{
"epoch": 1.0274725274725274,
"grad_norm": 0.10813715308904648,
"learning_rate": 9.283432969337672e-06,
"loss": 1.1752,
"step": 374
},
{
"epoch": 1.032967032967033,
"grad_norm": 0.11131364107131958,
"learning_rate": 9.271013132368799e-06,
"loss": 1.1714,
"step": 376
},
{
"epoch": 1.0384615384615385,
"grad_norm": 0.09973058849573135,
"learning_rate": 9.258495042083222e-06,
"loss": 1.1696,
"step": 378
},
{
"epoch": 1.043956043956044,
"grad_norm": 0.10416287928819656,
"learning_rate": 9.245878986455684e-06,
"loss": 1.1641,
"step": 380
},
{
"epoch": 1.0494505494505495,
"grad_norm": 0.08854486793279648,
"learning_rate": 9.23316525571458e-06,
"loss": 1.1649,
"step": 382
},
{
"epoch": 1.054945054945055,
"grad_norm": 0.1364666223526001,
"learning_rate": 9.2203541423353e-06,
"loss": 1.1769,
"step": 384
},
{
"epoch": 1.0604395604395604,
"grad_norm": 0.14936769008636475,
"learning_rate": 9.207445941033483e-06,
"loss": 1.1686,
"step": 386
},
{
"epoch": 1.065934065934066,
"grad_norm": 0.18906323611736298,
"learning_rate": 9.19444094875825e-06,
"loss": 1.1692,
"step": 388
},
{
"epoch": 1.0714285714285714,
"grad_norm": 0.44419270753860474,
"learning_rate": 9.18133946468537e-06,
"loss": 1.1746,
"step": 390
},
{
"epoch": 1.0769230769230769,
"grad_norm": 0.1541990488767624,
"learning_rate": 9.16814179021037e-06,
"loss": 1.1735,
"step": 392
},
{
"epoch": 1.0824175824175823,
"grad_norm": 0.10936509072780609,
"learning_rate": 9.154848228941607e-06,
"loss": 1.1734,
"step": 394
},
{
"epoch": 1.0879120879120878,
"grad_norm": 0.0885339230298996,
"learning_rate": 9.14145908669329e-06,
"loss": 1.1702,
"step": 396
},
{
"epoch": 1.0934065934065935,
"grad_norm": 0.07241977006196976,
"learning_rate": 9.127974671478432e-06,
"loss": 1.1714,
"step": 398
},
{
"epoch": 1.098901098901099,
"grad_norm": 0.08955956995487213,
"learning_rate": 9.114395293501775e-06,
"loss": 1.1712,
"step": 400
},
{
"epoch": 1.1043956043956045,
"grad_norm": 0.11948683112859726,
"learning_rate": 9.100721265152644e-06,
"loss": 1.1799,
"step": 402
},
{
"epoch": 1.10989010989011,
"grad_norm": 0.11843076348304749,
"learning_rate": 9.086952900997774e-06,
"loss": 1.1716,
"step": 404
},
{
"epoch": 1.1153846153846154,
"grad_norm": 0.08606066554784775,
"learning_rate": 9.073090517774057e-06,
"loss": 1.1777,
"step": 406
},
{
"epoch": 1.120879120879121,
"grad_norm": 0.08759862184524536,
"learning_rate": 9.059134434381274e-06,
"loss": 1.1698,
"step": 408
},
{
"epoch": 1.1263736263736264,
"grad_norm": 0.10086913406848907,
"learning_rate": 9.045084971874738e-06,
"loss": 1.1641,
"step": 410
},
{
"epoch": 1.1318681318681318,
"grad_norm": 0.11529091000556946,
"learning_rate": 9.030942453457928e-06,
"loss": 1.1655,
"step": 412
},
{
"epoch": 1.1373626373626373,
"grad_norm": 0.0781669095158577,
"learning_rate": 9.01670720447504e-06,
"loss": 1.1678,
"step": 414
},
{
"epoch": 1.1428571428571428,
"grad_norm": 0.09989538788795471,
"learning_rate": 9.00237955240351e-06,
"loss": 1.1689,
"step": 416
},
{
"epoch": 1.1483516483516483,
"grad_norm": 0.12073778361082077,
"learning_rate": 8.987959826846479e-06,
"loss": 1.1747,
"step": 418
},
{
"epoch": 1.1538461538461537,
"grad_norm": 0.12605726718902588,
"learning_rate": 8.973448359525207e-06,
"loss": 1.1725,
"step": 420
},
{
"epoch": 1.1593406593406592,
"grad_norm": 0.12503278255462646,
"learning_rate": 8.958845484271443e-06,
"loss": 1.1709,
"step": 422
},
{
"epoch": 1.164835164835165,
"grad_norm": 0.10487185418605804,
"learning_rate": 8.944151537019752e-06,
"loss": 1.1566,
"step": 424
},
{
"epoch": 1.1703296703296704,
"grad_norm": 0.09222570061683655,
"learning_rate": 8.929366855799777e-06,
"loss": 1.1808,
"step": 426
},
{
"epoch": 1.1758241758241759,
"grad_norm": 0.10268251597881317,
"learning_rate": 8.914491780728471e-06,
"loss": 1.1782,
"step": 428
},
{
"epoch": 1.1813186813186813,
"grad_norm": 0.08701249957084656,
"learning_rate": 8.899526654002268e-06,
"loss": 1.1716,
"step": 430
},
{
"epoch": 1.1868131868131868,
"grad_norm": 0.11233299225568771,
"learning_rate": 8.88447181988921e-06,
"loss": 1.1699,
"step": 432
},
{
"epoch": 1.1923076923076923,
"grad_norm": 0.09972134232521057,
"learning_rate": 8.869327624721033e-06,
"loss": 1.1687,
"step": 434
},
{
"epoch": 1.1978021978021978,
"grad_norm": 0.10448331385850906,
"learning_rate": 8.854094416885192e-06,
"loss": 1.1656,
"step": 436
},
{
"epoch": 1.2032967032967032,
"grad_norm": 0.11365893483161926,
"learning_rate": 8.838772546816857e-06,
"loss": 1.1764,
"step": 438
},
{
"epoch": 1.2032967032967032,
"eval_loss": 1.1709927320480347,
"eval_runtime": 299.4182,
"eval_samples_per_second": 17.287,
"eval_steps_per_second": 0.271,
"step": 438
},
{
"epoch": 1.2087912087912087,
"grad_norm": 0.1238979920744896,
"learning_rate": 8.823362366990833e-06,
"loss": 1.1718,
"step": 440
},
{
"epoch": 1.2142857142857142,
"grad_norm": 0.12039055675268173,
"learning_rate": 8.807864231913475e-06,
"loss": 1.1686,
"step": 442
},
{
"epoch": 1.2197802197802199,
"grad_norm": 0.09191784262657166,
"learning_rate": 8.792278498114517e-06,
"loss": 1.1702,
"step": 444
},
{
"epoch": 1.2252747252747254,
"grad_norm": 0.08927666395902634,
"learning_rate": 8.77660552413887e-06,
"loss": 1.1761,
"step": 446
},
{
"epoch": 1.2307692307692308,
"grad_norm": 0.08614211529493332,
"learning_rate": 8.760845670538387e-06,
"loss": 1.1743,
"step": 448
},
{
"epoch": 1.2362637362637363,
"grad_norm": 0.10282223671674728,
"learning_rate": 8.744999299863549e-06,
"loss": 1.1705,
"step": 450
},
{
"epoch": 1.2417582417582418,
"grad_norm": 0.10722998529672623,
"learning_rate": 8.729066776655144e-06,
"loss": 1.169,
"step": 452
},
{
"epoch": 1.2472527472527473,
"grad_norm": 0.09755399823188782,
"learning_rate": 8.713048467435865e-06,
"loss": 1.1732,
"step": 454
},
{
"epoch": 1.2527472527472527,
"grad_norm": 0.09675360471010208,
"learning_rate": 8.696944740701891e-06,
"loss": 1.1696,
"step": 456
},
{
"epoch": 1.2582417582417582,
"grad_norm": 0.08405736833810806,
"learning_rate": 8.6807559669144e-06,
"loss": 1.1678,
"step": 458
},
{
"epoch": 1.2637362637362637,
"grad_norm": 0.08441252261400223,
"learning_rate": 8.664482518491053e-06,
"loss": 1.1684,
"step": 460
},
{
"epoch": 1.2692307692307692,
"grad_norm": 0.11046714335680008,
"learning_rate": 8.648124769797424e-06,
"loss": 1.1747,
"step": 462
},
{
"epoch": 1.2747252747252746,
"grad_norm": 0.10532383620738983,
"learning_rate": 8.631683097138386e-06,
"loss": 1.1728,
"step": 464
},
{
"epoch": 1.2802197802197801,
"grad_norm": 0.09521500766277313,
"learning_rate": 8.615157878749462e-06,
"loss": 1.169,
"step": 466
},
{
"epoch": 1.2857142857142856,
"grad_norm": 0.09845519065856934,
"learning_rate": 8.598549494788111e-06,
"loss": 1.1667,
"step": 468
},
{
"epoch": 1.2912087912087913,
"grad_norm": 0.10502909123897552,
"learning_rate": 8.581858327324996e-06,
"loss": 1.1709,
"step": 470
},
{
"epoch": 1.2967032967032968,
"grad_norm": 0.0786222293972969,
"learning_rate": 8.565084760335188e-06,
"loss": 1.165,
"step": 472
},
{
"epoch": 1.3021978021978022,
"grad_norm": 0.10599285364151001,
"learning_rate": 8.548229179689325e-06,
"loss": 1.1718,
"step": 474
},
{
"epoch": 1.3076923076923077,
"grad_norm": 0.11965076625347137,
"learning_rate": 8.531291973144755e-06,
"loss": 1.1683,
"step": 476
},
{
"epoch": 1.3131868131868132,
"grad_norm": 0.10626693069934845,
"learning_rate": 8.5142735303366e-06,
"loss": 1.1676,
"step": 478
},
{
"epoch": 1.3186813186813187,
"grad_norm": 0.10863650590181351,
"learning_rate": 8.497174242768792e-06,
"loss": 1.17,
"step": 480
},
{
"epoch": 1.3241758241758241,
"grad_norm": 0.11272372305393219,
"learning_rate": 8.479994503805079e-06,
"loss": 1.1674,
"step": 482
},
{
"epoch": 1.3296703296703296,
"grad_norm": 0.1116040050983429,
"learning_rate": 8.462734708659959e-06,
"loss": 1.1711,
"step": 484
},
{
"epoch": 1.335164835164835,
"grad_norm": 0.1068933829665184,
"learning_rate": 8.445395254389605e-06,
"loss": 1.179,
"step": 486
},
{
"epoch": 1.3406593406593408,
"grad_norm": 0.11034112423658371,
"learning_rate": 8.427976539882725e-06,
"loss": 1.1617,
"step": 488
},
{
"epoch": 1.3461538461538463,
"grad_norm": 0.11086605489253998,
"learning_rate": 8.410478965851371e-06,
"loss": 1.171,
"step": 490
},
{
"epoch": 1.3516483516483517,
"grad_norm": 0.1053035780787468,
"learning_rate": 8.39290293482175e-06,
"loss": 1.1696,
"step": 492
},
{
"epoch": 1.3571428571428572,
"grad_norm": 0.11563800275325775,
"learning_rate": 8.375248851124937e-06,
"loss": 1.1817,
"step": 494
},
{
"epoch": 1.3626373626373627,
"grad_norm": 0.09770724177360535,
"learning_rate": 8.357517120887586e-06,
"loss": 1.1726,
"step": 496
},
{
"epoch": 1.3681318681318682,
"grad_norm": 0.10480888187885284,
"learning_rate": 8.339708152022586e-06,
"loss": 1.1716,
"step": 498
},
{
"epoch": 1.3736263736263736,
"grad_norm": 0.0983305424451828,
"learning_rate": 8.321822354219677e-06,
"loss": 1.1708,
"step": 500
},
{
"epoch": 1.379120879120879,
"grad_norm": 0.11572114378213882,
"learning_rate": 8.303860138936027e-06,
"loss": 1.1669,
"step": 502
},
{
"epoch": 1.3846153846153846,
"grad_norm": 0.15876273810863495,
"learning_rate": 8.285821919386758e-06,
"loss": 1.172,
"step": 504
},
{
"epoch": 1.39010989010989,
"grad_norm": 0.09302208572626114,
"learning_rate": 8.267708110535449e-06,
"loss": 1.1689,
"step": 506
},
{
"epoch": 1.3956043956043955,
"grad_norm": 0.0982600748538971,
"learning_rate": 8.24951912908459e-06,
"loss": 1.1683,
"step": 508
},
{
"epoch": 1.401098901098901,
"grad_norm": 0.1328704059123993,
"learning_rate": 8.231255393465993e-06,
"loss": 1.1686,
"step": 510
},
{
"epoch": 1.4065934065934065,
"grad_norm": 0.1403021216392517,
"learning_rate": 8.21291732383116e-06,
"loss": 1.1654,
"step": 512
},
{
"epoch": 1.412087912087912,
"grad_norm": 0.10671450942754745,
"learning_rate": 8.19450534204163e-06,
"loss": 1.1671,
"step": 514
},
{
"epoch": 1.4175824175824177,
"grad_norm": 0.10965994000434875,
"learning_rate": 8.176019871659263e-06,
"loss": 1.1791,
"step": 516
},
{
"epoch": 1.4230769230769231,
"grad_norm": 0.09535852819681168,
"learning_rate": 8.157461337936506e-06,
"loss": 1.1654,
"step": 518
},
{
"epoch": 1.4285714285714286,
"grad_norm": 0.09910845011472702,
"learning_rate": 8.138830167806601e-06,
"loss": 1.1612,
"step": 520
},
{
"epoch": 1.434065934065934,
"grad_norm": 0.08044712990522385,
"learning_rate": 8.120126789873775e-06,
"loss": 1.169,
"step": 522
},
{
"epoch": 1.4395604395604396,
"grad_norm": 0.11208353191614151,
"learning_rate": 8.10135163440336e-06,
"loss": 1.1676,
"step": 524
},
{
"epoch": 1.445054945054945,
"grad_norm": 0.09039778262376785,
"learning_rate": 8.08250513331192e-06,
"loss": 1.1647,
"step": 526
},
{
"epoch": 1.4505494505494505,
"grad_norm": 0.08799196779727936,
"learning_rate": 8.063587720157298e-06,
"loss": 1.1594,
"step": 528
},
{
"epoch": 1.456043956043956,
"grad_norm": 0.09703148156404495,
"learning_rate": 8.044599830128643e-06,
"loss": 1.1751,
"step": 530
},
{
"epoch": 1.4615384615384617,
"grad_norm": 0.10245784372091293,
"learning_rate": 8.02554190003641e-06,
"loss": 1.1746,
"step": 532
},
{
"epoch": 1.4670329670329672,
"grad_norm": 0.09949609637260437,
"learning_rate": 8.006414368302297e-06,
"loss": 1.1696,
"step": 534
},
{
"epoch": 1.4725274725274726,
"grad_norm": 0.11576993763446808,
"learning_rate": 7.98721767494917e-06,
"loss": 1.1662,
"step": 536
},
{
"epoch": 1.478021978021978,
"grad_norm": 0.09676701575517654,
"learning_rate": 7.967952261590936e-06,
"loss": 1.1652,
"step": 538
},
{
"epoch": 1.4835164835164836,
"grad_norm": 0.10620467364788055,
"learning_rate": 7.94861857142238e-06,
"loss": 1.1659,
"step": 540
},
{
"epoch": 1.489010989010989,
"grad_norm": 0.09917207062244415,
"learning_rate": 7.929217049208977e-06,
"loss": 1.175,
"step": 542
},
{
"epoch": 1.4945054945054945,
"grad_norm": 0.131545290350914,
"learning_rate": 7.90974814127666e-06,
"loss": 1.1619,
"step": 544
},
{
"epoch": 1.5,
"grad_norm": 0.16173675656318665,
"learning_rate": 7.890212295501542e-06,
"loss": 1.1718,
"step": 546
},
{
"epoch": 1.5054945054945055,
"grad_norm": 0.11190006136894226,
"learning_rate": 7.870609961299627e-06,
"loss": 1.1707,
"step": 548
},
{
"epoch": 1.510989010989011,
"grad_norm": 0.0947834774851799,
"learning_rate": 7.850941589616458e-06,
"loss": 1.1703,
"step": 550
},
{
"epoch": 1.5164835164835164,
"grad_norm": 0.12736476957798004,
"learning_rate": 7.831207632916757e-06,
"loss": 1.1743,
"step": 552
},
{
"epoch": 1.521978021978022,
"grad_norm": 0.17213790118694305,
"learning_rate": 7.811408545174001e-06,
"loss": 1.1763,
"step": 554
},
{
"epoch": 1.5274725274725274,
"grad_norm": 0.26195263862609863,
"learning_rate": 7.791544781859993e-06,
"loss": 1.1741,
"step": 556
},
{
"epoch": 1.5329670329670328,
"grad_norm": 0.1335473209619522,
"learning_rate": 7.771616799934372e-06,
"loss": 1.169,
"step": 558
},
{
"epoch": 1.5384615384615383,
"grad_norm": 0.11248943954706192,
"learning_rate": 7.751625057834107e-06,
"loss": 1.1611,
"step": 560
},
{
"epoch": 1.5439560439560438,
"grad_norm": 0.12497013807296753,
"learning_rate": 7.731570015462953e-06,
"loss": 1.1657,
"step": 562
},
{
"epoch": 1.5494505494505495,
"grad_norm": 0.10039670765399933,
"learning_rate": 7.711452134180865e-06,
"loss": 1.1689,
"step": 564
},
{
"epoch": 1.554945054945055,
"grad_norm": 0.09104856103658676,
"learning_rate": 7.691271876793387e-06,
"loss": 1.1691,
"step": 566
},
{
"epoch": 1.5604395604395604,
"grad_norm": 0.10671708732843399,
"learning_rate": 7.67102970754101e-06,
"loss": 1.1622,
"step": 568
},
{
"epoch": 1.565934065934066,
"grad_norm": 0.10622277110815048,
"learning_rate": 7.65072609208848e-06,
"loss": 1.1716,
"step": 570
},
{
"epoch": 1.5714285714285714,
"grad_norm": 0.09694940596818924,
"learning_rate": 7.630361497514104e-06,
"loss": 1.1681,
"step": 572
},
{
"epoch": 1.5769230769230769,
"grad_norm": 0.11600632965564728,
"learning_rate": 7.6099363922989845e-06,
"loss": 1.169,
"step": 574
},
{
"epoch": 1.5824175824175826,
"grad_norm": 0.10654427111148834,
"learning_rate": 7.5894512463162595e-06,
"loss": 1.1711,
"step": 576
},
{
"epoch": 1.587912087912088,
"grad_norm": 0.08246386051177979,
"learning_rate": 7.568906530820281e-06,
"loss": 1.1661,
"step": 578
},
{
"epoch": 1.5934065934065935,
"grad_norm": 0.09424301981925964,
"learning_rate": 7.5483027184357825e-06,
"loss": 1.166,
"step": 580
},
{
"epoch": 1.598901098901099,
"grad_norm": 0.08908534795045853,
"learning_rate": 7.527640283147003e-06,
"loss": 1.1698,
"step": 582
},
{
"epoch": 1.6043956043956045,
"grad_norm": 0.10692695528268814,
"learning_rate": 7.50691970028678e-06,
"loss": 1.1701,
"step": 584
},
{
"epoch": 1.6043956043956045,
"eval_loss": 1.1695584058761597,
"eval_runtime": 299.2734,
"eval_samples_per_second": 17.295,
"eval_steps_per_second": 0.271,
"step": 584
},
{
"epoch": 1.60989010989011,
"grad_norm": 0.09625241905450821,
"learning_rate": 7.486141446525619e-06,
"loss": 1.1734,
"step": 586
},
{
"epoch": 1.6153846153846154,
"grad_norm": 0.09392037242650986,
"learning_rate": 7.465305999860728e-06,
"loss": 1.1698,
"step": 588
},
{
"epoch": 1.620879120879121,
"grad_norm": 0.08401469886302948,
"learning_rate": 7.444413839605017e-06,
"loss": 1.1659,
"step": 590
},
{
"epoch": 1.6263736263736264,
"grad_norm": 0.11073325574398041,
"learning_rate": 7.423465446376079e-06,
"loss": 1.166,
"step": 592
},
{
"epoch": 1.6318681318681318,
"grad_norm": 0.09533084183931351,
"learning_rate": 7.402461302085121e-06,
"loss": 1.1734,
"step": 594
},
{
"epoch": 1.6373626373626373,
"grad_norm": 0.08371090888977051,
"learning_rate": 7.381401889925894e-06,
"loss": 1.167,
"step": 596
},
{
"epoch": 1.6428571428571428,
"grad_norm": 0.0968126654624939,
"learning_rate": 7.360287694363566e-06,
"loss": 1.1679,
"step": 598
},
{
"epoch": 1.6483516483516483,
"grad_norm": 0.09040292352437973,
"learning_rate": 7.3391192011235764e-06,
"loss": 1.1754,
"step": 600
},
{
"epoch": 1.6538461538461537,
"grad_norm": 0.09428130835294724,
"learning_rate": 7.317896897180472e-06,
"loss": 1.1713,
"step": 602
},
{
"epoch": 1.6593406593406592,
"grad_norm": 0.08703873306512833,
"learning_rate": 7.296621270746691e-06,
"loss": 1.1688,
"step": 604
},
{
"epoch": 1.6648351648351647,
"grad_norm": 0.09342585504055023,
"learning_rate": 7.275292811261346e-06,
"loss": 1.1651,
"step": 606
},
{
"epoch": 1.6703296703296702,
"grad_norm": 0.11370475590229034,
"learning_rate": 7.253912009378953e-06,
"loss": 1.1721,
"step": 608
},
{
"epoch": 1.6758241758241759,
"grad_norm": 0.0810871347784996,
"learning_rate": 7.2324793569581474e-06,
"loss": 1.174,
"step": 610
},
{
"epoch": 1.6813186813186813,
"grad_norm": 0.09331346303224564,
"learning_rate": 7.210995347050372e-06,
"loss": 1.1691,
"step": 612
},
{
"epoch": 1.6868131868131868,
"grad_norm": 0.1314326524734497,
"learning_rate": 7.189460473888535e-06,
"loss": 1.1672,
"step": 614
},
{
"epoch": 1.6923076923076923,
"grad_norm": 0.12525762617588043,
"learning_rate": 7.167875232875632e-06,
"loss": 1.1645,
"step": 616
},
{
"epoch": 1.6978021978021978,
"grad_norm": 0.13776274025440216,
"learning_rate": 7.146240120573358e-06,
"loss": 1.1693,
"step": 618
},
{
"epoch": 1.7032967032967035,
"grad_norm": 0.1336820125579834,
"learning_rate": 7.124555634690684e-06,
"loss": 1.1649,
"step": 620
},
{
"epoch": 1.708791208791209,
"grad_norm": 0.12508928775787354,
"learning_rate": 7.1028222740724e-06,
"loss": 1.169,
"step": 622
},
{
"epoch": 1.7142857142857144,
"grad_norm": 0.10577461123466492,
"learning_rate": 7.081040538687649e-06,
"loss": 1.1575,
"step": 624
},
{
"epoch": 1.7197802197802199,
"grad_norm": 0.11875636875629425,
"learning_rate": 7.059210929618416e-06,
"loss": 1.1666,
"step": 626
},
{
"epoch": 1.7252747252747254,
"grad_norm": 0.10840770602226257,
"learning_rate": 7.037333949048005e-06,
"loss": 1.1664,
"step": 628
},
{
"epoch": 1.7307692307692308,
"grad_norm": 0.11352021992206573,
"learning_rate": 7.0154101002494914e-06,
"loss": 1.1643,
"step": 630
},
{
"epoch": 1.7362637362637363,
"grad_norm": 0.08095169812440872,
"learning_rate": 6.993439887574133e-06,
"loss": 1.1691,
"step": 632
},
{
"epoch": 1.7417582417582418,
"grad_norm": 0.09175344556570053,
"learning_rate": 6.971423816439782e-06,
"loss": 1.1664,
"step": 634
},
{
"epoch": 1.7472527472527473,
"grad_norm": 0.07321044057607651,
"learning_rate": 6.949362393319239e-06,
"loss": 1.1649,
"step": 636
},
{
"epoch": 1.7527472527472527,
"grad_norm": 0.10566896945238113,
"learning_rate": 6.927256125728624e-06,
"loss": 1.167,
"step": 638
},
{
"epoch": 1.7582417582417582,
"grad_norm": 0.10230179876089096,
"learning_rate": 6.905105522215684e-06,
"loss": 1.1569,
"step": 640
},
{
"epoch": 1.7637362637362637,
"grad_norm": 0.09293391555547714,
"learning_rate": 6.8829110923481e-06,
"loss": 1.1677,
"step": 642
},
{
"epoch": 1.7692307692307692,
"grad_norm": 0.08087108284235,
"learning_rate": 6.8606733467017675e-06,
"loss": 1.1691,
"step": 644
},
{
"epoch": 1.7747252747252746,
"grad_norm": 0.08191601186990738,
"learning_rate": 6.838392796849042e-06,
"loss": 1.1683,
"step": 646
},
{
"epoch": 1.7802197802197801,
"grad_norm": 0.08746813237667084,
"learning_rate": 6.816069955346986e-06,
"loss": 1.1643,
"step": 648
},
{
"epoch": 1.7857142857142856,
"grad_norm": 0.10125511139631271,
"learning_rate": 6.7937053357255585e-06,
"loss": 1.1716,
"step": 650
},
{
"epoch": 1.791208791208791,
"grad_norm": 0.08800845593214035,
"learning_rate": 6.771299452475818e-06,
"loss": 1.1721,
"step": 652
},
{
"epoch": 1.7967032967032965,
"grad_norm": 0.09477917104959488,
"learning_rate": 6.748852821038075e-06,
"loss": 1.1649,
"step": 654
},
{
"epoch": 1.8021978021978022,
"grad_norm": 0.10864371061325073,
"learning_rate": 6.7263659577900375e-06,
"loss": 1.1621,
"step": 656
},
{
"epoch": 1.8076923076923077,
"grad_norm": 0.07818944752216339,
"learning_rate": 6.703839380034945e-06,
"loss": 1.1667,
"step": 658
},
{
"epoch": 1.8131868131868132,
"grad_norm": 0.0869673416018486,
"learning_rate": 6.681273605989643e-06,
"loss": 1.1692,
"step": 660
},
{
"epoch": 1.8186813186813187,
"grad_norm": 0.10159242898225784,
"learning_rate": 6.6586691547726855e-06,
"loss": 1.1708,
"step": 662
},
{
"epoch": 1.8241758241758241,
"grad_norm": 0.20149534940719604,
"learning_rate": 6.636026546392374e-06,
"loss": 1.175,
"step": 664
},
{
"epoch": 1.8296703296703298,
"grad_norm": 0.35707324743270874,
"learning_rate": 6.613346301734813e-06,
"loss": 1.1556,
"step": 666
},
{
"epoch": 1.8351648351648353,
"grad_norm": 0.10949152708053589,
"learning_rate": 6.590628942551909e-06,
"loss": 1.1696,
"step": 668
},
{
"epoch": 1.8406593406593408,
"grad_norm": 0.07581349462270737,
"learning_rate": 6.567874991449383e-06,
"loss": 1.1701,
"step": 670
},
{
"epoch": 1.8461538461538463,
"grad_norm": 0.11389974504709244,
"learning_rate": 6.545084971874738e-06,
"loss": 1.1721,
"step": 672
},
{
"epoch": 1.8516483516483517,
"grad_norm": 0.10378382354974747,
"learning_rate": 6.522259408105223e-06,
"loss": 1.1674,
"step": 674
},
{
"epoch": 1.8571428571428572,
"grad_norm": 0.09723298251628876,
"learning_rate": 6.499398825235767e-06,
"loss": 1.1696,
"step": 676
},
{
"epoch": 1.8626373626373627,
"grad_norm": 0.10554076731204987,
"learning_rate": 6.476503749166903e-06,
"loss": 1.1674,
"step": 678
},
{
"epoch": 1.8681318681318682,
"grad_norm": 0.09231211990118027,
"learning_rate": 6.453574706592676e-06,
"loss": 1.1652,
"step": 680
},
{
"epoch": 1.8736263736263736,
"grad_norm": 0.09046723693609238,
"learning_rate": 6.4306122249885105e-06,
"loss": 1.1716,
"step": 682
},
{
"epoch": 1.879120879120879,
"grad_norm": 0.09619590640068054,
"learning_rate": 6.407616832599091e-06,
"loss": 1.1699,
"step": 684
},
{
"epoch": 1.8846153846153846,
"grad_norm": 0.08878965675830841,
"learning_rate": 6.384589058426201e-06,
"loss": 1.1701,
"step": 686
},
{
"epoch": 1.89010989010989,
"grad_norm": 0.0882963240146637,
"learning_rate": 6.36152943221656e-06,
"loss": 1.1757,
"step": 688
},
{
"epoch": 1.8956043956043955,
"grad_norm": 0.10285267233848572,
"learning_rate": 6.338438484449632e-06,
"loss": 1.1612,
"step": 690
},
{
"epoch": 1.901098901098901,
"grad_norm": 0.08342345803976059,
"learning_rate": 6.31531674632542e-06,
"loss": 1.1696,
"step": 692
},
{
"epoch": 1.9065934065934065,
"grad_norm": 0.09322002530097961,
"learning_rate": 6.292164749752256e-06,
"loss": 1.1733,
"step": 694
},
{
"epoch": 1.912087912087912,
"grad_norm": 0.10594601929187775,
"learning_rate": 6.268983027334557e-06,
"loss": 1.1675,
"step": 696
},
{
"epoch": 1.9175824175824174,
"grad_norm": 0.10415042191743851,
"learning_rate": 6.245772112360568e-06,
"loss": 1.1706,
"step": 698
},
{
"epoch": 1.9230769230769231,
"grad_norm": 0.08474233746528625,
"learning_rate": 6.222532538790107e-06,
"loss": 1.1717,
"step": 700
},
{
"epoch": 1.9285714285714286,
"grad_norm": 0.09690573066473007,
"learning_rate": 6.199264841242267e-06,
"loss": 1.1579,
"step": 702
},
{
"epoch": 1.934065934065934,
"grad_norm": 0.09174709767103195,
"learning_rate": 6.17596955498313e-06,
"loss": 1.1707,
"step": 704
},
{
"epoch": 1.9395604395604396,
"grad_norm": 0.09256250411272049,
"learning_rate": 6.1526472159134454e-06,
"loss": 1.1675,
"step": 706
},
{
"epoch": 1.945054945054945,
"grad_norm": 0.115385040640831,
"learning_rate": 6.129298360556304e-06,
"loss": 1.1666,
"step": 708
},
{
"epoch": 1.9505494505494505,
"grad_norm": 0.09870624542236328,
"learning_rate": 6.105923526044794e-06,
"loss": 1.1712,
"step": 710
},
{
"epoch": 1.9560439560439562,
"grad_norm": 0.09220367670059204,
"learning_rate": 6.08252325010965e-06,
"loss": 1.1702,
"step": 712
},
{
"epoch": 1.9615384615384617,
"grad_norm": 0.08229216188192368,
"learning_rate": 6.059098071066874e-06,
"loss": 1.1641,
"step": 714
},
{
"epoch": 1.9670329670329672,
"grad_norm": 0.08601760119199753,
"learning_rate": 6.035648527805359e-06,
"loss": 1.1658,
"step": 716
},
{
"epoch": 1.9725274725274726,
"grad_norm": 0.09920581430196762,
"learning_rate": 6.012175159774488e-06,
"loss": 1.1627,
"step": 718
},
{
"epoch": 1.978021978021978,
"grad_norm": 0.09022627770900726,
"learning_rate": 5.988678506971726e-06,
"loss": 1.1723,
"step": 720
},
{
"epoch": 1.9835164835164836,
"grad_norm": 0.08053141087293625,
"learning_rate": 5.965159109930196e-06,
"loss": 1.1785,
"step": 722
},
{
"epoch": 1.989010989010989,
"grad_norm": 0.07670550793409348,
"learning_rate": 5.941617509706247e-06,
"loss": 1.1606,
"step": 724
},
{
"epoch": 1.9945054945054945,
"grad_norm": 0.0779750868678093,
"learning_rate": 5.9180542478670025e-06,
"loss": 1.1624,
"step": 726
},
{
"epoch": 2.0,
"grad_norm": 0.0955984815955162,
"learning_rate": 5.894469866477905e-06,
"loss": 1.1647,
"step": 728
},
{
"epoch": 2.0054945054945055,
"grad_norm": 0.08561510592699051,
"learning_rate": 5.87086490809025e-06,
"loss": 1.1675,
"step": 730
},
{
"epoch": 2.0054945054945055,
"eval_loss": 1.168563961982727,
"eval_runtime": 298.1262,
"eval_samples_per_second": 17.362,
"eval_steps_per_second": 0.272,
"step": 730
},
{
"epoch": 2.010989010989011,
"grad_norm": 0.08887989073991776,
"learning_rate": 5.847239915728695e-06,
"loss": 1.1711,
"step": 732
},
{
"epoch": 2.0164835164835164,
"grad_norm": 0.1793489307165146,
"learning_rate": 5.823595432878775e-06,
"loss": 1.1689,
"step": 734
},
{
"epoch": 2.021978021978022,
"grad_norm": 0.14795880019664764,
"learning_rate": 5.799932003474398e-06,
"loss": 1.1692,
"step": 736
},
{
"epoch": 2.0274725274725274,
"grad_norm": 0.08862569183111191,
"learning_rate": 5.776250171885329e-06,
"loss": 1.1653,
"step": 738
},
{
"epoch": 2.032967032967033,
"grad_norm": 0.10424422472715378,
"learning_rate": 5.752550482904674e-06,
"loss": 1.1651,
"step": 740
},
{
"epoch": 2.0384615384615383,
"grad_norm": 0.0930488184094429,
"learning_rate": 5.728833481736339e-06,
"loss": 1.1666,
"step": 742
},
{
"epoch": 2.043956043956044,
"grad_norm": 0.1007128432393074,
"learning_rate": 5.705099713982491e-06,
"loss": 1.1737,
"step": 744
},
{
"epoch": 2.0494505494505493,
"grad_norm": 0.0997442975640297,
"learning_rate": 5.6813497256310124e-06,
"loss": 1.1701,
"step": 746
},
{
"epoch": 2.0549450549450547,
"grad_norm": 0.09021608531475067,
"learning_rate": 5.6575840630429295e-06,
"loss": 1.1677,
"step": 748
},
{
"epoch": 2.0604395604395602,
"grad_norm": 0.08044509589672089,
"learning_rate": 5.633803272939851e-06,
"loss": 1.1696,
"step": 750
},
{
"epoch": 2.065934065934066,
"grad_norm": 0.08909393101930618,
"learning_rate": 5.610007902391387e-06,
"loss": 1.1616,
"step": 752
},
{
"epoch": 2.0714285714285716,
"grad_norm": 0.08932027965784073,
"learning_rate": 5.586198498802577e-06,
"loss": 1.1623,
"step": 754
},
{
"epoch": 2.076923076923077,
"grad_norm": 0.1681545227766037,
"learning_rate": 5.562375609901273e-06,
"loss": 1.1654,
"step": 756
},
{
"epoch": 2.0824175824175826,
"grad_norm": 0.1894630342721939,
"learning_rate": 5.538539783725556e-06,
"loss": 1.1681,
"step": 758
},
{
"epoch": 2.087912087912088,
"grad_norm": 0.11349444836378098,
"learning_rate": 5.51469156861113e-06,
"loss": 1.1688,
"step": 760
},
{
"epoch": 2.0934065934065935,
"grad_norm": 0.08764316141605377,
"learning_rate": 5.490831513178698e-06,
"loss": 1.1718,
"step": 762
},
{
"epoch": 2.098901098901099,
"grad_norm": 0.08182983100414276,
"learning_rate": 5.466960166321348e-06,
"loss": 1.1698,
"step": 764
},
{
"epoch": 2.1043956043956045,
"grad_norm": 0.08446374535560608,
"learning_rate": 5.44307807719192e-06,
"loss": 1.1605,
"step": 766
},
{
"epoch": 2.10989010989011,
"grad_norm": 0.06986381858587265,
"learning_rate": 5.4191857951903825e-06,
"loss": 1.1661,
"step": 768
},
{
"epoch": 2.1153846153846154,
"grad_norm": 0.09340260177850723,
"learning_rate": 5.395283869951184e-06,
"loss": 1.1685,
"step": 770
},
{
"epoch": 2.120879120879121,
"grad_norm": 0.10494975745677948,
"learning_rate": 5.371372851330612e-06,
"loss": 1.1748,
"step": 772
},
{
"epoch": 2.1263736263736264,
"grad_norm": 0.09624402970075607,
"learning_rate": 5.347453289394146e-06,
"loss": 1.1683,
"step": 774
},
{
"epoch": 2.131868131868132,
"grad_norm": 0.09158363193273544,
"learning_rate": 5.3235257344037996e-06,
"loss": 1.1693,
"step": 776
},
{
"epoch": 2.1373626373626373,
"grad_norm": 0.07997617870569229,
"learning_rate": 5.29959073680547e-06,
"loss": 1.1671,
"step": 778
},
{
"epoch": 2.142857142857143,
"grad_norm": 0.08307652175426483,
"learning_rate": 5.275648847216263e-06,
"loss": 1.1648,
"step": 780
},
{
"epoch": 2.1483516483516483,
"grad_norm": 0.09365742653608322,
"learning_rate": 5.251700616411836e-06,
"loss": 1.1738,
"step": 782
},
{
"epoch": 2.1538461538461537,
"grad_norm": 0.09766220301389694,
"learning_rate": 5.22774659531372e-06,
"loss": 1.1649,
"step": 784
},
{
"epoch": 2.159340659340659,
"grad_norm": 0.10807793587446213,
"learning_rate": 5.203787334976655e-06,
"loss": 1.1728,
"step": 786
},
{
"epoch": 2.1648351648351647,
"grad_norm": 0.07477546483278275,
"learning_rate": 5.179823386575908e-06,
"loss": 1.1701,
"step": 788
},
{
"epoch": 2.17032967032967,
"grad_norm": 0.08150489628314972,
"learning_rate": 5.155855301394585e-06,
"loss": 1.1672,
"step": 790
},
{
"epoch": 2.1758241758241756,
"grad_norm": 0.11139123886823654,
"learning_rate": 5.131883630810966e-06,
"loss": 1.172,
"step": 792
},
{
"epoch": 2.181318681318681,
"grad_norm": 0.09428671002388,
"learning_rate": 5.107908926285813e-06,
"loss": 1.1662,
"step": 794
},
{
"epoch": 2.186813186813187,
"grad_norm": 0.11421363055706024,
"learning_rate": 5.083931739349675e-06,
"loss": 1.1666,
"step": 796
},
{
"epoch": 2.1923076923076925,
"grad_norm": 0.08080089837312698,
"learning_rate": 5.059952621590216e-06,
"loss": 1.1681,
"step": 798
},
{
"epoch": 2.197802197802198,
"grad_norm": 0.09634223580360413,
"learning_rate": 5.035972124639511e-06,
"loss": 1.1654,
"step": 800
},
{
"epoch": 2.2032967032967035,
"grad_norm": 0.08151613175868988,
"learning_rate": 5.011990800161369e-06,
"loss": 1.1668,
"step": 802
},
{
"epoch": 2.208791208791209,
"grad_norm": 0.09357167035341263,
"learning_rate": 4.988009199838632e-06,
"loss": 1.1611,
"step": 804
},
{
"epoch": 2.2142857142857144,
"grad_norm": 0.08290430158376694,
"learning_rate": 4.96402787536049e-06,
"loss": 1.1562,
"step": 806
},
{
"epoch": 2.21978021978022,
"grad_norm": 0.09985602647066116,
"learning_rate": 4.940047378409786e-06,
"loss": 1.1716,
"step": 808
},
{
"epoch": 2.2252747252747254,
"grad_norm": 0.08841745555400848,
"learning_rate": 4.9160682606503255e-06,
"loss": 1.1723,
"step": 810
},
{
"epoch": 2.230769230769231,
"grad_norm": 0.08120245486497879,
"learning_rate": 4.892091073714189e-06,
"loss": 1.1642,
"step": 812
},
{
"epoch": 2.2362637362637363,
"grad_norm": 0.08893705904483795,
"learning_rate": 4.868116369189033e-06,
"loss": 1.1614,
"step": 814
},
{
"epoch": 2.241758241758242,
"grad_norm": 0.16162922978401184,
"learning_rate": 4.844144698605418e-06,
"loss": 1.1682,
"step": 816
},
{
"epoch": 2.2472527472527473,
"grad_norm": 0.08067131042480469,
"learning_rate": 4.820176613424095e-06,
"loss": 1.1676,
"step": 818
},
{
"epoch": 2.2527472527472527,
"grad_norm": 0.11158166825771332,
"learning_rate": 4.796212665023345e-06,
"loss": 1.1731,
"step": 820
},
{
"epoch": 2.258241758241758,
"grad_norm": 0.08565322309732437,
"learning_rate": 4.7722534046862805e-06,
"loss": 1.1624,
"step": 822
},
{
"epoch": 2.2637362637362637,
"grad_norm": 0.09021101146936417,
"learning_rate": 4.748299383588167e-06,
"loss": 1.1649,
"step": 824
},
{
"epoch": 2.269230769230769,
"grad_norm": 0.09416112303733826,
"learning_rate": 4.7243511527837374e-06,
"loss": 1.1643,
"step": 826
},
{
"epoch": 2.2747252747252746,
"grad_norm": 0.10305804759263992,
"learning_rate": 4.7004092631945315e-06,
"loss": 1.1709,
"step": 828
},
{
"epoch": 2.28021978021978,
"grad_norm": 0.11213865131139755,
"learning_rate": 4.6764742655962e-06,
"loss": 1.1684,
"step": 830
},
{
"epoch": 2.2857142857142856,
"grad_norm": 0.10108992457389832,
"learning_rate": 4.652546710605857e-06,
"loss": 1.1619,
"step": 832
},
{
"epoch": 2.291208791208791,
"grad_norm": 0.08697344362735748,
"learning_rate": 4.628627148669391e-06,
"loss": 1.1686,
"step": 834
},
{
"epoch": 2.2967032967032965,
"grad_norm": 0.10428237169981003,
"learning_rate": 4.604716130048818e-06,
"loss": 1.1668,
"step": 836
},
{
"epoch": 2.302197802197802,
"grad_norm": 0.08793749660253525,
"learning_rate": 4.580814204809618e-06,
"loss": 1.1713,
"step": 838
},
{
"epoch": 2.3076923076923075,
"grad_norm": 0.11877471953630447,
"learning_rate": 4.5569219228080805e-06,
"loss": 1.162,
"step": 840
},
{
"epoch": 2.313186813186813,
"grad_norm": 0.08249559253454208,
"learning_rate": 4.5330398336786526e-06,
"loss": 1.1678,
"step": 842
},
{
"epoch": 2.3186813186813184,
"grad_norm": 0.11602500081062317,
"learning_rate": 4.509168486821304e-06,
"loss": 1.164,
"step": 844
},
{
"epoch": 2.3241758241758244,
"grad_norm": 0.11740805953741074,
"learning_rate": 4.48530843138887e-06,
"loss": 1.1643,
"step": 846
},
{
"epoch": 2.32967032967033,
"grad_norm": 0.10009398311376572,
"learning_rate": 4.4614602162744455e-06,
"loss": 1.1629,
"step": 848
},
{
"epoch": 2.3351648351648353,
"grad_norm": 0.07898547500371933,
"learning_rate": 4.4376243900987296e-06,
"loss": 1.1568,
"step": 850
},
{
"epoch": 2.340659340659341,
"grad_norm": 0.08736059069633484,
"learning_rate": 4.413801501197424e-06,
"loss": 1.1703,
"step": 852
},
{
"epoch": 2.3461538461538463,
"grad_norm": 0.08960308879613876,
"learning_rate": 4.389992097608613e-06,
"loss": 1.1632,
"step": 854
},
{
"epoch": 2.3516483516483517,
"grad_norm": 0.10276441276073456,
"learning_rate": 4.366196727060152e-06,
"loss": 1.1634,
"step": 856
},
{
"epoch": 2.357142857142857,
"grad_norm": 0.07904151827096939,
"learning_rate": 4.342415936957073e-06,
"loss": 1.1709,
"step": 858
},
{
"epoch": 2.3626373626373627,
"grad_norm": 0.07296542823314667,
"learning_rate": 4.318650274368989e-06,
"loss": 1.1672,
"step": 860
},
{
"epoch": 2.368131868131868,
"grad_norm": 0.08600784838199615,
"learning_rate": 4.294900286017509e-06,
"loss": 1.1612,
"step": 862
},
{
"epoch": 2.3736263736263736,
"grad_norm": 0.08018805086612701,
"learning_rate": 4.271166518263662e-06,
"loss": 1.1722,
"step": 864
},
{
"epoch": 2.379120879120879,
"grad_norm": 0.07395070046186447,
"learning_rate": 4.247449517095329e-06,
"loss": 1.1711,
"step": 866
},
{
"epoch": 2.3846153846153846,
"grad_norm": 0.07601243257522583,
"learning_rate": 4.223749828114672e-06,
"loss": 1.1771,
"step": 868
},
{
"epoch": 2.39010989010989,
"grad_norm": 0.09423944354057312,
"learning_rate": 4.2000679965256045e-06,
"loss": 1.1603,
"step": 870
},
{
"epoch": 2.3956043956043955,
"grad_norm": 0.08968156576156616,
"learning_rate": 4.176404567121225e-06,
"loss": 1.1608,
"step": 872
},
{
"epoch": 2.401098901098901,
"grad_norm": 0.10625026375055313,
"learning_rate": 4.152760084271305e-06,
"loss": 1.1574,
"step": 874
},
{
"epoch": 2.4065934065934065,
"grad_norm": 0.18610498309135437,
"learning_rate": 4.129135091909752e-06,
"loss": 1.1747,
"step": 876
},
{
"epoch": 2.4065934065934065,
"eval_loss": 1.1678768396377563,
"eval_runtime": 299.619,
"eval_samples_per_second": 17.275,
"eval_steps_per_second": 0.27,
"step": 876
},
{
"epoch": 2.412087912087912,
"grad_norm": 0.15964192152023315,
"learning_rate": 4.105530133522096e-06,
"loss": 1.1678,
"step": 878
},
{
"epoch": 2.4175824175824174,
"grad_norm": 0.11698123812675476,
"learning_rate": 4.081945752133e-06,
"loss": 1.1678,
"step": 880
},
{
"epoch": 2.423076923076923,
"grad_norm": 0.16632910072803497,
"learning_rate": 4.058382490293755e-06,
"loss": 1.1722,
"step": 882
},
{
"epoch": 2.4285714285714284,
"grad_norm": 0.17577840387821198,
"learning_rate": 4.034840890069805e-06,
"loss": 1.1643,
"step": 884
},
{
"epoch": 2.4340659340659343,
"grad_norm": 0.17302070558071136,
"learning_rate": 4.0113214930282765e-06,
"loss": 1.1641,
"step": 886
},
{
"epoch": 2.4395604395604398,
"grad_norm": 0.2563433349132538,
"learning_rate": 3.987824840225512e-06,
"loss": 1.1678,
"step": 888
},
{
"epoch": 2.4450549450549453,
"grad_norm": 0.19088977575302124,
"learning_rate": 3.964351472194642e-06,
"loss": 1.174,
"step": 890
},
{
"epoch": 2.4505494505494507,
"grad_norm": 0.22029054164886475,
"learning_rate": 3.940901928933127e-06,
"loss": 1.1641,
"step": 892
},
{
"epoch": 2.456043956043956,
"grad_norm": 0.26073768734931946,
"learning_rate": 3.917476749890351e-06,
"loss": 1.1594,
"step": 894
},
{
"epoch": 2.4615384615384617,
"grad_norm": 0.14887195825576782,
"learning_rate": 3.894076473955207e-06,
"loss": 1.1696,
"step": 896
},
{
"epoch": 2.467032967032967,
"grad_norm": 0.09114201366901398,
"learning_rate": 3.8707016394436985e-06,
"loss": 1.1686,
"step": 898
},
{
"epoch": 2.4725274725274726,
"grad_norm": 0.12883096933364868,
"learning_rate": 3.847352784086556e-06,
"loss": 1.1614,
"step": 900
},
{
"epoch": 2.478021978021978,
"grad_norm": 0.10688427090644836,
"learning_rate": 3.8240304450168716e-06,
"loss": 1.1583,
"step": 902
},
{
"epoch": 2.4835164835164836,
"grad_norm": 0.08104757964611053,
"learning_rate": 3.8007351587577342e-06,
"loss": 1.1713,
"step": 904
},
{
"epoch": 2.489010989010989,
"grad_norm": 0.11496758460998535,
"learning_rate": 3.777467461209895e-06,
"loss": 1.1662,
"step": 906
},
{
"epoch": 2.4945054945054945,
"grad_norm": 0.08005277067422867,
"learning_rate": 3.754227887639434e-06,
"loss": 1.1683,
"step": 908
},
{
"epoch": 2.5,
"grad_norm": 0.08510690182447433,
"learning_rate": 3.7310169726654444e-06,
"loss": 1.1603,
"step": 910
},
{
"epoch": 2.5054945054945055,
"grad_norm": 0.08159387856721878,
"learning_rate": 3.707835250247745e-06,
"loss": 1.1703,
"step": 912
},
{
"epoch": 2.510989010989011,
"grad_norm": 0.08646809309720993,
"learning_rate": 3.684683253674583e-06,
"loss": 1.1744,
"step": 914
},
{
"epoch": 2.5164835164835164,
"grad_norm": 0.10196442157030106,
"learning_rate": 3.6615615155503703e-06,
"loss": 1.1607,
"step": 916
},
{
"epoch": 2.521978021978022,
"grad_norm": 0.07023598998785019,
"learning_rate": 3.638470567783442e-06,
"loss": 1.17,
"step": 918
},
{
"epoch": 2.5274725274725274,
"grad_norm": 0.09794013947248459,
"learning_rate": 3.615410941573799e-06,
"loss": 1.1713,
"step": 920
},
{
"epoch": 2.532967032967033,
"grad_norm": 0.07678301632404327,
"learning_rate": 3.59238316740091e-06,
"loss": 1.1673,
"step": 922
},
{
"epoch": 2.5384615384615383,
"grad_norm": 0.09355033934116364,
"learning_rate": 3.5693877750114903e-06,
"loss": 1.1639,
"step": 924
},
{
"epoch": 2.543956043956044,
"grad_norm": 0.08433817327022552,
"learning_rate": 3.546425293407324e-06,
"loss": 1.1614,
"step": 926
},
{
"epoch": 2.5494505494505493,
"grad_norm": 0.08144976943731308,
"learning_rate": 3.523496250833098e-06,
"loss": 1.1599,
"step": 928
},
{
"epoch": 2.5549450549450547,
"grad_norm": 0.0796588808298111,
"learning_rate": 3.5006011747642366e-06,
"loss": 1.1667,
"step": 930
},
{
"epoch": 2.5604395604395602,
"grad_norm": 0.08032579720020294,
"learning_rate": 3.4777405918947795e-06,
"loss": 1.1612,
"step": 932
},
{
"epoch": 2.5659340659340657,
"grad_norm": 0.08575133234262466,
"learning_rate": 3.4549150281252635e-06,
"loss": 1.1695,
"step": 934
},
{
"epoch": 2.571428571428571,
"grad_norm": 0.08064404129981995,
"learning_rate": 3.4321250085506174e-06,
"loss": 1.1698,
"step": 936
},
{
"epoch": 2.5769230769230766,
"grad_norm": 0.07455819100141525,
"learning_rate": 3.4093710574480926e-06,
"loss": 1.1643,
"step": 938
},
{
"epoch": 2.5824175824175826,
"grad_norm": 0.08538255095481873,
"learning_rate": 3.386653698265189e-06,
"loss": 1.1593,
"step": 940
},
{
"epoch": 2.587912087912088,
"grad_norm": 0.08684508502483368,
"learning_rate": 3.3639734536076263e-06,
"loss": 1.1651,
"step": 942
},
{
"epoch": 2.5934065934065935,
"grad_norm": 0.08089859038591385,
"learning_rate": 3.341330845227316e-06,
"loss": 1.165,
"step": 944
},
{
"epoch": 2.598901098901099,
"grad_norm": 0.07942517846822739,
"learning_rate": 3.3187263940103587e-06,
"loss": 1.1685,
"step": 946
},
{
"epoch": 2.6043956043956045,
"grad_norm": 0.08427475392818451,
"learning_rate": 3.296160619965056e-06,
"loss": 1.1547,
"step": 948
},
{
"epoch": 2.60989010989011,
"grad_norm": 0.08193778246641159,
"learning_rate": 3.2736340422099633e-06,
"loss": 1.1651,
"step": 950
},
{
"epoch": 2.6153846153846154,
"grad_norm": 0.0785454586148262,
"learning_rate": 3.2511471789619274e-06,
"loss": 1.169,
"step": 952
},
{
"epoch": 2.620879120879121,
"grad_norm": 0.08144666999578476,
"learning_rate": 3.228700547524184e-06,
"loss": 1.1656,
"step": 954
},
{
"epoch": 2.6263736263736264,
"grad_norm": 0.08215656876564026,
"learning_rate": 3.206294664274443e-06,
"loss": 1.163,
"step": 956
},
{
"epoch": 2.631868131868132,
"grad_norm": 0.07643819600343704,
"learning_rate": 3.183930044653014e-06,
"loss": 1.1609,
"step": 958
},
{
"epoch": 2.6373626373626373,
"grad_norm": 0.06851556152105331,
"learning_rate": 3.1616072031509594e-06,
"loss": 1.168,
"step": 960
},
{
"epoch": 2.642857142857143,
"grad_norm": 0.07472985982894897,
"learning_rate": 3.139326653298236e-06,
"loss": 1.1672,
"step": 962
},
{
"epoch": 2.6483516483516483,
"grad_norm": 0.08239593356847763,
"learning_rate": 3.117088907651902e-06,
"loss": 1.1668,
"step": 964
},
{
"epoch": 2.6538461538461537,
"grad_norm": 0.07118227332830429,
"learning_rate": 3.094894477784318e-06,
"loss": 1.163,
"step": 966
},
{
"epoch": 2.659340659340659,
"grad_norm": 0.08010434359312057,
"learning_rate": 3.0727438742713766e-06,
"loss": 1.1674,
"step": 968
},
{
"epoch": 2.6648351648351647,
"grad_norm": 0.06395678967237473,
"learning_rate": 3.0506376066807632e-06,
"loss": 1.1709,
"step": 970
},
{
"epoch": 2.67032967032967,
"grad_norm": 0.07488574087619781,
"learning_rate": 3.028576183560221e-06,
"loss": 1.1616,
"step": 972
},
{
"epoch": 2.675824175824176,
"grad_norm": 0.0717063769698143,
"learning_rate": 3.006560112425867e-06,
"loss": 1.159,
"step": 974
},
{
"epoch": 2.6813186813186816,
"grad_norm": 0.0665455088019371,
"learning_rate": 2.9845898997505102e-06,
"loss": 1.1717,
"step": 976
},
{
"epoch": 2.686813186813187,
"grad_norm": 0.07839926332235336,
"learning_rate": 2.962666050951997e-06,
"loss": 1.1635,
"step": 978
},
{
"epoch": 2.6923076923076925,
"grad_norm": 0.08876995742321014,
"learning_rate": 2.940789070381587e-06,
"loss": 1.1634,
"step": 980
},
{
"epoch": 2.697802197802198,
"grad_norm": 0.09154446423053741,
"learning_rate": 2.918959461312353e-06,
"loss": 1.1651,
"step": 982
},
{
"epoch": 2.7032967032967035,
"grad_norm": 0.08424372225999832,
"learning_rate": 2.897177725927599e-06,
"loss": 1.153,
"step": 984
},
{
"epoch": 2.708791208791209,
"grad_norm": 0.11759259551763535,
"learning_rate": 2.8754443653093186e-06,
"loss": 1.1662,
"step": 986
},
{
"epoch": 2.7142857142857144,
"grad_norm": 0.08856125921010971,
"learning_rate": 2.853759879426644e-06,
"loss": 1.1644,
"step": 988
},
{
"epoch": 2.71978021978022,
"grad_norm": 0.08224651217460632,
"learning_rate": 2.8321247671243695e-06,
"loss": 1.1628,
"step": 990
},
{
"epoch": 2.7252747252747254,
"grad_norm": 0.11355423927307129,
"learning_rate": 2.8105395261114666e-06,
"loss": 1.1663,
"step": 992
},
{
"epoch": 2.730769230769231,
"grad_norm": 0.06467260420322418,
"learning_rate": 2.7890046529496284e-06,
"loss": 1.1736,
"step": 994
},
{
"epoch": 2.7362637362637363,
"grad_norm": 0.08972840011119843,
"learning_rate": 2.7675206430418542e-06,
"loss": 1.1692,
"step": 996
},
{
"epoch": 2.741758241758242,
"grad_norm": 0.09491213411092758,
"learning_rate": 2.7460879906210485e-06,
"loss": 1.1707,
"step": 998
},
{
"epoch": 2.7472527472527473,
"grad_norm": 0.07063695043325424,
"learning_rate": 2.7247071887386544e-06,
"loss": 1.1605,
"step": 1000
},
{
"epoch": 2.7527472527472527,
"grad_norm": 0.07039818912744522,
"learning_rate": 2.70337872925331e-06,
"loss": 1.1648,
"step": 1002
},
{
"epoch": 2.758241758241758,
"grad_norm": 0.07515694946050644,
"learning_rate": 2.68210310281953e-06,
"loss": 1.1563,
"step": 1004
},
{
"epoch": 2.7637362637362637,
"grad_norm": 0.06818992644548416,
"learning_rate": 2.6608807988764252e-06,
"loss": 1.1735,
"step": 1006
},
{
"epoch": 2.769230769230769,
"grad_norm": 0.0827600434422493,
"learning_rate": 2.6397123056364364e-06,
"loss": 1.1698,
"step": 1008
},
{
"epoch": 2.7747252747252746,
"grad_norm": 0.10222798585891724,
"learning_rate": 2.618598110074105e-06,
"loss": 1.1689,
"step": 1010
},
{
"epoch": 2.78021978021978,
"grad_norm": 0.0823441818356514,
"learning_rate": 2.5975386979148792e-06,
"loss": 1.166,
"step": 1012
},
{
"epoch": 2.7857142857142856,
"grad_norm": 0.0735846534371376,
"learning_rate": 2.576534553623925e-06,
"loss": 1.1592,
"step": 1014
},
{
"epoch": 2.791208791208791,
"grad_norm": 0.07179060578346252,
"learning_rate": 2.5555861603949832e-06,
"loss": 1.1755,
"step": 1016
},
{
"epoch": 2.7967032967032965,
"grad_norm": 0.08374089747667313,
"learning_rate": 2.534694000139273e-06,
"loss": 1.1757,
"step": 1018
},
{
"epoch": 2.802197802197802,
"grad_norm": 0.07065007835626602,
"learning_rate": 2.513858553474382e-06,
"loss": 1.1698,
"step": 1020
},
{
"epoch": 2.8076923076923075,
"grad_norm": 0.08045931160449982,
"learning_rate": 2.4930802997132213e-06,
"loss": 1.1582,
"step": 1022
},
{
"epoch": 2.8076923076923075,
"eval_loss": 1.1672074794769287,
"eval_runtime": 299.934,
"eval_samples_per_second": 17.257,
"eval_steps_per_second": 0.27,
"step": 1022
},
{
"epoch": 2.813186813186813,
"grad_norm": 0.07823072373867035,
"learning_rate": 2.4723597168529984e-06,
"loss": 1.1596,
"step": 1024
},
{
"epoch": 2.8186813186813184,
"grad_norm": 0.08508284389972687,
"learning_rate": 2.4516972815642166e-06,
"loss": 1.1723,
"step": 1026
},
{
"epoch": 2.824175824175824,
"grad_norm": 0.07922125607728958,
"learning_rate": 2.4310934691797207e-06,
"loss": 1.1617,
"step": 1028
},
{
"epoch": 2.82967032967033,
"grad_norm": 0.0768875777721405,
"learning_rate": 2.410548753683743e-06,
"loss": 1.1664,
"step": 1030
},
{
"epoch": 2.8351648351648353,
"grad_norm": 0.07595735043287277,
"learning_rate": 2.390063607701016e-06,
"loss": 1.1731,
"step": 1032
},
{
"epoch": 2.840659340659341,
"grad_norm": 0.08126144856214523,
"learning_rate": 2.369638502485897e-06,
"loss": 1.163,
"step": 1034
},
{
"epoch": 2.8461538461538463,
"grad_norm": 0.06850147992372513,
"learning_rate": 2.3492739079115214e-06,
"loss": 1.1685,
"step": 1036
},
{
"epoch": 2.8516483516483517,
"grad_norm": 0.08522983640432358,
"learning_rate": 2.3289702924589914e-06,
"loss": 1.1686,
"step": 1038
},
{
"epoch": 2.857142857142857,
"grad_norm": 0.08155310899019241,
"learning_rate": 2.3087281232066134e-06,
"loss": 1.1653,
"step": 1040
},
{
"epoch": 2.8626373626373627,
"grad_norm": 0.07040958106517792,
"learning_rate": 2.2885478658191364e-06,
"loss": 1.1557,
"step": 1042
},
{
"epoch": 2.868131868131868,
"grad_norm": 0.0749848261475563,
"learning_rate": 2.268429984537048e-06,
"loss": 1.1724,
"step": 1044
},
{
"epoch": 2.8736263736263736,
"grad_norm": 0.06682237237691879,
"learning_rate": 2.248374942165894e-06,
"loss": 1.158,
"step": 1046
},
{
"epoch": 2.879120879120879,
"grad_norm": 0.0770927369594574,
"learning_rate": 2.2283832000656304e-06,
"loss": 1.1676,
"step": 1048
},
{
"epoch": 2.8846153846153846,
"grad_norm": 0.08294253051280975,
"learning_rate": 2.2084552181400087e-06,
"loss": 1.1654,
"step": 1050
},
{
"epoch": 2.89010989010989,
"grad_norm": 0.08596468716859818,
"learning_rate": 2.188591454826e-06,
"loss": 1.1689,
"step": 1052
},
{
"epoch": 2.8956043956043955,
"grad_norm": 0.08016978204250336,
"learning_rate": 2.168792367083243e-06,
"loss": 1.1614,
"step": 1054
},
{
"epoch": 2.901098901098901,
"grad_norm": 0.0838996022939682,
"learning_rate": 2.1490584103835433e-06,
"loss": 1.1658,
"step": 1056
},
{
"epoch": 2.9065934065934065,
"grad_norm": 0.07020293921232224,
"learning_rate": 2.1293900387003742e-06,
"loss": 1.1594,
"step": 1058
},
{
"epoch": 2.912087912087912,
"grad_norm": 0.07320253551006317,
"learning_rate": 2.109787704498459e-06,
"loss": 1.1635,
"step": 1060
},
{
"epoch": 2.9175824175824174,
"grad_norm": 0.0821409672498703,
"learning_rate": 2.0902518587233418e-06,
"loss": 1.1628,
"step": 1062
},
{
"epoch": 2.9230769230769234,
"grad_norm": 0.07211437821388245,
"learning_rate": 2.0707829507910237e-06,
"loss": 1.1689,
"step": 1064
},
{
"epoch": 2.928571428571429,
"grad_norm": 0.08253902941942215,
"learning_rate": 2.051381428577622e-06,
"loss": 1.1654,
"step": 1066
},
{
"epoch": 2.9340659340659343,
"grad_norm": 0.080472432076931,
"learning_rate": 2.0320477384090665e-06,
"loss": 1.1707,
"step": 1068
},
{
"epoch": 2.9395604395604398,
"grad_norm": 0.08331170678138733,
"learning_rate": 2.012782325050831e-06,
"loss": 1.1729,
"step": 1070
},
{
"epoch": 2.9450549450549453,
"grad_norm": 0.07666671276092529,
"learning_rate": 1.9935856316977044e-06,
"loss": 1.1715,
"step": 1072
},
{
"epoch": 2.9505494505494507,
"grad_norm": 0.09519699960947037,
"learning_rate": 1.9744580999635902e-06,
"loss": 1.1629,
"step": 1074
},
{
"epoch": 2.956043956043956,
"grad_norm": 0.0872369259595871,
"learning_rate": 1.9554001698713572e-06,
"loss": 1.1683,
"step": 1076
},
{
"epoch": 2.9615384615384617,
"grad_norm": 0.07113870233297348,
"learning_rate": 1.936412279842705e-06,
"loss": 1.1637,
"step": 1078
},
{
"epoch": 2.967032967032967,
"grad_norm": 0.08545450866222382,
"learning_rate": 1.9174948666880805e-06,
"loss": 1.1627,
"step": 1080
},
{
"epoch": 2.9725274725274726,
"grad_norm": 0.08711759746074677,
"learning_rate": 1.8986483655966408e-06,
"loss": 1.1582,
"step": 1082
},
{
"epoch": 2.978021978021978,
"grad_norm": 0.08870179951190948,
"learning_rate": 1.879873210126229e-06,
"loss": 1.167,
"step": 1084
},
{
"epoch": 2.9835164835164836,
"grad_norm": 0.1006346121430397,
"learning_rate": 1.8611698321933991e-06,
"loss": 1.1724,
"step": 1086
},
{
"epoch": 2.989010989010989,
"grad_norm": 0.07408854365348816,
"learning_rate": 1.8425386620634961e-06,
"loss": 1.1705,
"step": 1088
},
{
"epoch": 2.9945054945054945,
"grad_norm": 0.09146919846534729,
"learning_rate": 1.8239801283407393e-06,
"loss": 1.16,
"step": 1090
},
{
"epoch": 3.0,
"grad_norm": 0.0709662064909935,
"learning_rate": 1.8054946579583732e-06,
"loss": 1.1636,
"step": 1092
},
{
"epoch": 3.0054945054945055,
"grad_norm": 0.10273056477308273,
"learning_rate": 1.787082676168842e-06,
"loss": 1.1647,
"step": 1094
},
{
"epoch": 3.010989010989011,
"grad_norm": 0.07023092359304428,
"learning_rate": 1.7687446065340074e-06,
"loss": 1.162,
"step": 1096
},
{
"epoch": 3.0164835164835164,
"grad_norm": 0.08073507994413376,
"learning_rate": 1.7504808709154104e-06,
"loss": 1.1697,
"step": 1098
},
{
"epoch": 3.021978021978022,
"grad_norm": 0.07398983091115952,
"learning_rate": 1.7322918894645525e-06,
"loss": 1.1637,
"step": 1100
},
{
"epoch": 3.0274725274725274,
"grad_norm": 0.08476796001195908,
"learning_rate": 1.7141780806132429e-06,
"loss": 1.161,
"step": 1102
},
{
"epoch": 3.032967032967033,
"grad_norm": 0.07763465493917465,
"learning_rate": 1.696139861063974e-06,
"loss": 1.1685,
"step": 1104
},
{
"epoch": 3.0384615384615383,
"grad_norm": 0.08244740962982178,
"learning_rate": 1.6781776457803227e-06,
"loss": 1.1683,
"step": 1106
},
{
"epoch": 3.043956043956044,
"grad_norm": 0.09578699618577957,
"learning_rate": 1.660291847977415e-06,
"loss": 1.1677,
"step": 1108
},
{
"epoch": 3.0494505494505493,
"grad_norm": 0.07818014919757843,
"learning_rate": 1.6424828791124159e-06,
"loss": 1.166,
"step": 1110
},
{
"epoch": 3.0549450549450547,
"grad_norm": 0.07996879518032074,
"learning_rate": 1.624751148875065e-06,
"loss": 1.1732,
"step": 1112
},
{
"epoch": 3.0604395604395602,
"grad_norm": 0.06753943115472794,
"learning_rate": 1.6070970651782514e-06,
"loss": 1.1628,
"step": 1114
},
{
"epoch": 3.065934065934066,
"grad_norm": 0.06478522717952728,
"learning_rate": 1.5895210341486279e-06,
"loss": 1.1607,
"step": 1116
},
{
"epoch": 3.0714285714285716,
"grad_norm": 0.07904413342475891,
"learning_rate": 1.5720234601172767e-06,
"loss": 1.1655,
"step": 1118
},
{
"epoch": 3.076923076923077,
"grad_norm": 0.07776004076004028,
"learning_rate": 1.5546047456103964e-06,
"loss": 1.1677,
"step": 1120
},
{
"epoch": 3.0824175824175826,
"grad_norm": 0.07198958098888397,
"learning_rate": 1.537265291340042e-06,
"loss": 1.1636,
"step": 1122
},
{
"epoch": 3.087912087912088,
"grad_norm": 0.07310649752616882,
"learning_rate": 1.5200054961949233e-06,
"loss": 1.1601,
"step": 1124
},
{
"epoch": 3.0934065934065935,
"grad_norm": 0.07591935992240906,
"learning_rate": 1.5028257572312105e-06,
"loss": 1.1649,
"step": 1126
},
{
"epoch": 3.098901098901099,
"grad_norm": 0.07674799114465714,
"learning_rate": 1.485726469663401e-06,
"loss": 1.1704,
"step": 1128
},
{
"epoch": 3.1043956043956045,
"grad_norm": 0.09393850713968277,
"learning_rate": 1.468708026855245e-06,
"loss": 1.1692,
"step": 1130
},
{
"epoch": 3.10989010989011,
"grad_norm": 0.07599301636219025,
"learning_rate": 1.4517708203106763e-06,
"loss": 1.1645,
"step": 1132
},
{
"epoch": 3.1153846153846154,
"grad_norm": 0.08221649378538132,
"learning_rate": 1.4349152396648153e-06,
"loss": 1.1631,
"step": 1134
},
{
"epoch": 3.120879120879121,
"grad_norm": 0.07295921444892883,
"learning_rate": 1.4181416726750052e-06,
"loss": 1.1612,
"step": 1136
},
{
"epoch": 3.1263736263736264,
"grad_norm": 0.079580157995224,
"learning_rate": 1.4014505052118893e-06,
"loss": 1.1599,
"step": 1138
},
{
"epoch": 3.131868131868132,
"grad_norm": 0.09730138629674911,
"learning_rate": 1.3848421212505404e-06,
"loss": 1.1632,
"step": 1140
},
{
"epoch": 3.1373626373626373,
"grad_norm": 0.09287888556718826,
"learning_rate": 1.3683169028616155e-06,
"loss": 1.1602,
"step": 1142
},
{
"epoch": 3.142857142857143,
"grad_norm": 0.07959942519664764,
"learning_rate": 1.3518752302025773e-06,
"loss": 1.1629,
"step": 1144
},
{
"epoch": 3.1483516483516483,
"grad_norm": 0.07988713681697845,
"learning_rate": 1.3355174815089477e-06,
"loss": 1.1641,
"step": 1146
},
{
"epoch": 3.1538461538461537,
"grad_norm": 0.06947878748178482,
"learning_rate": 1.3192440330856005e-06,
"loss": 1.1614,
"step": 1148
},
{
"epoch": 3.159340659340659,
"grad_norm": 0.06804593652486801,
"learning_rate": 1.30305525929811e-06,
"loss": 1.1666,
"step": 1150
},
{
"epoch": 3.1648351648351647,
"grad_norm": 0.07112333178520203,
"learning_rate": 1.2869515325641357e-06,
"loss": 1.1592,
"step": 1152
},
{
"epoch": 3.17032967032967,
"grad_norm": 0.07313236594200134,
"learning_rate": 1.2709332233448573e-06,
"loss": 1.1686,
"step": 1154
},
{
"epoch": 3.1758241758241756,
"grad_norm": 0.07543834298849106,
"learning_rate": 1.2550007001364518e-06,
"loss": 1.1626,
"step": 1156
},
{
"epoch": 3.181318681318681,
"grad_norm": 0.07730558514595032,
"learning_rate": 1.239154329461615e-06,
"loss": 1.164,
"step": 1158
},
{
"epoch": 3.186813186813187,
"grad_norm": 0.07528182864189148,
"learning_rate": 1.223394475861131e-06,
"loss": 1.1621,
"step": 1160
},
{
"epoch": 3.1923076923076925,
"grad_norm": 0.07378531992435455,
"learning_rate": 1.207721501885486e-06,
"loss": 1.162,
"step": 1162
},
{
"epoch": 3.197802197802198,
"grad_norm": 0.07334302365779877,
"learning_rate": 1.1921357680865258e-06,
"loss": 1.1642,
"step": 1164
},
{
"epoch": 3.2032967032967035,
"grad_norm": 0.07697130739688873,
"learning_rate": 1.1766376330091684e-06,
"loss": 1.1647,
"step": 1166
},
{
"epoch": 3.208791208791209,
"grad_norm": 0.0726918876171112,
"learning_rate": 1.1612274531831463e-06,
"loss": 1.1719,
"step": 1168
},
{
"epoch": 3.208791208791209,
"eval_loss": 1.1669209003448486,
"eval_runtime": 299.9103,
"eval_samples_per_second": 17.258,
"eval_steps_per_second": 0.27,
"step": 1168
},
{
"epoch": 3.2142857142857144,
"grad_norm": 0.07847239077091217,
"learning_rate": 1.1459055831148074e-06,
"loss": 1.1651,
"step": 1170
},
{
"epoch": 3.21978021978022,
"grad_norm": 0.06886722892522812,
"learning_rate": 1.1306723752789672e-06,
"loss": 1.1648,
"step": 1172
},
{
"epoch": 3.2252747252747254,
"grad_norm": 0.0860794335603714,
"learning_rate": 1.1155281801107897e-06,
"loss": 1.168,
"step": 1174
},
{
"epoch": 3.230769230769231,
"grad_norm": 0.07369523495435715,
"learning_rate": 1.1004733459977325e-06,
"loss": 1.1669,
"step": 1176
},
{
"epoch": 3.2362637362637363,
"grad_norm": 0.07247929275035858,
"learning_rate": 1.0855082192715294e-06,
"loss": 1.168,
"step": 1178
},
{
"epoch": 3.241758241758242,
"grad_norm": 0.07475174218416214,
"learning_rate": 1.0706331442002226e-06,
"loss": 1.1622,
"step": 1180
},
{
"epoch": 3.2472527472527473,
"grad_norm": 0.07838430255651474,
"learning_rate": 1.0558484629802502e-06,
"loss": 1.171,
"step": 1182
},
{
"epoch": 3.2527472527472527,
"grad_norm": 0.0696970596909523,
"learning_rate": 1.041154515728559e-06,
"loss": 1.1621,
"step": 1184
},
{
"epoch": 3.258241758241758,
"grad_norm": 0.07396616786718369,
"learning_rate": 1.0265516404747943e-06,
"loss": 1.1641,
"step": 1186
},
{
"epoch": 3.2637362637362637,
"grad_norm": 0.08104917407035828,
"learning_rate": 1.0120401731535213e-06,
"loss": 1.1663,
"step": 1188
},
{
"epoch": 3.269230769230769,
"grad_norm": 0.0944487676024437,
"learning_rate": 9.976204475964907e-07,
"loss": 1.1618,
"step": 1190
},
{
"epoch": 3.2747252747252746,
"grad_norm": 0.07517724484205246,
"learning_rate": 9.832927955249605e-07,
"loss": 1.1726,
"step": 1192
},
{
"epoch": 3.28021978021978,
"grad_norm": 0.08082670718431473,
"learning_rate": 9.690575465420733e-07,
"loss": 1.1689,
"step": 1194
},
{
"epoch": 3.2857142857142856,
"grad_norm": 0.07103955000638962,
"learning_rate": 9.549150281252633e-07,
"loss": 1.1724,
"step": 1196
},
{
"epoch": 3.291208791208791,
"grad_norm": 0.08387453854084015,
"learning_rate": 9.408655656187282e-07,
"loss": 1.1598,
"step": 1198
},
{
"epoch": 3.2967032967032965,
"grad_norm": 0.07405807077884674,
"learning_rate": 9.269094822259439e-07,
"loss": 1.1648,
"step": 1200
},
{
"epoch": 3.302197802197802,
"grad_norm": 0.07046142220497131,
"learning_rate": 9.130470990022283e-07,
"loss": 1.1707,
"step": 1202
},
{
"epoch": 3.3076923076923075,
"grad_norm": 0.07296803593635559,
"learning_rate": 8.992787348473575e-07,
"loss": 1.1642,
"step": 1204
},
{
"epoch": 3.313186813186813,
"grad_norm": 0.0734080895781517,
"learning_rate": 8.856047064982276e-07,
"loss": 1.1558,
"step": 1206
},
{
"epoch": 3.3186813186813184,
"grad_norm": 0.07395216077566147,
"learning_rate": 8.720253285215685e-07,
"loss": 1.1721,
"step": 1208
},
{
"epoch": 3.3241758241758244,
"grad_norm": 0.07748089730739594,
"learning_rate": 8.585409133067119e-07,
"loss": 1.1653,
"step": 1210
},
{
"epoch": 3.32967032967033,
"grad_norm": 0.07397377490997314,
"learning_rate": 8.451517710583934e-07,
"loss": 1.1623,
"step": 1212
},
{
"epoch": 3.3351648351648353,
"grad_norm": 0.07272295653820038,
"learning_rate": 8.318582097896316e-07,
"loss": 1.1643,
"step": 1214
},
{
"epoch": 3.340659340659341,
"grad_norm": 0.07338716834783554,
"learning_rate": 8.18660535314631e-07,
"loss": 1.1612,
"step": 1216
},
{
"epoch": 3.3461538461538463,
"grad_norm": 0.06233609840273857,
"learning_rate": 8.055590512417499e-07,
"loss": 1.1642,
"step": 1218
},
{
"epoch": 3.3516483516483517,
"grad_norm": 0.07398121803998947,
"learning_rate": 7.925540589665187e-07,
"loss": 1.1719,
"step": 1220
},
{
"epoch": 3.357142857142857,
"grad_norm": 0.0798347070813179,
"learning_rate": 7.796458576647015e-07,
"loss": 1.1669,
"step": 1222
},
{
"epoch": 3.3626373626373627,
"grad_norm": 0.06861409544944763,
"learning_rate": 7.668347442854218e-07,
"loss": 1.1698,
"step": 1224
},
{
"epoch": 3.368131868131868,
"grad_norm": 0.06391950696706772,
"learning_rate": 7.541210135443188e-07,
"loss": 1.166,
"step": 1226
},
{
"epoch": 3.3736263736263736,
"grad_norm": 0.07030785083770752,
"learning_rate": 7.415049579167783e-07,
"loss": 1.166,
"step": 1228
},
{
"epoch": 3.379120879120879,
"grad_norm": 0.07540637254714966,
"learning_rate": 7.289868676312023e-07,
"loss": 1.1643,
"step": 1230
},
{
"epoch": 3.3846153846153846,
"grad_norm": 0.06893002241849899,
"learning_rate": 7.165670306623296e-07,
"loss": 1.163,
"step": 1232
},
{
"epoch": 3.39010989010989,
"grad_norm": 0.07210598886013031,
"learning_rate": 7.042457327246088e-07,
"loss": 1.1648,
"step": 1234
},
{
"epoch": 3.3956043956043955,
"grad_norm": 0.076958067715168,
"learning_rate": 6.920232572656349e-07,
"loss": 1.1635,
"step": 1236
},
{
"epoch": 3.401098901098901,
"grad_norm": 0.07737283408641815,
"learning_rate": 6.79899885459619e-07,
"loss": 1.161,
"step": 1238
},
{
"epoch": 3.4065934065934065,
"grad_norm": 0.07089677453041077,
"learning_rate": 6.678758962009241e-07,
"loss": 1.1595,
"step": 1240
},
{
"epoch": 3.412087912087912,
"grad_norm": 0.06980301439762115,
"learning_rate": 6.559515660976506e-07,
"loss": 1.1646,
"step": 1242
},
{
"epoch": 3.4175824175824174,
"grad_norm": 0.0859365463256836,
"learning_rate": 6.441271694652701e-07,
"loss": 1.1653,
"step": 1244
},
{
"epoch": 3.423076923076923,
"grad_norm": 0.0688340812921524,
"learning_rate": 6.32402978320315e-07,
"loss": 1.171,
"step": 1246
},
{
"epoch": 3.4285714285714284,
"grad_norm": 0.07327734678983688,
"learning_rate": 6.207792623741249e-07,
"loss": 1.1664,
"step": 1248
},
{
"epoch": 3.4340659340659343,
"grad_norm": 0.06622574478387833,
"learning_rate": 6.092562890266341e-07,
"loss": 1.162,
"step": 1250
},
{
"epoch": 3.4395604395604398,
"grad_norm": 0.06825845688581467,
"learning_rate": 5.97834323360233e-07,
"loss": 1.1601,
"step": 1252
},
{
"epoch": 3.4450549450549453,
"grad_norm": 0.0715174600481987,
"learning_rate": 5.86513628133652e-07,
"loss": 1.1653,
"step": 1254
},
{
"epoch": 3.4505494505494507,
"grad_norm": 0.06850449740886688,
"learning_rate": 5.75294463775935e-07,
"loss": 1.1625,
"step": 1256
},
{
"epoch": 3.456043956043956,
"grad_norm": 0.061297595500946045,
"learning_rate": 5.641770883804365e-07,
"loss": 1.168,
"step": 1258
},
{
"epoch": 3.4615384615384617,
"grad_norm": 0.07319982349872589,
"learning_rate": 5.531617576988879e-07,
"loss": 1.1693,
"step": 1260
},
{
"epoch": 3.467032967032967,
"grad_norm": 0.07959607243537903,
"learning_rate": 5.422487251355146e-07,
"loss": 1.1625,
"step": 1262
},
{
"epoch": 3.4725274725274726,
"grad_norm": 0.06803479790687561,
"learning_rate": 5.314382417412062e-07,
"loss": 1.162,
"step": 1264
},
{
"epoch": 3.478021978021978,
"grad_norm": 0.07737120240926743,
"learning_rate": 5.207305562077403e-07,
"loss": 1.1705,
"step": 1266
},
{
"epoch": 3.4835164835164836,
"grad_norm": 0.06910370290279388,
"learning_rate": 5.101259148620618e-07,
"loss": 1.1619,
"step": 1268
},
{
"epoch": 3.489010989010989,
"grad_norm": 0.06398583203554153,
"learning_rate": 4.99624561660616e-07,
"loss": 1.1659,
"step": 1270
},
{
"epoch": 3.4945054945054945,
"grad_norm": 0.06311172246932983,
"learning_rate": 4.892267381837396e-07,
"loss": 1.1595,
"step": 1272
},
{
"epoch": 3.5,
"grad_norm": 0.06687135994434357,
"learning_rate": 4.789326836300983e-07,
"loss": 1.1639,
"step": 1274
},
{
"epoch": 3.5054945054945055,
"grad_norm": 0.06371094286441803,
"learning_rate": 4.687426348111834e-07,
"loss": 1.1644,
"step": 1276
},
{
"epoch": 3.510989010989011,
"grad_norm": 0.0776761993765831,
"learning_rate": 4.586568261458729e-07,
"loss": 1.1745,
"step": 1278
},
{
"epoch": 3.5164835164835164,
"grad_norm": 0.06816533952951431,
"learning_rate": 4.486754896550288e-07,
"loss": 1.1636,
"step": 1280
},
{
"epoch": 3.521978021978022,
"grad_norm": 0.07598984241485596,
"learning_rate": 4.3879885495616505e-07,
"loss": 1.1657,
"step": 1282
},
{
"epoch": 3.5274725274725274,
"grad_norm": 0.06942006200551987,
"learning_rate": 4.290271492581627e-07,
"loss": 1.1638,
"step": 1284
},
{
"epoch": 3.532967032967033,
"grad_norm": 0.06620027124881744,
"learning_rate": 4.1936059735604497e-07,
"loss": 1.1698,
"step": 1286
},
{
"epoch": 3.5384615384615383,
"grad_norm": 0.07451992481946945,
"learning_rate": 4.0979942162580387e-07,
"loss": 1.1663,
"step": 1288
},
{
"epoch": 3.543956043956044,
"grad_norm": 0.07096972316503525,
"learning_rate": 4.003438420192873e-07,
"loss": 1.1627,
"step": 1290
},
{
"epoch": 3.5494505494505493,
"grad_norm": 0.0655079111456871,
"learning_rate": 3.9099407605913576e-07,
"loss": 1.1645,
"step": 1292
},
{
"epoch": 3.5549450549450547,
"grad_norm": 0.06677290052175522,
"learning_rate": 3.8175033883378233e-07,
"loss": 1.1642,
"step": 1294
},
{
"epoch": 3.5604395604395602,
"grad_norm": 0.06589250266551971,
"learning_rate": 3.7261284299249967e-07,
"loss": 1.1661,
"step": 1296
},
{
"epoch": 3.5659340659340657,
"grad_norm": 0.06369265913963318,
"learning_rate": 3.63581798740511e-07,
"loss": 1.1591,
"step": 1298
},
{
"epoch": 3.571428571428571,
"grad_norm": 0.06584078073501587,
"learning_rate": 3.5465741383415684e-07,
"loss": 1.1669,
"step": 1300
},
{
"epoch": 3.5769230769230766,
"grad_norm": 0.07220979779958725,
"learning_rate": 3.4583989357611037e-07,
"loss": 1.1637,
"step": 1302
},
{
"epoch": 3.5824175824175826,
"grad_norm": 0.07124118506908417,
"learning_rate": 3.371294408106585e-07,
"loss": 1.1685,
"step": 1304
},
{
"epoch": 3.587912087912088,
"grad_norm": 0.0632823258638382,
"learning_rate": 3.285262559190322e-07,
"loss": 1.1646,
"step": 1306
},
{
"epoch": 3.5934065934065935,
"grad_norm": 0.07331771403551102,
"learning_rate": 3.20030536814801e-07,
"loss": 1.1639,
"step": 1308
},
{
"epoch": 3.598901098901099,
"grad_norm": 0.07538831979036331,
"learning_rate": 3.1164247893931575e-07,
"loss": 1.1642,
"step": 1310
},
{
"epoch": 3.6043956043956045,
"grad_norm": 0.06919172406196594,
"learning_rate": 3.033622752572157e-07,
"loss": 1.165,
"step": 1312
},
{
"epoch": 3.60989010989011,
"grad_norm": 0.0683453157544136,
"learning_rate": 2.951901162519877e-07,
"loss": 1.1657,
"step": 1314
},
{
"epoch": 3.60989010989011,
"eval_loss": 1.1666840314865112,
"eval_runtime": 300.2612,
"eval_samples_per_second": 17.238,
"eval_steps_per_second": 0.27,
"step": 1314
},
{
"epoch": 3.6153846153846154,
"grad_norm": 0.07349937409162521,
"learning_rate": 2.8712618992158656e-07,
"loss": 1.167,
"step": 1316
},
{
"epoch": 3.620879120879121,
"grad_norm": 0.0776103138923645,
"learning_rate": 2.791706817741041e-07,
"loss": 1.1628,
"step": 1318
},
{
"epoch": 3.6263736263736264,
"grad_norm": 0.06213715299963951,
"learning_rate": 2.7132377482351037e-07,
"loss": 1.1623,
"step": 1320
},
{
"epoch": 3.631868131868132,
"grad_norm": 0.06590726226568222,
"learning_rate": 2.635856495854372e-07,
"loss": 1.1675,
"step": 1322
},
{
"epoch": 3.6373626373626373,
"grad_norm": 0.06559966504573822,
"learning_rate": 2.5595648407302496e-07,
"loss": 1.1692,
"step": 1324
},
{
"epoch": 3.642857142857143,
"grad_norm": 0.0632384866476059,
"learning_rate": 2.484364537928341e-07,
"loss": 1.1632,
"step": 1326
},
{
"epoch": 3.6483516483516483,
"grad_norm": 0.07011737674474716,
"learning_rate": 2.41025731740801e-07,
"loss": 1.1626,
"step": 1328
},
{
"epoch": 3.6538461538461537,
"grad_norm": 0.06871969997882843,
"learning_rate": 2.3372448839825978e-07,
"loss": 1.1597,
"step": 1330
},
{
"epoch": 3.659340659340659,
"grad_norm": 0.06817732751369476,
"learning_rate": 2.2653289172802295e-07,
"loss": 1.1714,
"step": 1332
},
{
"epoch": 3.6648351648351647,
"grad_norm": 0.06385420262813568,
"learning_rate": 2.194511071705141e-07,
"loss": 1.1747,
"step": 1334
},
{
"epoch": 3.67032967032967,
"grad_norm": 0.06753863394260406,
"learning_rate": 2.1247929763996534e-07,
"loss": 1.1602,
"step": 1336
},
{
"epoch": 3.675824175824176,
"grad_norm": 0.0678897351026535,
"learning_rate": 2.0561762352066638e-07,
"loss": 1.1685,
"step": 1338
},
{
"epoch": 3.6813186813186816,
"grad_norm": 0.06596899032592773,
"learning_rate": 1.988662426632765e-07,
"loss": 1.1657,
"step": 1340
},
{
"epoch": 3.686813186813187,
"grad_norm": 0.0690767765045166,
"learning_rate": 1.922253103811944e-07,
"loss": 1.167,
"step": 1342
},
{
"epoch": 3.6923076923076925,
"grad_norm": 0.07675404101610184,
"learning_rate": 1.85694979446982e-07,
"loss": 1.1661,
"step": 1344
},
{
"epoch": 3.697802197802198,
"grad_norm": 0.06820071488618851,
"learning_rate": 1.7927540008885414e-07,
"loss": 1.1635,
"step": 1346
},
{
"epoch": 3.7032967032967035,
"grad_norm": 0.07236689329147339,
"learning_rate": 1.729667199872187e-07,
"loss": 1.1581,
"step": 1348
},
{
"epoch": 3.708791208791209,
"grad_norm": 0.06689772009849548,
"learning_rate": 1.6676908427128103e-07,
"loss": 1.1617,
"step": 1350
},
{
"epoch": 3.7142857142857144,
"grad_norm": 0.06681732088327408,
"learning_rate": 1.6068263551570596e-07,
"loss": 1.1732,
"step": 1352
},
{
"epoch": 3.71978021978022,
"grad_norm": 0.07003802061080933,
"learning_rate": 1.5470751373733773e-07,
"loss": 1.1631,
"step": 1354
},
{
"epoch": 3.7252747252747254,
"grad_norm": 0.06234800070524216,
"learning_rate": 1.488438563919764e-07,
"loss": 1.1696,
"step": 1356
},
{
"epoch": 3.730769230769231,
"grad_norm": 0.0727643072605133,
"learning_rate": 1.4309179837122045e-07,
"loss": 1.1654,
"step": 1358
},
{
"epoch": 3.7362637362637363,
"grad_norm": 0.0751447007060051,
"learning_rate": 1.374514719993575e-07,
"loss": 1.1647,
"step": 1360
},
{
"epoch": 3.741758241758242,
"grad_norm": 0.061782509088516235,
"learning_rate": 1.3192300703032733e-07,
"loss": 1.1672,
"step": 1362
},
{
"epoch": 3.7472527472527473,
"grad_norm": 0.06823485344648361,
"learning_rate": 1.2650653064473106e-07,
"loss": 1.1682,
"step": 1364
},
{
"epoch": 3.7527472527472527,
"grad_norm": 0.061277277767658234,
"learning_rate": 1.2120216744690716e-07,
"loss": 1.1671,
"step": 1366
},
{
"epoch": 3.758241758241758,
"grad_norm": 0.06526022404432297,
"learning_rate": 1.1601003946206723e-07,
"loss": 1.1643,
"step": 1368
},
{
"epoch": 3.7637362637362637,
"grad_norm": 0.06104410067200661,
"learning_rate": 1.1093026613348601e-07,
"loss": 1.1574,
"step": 1370
},
{
"epoch": 3.769230769230769,
"grad_norm": 0.07134535163640976,
"learning_rate": 1.0596296431975406e-07,
"loss": 1.1614,
"step": 1372
},
{
"epoch": 3.7747252747252746,
"grad_norm": 0.06213730573654175,
"learning_rate": 1.0110824829209164e-07,
"loss": 1.1657,
"step": 1374
},
{
"epoch": 3.78021978021978,
"grad_norm": 0.06847725808620453,
"learning_rate": 9.636622973171583e-08,
"loss": 1.1663,
"step": 1376
},
{
"epoch": 3.7857142857142856,
"grad_norm": 0.07414616644382477,
"learning_rate": 9.17370177272775e-08,
"loss": 1.1668,
"step": 1378
},
{
"epoch": 3.791208791208791,
"grad_norm": 0.06581508368253708,
"learning_rate": 8.72207187723445e-08,
"loss": 1.1693,
"step": 1380
},
{
"epoch": 3.7967032967032965,
"grad_norm": 0.06397808343172073,
"learning_rate": 8.281743676295639e-08,
"loss": 1.159,
"step": 1382
},
{
"epoch": 3.802197802197802,
"grad_norm": 0.06523909419775009,
"learning_rate": 7.852727299523577e-08,
"loss": 1.1562,
"step": 1384
},
{
"epoch": 3.8076923076923075,
"grad_norm": 0.06779211014509201,
"learning_rate": 7.435032616305238e-08,
"loss": 1.1727,
"step": 1386
},
{
"epoch": 3.813186813186813,
"grad_norm": 0.06538354605436325,
"learning_rate": 7.028669235575714e-08,
"loss": 1.16,
"step": 1388
},
{
"epoch": 3.8186813186813184,
"grad_norm": 0.06357073783874512,
"learning_rate": 6.633646505597113e-08,
"loss": 1.1637,
"step": 1390
},
{
"epoch": 3.824175824175824,
"grad_norm": 0.0635320246219635,
"learning_rate": 6.249973513743345e-08,
"loss": 1.1639,
"step": 1392
},
{
"epoch": 3.82967032967033,
"grad_norm": 0.06830117851495743,
"learning_rate": 5.8776590862911764e-08,
"loss": 1.1585,
"step": 1394
},
{
"epoch": 3.8351648351648353,
"grad_norm": 0.06444601714611053,
"learning_rate": 5.5167117882171104e-08,
"loss": 1.1655,
"step": 1396
},
{
"epoch": 3.840659340659341,
"grad_norm": 0.06375352293252945,
"learning_rate": 5.167139923000553e-08,
"loss": 1.167,
"step": 1398
},
{
"epoch": 3.8461538461538463,
"grad_norm": 0.06707257032394409,
"learning_rate": 4.828951532432457e-08,
"loss": 1.171,
"step": 1400
},
{
"epoch": 3.8516483516483517,
"grad_norm": 0.06371629983186722,
"learning_rate": 4.5021543964306466e-08,
"loss": 1.1648,
"step": 1402
},
{
"epoch": 3.857142857142857,
"grad_norm": 0.0662907212972641,
"learning_rate": 4.186756032860728e-08,
"loss": 1.1609,
"step": 1404
},
{
"epoch": 3.8626373626373627,
"grad_norm": 0.06402067840099335,
"learning_rate": 3.8827636973630126e-08,
"loss": 1.1653,
"step": 1406
},
{
"epoch": 3.868131868131868,
"grad_norm": 0.07026324421167374,
"learning_rate": 3.590184383185758e-08,
"loss": 1.1583,
"step": 1408
},
{
"epoch": 3.8736263736263736,
"grad_norm": 0.06799634546041489,
"learning_rate": 3.309024821024354e-08,
"loss": 1.1688,
"step": 1410
},
{
"epoch": 3.879120879120879,
"grad_norm": 0.06444913148880005,
"learning_rate": 3.039291478866169e-08,
"loss": 1.1616,
"step": 1412
},
{
"epoch": 3.8846153846153846,
"grad_norm": 0.06899133324623108,
"learning_rate": 2.7809905618422227e-08,
"loss": 1.1598,
"step": 1414
},
{
"epoch": 3.89010989010989,
"grad_norm": 0.06785506010055542,
"learning_rate": 2.534128012083914e-08,
"loss": 1.1532,
"step": 1416
},
{
"epoch": 3.8956043956043955,
"grad_norm": 0.06701900064945221,
"learning_rate": 2.298709508586794e-08,
"loss": 1.1654,
"step": 1418
},
{
"epoch": 3.901098901098901,
"grad_norm": 0.06285455822944641,
"learning_rate": 2.074740467079672e-08,
"loss": 1.157,
"step": 1420
},
{
"epoch": 3.9065934065934065,
"grad_norm": 0.06540726125240326,
"learning_rate": 1.862226039899995e-08,
"loss": 1.1608,
"step": 1422
},
{
"epoch": 3.912087912087912,
"grad_norm": 0.07258068770170212,
"learning_rate": 1.661171115875493e-08,
"loss": 1.1696,
"step": 1424
},
{
"epoch": 3.9175824175824174,
"grad_norm": 0.06025758385658264,
"learning_rate": 1.4715803202116075e-08,
"loss": 1.1644,
"step": 1426
},
{
"epoch": 3.9230769230769234,
"grad_norm": 0.07302309572696686,
"learning_rate": 1.2934580143851294e-08,
"loss": 1.1621,
"step": 1428
},
{
"epoch": 3.928571428571429,
"grad_norm": 0.05992519110441208,
"learning_rate": 1.1268082960436688e-08,
"loss": 1.1689,
"step": 1430
},
{
"epoch": 3.9340659340659343,
"grad_norm": 0.06481453776359558,
"learning_rate": 9.716349989118412e-09,
"loss": 1.166,
"step": 1432
},
{
"epoch": 3.9395604395604398,
"grad_norm": 0.07549932599067688,
"learning_rate": 8.279416927026163e-09,
"loss": 1.159,
"step": 1434
},
{
"epoch": 3.9450549450549453,
"grad_norm": 0.06642387062311172,
"learning_rate": 6.9573168303532775e-09,
"loss": 1.1602,
"step": 1436
},
{
"epoch": 3.9505494505494507,
"grad_norm": 0.0685223713517189,
"learning_rate": 5.750080113598455e-09,
"loss": 1.1659,
"step": 1438
},
{
"epoch": 3.956043956043956,
"grad_norm": 0.06598740071058273,
"learning_rate": 4.65773454886298e-09,
"loss": 1.1607,
"step": 1440
},
{
"epoch": 3.9615384615384617,
"grad_norm": 0.06653392314910889,
"learning_rate": 3.6803052652134572e-09,
"loss": 1.1545,
"step": 1442
},
{
"epoch": 3.967032967032967,
"grad_norm": 0.06884676963090897,
"learning_rate": 2.817814748104497e-09,
"loss": 1.1557,
"step": 1444
},
{
"epoch": 3.9725274725274726,
"grad_norm": 0.06552927196025848,
"learning_rate": 2.070282838859683e-09,
"loss": 1.1591,
"step": 1446
},
{
"epoch": 3.978021978021978,
"grad_norm": 0.06410259753465652,
"learning_rate": 1.4377267342158274e-09,
"loss": 1.169,
"step": 1448
},
{
"epoch": 3.9835164835164836,
"grad_norm": 0.06548061966896057,
"learning_rate": 9.201609859271765e-10,
"loss": 1.1665,
"step": 1450
},
{
"epoch": 3.989010989010989,
"grad_norm": 0.06399548798799515,
"learning_rate": 5.17597500432343e-10,
"loss": 1.1717,
"step": 1452
},
{
"epoch": 3.9945054945054945,
"grad_norm": 0.06421104073524475,
"learning_rate": 2.3004553857675082e-10,
"loss": 1.1725,
"step": 1454
},
{
"epoch": 4.0,
"grad_norm": 0.06589093804359436,
"learning_rate": 5.751171540391287e-11,
"loss": 1.1709,
"step": 1456
}
],
"logging_steps": 2,
"max_steps": 1456,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 364,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 3.7105900771564585e+18,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}