train_mrpc_1753094152 / trainer_state.json
rbelanec's picture
End of training
9e25de4 verified
{
"best_global_step": 2891,
"best_metric": 0.137603759765625,
"best_model_checkpoint": "saves/lntuning/llama-3-8b-instruct/train_mrpc_1753094152/checkpoint-2891",
"epoch": 10.0,
"eval_steps": 413,
"global_step": 8260,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.006053268765133172,
"grad_norm": 4.481978893280029,
"learning_rate": 2.421307506053269e-07,
"loss": 0.2851,
"num_input_tokens_seen": 1920,
"step": 5
},
{
"epoch": 0.012106537530266344,
"grad_norm": 4.846698760986328,
"learning_rate": 5.447941888619855e-07,
"loss": 0.1664,
"num_input_tokens_seen": 3808,
"step": 10
},
{
"epoch": 0.018159806295399514,
"grad_norm": 1.9483075141906738,
"learning_rate": 8.474576271186441e-07,
"loss": 0.2561,
"num_input_tokens_seen": 5856,
"step": 15
},
{
"epoch": 0.024213075060532687,
"grad_norm": 10.700196266174316,
"learning_rate": 1.1501210653753028e-06,
"loss": 0.3437,
"num_input_tokens_seen": 7840,
"step": 20
},
{
"epoch": 0.03026634382566586,
"grad_norm": 2.950713872909546,
"learning_rate": 1.4527845036319614e-06,
"loss": 0.3354,
"num_input_tokens_seen": 10080,
"step": 25
},
{
"epoch": 0.03631961259079903,
"grad_norm": 1.4796098470687866,
"learning_rate": 1.7554479418886198e-06,
"loss": 0.1572,
"num_input_tokens_seen": 12064,
"step": 30
},
{
"epoch": 0.0423728813559322,
"grad_norm": 6.562788963317871,
"learning_rate": 2.0581113801452785e-06,
"loss": 0.2741,
"num_input_tokens_seen": 14144,
"step": 35
},
{
"epoch": 0.048426150121065374,
"grad_norm": 2.534764528274536,
"learning_rate": 2.3607748184019373e-06,
"loss": 0.3235,
"num_input_tokens_seen": 16128,
"step": 40
},
{
"epoch": 0.05447941888619855,
"grad_norm": 2.070178508758545,
"learning_rate": 2.6634382566585957e-06,
"loss": 0.1762,
"num_input_tokens_seen": 18112,
"step": 45
},
{
"epoch": 0.06053268765133172,
"grad_norm": 3.431225061416626,
"learning_rate": 2.9661016949152545e-06,
"loss": 0.1964,
"num_input_tokens_seen": 20288,
"step": 50
},
{
"epoch": 0.06658595641646489,
"grad_norm": 2.204240083694458,
"learning_rate": 3.268765133171913e-06,
"loss": 0.2061,
"num_input_tokens_seen": 22368,
"step": 55
},
{
"epoch": 0.07263922518159806,
"grad_norm": 2.7140026092529297,
"learning_rate": 3.5714285714285714e-06,
"loss": 0.1718,
"num_input_tokens_seen": 24256,
"step": 60
},
{
"epoch": 0.07869249394673124,
"grad_norm": 1.759737491607666,
"learning_rate": 3.87409200968523e-06,
"loss": 0.1738,
"num_input_tokens_seen": 26464,
"step": 65
},
{
"epoch": 0.0847457627118644,
"grad_norm": 4.829991340637207,
"learning_rate": 4.176755447941889e-06,
"loss": 0.2368,
"num_input_tokens_seen": 28736,
"step": 70
},
{
"epoch": 0.09079903147699758,
"grad_norm": 2.778394937515259,
"learning_rate": 4.479418886198548e-06,
"loss": 0.2533,
"num_input_tokens_seen": 30848,
"step": 75
},
{
"epoch": 0.09685230024213075,
"grad_norm": 4.135257244110107,
"learning_rate": 4.782082324455206e-06,
"loss": 0.1632,
"num_input_tokens_seen": 32992,
"step": 80
},
{
"epoch": 0.10290556900726393,
"grad_norm": 5.459167957305908,
"learning_rate": 5.084745762711865e-06,
"loss": 0.2906,
"num_input_tokens_seen": 34976,
"step": 85
},
{
"epoch": 0.1089588377723971,
"grad_norm": 2.954826831817627,
"learning_rate": 5.3874092009685235e-06,
"loss": 0.2841,
"num_input_tokens_seen": 36896,
"step": 90
},
{
"epoch": 0.11501210653753027,
"grad_norm": 1.3170984983444214,
"learning_rate": 5.6900726392251815e-06,
"loss": 0.1628,
"num_input_tokens_seen": 38848,
"step": 95
},
{
"epoch": 0.12106537530266344,
"grad_norm": 1.588362455368042,
"learning_rate": 5.99273607748184e-06,
"loss": 0.1954,
"num_input_tokens_seen": 40960,
"step": 100
},
{
"epoch": 0.1271186440677966,
"grad_norm": 2.257688522338867,
"learning_rate": 6.295399515738499e-06,
"loss": 0.1837,
"num_input_tokens_seen": 43008,
"step": 105
},
{
"epoch": 0.13317191283292978,
"grad_norm": 2.7970499992370605,
"learning_rate": 6.598062953995157e-06,
"loss": 0.213,
"num_input_tokens_seen": 44992,
"step": 110
},
{
"epoch": 0.13922518159806296,
"grad_norm": 2.929041862487793,
"learning_rate": 6.900726392251816e-06,
"loss": 0.2752,
"num_input_tokens_seen": 47072,
"step": 115
},
{
"epoch": 0.14527845036319612,
"grad_norm": 5.870335578918457,
"learning_rate": 7.203389830508475e-06,
"loss": 0.2604,
"num_input_tokens_seen": 49152,
"step": 120
},
{
"epoch": 0.1513317191283293,
"grad_norm": 1.5183007717132568,
"learning_rate": 7.5060532687651345e-06,
"loss": 0.2287,
"num_input_tokens_seen": 51232,
"step": 125
},
{
"epoch": 0.15738498789346247,
"grad_norm": 4.100545883178711,
"learning_rate": 7.808716707021792e-06,
"loss": 0.2572,
"num_input_tokens_seen": 53312,
"step": 130
},
{
"epoch": 0.16343825665859565,
"grad_norm": 7.2280168533325195,
"learning_rate": 8.111380145278451e-06,
"loss": 0.2749,
"num_input_tokens_seen": 55520,
"step": 135
},
{
"epoch": 0.1694915254237288,
"grad_norm": 0.9827978014945984,
"learning_rate": 8.41404358353511e-06,
"loss": 0.1523,
"num_input_tokens_seen": 57440,
"step": 140
},
{
"epoch": 0.17554479418886199,
"grad_norm": 1.256384253501892,
"learning_rate": 8.716707021791767e-06,
"loss": 0.1786,
"num_input_tokens_seen": 59584,
"step": 145
},
{
"epoch": 0.18159806295399517,
"grad_norm": 2.8723526000976562,
"learning_rate": 9.019370460048427e-06,
"loss": 0.199,
"num_input_tokens_seen": 61760,
"step": 150
},
{
"epoch": 0.18765133171912832,
"grad_norm": 1.4389151334762573,
"learning_rate": 9.322033898305085e-06,
"loss": 0.161,
"num_input_tokens_seen": 63744,
"step": 155
},
{
"epoch": 0.1937046004842615,
"grad_norm": 5.578898906707764,
"learning_rate": 9.624697336561745e-06,
"loss": 0.1934,
"num_input_tokens_seen": 65856,
"step": 160
},
{
"epoch": 0.19975786924939468,
"grad_norm": 3.256887435913086,
"learning_rate": 9.927360774818403e-06,
"loss": 0.2321,
"num_input_tokens_seen": 67840,
"step": 165
},
{
"epoch": 0.20581113801452786,
"grad_norm": 6.120675086975098,
"learning_rate": 1.023002421307506e-05,
"loss": 0.242,
"num_input_tokens_seen": 69920,
"step": 170
},
{
"epoch": 0.211864406779661,
"grad_norm": 2.3415868282318115,
"learning_rate": 1.053268765133172e-05,
"loss": 0.2627,
"num_input_tokens_seen": 72032,
"step": 175
},
{
"epoch": 0.2179176755447942,
"grad_norm": 2.5677762031555176,
"learning_rate": 1.0835351089588378e-05,
"loss": 0.1538,
"num_input_tokens_seen": 74112,
"step": 180
},
{
"epoch": 0.22397094430992737,
"grad_norm": 3.503406047821045,
"learning_rate": 1.1138014527845036e-05,
"loss": 0.2057,
"num_input_tokens_seen": 76288,
"step": 185
},
{
"epoch": 0.23002421307506055,
"grad_norm": 1.0483795404434204,
"learning_rate": 1.1440677966101696e-05,
"loss": 0.1511,
"num_input_tokens_seen": 78496,
"step": 190
},
{
"epoch": 0.2360774818401937,
"grad_norm": 2.0007925033569336,
"learning_rate": 1.1743341404358354e-05,
"loss": 0.1411,
"num_input_tokens_seen": 80416,
"step": 195
},
{
"epoch": 0.24213075060532688,
"grad_norm": 1.8174257278442383,
"learning_rate": 1.2046004842615012e-05,
"loss": 0.1116,
"num_input_tokens_seen": 82432,
"step": 200
},
{
"epoch": 0.24818401937046006,
"grad_norm": 1.2617318630218506,
"learning_rate": 1.2348668280871672e-05,
"loss": 0.1715,
"num_input_tokens_seen": 84416,
"step": 205
},
{
"epoch": 0.2542372881355932,
"grad_norm": 0.981200635433197,
"learning_rate": 1.2651331719128328e-05,
"loss": 0.1204,
"num_input_tokens_seen": 86336,
"step": 210
},
{
"epoch": 0.2602905569007264,
"grad_norm": 2.5177159309387207,
"learning_rate": 1.2953995157384988e-05,
"loss": 0.2313,
"num_input_tokens_seen": 88352,
"step": 215
},
{
"epoch": 0.26634382566585957,
"grad_norm": 5.912724018096924,
"learning_rate": 1.3256658595641647e-05,
"loss": 0.1681,
"num_input_tokens_seen": 90432,
"step": 220
},
{
"epoch": 0.27239709443099275,
"grad_norm": 1.0392547845840454,
"learning_rate": 1.3559322033898305e-05,
"loss": 0.1661,
"num_input_tokens_seen": 92512,
"step": 225
},
{
"epoch": 0.2784503631961259,
"grad_norm": 1.2187656164169312,
"learning_rate": 1.3861985472154965e-05,
"loss": 0.1194,
"num_input_tokens_seen": 94528,
"step": 230
},
{
"epoch": 0.2845036319612591,
"grad_norm": 2.3299198150634766,
"learning_rate": 1.4164648910411623e-05,
"loss": 0.1437,
"num_input_tokens_seen": 96576,
"step": 235
},
{
"epoch": 0.29055690072639223,
"grad_norm": 1.3196054697036743,
"learning_rate": 1.4467312348668283e-05,
"loss": 0.2938,
"num_input_tokens_seen": 98624,
"step": 240
},
{
"epoch": 0.2966101694915254,
"grad_norm": 1.6813981533050537,
"learning_rate": 1.4769975786924939e-05,
"loss": 0.2396,
"num_input_tokens_seen": 100832,
"step": 245
},
{
"epoch": 0.3026634382566586,
"grad_norm": 6.751827716827393,
"learning_rate": 1.5072639225181599e-05,
"loss": 0.3256,
"num_input_tokens_seen": 102784,
"step": 250
},
{
"epoch": 0.30871670702179177,
"grad_norm": 3.1990904808044434,
"learning_rate": 1.5375302663438258e-05,
"loss": 0.1266,
"num_input_tokens_seen": 104800,
"step": 255
},
{
"epoch": 0.31476997578692495,
"grad_norm": 7.041596412658691,
"learning_rate": 1.5677966101694916e-05,
"loss": 0.2191,
"num_input_tokens_seen": 106880,
"step": 260
},
{
"epoch": 0.32082324455205813,
"grad_norm": 2.598926305770874,
"learning_rate": 1.5980629539951574e-05,
"loss": 0.2082,
"num_input_tokens_seen": 108960,
"step": 265
},
{
"epoch": 0.3268765133171913,
"grad_norm": 1.5992330312728882,
"learning_rate": 1.6283292978208232e-05,
"loss": 0.1259,
"num_input_tokens_seen": 110944,
"step": 270
},
{
"epoch": 0.33292978208232443,
"grad_norm": 2.167006731033325,
"learning_rate": 1.6585956416464894e-05,
"loss": 0.1686,
"num_input_tokens_seen": 113088,
"step": 275
},
{
"epoch": 0.3389830508474576,
"grad_norm": 2.5932061672210693,
"learning_rate": 1.6888619854721548e-05,
"loss": 0.184,
"num_input_tokens_seen": 115072,
"step": 280
},
{
"epoch": 0.3450363196125908,
"grad_norm": 2.1988534927368164,
"learning_rate": 1.719128329297821e-05,
"loss": 0.1425,
"num_input_tokens_seen": 117280,
"step": 285
},
{
"epoch": 0.35108958837772397,
"grad_norm": 1.3688409328460693,
"learning_rate": 1.7493946731234868e-05,
"loss": 0.2165,
"num_input_tokens_seen": 119456,
"step": 290
},
{
"epoch": 0.35714285714285715,
"grad_norm": 2.1994192600250244,
"learning_rate": 1.7796610169491526e-05,
"loss": 0.1354,
"num_input_tokens_seen": 121472,
"step": 295
},
{
"epoch": 0.36319612590799033,
"grad_norm": 2.168099880218506,
"learning_rate": 1.8099273607748184e-05,
"loss": 0.1149,
"num_input_tokens_seen": 123648,
"step": 300
},
{
"epoch": 0.3692493946731235,
"grad_norm": 1.432853102684021,
"learning_rate": 1.8401937046004845e-05,
"loss": 0.2357,
"num_input_tokens_seen": 125632,
"step": 305
},
{
"epoch": 0.37530266343825663,
"grad_norm": 3.39896297454834,
"learning_rate": 1.8704600484261503e-05,
"loss": 0.1899,
"num_input_tokens_seen": 127680,
"step": 310
},
{
"epoch": 0.3813559322033898,
"grad_norm": 3.7564778327941895,
"learning_rate": 1.900726392251816e-05,
"loss": 0.1355,
"num_input_tokens_seen": 129856,
"step": 315
},
{
"epoch": 0.387409200968523,
"grad_norm": 3.254880905151367,
"learning_rate": 1.930992736077482e-05,
"loss": 0.1704,
"num_input_tokens_seen": 131872,
"step": 320
},
{
"epoch": 0.3934624697336562,
"grad_norm": 1.0621639490127563,
"learning_rate": 1.9612590799031477e-05,
"loss": 0.262,
"num_input_tokens_seen": 133984,
"step": 325
},
{
"epoch": 0.39951573849878935,
"grad_norm": 4.072643280029297,
"learning_rate": 1.9915254237288135e-05,
"loss": 0.155,
"num_input_tokens_seen": 136128,
"step": 330
},
{
"epoch": 0.40556900726392253,
"grad_norm": 1.1124200820922852,
"learning_rate": 2.0217917675544796e-05,
"loss": 0.179,
"num_input_tokens_seen": 138208,
"step": 335
},
{
"epoch": 0.4116222760290557,
"grad_norm": 2.559614896774292,
"learning_rate": 2.0520581113801454e-05,
"loss": 0.1799,
"num_input_tokens_seen": 140288,
"step": 340
},
{
"epoch": 0.41767554479418884,
"grad_norm": 5.550265789031982,
"learning_rate": 2.0823244552058112e-05,
"loss": 0.212,
"num_input_tokens_seen": 142336,
"step": 345
},
{
"epoch": 0.423728813559322,
"grad_norm": 4.045815944671631,
"learning_rate": 2.1125907990314774e-05,
"loss": 0.1164,
"num_input_tokens_seen": 144448,
"step": 350
},
{
"epoch": 0.4297820823244552,
"grad_norm": 2.1211326122283936,
"learning_rate": 2.1428571428571428e-05,
"loss": 0.2411,
"num_input_tokens_seen": 146464,
"step": 355
},
{
"epoch": 0.4358353510895884,
"grad_norm": 0.916531503200531,
"learning_rate": 2.1731234866828086e-05,
"loss": 0.1893,
"num_input_tokens_seen": 148512,
"step": 360
},
{
"epoch": 0.44188861985472155,
"grad_norm": 1.21219003200531,
"learning_rate": 2.2033898305084748e-05,
"loss": 0.1485,
"num_input_tokens_seen": 150496,
"step": 365
},
{
"epoch": 0.44794188861985473,
"grad_norm": 1.8068974018096924,
"learning_rate": 2.2336561743341405e-05,
"loss": 0.2618,
"num_input_tokens_seen": 152608,
"step": 370
},
{
"epoch": 0.4539951573849879,
"grad_norm": 2.374595880508423,
"learning_rate": 2.2639225181598063e-05,
"loss": 0.2532,
"num_input_tokens_seen": 154624,
"step": 375
},
{
"epoch": 0.4600484261501211,
"grad_norm": 0.7082160711288452,
"learning_rate": 2.2941888619854725e-05,
"loss": 0.1748,
"num_input_tokens_seen": 156608,
"step": 380
},
{
"epoch": 0.4661016949152542,
"grad_norm": 2.4511871337890625,
"learning_rate": 2.3244552058111383e-05,
"loss": 0.2267,
"num_input_tokens_seen": 158560,
"step": 385
},
{
"epoch": 0.4721549636803874,
"grad_norm": 5.554380416870117,
"learning_rate": 2.3547215496368037e-05,
"loss": 0.1376,
"num_input_tokens_seen": 160640,
"step": 390
},
{
"epoch": 0.4782082324455206,
"grad_norm": 2.231539011001587,
"learning_rate": 2.38498789346247e-05,
"loss": 0.1913,
"num_input_tokens_seen": 162784,
"step": 395
},
{
"epoch": 0.48426150121065376,
"grad_norm": 1.5285236835479736,
"learning_rate": 2.4152542372881357e-05,
"loss": 0.1789,
"num_input_tokens_seen": 164832,
"step": 400
},
{
"epoch": 0.49031476997578693,
"grad_norm": 3.0688276290893555,
"learning_rate": 2.4455205811138015e-05,
"loss": 0.1788,
"num_input_tokens_seen": 166880,
"step": 405
},
{
"epoch": 0.4963680387409201,
"grad_norm": 1.0780361890792847,
"learning_rate": 2.4757869249394676e-05,
"loss": 0.1648,
"num_input_tokens_seen": 169088,
"step": 410
},
{
"epoch": 0.5,
"eval_loss": 0.17582012712955475,
"eval_runtime": 4.5985,
"eval_samples_per_second": 79.808,
"eval_steps_per_second": 20.006,
"num_input_tokens_seen": 170336,
"step": 413
},
{
"epoch": 0.5024213075060533,
"grad_norm": 2.741290330886841,
"learning_rate": 2.5060532687651334e-05,
"loss": 0.1725,
"num_input_tokens_seen": 171232,
"step": 415
},
{
"epoch": 0.5084745762711864,
"grad_norm": 1.3940932750701904,
"learning_rate": 2.536319612590799e-05,
"loss": 0.2014,
"num_input_tokens_seen": 173376,
"step": 420
},
{
"epoch": 0.5145278450363197,
"grad_norm": 1.2796820402145386,
"learning_rate": 2.566585956416465e-05,
"loss": 0.1241,
"num_input_tokens_seen": 175424,
"step": 425
},
{
"epoch": 0.5205811138014528,
"grad_norm": 3.262789487838745,
"learning_rate": 2.5968523002421308e-05,
"loss": 0.1782,
"num_input_tokens_seen": 177664,
"step": 430
},
{
"epoch": 0.5266343825665859,
"grad_norm": 3.2393486499786377,
"learning_rate": 2.627118644067797e-05,
"loss": 0.2036,
"num_input_tokens_seen": 179648,
"step": 435
},
{
"epoch": 0.5326876513317191,
"grad_norm": 2.748382329940796,
"learning_rate": 2.6573849878934624e-05,
"loss": 0.2153,
"num_input_tokens_seen": 181664,
"step": 440
},
{
"epoch": 0.5387409200968523,
"grad_norm": 1.3553707599639893,
"learning_rate": 2.6876513317191282e-05,
"loss": 0.2238,
"num_input_tokens_seen": 183840,
"step": 445
},
{
"epoch": 0.5447941888619855,
"grad_norm": 0.9500827789306641,
"learning_rate": 2.7179176755447943e-05,
"loss": 0.1729,
"num_input_tokens_seen": 185696,
"step": 450
},
{
"epoch": 0.5508474576271186,
"grad_norm": 1.636928677558899,
"learning_rate": 2.74818401937046e-05,
"loss": 0.1541,
"num_input_tokens_seen": 187712,
"step": 455
},
{
"epoch": 0.5569007263922519,
"grad_norm": 1.2328100204467773,
"learning_rate": 2.7784503631961263e-05,
"loss": 0.172,
"num_input_tokens_seen": 189728,
"step": 460
},
{
"epoch": 0.562953995157385,
"grad_norm": 1.4917689561843872,
"learning_rate": 2.8087167070217917e-05,
"loss": 0.1796,
"num_input_tokens_seen": 191744,
"step": 465
},
{
"epoch": 0.5690072639225182,
"grad_norm": 1.818116307258606,
"learning_rate": 2.838983050847458e-05,
"loss": 0.1308,
"num_input_tokens_seen": 193888,
"step": 470
},
{
"epoch": 0.5750605326876513,
"grad_norm": 2.183380126953125,
"learning_rate": 2.8692493946731237e-05,
"loss": 0.2065,
"num_input_tokens_seen": 195904,
"step": 475
},
{
"epoch": 0.5811138014527845,
"grad_norm": 2.1599833965301514,
"learning_rate": 2.899515738498789e-05,
"loss": 0.2173,
"num_input_tokens_seen": 197952,
"step": 480
},
{
"epoch": 0.5871670702179177,
"grad_norm": 3.2142691612243652,
"learning_rate": 2.9297820823244553e-05,
"loss": 0.1674,
"num_input_tokens_seen": 199968,
"step": 485
},
{
"epoch": 0.5932203389830508,
"grad_norm": 9.413737297058105,
"learning_rate": 2.960048426150121e-05,
"loss": 0.2289,
"num_input_tokens_seen": 201984,
"step": 490
},
{
"epoch": 0.5992736077481841,
"grad_norm": 2.025810956954956,
"learning_rate": 2.9903147699757872e-05,
"loss": 0.1412,
"num_input_tokens_seen": 204000,
"step": 495
},
{
"epoch": 0.6053268765133172,
"grad_norm": 3.99399471282959,
"learning_rate": 3.0205811138014527e-05,
"loss": 0.168,
"num_input_tokens_seen": 206080,
"step": 500
},
{
"epoch": 0.6113801452784504,
"grad_norm": 3.777092933654785,
"learning_rate": 3.050847457627119e-05,
"loss": 0.1914,
"num_input_tokens_seen": 208000,
"step": 505
},
{
"epoch": 0.6174334140435835,
"grad_norm": 1.1595715284347534,
"learning_rate": 3.0811138014527846e-05,
"loss": 0.1263,
"num_input_tokens_seen": 210208,
"step": 510
},
{
"epoch": 0.6234866828087167,
"grad_norm": 2.891212224960327,
"learning_rate": 3.111380145278451e-05,
"loss": 0.1541,
"num_input_tokens_seen": 212288,
"step": 515
},
{
"epoch": 0.6295399515738499,
"grad_norm": 2.809659957885742,
"learning_rate": 3.141646489104116e-05,
"loss": 0.1527,
"num_input_tokens_seen": 214400,
"step": 520
},
{
"epoch": 0.635593220338983,
"grad_norm": 2.3840301036834717,
"learning_rate": 3.1719128329297823e-05,
"loss": 0.1366,
"num_input_tokens_seen": 216512,
"step": 525
},
{
"epoch": 0.6416464891041163,
"grad_norm": 1.3832828998565674,
"learning_rate": 3.2021791767554485e-05,
"loss": 0.1665,
"num_input_tokens_seen": 218496,
"step": 530
},
{
"epoch": 0.6476997578692494,
"grad_norm": 2.0845787525177,
"learning_rate": 3.232445520581114e-05,
"loss": 0.1143,
"num_input_tokens_seen": 220512,
"step": 535
},
{
"epoch": 0.6537530266343826,
"grad_norm": 1.3605108261108398,
"learning_rate": 3.26271186440678e-05,
"loss": 0.1541,
"num_input_tokens_seen": 222560,
"step": 540
},
{
"epoch": 0.6598062953995157,
"grad_norm": 2.2927634716033936,
"learning_rate": 3.2929782082324455e-05,
"loss": 0.1235,
"num_input_tokens_seen": 224480,
"step": 545
},
{
"epoch": 0.6658595641646489,
"grad_norm": 7.201409339904785,
"learning_rate": 3.323244552058112e-05,
"loss": 0.2002,
"num_input_tokens_seen": 226560,
"step": 550
},
{
"epoch": 0.6719128329297821,
"grad_norm": 2.5615785121917725,
"learning_rate": 3.353510895883777e-05,
"loss": 0.1185,
"num_input_tokens_seen": 228512,
"step": 555
},
{
"epoch": 0.6779661016949152,
"grad_norm": 5.765679359436035,
"learning_rate": 3.383777239709443e-05,
"loss": 0.1894,
"num_input_tokens_seen": 230528,
"step": 560
},
{
"epoch": 0.6840193704600485,
"grad_norm": 1.7863295078277588,
"learning_rate": 3.4140435835351094e-05,
"loss": 0.1473,
"num_input_tokens_seen": 232768,
"step": 565
},
{
"epoch": 0.6900726392251816,
"grad_norm": 3.400667428970337,
"learning_rate": 3.444309927360775e-05,
"loss": 0.1475,
"num_input_tokens_seen": 234784,
"step": 570
},
{
"epoch": 0.6961259079903148,
"grad_norm": 3.5374653339385986,
"learning_rate": 3.474576271186441e-05,
"loss": 0.3046,
"num_input_tokens_seen": 236832,
"step": 575
},
{
"epoch": 0.7021791767554479,
"grad_norm": 2.3441779613494873,
"learning_rate": 3.5048426150121065e-05,
"loss": 0.1825,
"num_input_tokens_seen": 238784,
"step": 580
},
{
"epoch": 0.7082324455205811,
"grad_norm": 3.510235071182251,
"learning_rate": 3.5351089588377726e-05,
"loss": 0.236,
"num_input_tokens_seen": 240768,
"step": 585
},
{
"epoch": 0.7142857142857143,
"grad_norm": 3.022887706756592,
"learning_rate": 3.565375302663439e-05,
"loss": 0.205,
"num_input_tokens_seen": 242848,
"step": 590
},
{
"epoch": 0.7203389830508474,
"grad_norm": 5.518234729766846,
"learning_rate": 3.595641646489104e-05,
"loss": 0.1348,
"num_input_tokens_seen": 244992,
"step": 595
},
{
"epoch": 0.7263922518159807,
"grad_norm": 4.24160623550415,
"learning_rate": 3.62590799031477e-05,
"loss": 0.1509,
"num_input_tokens_seen": 247040,
"step": 600
},
{
"epoch": 0.7324455205811138,
"grad_norm": 4.148669242858887,
"learning_rate": 3.656174334140436e-05,
"loss": 0.2164,
"num_input_tokens_seen": 249184,
"step": 605
},
{
"epoch": 0.738498789346247,
"grad_norm": 3.636296033859253,
"learning_rate": 3.686440677966102e-05,
"loss": 0.1869,
"num_input_tokens_seen": 251104,
"step": 610
},
{
"epoch": 0.7445520581113801,
"grad_norm": 4.738290786743164,
"learning_rate": 3.7167070217917674e-05,
"loss": 0.2009,
"num_input_tokens_seen": 253056,
"step": 615
},
{
"epoch": 0.7506053268765133,
"grad_norm": 1.276099443435669,
"learning_rate": 3.7469733656174335e-05,
"loss": 0.1407,
"num_input_tokens_seen": 255008,
"step": 620
},
{
"epoch": 0.7566585956416465,
"grad_norm": 6.177402973175049,
"learning_rate": 3.7772397094431e-05,
"loss": 0.1873,
"num_input_tokens_seen": 257120,
"step": 625
},
{
"epoch": 0.7627118644067796,
"grad_norm": 1.6612720489501953,
"learning_rate": 3.807506053268765e-05,
"loss": 0.1251,
"num_input_tokens_seen": 259168,
"step": 630
},
{
"epoch": 0.7687651331719129,
"grad_norm": 4.596778392791748,
"learning_rate": 3.837772397094431e-05,
"loss": 0.1175,
"num_input_tokens_seen": 261216,
"step": 635
},
{
"epoch": 0.774818401937046,
"grad_norm": 5.147200584411621,
"learning_rate": 3.868038740920097e-05,
"loss": 0.1912,
"num_input_tokens_seen": 263136,
"step": 640
},
{
"epoch": 0.7808716707021792,
"grad_norm": 3.2150142192840576,
"learning_rate": 3.898305084745763e-05,
"loss": 0.2389,
"num_input_tokens_seen": 265184,
"step": 645
},
{
"epoch": 0.7869249394673123,
"grad_norm": 1.7957394123077393,
"learning_rate": 3.928571428571429e-05,
"loss": 0.156,
"num_input_tokens_seen": 267264,
"step": 650
},
{
"epoch": 0.7929782082324455,
"grad_norm": 1.8125636577606201,
"learning_rate": 3.958837772397095e-05,
"loss": 0.3316,
"num_input_tokens_seen": 269216,
"step": 655
},
{
"epoch": 0.7990314769975787,
"grad_norm": 1.157749056816101,
"learning_rate": 3.9891041162227606e-05,
"loss": 0.1813,
"num_input_tokens_seen": 271264,
"step": 660
},
{
"epoch": 0.8050847457627118,
"grad_norm": 1.4555474519729614,
"learning_rate": 4.019370460048426e-05,
"loss": 0.1342,
"num_input_tokens_seen": 273280,
"step": 665
},
{
"epoch": 0.8111380145278451,
"grad_norm": 1.5828756093978882,
"learning_rate": 4.049636803874092e-05,
"loss": 0.1223,
"num_input_tokens_seen": 275392,
"step": 670
},
{
"epoch": 0.8171912832929782,
"grad_norm": 3.629384994506836,
"learning_rate": 4.0799031476997577e-05,
"loss": 0.1947,
"num_input_tokens_seen": 277440,
"step": 675
},
{
"epoch": 0.8232445520581114,
"grad_norm": 2.305537223815918,
"learning_rate": 4.110169491525424e-05,
"loss": 0.2098,
"num_input_tokens_seen": 279456,
"step": 680
},
{
"epoch": 0.8292978208232445,
"grad_norm": 1.5820642709732056,
"learning_rate": 4.14043583535109e-05,
"loss": 0.1602,
"num_input_tokens_seen": 281632,
"step": 685
},
{
"epoch": 0.8353510895883777,
"grad_norm": 3.7225077152252197,
"learning_rate": 4.170702179176756e-05,
"loss": 0.2414,
"num_input_tokens_seen": 283680,
"step": 690
},
{
"epoch": 0.8414043583535109,
"grad_norm": 2.8825156688690186,
"learning_rate": 4.2009685230024215e-05,
"loss": 0.1334,
"num_input_tokens_seen": 285760,
"step": 695
},
{
"epoch": 0.847457627118644,
"grad_norm": 6.222320556640625,
"learning_rate": 4.231234866828087e-05,
"loss": 0.1221,
"num_input_tokens_seen": 287904,
"step": 700
},
{
"epoch": 0.8535108958837773,
"grad_norm": 1.3990041017532349,
"learning_rate": 4.261501210653753e-05,
"loss": 0.1937,
"num_input_tokens_seen": 289952,
"step": 705
},
{
"epoch": 0.8595641646489104,
"grad_norm": 3.03328800201416,
"learning_rate": 4.2917675544794186e-05,
"loss": 0.129,
"num_input_tokens_seen": 291968,
"step": 710
},
{
"epoch": 0.8656174334140436,
"grad_norm": 2.7313449382781982,
"learning_rate": 4.3220338983050854e-05,
"loss": 0.2106,
"num_input_tokens_seen": 293952,
"step": 715
},
{
"epoch": 0.8716707021791767,
"grad_norm": 9.207576751708984,
"learning_rate": 4.352300242130751e-05,
"loss": 0.2177,
"num_input_tokens_seen": 296192,
"step": 720
},
{
"epoch": 0.8777239709443099,
"grad_norm": 2.5809319019317627,
"learning_rate": 4.382566585956417e-05,
"loss": 0.1338,
"num_input_tokens_seen": 298336,
"step": 725
},
{
"epoch": 0.8837772397094431,
"grad_norm": 1.9844462871551514,
"learning_rate": 4.4128329297820825e-05,
"loss": 0.1614,
"num_input_tokens_seen": 300352,
"step": 730
},
{
"epoch": 0.8898305084745762,
"grad_norm": 3.2159414291381836,
"learning_rate": 4.443099273607748e-05,
"loss": 0.1122,
"num_input_tokens_seen": 302432,
"step": 735
},
{
"epoch": 0.8958837772397095,
"grad_norm": 9.370662689208984,
"learning_rate": 4.473365617433414e-05,
"loss": 0.1848,
"num_input_tokens_seen": 304544,
"step": 740
},
{
"epoch": 0.9019370460048426,
"grad_norm": 0.8076304197311401,
"learning_rate": 4.50363196125908e-05,
"loss": 0.1298,
"num_input_tokens_seen": 306592,
"step": 745
},
{
"epoch": 0.9079903147699758,
"grad_norm": 4.535274982452393,
"learning_rate": 4.533898305084746e-05,
"loss": 0.16,
"num_input_tokens_seen": 308672,
"step": 750
},
{
"epoch": 0.914043583535109,
"grad_norm": 0.7559286952018738,
"learning_rate": 4.564164648910412e-05,
"loss": 0.0545,
"num_input_tokens_seen": 310656,
"step": 755
},
{
"epoch": 0.9200968523002422,
"grad_norm": 2.4823074340820312,
"learning_rate": 4.594430992736078e-05,
"loss": 0.2359,
"num_input_tokens_seen": 312768,
"step": 760
},
{
"epoch": 0.9261501210653753,
"grad_norm": 0.8833109736442566,
"learning_rate": 4.6246973365617434e-05,
"loss": 0.0402,
"num_input_tokens_seen": 314880,
"step": 765
},
{
"epoch": 0.9322033898305084,
"grad_norm": 2.477018117904663,
"learning_rate": 4.654963680387409e-05,
"loss": 0.0613,
"num_input_tokens_seen": 316960,
"step": 770
},
{
"epoch": 0.9382566585956417,
"grad_norm": 5.7945685386657715,
"learning_rate": 4.685230024213076e-05,
"loss": 0.1336,
"num_input_tokens_seen": 319072,
"step": 775
},
{
"epoch": 0.9443099273607748,
"grad_norm": 1.619981050491333,
"learning_rate": 4.715496368038741e-05,
"loss": 0.1738,
"num_input_tokens_seen": 321248,
"step": 780
},
{
"epoch": 0.950363196125908,
"grad_norm": 1.6304205656051636,
"learning_rate": 4.745762711864407e-05,
"loss": 0.2943,
"num_input_tokens_seen": 323168,
"step": 785
},
{
"epoch": 0.9564164648910412,
"grad_norm": 1.6372556686401367,
"learning_rate": 4.776029055690073e-05,
"loss": 0.1316,
"num_input_tokens_seen": 325280,
"step": 790
},
{
"epoch": 0.9624697336561744,
"grad_norm": 2.625636339187622,
"learning_rate": 4.806295399515739e-05,
"loss": 0.1265,
"num_input_tokens_seen": 327264,
"step": 795
},
{
"epoch": 0.9685230024213075,
"grad_norm": 2.179762363433838,
"learning_rate": 4.836561743341404e-05,
"loss": 0.1583,
"num_input_tokens_seen": 329312,
"step": 800
},
{
"epoch": 0.9745762711864406,
"grad_norm": 0.8714865446090698,
"learning_rate": 4.8668280871670705e-05,
"loss": 0.1965,
"num_input_tokens_seen": 331296,
"step": 805
},
{
"epoch": 0.9806295399515739,
"grad_norm": 0.9597650766372681,
"learning_rate": 4.8970944309927366e-05,
"loss": 0.0882,
"num_input_tokens_seen": 333280,
"step": 810
},
{
"epoch": 0.986682808716707,
"grad_norm": 1.2910470962524414,
"learning_rate": 4.927360774818402e-05,
"loss": 0.1259,
"num_input_tokens_seen": 335392,
"step": 815
},
{
"epoch": 0.9927360774818402,
"grad_norm": 1.1439247131347656,
"learning_rate": 4.957627118644068e-05,
"loss": 0.2821,
"num_input_tokens_seen": 337504,
"step": 820
},
{
"epoch": 0.9987893462469734,
"grad_norm": 1.6988706588745117,
"learning_rate": 4.9878934624697336e-05,
"loss": 0.1132,
"num_input_tokens_seen": 339488,
"step": 825
},
{
"epoch": 1.0,
"eval_loss": 0.17887082695960999,
"eval_runtime": 4.5703,
"eval_samples_per_second": 80.302,
"eval_steps_per_second": 20.13,
"num_input_tokens_seen": 339568,
"step": 826
},
{
"epoch": 1.0048426150121066,
"grad_norm": 0.8017047643661499,
"learning_rate": 4.99999799087436e-05,
"loss": 0.1615,
"num_input_tokens_seen": 341168,
"step": 830
},
{
"epoch": 1.0108958837772397,
"grad_norm": 0.8913490772247314,
"learning_rate": 4.999985712896029e-05,
"loss": 0.072,
"num_input_tokens_seen": 343216,
"step": 835
},
{
"epoch": 1.0169491525423728,
"grad_norm": 4.105673313140869,
"learning_rate": 4.9999622731750315e-05,
"loss": 0.1312,
"num_input_tokens_seen": 345200,
"step": 840
},
{
"epoch": 1.023002421307506,
"grad_norm": 0.4698481559753418,
"learning_rate": 4.999927671816018e-05,
"loss": 0.0782,
"num_input_tokens_seen": 347408,
"step": 845
},
{
"epoch": 1.0290556900726393,
"grad_norm": 4.901593208312988,
"learning_rate": 4.999881908973474e-05,
"loss": 0.219,
"num_input_tokens_seen": 349488,
"step": 850
},
{
"epoch": 1.0351089588377724,
"grad_norm": 3.6952834129333496,
"learning_rate": 4.9998249848517185e-05,
"loss": 0.3065,
"num_input_tokens_seen": 351536,
"step": 855
},
{
"epoch": 1.0411622276029056,
"grad_norm": 2.738077402114868,
"learning_rate": 4.999756899704902e-05,
"loss": 0.2376,
"num_input_tokens_seen": 353488,
"step": 860
},
{
"epoch": 1.0472154963680387,
"grad_norm": 3.399280071258545,
"learning_rate": 4.999677653837004e-05,
"loss": 0.0926,
"num_input_tokens_seen": 355504,
"step": 865
},
{
"epoch": 1.053268765133172,
"grad_norm": 9.515294075012207,
"learning_rate": 4.999587247601837e-05,
"loss": 0.2354,
"num_input_tokens_seen": 357456,
"step": 870
},
{
"epoch": 1.0593220338983051,
"grad_norm": 2.640784502029419,
"learning_rate": 4.99948568140304e-05,
"loss": 0.201,
"num_input_tokens_seen": 359568,
"step": 875
},
{
"epoch": 1.0653753026634383,
"grad_norm": 1.2339344024658203,
"learning_rate": 4.999372955694077e-05,
"loss": 0.2238,
"num_input_tokens_seen": 361616,
"step": 880
},
{
"epoch": 1.0714285714285714,
"grad_norm": 2.403195381164551,
"learning_rate": 4.999249070978237e-05,
"loss": 0.1327,
"num_input_tokens_seen": 363696,
"step": 885
},
{
"epoch": 1.0774818401937045,
"grad_norm": 2.847059965133667,
"learning_rate": 4.9991140278086316e-05,
"loss": 0.1774,
"num_input_tokens_seen": 365712,
"step": 890
},
{
"epoch": 1.0835351089588379,
"grad_norm": 0.30369001626968384,
"learning_rate": 4.998967826788191e-05,
"loss": 0.0374,
"num_input_tokens_seen": 367824,
"step": 895
},
{
"epoch": 1.089588377723971,
"grad_norm": 2.6305861473083496,
"learning_rate": 4.998810468569661e-05,
"loss": 0.1665,
"num_input_tokens_seen": 369840,
"step": 900
},
{
"epoch": 1.0956416464891041,
"grad_norm": 0.4452836513519287,
"learning_rate": 4.998641953855604e-05,
"loss": 0.1515,
"num_input_tokens_seen": 371952,
"step": 905
},
{
"epoch": 1.1016949152542372,
"grad_norm": 1.6030263900756836,
"learning_rate": 4.9984622833983906e-05,
"loss": 0.136,
"num_input_tokens_seen": 374032,
"step": 910
},
{
"epoch": 1.1077481840193704,
"grad_norm": 3.4988062381744385,
"learning_rate": 4.9982714580002e-05,
"loss": 0.1357,
"num_input_tokens_seen": 376176,
"step": 915
},
{
"epoch": 1.1138014527845037,
"grad_norm": 3.163297653198242,
"learning_rate": 4.998069478513013e-05,
"loss": 0.1427,
"num_input_tokens_seen": 378224,
"step": 920
},
{
"epoch": 1.1198547215496368,
"grad_norm": 1.0224205255508423,
"learning_rate": 4.997856345838615e-05,
"loss": 0.1399,
"num_input_tokens_seen": 380304,
"step": 925
},
{
"epoch": 1.12590799031477,
"grad_norm": 1.4912834167480469,
"learning_rate": 4.997632060928582e-05,
"loss": 0.0602,
"num_input_tokens_seen": 382416,
"step": 930
},
{
"epoch": 1.131961259079903,
"grad_norm": 1.3977404832839966,
"learning_rate": 4.997396624784284e-05,
"loss": 0.1835,
"num_input_tokens_seen": 384592,
"step": 935
},
{
"epoch": 1.1380145278450362,
"grad_norm": 0.7628365755081177,
"learning_rate": 4.9971500384568795e-05,
"loss": 0.1278,
"num_input_tokens_seen": 386640,
"step": 940
},
{
"epoch": 1.1440677966101696,
"grad_norm": 1.2870368957519531,
"learning_rate": 4.996892303047306e-05,
"loss": 0.13,
"num_input_tokens_seen": 388752,
"step": 945
},
{
"epoch": 1.1501210653753027,
"grad_norm": 12.154106140136719,
"learning_rate": 4.996623419706282e-05,
"loss": 0.225,
"num_input_tokens_seen": 390928,
"step": 950
},
{
"epoch": 1.1561743341404358,
"grad_norm": 0.07500404119491577,
"learning_rate": 4.996343389634298e-05,
"loss": 0.0603,
"num_input_tokens_seen": 393104,
"step": 955
},
{
"epoch": 1.162227602905569,
"grad_norm": 2.0933313369750977,
"learning_rate": 4.996052214081608e-05,
"loss": 0.1878,
"num_input_tokens_seen": 395184,
"step": 960
},
{
"epoch": 1.1682808716707023,
"grad_norm": 7.35886812210083,
"learning_rate": 4.995749894348232e-05,
"loss": 0.1715,
"num_input_tokens_seen": 397264,
"step": 965
},
{
"epoch": 1.1743341404358354,
"grad_norm": 4.131441593170166,
"learning_rate": 4.995436431783942e-05,
"loss": 0.1233,
"num_input_tokens_seen": 399376,
"step": 970
},
{
"epoch": 1.1803874092009685,
"grad_norm": 2.0855846405029297,
"learning_rate": 4.9951118277882636e-05,
"loss": 0.1681,
"num_input_tokens_seen": 401296,
"step": 975
},
{
"epoch": 1.1864406779661016,
"grad_norm": 4.022651672363281,
"learning_rate": 4.994776083810463e-05,
"loss": 0.1455,
"num_input_tokens_seen": 403344,
"step": 980
},
{
"epoch": 1.192493946731235,
"grad_norm": 2.833282709121704,
"learning_rate": 4.994429201349542e-05,
"loss": 0.1993,
"num_input_tokens_seen": 405328,
"step": 985
},
{
"epoch": 1.1985472154963681,
"grad_norm": 3.453665256500244,
"learning_rate": 4.994071181954237e-05,
"loss": 0.1205,
"num_input_tokens_seen": 407344,
"step": 990
},
{
"epoch": 1.2046004842615012,
"grad_norm": 4.6256422996521,
"learning_rate": 4.993702027223004e-05,
"loss": 0.142,
"num_input_tokens_seen": 409296,
"step": 995
},
{
"epoch": 1.2106537530266344,
"grad_norm": 2.930302381515503,
"learning_rate": 4.9933217388040164e-05,
"loss": 0.2749,
"num_input_tokens_seen": 411536,
"step": 1000
},
{
"epoch": 1.2167070217917675,
"grad_norm": 7.9702043533325195,
"learning_rate": 4.992930318395157e-05,
"loss": 0.0887,
"num_input_tokens_seen": 413712,
"step": 1005
},
{
"epoch": 1.2227602905569008,
"grad_norm": 8.404480934143066,
"learning_rate": 4.99252776774401e-05,
"loss": 0.1615,
"num_input_tokens_seen": 415856,
"step": 1010
},
{
"epoch": 1.228813559322034,
"grad_norm": 3.982208728790283,
"learning_rate": 4.99211408864785e-05,
"loss": 0.2316,
"num_input_tokens_seen": 417808,
"step": 1015
},
{
"epoch": 1.234866828087167,
"grad_norm": 0.8618561625480652,
"learning_rate": 4.991689282953642e-05,
"loss": 0.0428,
"num_input_tokens_seen": 419760,
"step": 1020
},
{
"epoch": 1.2409200968523002,
"grad_norm": 3.26019287109375,
"learning_rate": 4.991253352558025e-05,
"loss": 0.3529,
"num_input_tokens_seen": 421840,
"step": 1025
},
{
"epoch": 1.2469733656174333,
"grad_norm": 7.507120609283447,
"learning_rate": 4.9908062994073056e-05,
"loss": 0.2839,
"num_input_tokens_seen": 424112,
"step": 1030
},
{
"epoch": 1.2530266343825667,
"grad_norm": 2.8803584575653076,
"learning_rate": 4.990348125497454e-05,
"loss": 0.1223,
"num_input_tokens_seen": 426128,
"step": 1035
},
{
"epoch": 1.2590799031476998,
"grad_norm": 0.847564697265625,
"learning_rate": 4.9898788328740884e-05,
"loss": 0.118,
"num_input_tokens_seen": 428272,
"step": 1040
},
{
"epoch": 1.265133171912833,
"grad_norm": 0.8926573395729065,
"learning_rate": 4.989398423632471e-05,
"loss": 0.2175,
"num_input_tokens_seen": 430320,
"step": 1045
},
{
"epoch": 1.271186440677966,
"grad_norm": 1.0349339246749878,
"learning_rate": 4.988906899917496e-05,
"loss": 0.1073,
"num_input_tokens_seen": 432272,
"step": 1050
},
{
"epoch": 1.2772397094430992,
"grad_norm": 1.2034751176834106,
"learning_rate": 4.988404263923679e-05,
"loss": 0.0985,
"num_input_tokens_seen": 434384,
"step": 1055
},
{
"epoch": 1.2832929782082325,
"grad_norm": 3.1568541526794434,
"learning_rate": 4.987890517895152e-05,
"loss": 0.1696,
"num_input_tokens_seen": 436336,
"step": 1060
},
{
"epoch": 1.2893462469733656,
"grad_norm": 4.365025520324707,
"learning_rate": 4.987365664125647e-05,
"loss": 0.1524,
"num_input_tokens_seen": 438352,
"step": 1065
},
{
"epoch": 1.2953995157384988,
"grad_norm": 1.4751983880996704,
"learning_rate": 4.986829704958491e-05,
"loss": 0.1576,
"num_input_tokens_seen": 440336,
"step": 1070
},
{
"epoch": 1.3014527845036319,
"grad_norm": 2.4118525981903076,
"learning_rate": 4.986282642786594e-05,
"loss": 0.1163,
"num_input_tokens_seen": 442352,
"step": 1075
},
{
"epoch": 1.307506053268765,
"grad_norm": 0.789813756942749,
"learning_rate": 4.985724480052435e-05,
"loss": 0.0842,
"num_input_tokens_seen": 444496,
"step": 1080
},
{
"epoch": 1.3135593220338984,
"grad_norm": 1.4542982578277588,
"learning_rate": 4.985155219248057e-05,
"loss": 0.105,
"num_input_tokens_seen": 446576,
"step": 1085
},
{
"epoch": 1.3196125907990315,
"grad_norm": 6.07450532913208,
"learning_rate": 4.9845748629150514e-05,
"loss": 0.1469,
"num_input_tokens_seen": 448656,
"step": 1090
},
{
"epoch": 1.3256658595641646,
"grad_norm": 1.8046495914459229,
"learning_rate": 4.9839834136445485e-05,
"loss": 0.1416,
"num_input_tokens_seen": 450672,
"step": 1095
},
{
"epoch": 1.331719128329298,
"grad_norm": 1.6317821741104126,
"learning_rate": 4.983380874077204e-05,
"loss": 0.1155,
"num_input_tokens_seen": 452784,
"step": 1100
},
{
"epoch": 1.3377723970944309,
"grad_norm": 0.9619260430335999,
"learning_rate": 4.982767246903192e-05,
"loss": 0.1092,
"num_input_tokens_seen": 454864,
"step": 1105
},
{
"epoch": 1.3438256658595642,
"grad_norm": 4.070939540863037,
"learning_rate": 4.982142534862185e-05,
"loss": 0.2142,
"num_input_tokens_seen": 457072,
"step": 1110
},
{
"epoch": 1.3498789346246973,
"grad_norm": 10.078119277954102,
"learning_rate": 4.981506740743351e-05,
"loss": 0.1481,
"num_input_tokens_seen": 459184,
"step": 1115
},
{
"epoch": 1.3559322033898304,
"grad_norm": 5.898523330688477,
"learning_rate": 4.980859867385334e-05,
"loss": 0.1137,
"num_input_tokens_seen": 461136,
"step": 1120
},
{
"epoch": 1.3619854721549638,
"grad_norm": 5.684123516082764,
"learning_rate": 4.9802019176762434e-05,
"loss": 0.1881,
"num_input_tokens_seen": 463152,
"step": 1125
},
{
"epoch": 1.368038740920097,
"grad_norm": 2.8209903240203857,
"learning_rate": 4.9795328945536425e-05,
"loss": 0.1448,
"num_input_tokens_seen": 465232,
"step": 1130
},
{
"epoch": 1.37409200968523,
"grad_norm": 1.7147382497787476,
"learning_rate": 4.978852801004534e-05,
"loss": 0.1441,
"num_input_tokens_seen": 467248,
"step": 1135
},
{
"epoch": 1.3801452784503632,
"grad_norm": 3.1413381099700928,
"learning_rate": 4.9781616400653464e-05,
"loss": 0.1192,
"num_input_tokens_seen": 469424,
"step": 1140
},
{
"epoch": 1.3861985472154963,
"grad_norm": 6.792686462402344,
"learning_rate": 4.9774594148219225e-05,
"loss": 0.2559,
"num_input_tokens_seen": 471344,
"step": 1145
},
{
"epoch": 1.3922518159806296,
"grad_norm": 1.0777902603149414,
"learning_rate": 4.976746128409502e-05,
"loss": 0.1163,
"num_input_tokens_seen": 473296,
"step": 1150
},
{
"epoch": 1.3983050847457628,
"grad_norm": 5.556042194366455,
"learning_rate": 4.9760217840127126e-05,
"loss": 0.1898,
"num_input_tokens_seen": 475312,
"step": 1155
},
{
"epoch": 1.4043583535108959,
"grad_norm": 5.81771993637085,
"learning_rate": 4.97528638486555e-05,
"loss": 0.154,
"num_input_tokens_seen": 477360,
"step": 1160
},
{
"epoch": 1.410411622276029,
"grad_norm": 0.9776678681373596,
"learning_rate": 4.9745399342513666e-05,
"loss": 0.1191,
"num_input_tokens_seen": 479376,
"step": 1165
},
{
"epoch": 1.4164648910411621,
"grad_norm": 2.5209102630615234,
"learning_rate": 4.9737824355028584e-05,
"loss": 0.1837,
"num_input_tokens_seen": 481456,
"step": 1170
},
{
"epoch": 1.4225181598062955,
"grad_norm": 0.49560511112213135,
"learning_rate": 4.973013892002047e-05,
"loss": 0.1288,
"num_input_tokens_seen": 483632,
"step": 1175
},
{
"epoch": 1.4285714285714286,
"grad_norm": 7.962185859680176,
"learning_rate": 4.9722343071802665e-05,
"loss": 0.1998,
"num_input_tokens_seen": 485680,
"step": 1180
},
{
"epoch": 1.4346246973365617,
"grad_norm": 2.495086908340454,
"learning_rate": 4.971443684518145e-05,
"loss": 0.1084,
"num_input_tokens_seen": 487792,
"step": 1185
},
{
"epoch": 1.4406779661016949,
"grad_norm": 3.3710200786590576,
"learning_rate": 4.970642027545596e-05,
"loss": 0.2048,
"num_input_tokens_seen": 489840,
"step": 1190
},
{
"epoch": 1.446731234866828,
"grad_norm": 1.2718791961669922,
"learning_rate": 4.969829339841793e-05,
"loss": 0.0573,
"num_input_tokens_seen": 491824,
"step": 1195
},
{
"epoch": 1.4527845036319613,
"grad_norm": 3.988370180130005,
"learning_rate": 4.9690056250351626e-05,
"loss": 0.1191,
"num_input_tokens_seen": 493872,
"step": 1200
},
{
"epoch": 1.4588377723970944,
"grad_norm": 2.300257921218872,
"learning_rate": 4.9681708868033616e-05,
"loss": 0.1173,
"num_input_tokens_seen": 495952,
"step": 1205
},
{
"epoch": 1.4648910411622276,
"grad_norm": 1.824439287185669,
"learning_rate": 4.967325128873266e-05,
"loss": 0.25,
"num_input_tokens_seen": 497936,
"step": 1210
},
{
"epoch": 1.4709443099273607,
"grad_norm": 0.630806565284729,
"learning_rate": 4.96646835502095e-05,
"loss": 0.1407,
"num_input_tokens_seen": 499856,
"step": 1215
},
{
"epoch": 1.4769975786924938,
"grad_norm": 2.253023147583008,
"learning_rate": 4.965600569071671e-05,
"loss": 0.1138,
"num_input_tokens_seen": 501840,
"step": 1220
},
{
"epoch": 1.4830508474576272,
"grad_norm": 1.211021065711975,
"learning_rate": 4.9647217748998534e-05,
"loss": 0.0632,
"num_input_tokens_seen": 503952,
"step": 1225
},
{
"epoch": 1.4891041162227603,
"grad_norm": 4.336838722229004,
"learning_rate": 4.963831976429067e-05,
"loss": 0.1787,
"num_input_tokens_seen": 505936,
"step": 1230
},
{
"epoch": 1.4951573849878934,
"grad_norm": 4.364633560180664,
"learning_rate": 4.9629311776320176e-05,
"loss": 0.144,
"num_input_tokens_seen": 507984,
"step": 1235
},
{
"epoch": 1.5,
"eval_loss": 0.1783103495836258,
"eval_runtime": 4.5928,
"eval_samples_per_second": 79.908,
"eval_steps_per_second": 20.031,
"num_input_tokens_seen": 509456,
"step": 1239
},
{
"epoch": 1.5012106537530268,
"grad_norm": 0.9117156863212585,
"learning_rate": 4.962019382530521e-05,
"loss": 0.2137,
"num_input_tokens_seen": 509840,
"step": 1240
},
{
"epoch": 1.5072639225181597,
"grad_norm": 1.2304811477661133,
"learning_rate": 4.9610965951954885e-05,
"loss": 0.169,
"num_input_tokens_seen": 511856,
"step": 1245
},
{
"epoch": 1.513317191283293,
"grad_norm": 2.3721790313720703,
"learning_rate": 4.960162819746911e-05,
"loss": 0.211,
"num_input_tokens_seen": 513904,
"step": 1250
},
{
"epoch": 1.5193704600484261,
"grad_norm": 4.308636665344238,
"learning_rate": 4.9592180603538366e-05,
"loss": 0.1163,
"num_input_tokens_seen": 515824,
"step": 1255
},
{
"epoch": 1.5254237288135593,
"grad_norm": 3.0907084941864014,
"learning_rate": 4.958262321234353e-05,
"loss": 0.0857,
"num_input_tokens_seen": 517744,
"step": 1260
},
{
"epoch": 1.5314769975786926,
"grad_norm": 0.9459354877471924,
"learning_rate": 4.95729560665557e-05,
"loss": 0.1442,
"num_input_tokens_seen": 519824,
"step": 1265
},
{
"epoch": 1.5375302663438255,
"grad_norm": 4.082531452178955,
"learning_rate": 4.956317920933602e-05,
"loss": 0.1981,
"num_input_tokens_seen": 521936,
"step": 1270
},
{
"epoch": 1.5435835351089588,
"grad_norm": 3.863461971282959,
"learning_rate": 4.955329268433543e-05,
"loss": 0.1058,
"num_input_tokens_seen": 524176,
"step": 1275
},
{
"epoch": 1.549636803874092,
"grad_norm": 6.776220321655273,
"learning_rate": 4.954329653569452e-05,
"loss": 0.1502,
"num_input_tokens_seen": 526256,
"step": 1280
},
{
"epoch": 1.555690072639225,
"grad_norm": 3.519094467163086,
"learning_rate": 4.953319080804333e-05,
"loss": 0.1133,
"num_input_tokens_seen": 528304,
"step": 1285
},
{
"epoch": 1.5617433414043584,
"grad_norm": 4.169881820678711,
"learning_rate": 4.952297554650113e-05,
"loss": 0.2624,
"num_input_tokens_seen": 530320,
"step": 1290
},
{
"epoch": 1.5677966101694916,
"grad_norm": 1.6443334817886353,
"learning_rate": 4.9512650796676216e-05,
"loss": 0.0577,
"num_input_tokens_seen": 532400,
"step": 1295
},
{
"epoch": 1.5738498789346247,
"grad_norm": 0.9508858323097229,
"learning_rate": 4.9502216604665744e-05,
"loss": 0.1549,
"num_input_tokens_seen": 534416,
"step": 1300
},
{
"epoch": 1.5799031476997578,
"grad_norm": 2.909374952316284,
"learning_rate": 4.949167301705548e-05,
"loss": 0.1042,
"num_input_tokens_seen": 536592,
"step": 1305
},
{
"epoch": 1.585956416464891,
"grad_norm": 0.7363409996032715,
"learning_rate": 4.948102008091962e-05,
"loss": 0.0307,
"num_input_tokens_seen": 538704,
"step": 1310
},
{
"epoch": 1.5920096852300243,
"grad_norm": 5.999347686767578,
"learning_rate": 4.947025784382057e-05,
"loss": 0.2497,
"num_input_tokens_seen": 540912,
"step": 1315
},
{
"epoch": 1.5980629539951574,
"grad_norm": 4.051459789276123,
"learning_rate": 4.9459386353808736e-05,
"loss": 0.2243,
"num_input_tokens_seen": 542864,
"step": 1320
},
{
"epoch": 1.6041162227602905,
"grad_norm": 9.532818794250488,
"learning_rate": 4.944840565942229e-05,
"loss": 0.3938,
"num_input_tokens_seen": 544816,
"step": 1325
},
{
"epoch": 1.6101694915254239,
"grad_norm": 7.469811916351318,
"learning_rate": 4.943731580968701e-05,
"loss": 0.2847,
"num_input_tokens_seen": 546864,
"step": 1330
},
{
"epoch": 1.6162227602905568,
"grad_norm": 3.413723945617676,
"learning_rate": 4.942611685411598e-05,
"loss": 0.2188,
"num_input_tokens_seen": 548912,
"step": 1335
},
{
"epoch": 1.6222760290556901,
"grad_norm": 4.4016432762146,
"learning_rate": 4.9414808842709435e-05,
"loss": 0.2475,
"num_input_tokens_seen": 551056,
"step": 1340
},
{
"epoch": 1.6283292978208233,
"grad_norm": 1.3102144002914429,
"learning_rate": 4.940339182595451e-05,
"loss": 0.0887,
"num_input_tokens_seen": 553072,
"step": 1345
},
{
"epoch": 1.6343825665859564,
"grad_norm": 1.6671048402786255,
"learning_rate": 4.9391865854825015e-05,
"loss": 0.1674,
"num_input_tokens_seen": 555248,
"step": 1350
},
{
"epoch": 1.6404358353510897,
"grad_norm": 1.2309081554412842,
"learning_rate": 4.938023098078122e-05,
"loss": 0.0685,
"num_input_tokens_seen": 557360,
"step": 1355
},
{
"epoch": 1.6464891041162226,
"grad_norm": 3.3514671325683594,
"learning_rate": 4.93684872557696e-05,
"loss": 0.0737,
"num_input_tokens_seen": 559376,
"step": 1360
},
{
"epoch": 1.652542372881356,
"grad_norm": 4.470357418060303,
"learning_rate": 4.935663473222264e-05,
"loss": 0.1702,
"num_input_tokens_seen": 561296,
"step": 1365
},
{
"epoch": 1.658595641646489,
"grad_norm": 1.129664421081543,
"learning_rate": 4.934467346305856e-05,
"loss": 0.1448,
"num_input_tokens_seen": 563376,
"step": 1370
},
{
"epoch": 1.6646489104116222,
"grad_norm": 0.7362378835678101,
"learning_rate": 4.933260350168112e-05,
"loss": 0.0589,
"num_input_tokens_seen": 565424,
"step": 1375
},
{
"epoch": 1.6707021791767556,
"grad_norm": 0.631158173084259,
"learning_rate": 4.932042490197933e-05,
"loss": 0.0482,
"num_input_tokens_seen": 567504,
"step": 1380
},
{
"epoch": 1.6767554479418885,
"grad_norm": 3.6880412101745605,
"learning_rate": 4.930813771832728e-05,
"loss": 0.2682,
"num_input_tokens_seen": 569744,
"step": 1385
},
{
"epoch": 1.6828087167070218,
"grad_norm": 1.3964409828186035,
"learning_rate": 4.929574200558382e-05,
"loss": 0.1129,
"num_input_tokens_seen": 571888,
"step": 1390
},
{
"epoch": 1.688861985472155,
"grad_norm": 2.9864020347595215,
"learning_rate": 4.928323781909239e-05,
"loss": 0.1442,
"num_input_tokens_seen": 574000,
"step": 1395
},
{
"epoch": 1.694915254237288,
"grad_norm": 4.279382705688477,
"learning_rate": 4.927062521468068e-05,
"loss": 0.212,
"num_input_tokens_seen": 575984,
"step": 1400
},
{
"epoch": 1.7009685230024214,
"grad_norm": 1.5373003482818604,
"learning_rate": 4.92579042486605e-05,
"loss": 0.0917,
"num_input_tokens_seen": 577968,
"step": 1405
},
{
"epoch": 1.7070217917675545,
"grad_norm": 1.7001993656158447,
"learning_rate": 4.924507497782743e-05,
"loss": 0.137,
"num_input_tokens_seen": 580112,
"step": 1410
},
{
"epoch": 1.7130750605326877,
"grad_norm": 1.1870551109313965,
"learning_rate": 4.923213745946059e-05,
"loss": 0.1927,
"num_input_tokens_seen": 582160,
"step": 1415
},
{
"epoch": 1.7191283292978208,
"grad_norm": 0.5280523300170898,
"learning_rate": 4.921909175132242e-05,
"loss": 0.1633,
"num_input_tokens_seen": 584304,
"step": 1420
},
{
"epoch": 1.725181598062954,
"grad_norm": 1.6161837577819824,
"learning_rate": 4.920593791165839e-05,
"loss": 0.1462,
"num_input_tokens_seen": 586096,
"step": 1425
},
{
"epoch": 1.7312348668280872,
"grad_norm": 3.1209022998809814,
"learning_rate": 4.919267599919674e-05,
"loss": 0.1477,
"num_input_tokens_seen": 588240,
"step": 1430
},
{
"epoch": 1.7372881355932204,
"grad_norm": 1.9811747074127197,
"learning_rate": 4.917930607314823e-05,
"loss": 0.2459,
"num_input_tokens_seen": 590480,
"step": 1435
},
{
"epoch": 1.7433414043583535,
"grad_norm": 1.1726977825164795,
"learning_rate": 4.916582819320588e-05,
"loss": 0.1896,
"num_input_tokens_seen": 592528,
"step": 1440
},
{
"epoch": 1.7493946731234868,
"grad_norm": 3.542783737182617,
"learning_rate": 4.915224241954467e-05,
"loss": 0.2088,
"num_input_tokens_seen": 594416,
"step": 1445
},
{
"epoch": 1.7554479418886197,
"grad_norm": 1.4151628017425537,
"learning_rate": 4.9138548812821316e-05,
"loss": 0.1804,
"num_input_tokens_seen": 596432,
"step": 1450
},
{
"epoch": 1.761501210653753,
"grad_norm": 1.6290380954742432,
"learning_rate": 4.912474743417399e-05,
"loss": 0.0787,
"num_input_tokens_seen": 598352,
"step": 1455
},
{
"epoch": 1.7675544794188862,
"grad_norm": 1.6579142808914185,
"learning_rate": 4.911083834522199e-05,
"loss": 0.0636,
"num_input_tokens_seen": 600432,
"step": 1460
},
{
"epoch": 1.7736077481840193,
"grad_norm": 5.867772579193115,
"learning_rate": 4.909682160806556e-05,
"loss": 0.2342,
"num_input_tokens_seen": 602480,
"step": 1465
},
{
"epoch": 1.7796610169491527,
"grad_norm": 4.180382251739502,
"learning_rate": 4.908269728528553e-05,
"loss": 0.1955,
"num_input_tokens_seen": 604496,
"step": 1470
},
{
"epoch": 1.7857142857142856,
"grad_norm": 3.5099289417266846,
"learning_rate": 4.90684654399431e-05,
"loss": 0.1021,
"num_input_tokens_seen": 606672,
"step": 1475
},
{
"epoch": 1.791767554479419,
"grad_norm": 0.4963616728782654,
"learning_rate": 4.9054126135579495e-05,
"loss": 0.1071,
"num_input_tokens_seen": 608752,
"step": 1480
},
{
"epoch": 1.797820823244552,
"grad_norm": 2.4257378578186035,
"learning_rate": 4.9039679436215734e-05,
"loss": 0.129,
"num_input_tokens_seen": 610608,
"step": 1485
},
{
"epoch": 1.8038740920096852,
"grad_norm": 3.747525215148926,
"learning_rate": 4.9025125406352335e-05,
"loss": 0.2842,
"num_input_tokens_seen": 612688,
"step": 1490
},
{
"epoch": 1.8099273607748185,
"grad_norm": 0.5966554880142212,
"learning_rate": 4.9010464110968976e-05,
"loss": 0.0584,
"num_input_tokens_seen": 614704,
"step": 1495
},
{
"epoch": 1.8159806295399514,
"grad_norm": 2.613186836242676,
"learning_rate": 4.89956956155243e-05,
"loss": 0.1314,
"num_input_tokens_seen": 616784,
"step": 1500
},
{
"epoch": 1.8220338983050848,
"grad_norm": 3.0699353218078613,
"learning_rate": 4.898081998595555e-05,
"loss": 0.1275,
"num_input_tokens_seen": 618736,
"step": 1505
},
{
"epoch": 1.828087167070218,
"grad_norm": 2.9778733253479004,
"learning_rate": 4.8965837288678253e-05,
"loss": 0.1038,
"num_input_tokens_seen": 620944,
"step": 1510
},
{
"epoch": 1.834140435835351,
"grad_norm": 2.6867446899414062,
"learning_rate": 4.895074759058601e-05,
"loss": 0.1367,
"num_input_tokens_seen": 623120,
"step": 1515
},
{
"epoch": 1.8401937046004844,
"grad_norm": 1.885842204093933,
"learning_rate": 4.893555095905014e-05,
"loss": 0.074,
"num_input_tokens_seen": 625136,
"step": 1520
},
{
"epoch": 1.8462469733656173,
"grad_norm": 4.191492557525635,
"learning_rate": 4.892024746191939e-05,
"loss": 0.1355,
"num_input_tokens_seen": 627216,
"step": 1525
},
{
"epoch": 1.8523002421307506,
"grad_norm": 0.14360515773296356,
"learning_rate": 4.890483716751961e-05,
"loss": 0.0959,
"num_input_tokens_seen": 629136,
"step": 1530
},
{
"epoch": 1.8583535108958837,
"grad_norm": 2.302429676055908,
"learning_rate": 4.888932014465352e-05,
"loss": 0.1219,
"num_input_tokens_seen": 631216,
"step": 1535
},
{
"epoch": 1.8644067796610169,
"grad_norm": 2.125596046447754,
"learning_rate": 4.8873696462600303e-05,
"loss": 0.0763,
"num_input_tokens_seen": 633328,
"step": 1540
},
{
"epoch": 1.8704600484261502,
"grad_norm": 5.019554615020752,
"learning_rate": 4.8857966191115365e-05,
"loss": 0.2288,
"num_input_tokens_seen": 635440,
"step": 1545
},
{
"epoch": 1.8765133171912833,
"grad_norm": 0.8047354221343994,
"learning_rate": 4.884212940043001e-05,
"loss": 0.132,
"num_input_tokens_seen": 637360,
"step": 1550
},
{
"epoch": 1.8825665859564165,
"grad_norm": 2.835331678390503,
"learning_rate": 4.882618616125111e-05,
"loss": 0.243,
"num_input_tokens_seen": 639280,
"step": 1555
},
{
"epoch": 1.8886198547215496,
"grad_norm": 0.7895589470863342,
"learning_rate": 4.881013654476081e-05,
"loss": 0.1762,
"num_input_tokens_seen": 641328,
"step": 1560
},
{
"epoch": 1.8946731234866827,
"grad_norm": 1.5545868873596191,
"learning_rate": 4.8793980622616195e-05,
"loss": 0.2117,
"num_input_tokens_seen": 643376,
"step": 1565
},
{
"epoch": 1.900726392251816,
"grad_norm": 1.4125303030014038,
"learning_rate": 4.877771846694897e-05,
"loss": 0.0568,
"num_input_tokens_seen": 645488,
"step": 1570
},
{
"epoch": 1.9067796610169492,
"grad_norm": 4.356451988220215,
"learning_rate": 4.876135015036515e-05,
"loss": 0.1828,
"num_input_tokens_seen": 647600,
"step": 1575
},
{
"epoch": 1.9128329297820823,
"grad_norm": 0.882681667804718,
"learning_rate": 4.874487574594473e-05,
"loss": 0.1048,
"num_input_tokens_seen": 649648,
"step": 1580
},
{
"epoch": 1.9188861985472156,
"grad_norm": 2.282263994216919,
"learning_rate": 4.872829532724136e-05,
"loss": 0.1744,
"num_input_tokens_seen": 651760,
"step": 1585
},
{
"epoch": 1.9249394673123486,
"grad_norm": 1.4083192348480225,
"learning_rate": 4.8711608968282e-05,
"loss": 0.2004,
"num_input_tokens_seen": 653840,
"step": 1590
},
{
"epoch": 1.930992736077482,
"grad_norm": 3.696666955947876,
"learning_rate": 4.8694816743566616e-05,
"loss": 0.2136,
"num_input_tokens_seen": 655888,
"step": 1595
},
{
"epoch": 1.937046004842615,
"grad_norm": 0.2428300827741623,
"learning_rate": 4.867791872806785e-05,
"loss": 0.0729,
"num_input_tokens_seen": 657776,
"step": 1600
},
{
"epoch": 1.9430992736077481,
"grad_norm": 1.8835179805755615,
"learning_rate": 4.8660914997230624e-05,
"loss": 0.1833,
"num_input_tokens_seen": 659952,
"step": 1605
},
{
"epoch": 1.9491525423728815,
"grad_norm": 5.549609184265137,
"learning_rate": 4.8643805626971894e-05,
"loss": 0.1643,
"num_input_tokens_seen": 661968,
"step": 1610
},
{
"epoch": 1.9552058111380144,
"grad_norm": 0.962121844291687,
"learning_rate": 4.862659069368026e-05,
"loss": 0.1775,
"num_input_tokens_seen": 664048,
"step": 1615
},
{
"epoch": 1.9612590799031477,
"grad_norm": 0.7466434836387634,
"learning_rate": 4.8609270274215614e-05,
"loss": 0.2465,
"num_input_tokens_seen": 665968,
"step": 1620
},
{
"epoch": 1.9673123486682809,
"grad_norm": 1.28084135055542,
"learning_rate": 4.859184444590882e-05,
"loss": 0.1881,
"num_input_tokens_seen": 667920,
"step": 1625
},
{
"epoch": 1.973365617433414,
"grad_norm": 0.41623222827911377,
"learning_rate": 4.857431328656137e-05,
"loss": 0.0832,
"num_input_tokens_seen": 669968,
"step": 1630
},
{
"epoch": 1.9794188861985473,
"grad_norm": 1.7877904176712036,
"learning_rate": 4.855667687444504e-05,
"loss": 0.1038,
"num_input_tokens_seen": 672016,
"step": 1635
},
{
"epoch": 1.9854721549636802,
"grad_norm": 0.9274682402610779,
"learning_rate": 4.85389352883015e-05,
"loss": 0.0918,
"num_input_tokens_seen": 674096,
"step": 1640
},
{
"epoch": 1.9915254237288136,
"grad_norm": 2.654660940170288,
"learning_rate": 4.8521088607342016e-05,
"loss": 0.123,
"num_input_tokens_seen": 676208,
"step": 1645
},
{
"epoch": 1.9975786924939467,
"grad_norm": 0.8411111235618591,
"learning_rate": 4.850313691124707e-05,
"loss": 0.0424,
"num_input_tokens_seen": 678256,
"step": 1650
},
{
"epoch": 2.0,
"eval_loss": 0.1412314772605896,
"eval_runtime": 4.5644,
"eval_samples_per_second": 80.404,
"eval_steps_per_second": 20.156,
"num_input_tokens_seen": 678688,
"step": 1652
},
{
"epoch": 2.00363196125908,
"grad_norm": 2.427675485610962,
"learning_rate": 4.8485080280166006e-05,
"loss": 0.1351,
"num_input_tokens_seen": 679936,
"step": 1655
},
{
"epoch": 2.009685230024213,
"grad_norm": 3.1601905822753906,
"learning_rate": 4.8466918794716666e-05,
"loss": 0.1074,
"num_input_tokens_seen": 681984,
"step": 1660
},
{
"epoch": 2.015738498789346,
"grad_norm": 2.2332310676574707,
"learning_rate": 4.8448652535985045e-05,
"loss": 0.0816,
"num_input_tokens_seen": 684096,
"step": 1665
},
{
"epoch": 2.0217917675544794,
"grad_norm": 1.4707834720611572,
"learning_rate": 4.8430281585524926e-05,
"loss": 0.0863,
"num_input_tokens_seen": 686016,
"step": 1670
},
{
"epoch": 2.0278450363196128,
"grad_norm": 3.2379772663116455,
"learning_rate": 4.841180602535751e-05,
"loss": 0.0878,
"num_input_tokens_seen": 687904,
"step": 1675
},
{
"epoch": 2.0338983050847457,
"grad_norm": 5.141901969909668,
"learning_rate": 4.839322593797104e-05,
"loss": 0.2172,
"num_input_tokens_seen": 689888,
"step": 1680
},
{
"epoch": 2.039951573849879,
"grad_norm": 3.7298524379730225,
"learning_rate": 4.837454140632045e-05,
"loss": 0.1119,
"num_input_tokens_seen": 691808,
"step": 1685
},
{
"epoch": 2.046004842615012,
"grad_norm": 1.6675068140029907,
"learning_rate": 4.8355752513826995e-05,
"loss": 0.2492,
"num_input_tokens_seen": 693856,
"step": 1690
},
{
"epoch": 2.0520581113801453,
"grad_norm": 2.8532845973968506,
"learning_rate": 4.833685934437787e-05,
"loss": 0.1375,
"num_input_tokens_seen": 695680,
"step": 1695
},
{
"epoch": 2.0581113801452786,
"grad_norm": 1.0494747161865234,
"learning_rate": 4.831786198232583e-05,
"loss": 0.0764,
"num_input_tokens_seen": 697792,
"step": 1700
},
{
"epoch": 2.0641646489104115,
"grad_norm": 3.190880537033081,
"learning_rate": 4.8298760512488824e-05,
"loss": 0.155,
"num_input_tokens_seen": 699712,
"step": 1705
},
{
"epoch": 2.070217917675545,
"grad_norm": 2.6185317039489746,
"learning_rate": 4.827955502014963e-05,
"loss": 0.1041,
"num_input_tokens_seen": 701792,
"step": 1710
},
{
"epoch": 2.0762711864406778,
"grad_norm": 2.9850172996520996,
"learning_rate": 4.826024559105542e-05,
"loss": 0.206,
"num_input_tokens_seen": 703936,
"step": 1715
},
{
"epoch": 2.082324455205811,
"grad_norm": 3.1610946655273438,
"learning_rate": 4.8240832311417465e-05,
"loss": 0.092,
"num_input_tokens_seen": 705952,
"step": 1720
},
{
"epoch": 2.0883777239709445,
"grad_norm": 3.6915476322174072,
"learning_rate": 4.822131526791065e-05,
"loss": 0.1464,
"num_input_tokens_seen": 708064,
"step": 1725
},
{
"epoch": 2.0944309927360774,
"grad_norm": 3.869178056716919,
"learning_rate": 4.820169454767318e-05,
"loss": 0.092,
"num_input_tokens_seen": 709952,
"step": 1730
},
{
"epoch": 2.1004842615012107,
"grad_norm": 3.4117565155029297,
"learning_rate": 4.8181970238306104e-05,
"loss": 0.127,
"num_input_tokens_seen": 712032,
"step": 1735
},
{
"epoch": 2.106537530266344,
"grad_norm": 1.2668098211288452,
"learning_rate": 4.816214242787302e-05,
"loss": 0.0634,
"num_input_tokens_seen": 714048,
"step": 1740
},
{
"epoch": 2.112590799031477,
"grad_norm": 1.9988105297088623,
"learning_rate": 4.814221120489958e-05,
"loss": 0.1309,
"num_input_tokens_seen": 716224,
"step": 1745
},
{
"epoch": 2.1186440677966103,
"grad_norm": 5.828415870666504,
"learning_rate": 4.812217665837316e-05,
"loss": 0.149,
"num_input_tokens_seen": 718144,
"step": 1750
},
{
"epoch": 2.124697336561743,
"grad_norm": 1.041694164276123,
"learning_rate": 4.810203887774247e-05,
"loss": 0.1105,
"num_input_tokens_seen": 720288,
"step": 1755
},
{
"epoch": 2.1307506053268765,
"grad_norm": 2.428319215774536,
"learning_rate": 4.808179795291712e-05,
"loss": 0.2155,
"num_input_tokens_seen": 722368,
"step": 1760
},
{
"epoch": 2.13680387409201,
"grad_norm": 2.7118420600891113,
"learning_rate": 4.8061453974267195e-05,
"loss": 0.1619,
"num_input_tokens_seen": 724480,
"step": 1765
},
{
"epoch": 2.142857142857143,
"grad_norm": 1.4349424839019775,
"learning_rate": 4.804100703262294e-05,
"loss": 0.1896,
"num_input_tokens_seen": 726784,
"step": 1770
},
{
"epoch": 2.148910411622276,
"grad_norm": 3.1386163234710693,
"learning_rate": 4.8020457219274266e-05,
"loss": 0.1684,
"num_input_tokens_seen": 728768,
"step": 1775
},
{
"epoch": 2.154963680387409,
"grad_norm": 5.631900310516357,
"learning_rate": 4.799980462597039e-05,
"loss": 0.2013,
"num_input_tokens_seen": 730816,
"step": 1780
},
{
"epoch": 2.1610169491525424,
"grad_norm": 1.2051900625228882,
"learning_rate": 4.7979049344919416e-05,
"loss": 0.0672,
"num_input_tokens_seen": 732896,
"step": 1785
},
{
"epoch": 2.1670702179176757,
"grad_norm": 0.28911882638931274,
"learning_rate": 4.795819146878792e-05,
"loss": 0.1565,
"num_input_tokens_seen": 734816,
"step": 1790
},
{
"epoch": 2.1731234866828086,
"grad_norm": 6.5837788581848145,
"learning_rate": 4.7937231090700516e-05,
"loss": 0.2135,
"num_input_tokens_seen": 736800,
"step": 1795
},
{
"epoch": 2.179176755447942,
"grad_norm": 0.9957193732261658,
"learning_rate": 4.7916168304239496e-05,
"loss": 0.1847,
"num_input_tokens_seen": 738944,
"step": 1800
},
{
"epoch": 2.185230024213075,
"grad_norm": 0.9871184229850769,
"learning_rate": 4.789500320344435e-05,
"loss": 0.0843,
"num_input_tokens_seen": 741184,
"step": 1805
},
{
"epoch": 2.1912832929782082,
"grad_norm": 2.35097599029541,
"learning_rate": 4.787373588281138e-05,
"loss": 0.0992,
"num_input_tokens_seen": 743136,
"step": 1810
},
{
"epoch": 2.1973365617433416,
"grad_norm": 4.3666181564331055,
"learning_rate": 4.785236643729327e-05,
"loss": 0.1001,
"num_input_tokens_seen": 745152,
"step": 1815
},
{
"epoch": 2.2033898305084745,
"grad_norm": 2.161907196044922,
"learning_rate": 4.7830894962298675e-05,
"loss": 0.3108,
"num_input_tokens_seen": 747200,
"step": 1820
},
{
"epoch": 2.209443099273608,
"grad_norm": 3.8931241035461426,
"learning_rate": 4.7809321553691764e-05,
"loss": 0.2053,
"num_input_tokens_seen": 749344,
"step": 1825
},
{
"epoch": 2.2154963680387407,
"grad_norm": 0.4339629113674164,
"learning_rate": 4.778764630779183e-05,
"loss": 0.0899,
"num_input_tokens_seen": 751360,
"step": 1830
},
{
"epoch": 2.221549636803874,
"grad_norm": 3.072408676147461,
"learning_rate": 4.7765869321372836e-05,
"loss": 0.1655,
"num_input_tokens_seen": 753408,
"step": 1835
},
{
"epoch": 2.2276029055690074,
"grad_norm": 2.9156787395477295,
"learning_rate": 4.774399069166296e-05,
"loss": 0.1509,
"num_input_tokens_seen": 755456,
"step": 1840
},
{
"epoch": 2.2336561743341403,
"grad_norm": 2.098881244659424,
"learning_rate": 4.772201051634426e-05,
"loss": 0.1335,
"num_input_tokens_seen": 757504,
"step": 1845
},
{
"epoch": 2.2397094430992737,
"grad_norm": 0.5506150126457214,
"learning_rate": 4.769992889355208e-05,
"loss": 0.1555,
"num_input_tokens_seen": 759584,
"step": 1850
},
{
"epoch": 2.2457627118644066,
"grad_norm": 1.9892140626907349,
"learning_rate": 4.767774592187475e-05,
"loss": 0.11,
"num_input_tokens_seen": 761632,
"step": 1855
},
{
"epoch": 2.25181598062954,
"grad_norm": 3.1103200912475586,
"learning_rate": 4.76554617003531e-05,
"loss": 0.1216,
"num_input_tokens_seen": 763648,
"step": 1860
},
{
"epoch": 2.2578692493946733,
"grad_norm": 3.333366632461548,
"learning_rate": 4.763307632847998e-05,
"loss": 0.1664,
"num_input_tokens_seen": 765696,
"step": 1865
},
{
"epoch": 2.263922518159806,
"grad_norm": 5.307172775268555,
"learning_rate": 4.761058990619986e-05,
"loss": 0.1662,
"num_input_tokens_seen": 767776,
"step": 1870
},
{
"epoch": 2.2699757869249395,
"grad_norm": 3.0664305686950684,
"learning_rate": 4.7588002533908405e-05,
"loss": 0.1085,
"num_input_tokens_seen": 769856,
"step": 1875
},
{
"epoch": 2.2760290556900724,
"grad_norm": 3.5791735649108887,
"learning_rate": 4.756531431245195e-05,
"loss": 0.1527,
"num_input_tokens_seen": 771904,
"step": 1880
},
{
"epoch": 2.2820823244552058,
"grad_norm": 1.5854547023773193,
"learning_rate": 4.75425253431271e-05,
"loss": 0.1064,
"num_input_tokens_seen": 774016,
"step": 1885
},
{
"epoch": 2.288135593220339,
"grad_norm": 1.7430024147033691,
"learning_rate": 4.7519635727680286e-05,
"loss": 0.215,
"num_input_tokens_seen": 776000,
"step": 1890
},
{
"epoch": 2.294188861985472,
"grad_norm": 2.2312424182891846,
"learning_rate": 4.749664556830731e-05,
"loss": 0.1338,
"num_input_tokens_seen": 778208,
"step": 1895
},
{
"epoch": 2.3002421307506054,
"grad_norm": 2.4188883304595947,
"learning_rate": 4.747355496765283e-05,
"loss": 0.1155,
"num_input_tokens_seen": 780128,
"step": 1900
},
{
"epoch": 2.3062953995157383,
"grad_norm": 1.0266026258468628,
"learning_rate": 4.745036402880999e-05,
"loss": 0.0778,
"num_input_tokens_seen": 782112,
"step": 1905
},
{
"epoch": 2.3123486682808716,
"grad_norm": 1.269418478012085,
"learning_rate": 4.7427072855319886e-05,
"loss": 0.1447,
"num_input_tokens_seen": 784256,
"step": 1910
},
{
"epoch": 2.318401937046005,
"grad_norm": 2.8256311416625977,
"learning_rate": 4.740368155117116e-05,
"loss": 0.1182,
"num_input_tokens_seen": 786272,
"step": 1915
},
{
"epoch": 2.324455205811138,
"grad_norm": 2.589238166809082,
"learning_rate": 4.7380190220799484e-05,
"loss": 0.045,
"num_input_tokens_seen": 788352,
"step": 1920
},
{
"epoch": 2.330508474576271,
"grad_norm": 0.9821070432662964,
"learning_rate": 4.735659896908713e-05,
"loss": 0.1627,
"num_input_tokens_seen": 790368,
"step": 1925
},
{
"epoch": 2.3365617433414045,
"grad_norm": 0.7688742876052856,
"learning_rate": 4.73329079013625e-05,
"loss": 0.1664,
"num_input_tokens_seen": 792320,
"step": 1930
},
{
"epoch": 2.3426150121065374,
"grad_norm": 0.5684288740158081,
"learning_rate": 4.730911712339964e-05,
"loss": 0.0689,
"num_input_tokens_seen": 794272,
"step": 1935
},
{
"epoch": 2.348668280871671,
"grad_norm": 4.761138439178467,
"learning_rate": 4.728522674141776e-05,
"loss": 0.2063,
"num_input_tokens_seen": 796416,
"step": 1940
},
{
"epoch": 2.3547215496368037,
"grad_norm": 1.1905338764190674,
"learning_rate": 4.7261236862080805e-05,
"loss": 0.0513,
"num_input_tokens_seen": 798432,
"step": 1945
},
{
"epoch": 2.360774818401937,
"grad_norm": 3.9322619438171387,
"learning_rate": 4.723714759249692e-05,
"loss": 0.15,
"num_input_tokens_seen": 800480,
"step": 1950
},
{
"epoch": 2.3668280871670704,
"grad_norm": 4.847472667694092,
"learning_rate": 4.721295904021802e-05,
"loss": 0.1752,
"num_input_tokens_seen": 802432,
"step": 1955
},
{
"epoch": 2.3728813559322033,
"grad_norm": 0.6638381481170654,
"learning_rate": 4.718867131323927e-05,
"loss": 0.1272,
"num_input_tokens_seen": 804352,
"step": 1960
},
{
"epoch": 2.3789346246973366,
"grad_norm": 3.257404327392578,
"learning_rate": 4.7164284519998644e-05,
"loss": 0.1885,
"num_input_tokens_seen": 806272,
"step": 1965
},
{
"epoch": 2.38498789346247,
"grad_norm": 1.8241240978240967,
"learning_rate": 4.71397987693764e-05,
"loss": 0.1153,
"num_input_tokens_seen": 808352,
"step": 1970
},
{
"epoch": 2.391041162227603,
"grad_norm": 1.4032349586486816,
"learning_rate": 4.711521417069462e-05,
"loss": 0.1143,
"num_input_tokens_seen": 810304,
"step": 1975
},
{
"epoch": 2.3970944309927362,
"grad_norm": 0.7167484760284424,
"learning_rate": 4.709053083371672e-05,
"loss": 0.0683,
"num_input_tokens_seen": 812384,
"step": 1980
},
{
"epoch": 2.403147699757869,
"grad_norm": 1.5216193199157715,
"learning_rate": 4.706574886864696e-05,
"loss": 0.1331,
"num_input_tokens_seen": 814400,
"step": 1985
},
{
"epoch": 2.4092009685230025,
"grad_norm": 0.35691097378730774,
"learning_rate": 4.7040868386129935e-05,
"loss": 0.0971,
"num_input_tokens_seen": 816448,
"step": 1990
},
{
"epoch": 2.415254237288136,
"grad_norm": 0.27627068758010864,
"learning_rate": 4.701588949725009e-05,
"loss": 0.1488,
"num_input_tokens_seen": 818624,
"step": 1995
},
{
"epoch": 2.4213075060532687,
"grad_norm": 3.6620829105377197,
"learning_rate": 4.699081231353124e-05,
"loss": 0.1598,
"num_input_tokens_seen": 820640,
"step": 2000
},
{
"epoch": 2.427360774818402,
"grad_norm": 4.2808837890625,
"learning_rate": 4.696563694693605e-05,
"loss": 0.2381,
"num_input_tokens_seen": 822752,
"step": 2005
},
{
"epoch": 2.433414043583535,
"grad_norm": 4.718568325042725,
"learning_rate": 4.694036350986556e-05,
"loss": 0.1828,
"num_input_tokens_seen": 824832,
"step": 2010
},
{
"epoch": 2.4394673123486683,
"grad_norm": 2.1222352981567383,
"learning_rate": 4.6914992115158634e-05,
"loss": 0.1231,
"num_input_tokens_seen": 826816,
"step": 2015
},
{
"epoch": 2.4455205811138017,
"grad_norm": 3.19950270652771,
"learning_rate": 4.688952287609152e-05,
"loss": 0.1957,
"num_input_tokens_seen": 828672,
"step": 2020
},
{
"epoch": 2.4515738498789346,
"grad_norm": 1.1805663108825684,
"learning_rate": 4.686395590637732e-05,
"loss": 0.1136,
"num_input_tokens_seen": 830752,
"step": 2025
},
{
"epoch": 2.457627118644068,
"grad_norm": 4.136936187744141,
"learning_rate": 4.683829132016544e-05,
"loss": 0.1257,
"num_input_tokens_seen": 832832,
"step": 2030
},
{
"epoch": 2.463680387409201,
"grad_norm": 1.4625964164733887,
"learning_rate": 4.6812529232041144e-05,
"loss": 0.129,
"num_input_tokens_seen": 834848,
"step": 2035
},
{
"epoch": 2.469733656174334,
"grad_norm": 1.1322230100631714,
"learning_rate": 4.6786669757025016e-05,
"loss": 0.0553,
"num_input_tokens_seen": 836864,
"step": 2040
},
{
"epoch": 2.4757869249394675,
"grad_norm": 3.3223824501037598,
"learning_rate": 4.676071301057243e-05,
"loss": 0.1469,
"num_input_tokens_seen": 838784,
"step": 2045
},
{
"epoch": 2.4818401937046004,
"grad_norm": 2.883943796157837,
"learning_rate": 4.673465910857306e-05,
"loss": 0.0844,
"num_input_tokens_seen": 840832,
"step": 2050
},
{
"epoch": 2.4878934624697338,
"grad_norm": 0.32013097405433655,
"learning_rate": 4.670850816735035e-05,
"loss": 0.1226,
"num_input_tokens_seen": 842752,
"step": 2055
},
{
"epoch": 2.4939467312348667,
"grad_norm": 3.8903720378875732,
"learning_rate": 4.668226030366101e-05,
"loss": 0.2109,
"num_input_tokens_seen": 844928,
"step": 2060
},
{
"epoch": 2.5,
"grad_norm": 1.0813623666763306,
"learning_rate": 4.665591563469445e-05,
"loss": 0.1393,
"num_input_tokens_seen": 847104,
"step": 2065
},
{
"epoch": 2.5,
"eval_loss": 0.1389196813106537,
"eval_runtime": 4.5634,
"eval_samples_per_second": 80.422,
"eval_steps_per_second": 20.16,
"num_input_tokens_seen": 847104,
"step": 2065
},
{
"epoch": 2.5060532687651333,
"grad_norm": 0.8157436847686768,
"learning_rate": 4.662947427807231e-05,
"loss": 0.0783,
"num_input_tokens_seen": 849120,
"step": 2070
},
{
"epoch": 2.5121065375302662,
"grad_norm": 3.726567268371582,
"learning_rate": 4.6602936351847924e-05,
"loss": 0.0737,
"num_input_tokens_seen": 851296,
"step": 2075
},
{
"epoch": 2.5181598062953996,
"grad_norm": 6.2762603759765625,
"learning_rate": 4.657630197450577e-05,
"loss": 0.2042,
"num_input_tokens_seen": 853344,
"step": 2080
},
{
"epoch": 2.5242130750605325,
"grad_norm": 2.1367788314819336,
"learning_rate": 4.6549571264960945e-05,
"loss": 0.1232,
"num_input_tokens_seen": 855264,
"step": 2085
},
{
"epoch": 2.530266343825666,
"grad_norm": 1.8090111017227173,
"learning_rate": 4.652274434255866e-05,
"loss": 0.1651,
"num_input_tokens_seen": 857152,
"step": 2090
},
{
"epoch": 2.536319612590799,
"grad_norm": 2.769136667251587,
"learning_rate": 4.6495821327073675e-05,
"loss": 0.162,
"num_input_tokens_seen": 859232,
"step": 2095
},
{
"epoch": 2.542372881355932,
"grad_norm": 4.568650722503662,
"learning_rate": 4.6468802338709783e-05,
"loss": 0.1262,
"num_input_tokens_seen": 861312,
"step": 2100
},
{
"epoch": 2.5484261501210654,
"grad_norm": 4.672850131988525,
"learning_rate": 4.644168749809929e-05,
"loss": 0.1883,
"num_input_tokens_seen": 863552,
"step": 2105
},
{
"epoch": 2.5544794188861983,
"grad_norm": 0.21458782255649567,
"learning_rate": 4.6414476926302406e-05,
"loss": 0.118,
"num_input_tokens_seen": 865568,
"step": 2110
},
{
"epoch": 2.5605326876513317,
"grad_norm": 3.470489501953125,
"learning_rate": 4.638717074480682e-05,
"loss": 0.1168,
"num_input_tokens_seen": 867488,
"step": 2115
},
{
"epoch": 2.566585956416465,
"grad_norm": 2.928154468536377,
"learning_rate": 4.6359769075527026e-05,
"loss": 0.0644,
"num_input_tokens_seen": 869504,
"step": 2120
},
{
"epoch": 2.572639225181598,
"grad_norm": 1.570302963256836,
"learning_rate": 4.6332272040803895e-05,
"loss": 0.0429,
"num_input_tokens_seen": 871680,
"step": 2125
},
{
"epoch": 2.5786924939467313,
"grad_norm": 1.8090287446975708,
"learning_rate": 4.630467976340405e-05,
"loss": 0.0671,
"num_input_tokens_seen": 873856,
"step": 2130
},
{
"epoch": 2.584745762711864,
"grad_norm": 3.67999005317688,
"learning_rate": 4.6276992366519365e-05,
"loss": 0.1545,
"num_input_tokens_seen": 875968,
"step": 2135
},
{
"epoch": 2.5907990314769975,
"grad_norm": 5.59567403793335,
"learning_rate": 4.624920997376637e-05,
"loss": 0.1714,
"num_input_tokens_seen": 878144,
"step": 2140
},
{
"epoch": 2.596852300242131,
"grad_norm": 3.0647568702697754,
"learning_rate": 4.622133270918576e-05,
"loss": 0.1627,
"num_input_tokens_seen": 880320,
"step": 2145
},
{
"epoch": 2.6029055690072638,
"grad_norm": 1.2152265310287476,
"learning_rate": 4.619336069724177e-05,
"loss": 0.1276,
"num_input_tokens_seen": 882400,
"step": 2150
},
{
"epoch": 2.608958837772397,
"grad_norm": 2.7467970848083496,
"learning_rate": 4.6165294062821696e-05,
"loss": 0.3168,
"num_input_tokens_seen": 884512,
"step": 2155
},
{
"epoch": 2.61501210653753,
"grad_norm": 1.246924638748169,
"learning_rate": 4.613713293123525e-05,
"loss": 0.0729,
"num_input_tokens_seen": 886592,
"step": 2160
},
{
"epoch": 2.6210653753026634,
"grad_norm": 3.9440741539001465,
"learning_rate": 4.610887742821408e-05,
"loss": 0.1066,
"num_input_tokens_seen": 888640,
"step": 2165
},
{
"epoch": 2.6271186440677967,
"grad_norm": 2.369666814804077,
"learning_rate": 4.608052767991118e-05,
"loss": 0.0523,
"num_input_tokens_seen": 890880,
"step": 2170
},
{
"epoch": 2.6331719128329296,
"grad_norm": 2.7659249305725098,
"learning_rate": 4.605208381290029e-05,
"loss": 0.0824,
"num_input_tokens_seen": 892832,
"step": 2175
},
{
"epoch": 2.639225181598063,
"grad_norm": 4.095376491546631,
"learning_rate": 4.6023545954175406e-05,
"loss": 0.1441,
"num_input_tokens_seen": 894912,
"step": 2180
},
{
"epoch": 2.645278450363196,
"grad_norm": 4.215013027191162,
"learning_rate": 4.599491423115014e-05,
"loss": 0.1021,
"num_input_tokens_seen": 896960,
"step": 2185
},
{
"epoch": 2.651331719128329,
"grad_norm": 0.6620141863822937,
"learning_rate": 4.59661887716572e-05,
"loss": 0.0891,
"num_input_tokens_seen": 898816,
"step": 2190
},
{
"epoch": 2.6573849878934626,
"grad_norm": 0.9556487798690796,
"learning_rate": 4.5937369703947785e-05,
"loss": 0.0992,
"num_input_tokens_seen": 900832,
"step": 2195
},
{
"epoch": 2.663438256658596,
"grad_norm": 1.2059531211853027,
"learning_rate": 4.590845715669104e-05,
"loss": 0.0924,
"num_input_tokens_seen": 902944,
"step": 2200
},
{
"epoch": 2.669491525423729,
"grad_norm": 1.3572944402694702,
"learning_rate": 4.5879451258973465e-05,
"loss": 0.1106,
"num_input_tokens_seen": 905056,
"step": 2205
},
{
"epoch": 2.6755447941888617,
"grad_norm": 3.2731289863586426,
"learning_rate": 4.5850352140298356e-05,
"loss": 0.1858,
"num_input_tokens_seen": 907232,
"step": 2210
},
{
"epoch": 2.681598062953995,
"grad_norm": 4.388832092285156,
"learning_rate": 4.582115993058519e-05,
"loss": 0.1382,
"num_input_tokens_seen": 909248,
"step": 2215
},
{
"epoch": 2.6876513317191284,
"grad_norm": 0.5173718333244324,
"learning_rate": 4.5791874760169095e-05,
"loss": 0.0703,
"num_input_tokens_seen": 911264,
"step": 2220
},
{
"epoch": 2.6937046004842617,
"grad_norm": 0.8536763191223145,
"learning_rate": 4.5762496759800246e-05,
"loss": 0.1439,
"num_input_tokens_seen": 913216,
"step": 2225
},
{
"epoch": 2.6997578692493946,
"grad_norm": 0.34403684735298157,
"learning_rate": 4.573302606064324e-05,
"loss": 0.1254,
"num_input_tokens_seen": 915328,
"step": 2230
},
{
"epoch": 2.705811138014528,
"grad_norm": 5.366670608520508,
"learning_rate": 4.5703462794276574e-05,
"loss": 0.1947,
"num_input_tokens_seen": 917568,
"step": 2235
},
{
"epoch": 2.711864406779661,
"grad_norm": 1.7030010223388672,
"learning_rate": 4.567380709269205e-05,
"loss": 0.1793,
"num_input_tokens_seen": 919424,
"step": 2240
},
{
"epoch": 2.7179176755447942,
"grad_norm": 4.686395645141602,
"learning_rate": 4.5644059088294145e-05,
"loss": 0.0785,
"num_input_tokens_seen": 921376,
"step": 2245
},
{
"epoch": 2.7239709443099276,
"grad_norm": 3.0779364109039307,
"learning_rate": 4.561421891389943e-05,
"loss": 0.0824,
"num_input_tokens_seen": 923456,
"step": 2250
},
{
"epoch": 2.7300242130750605,
"grad_norm": 5.996677875518799,
"learning_rate": 4.558428670273601e-05,
"loss": 0.1826,
"num_input_tokens_seen": 925504,
"step": 2255
},
{
"epoch": 2.736077481840194,
"grad_norm": 3.9204232692718506,
"learning_rate": 4.555426258844292e-05,
"loss": 0.1157,
"num_input_tokens_seen": 927456,
"step": 2260
},
{
"epoch": 2.7421307506053267,
"grad_norm": 3.752305507659912,
"learning_rate": 4.552414670506949e-05,
"loss": 0.0627,
"num_input_tokens_seen": 929440,
"step": 2265
},
{
"epoch": 2.74818401937046,
"grad_norm": 2.5699219703674316,
"learning_rate": 4.5493939187074784e-05,
"loss": 0.1061,
"num_input_tokens_seen": 931712,
"step": 2270
},
{
"epoch": 2.7542372881355934,
"grad_norm": 3.8020174503326416,
"learning_rate": 4.5463640169326994e-05,
"loss": 0.246,
"num_input_tokens_seen": 933664,
"step": 2275
},
{
"epoch": 2.7602905569007263,
"grad_norm": 2.4262406826019287,
"learning_rate": 4.5433249787102816e-05,
"loss": 0.1468,
"num_input_tokens_seen": 935840,
"step": 2280
},
{
"epoch": 2.7663438256658597,
"grad_norm": 2.018612861633301,
"learning_rate": 4.54027681760869e-05,
"loss": 0.1444,
"num_input_tokens_seen": 937920,
"step": 2285
},
{
"epoch": 2.7723970944309926,
"grad_norm": 4.552961349487305,
"learning_rate": 4.537219547237115e-05,
"loss": 0.1195,
"num_input_tokens_seen": 940160,
"step": 2290
},
{
"epoch": 2.778450363196126,
"grad_norm": 1.2480086088180542,
"learning_rate": 4.5341531812454234e-05,
"loss": 0.0581,
"num_input_tokens_seen": 942304,
"step": 2295
},
{
"epoch": 2.7845036319612593,
"grad_norm": 4.706632137298584,
"learning_rate": 4.5310777333240885e-05,
"loss": 0.128,
"num_input_tokens_seen": 944288,
"step": 2300
},
{
"epoch": 2.790556900726392,
"grad_norm": 3.13391375541687,
"learning_rate": 4.52799321720413e-05,
"loss": 0.1094,
"num_input_tokens_seen": 946368,
"step": 2305
},
{
"epoch": 2.7966101694915255,
"grad_norm": 0.19843530654907227,
"learning_rate": 4.524899646657059e-05,
"loss": 0.1588,
"num_input_tokens_seen": 948512,
"step": 2310
},
{
"epoch": 2.8026634382566584,
"grad_norm": 5.570863246917725,
"learning_rate": 4.521797035494809e-05,
"loss": 0.1722,
"num_input_tokens_seen": 950624,
"step": 2315
},
{
"epoch": 2.8087167070217918,
"grad_norm": 0.583500862121582,
"learning_rate": 4.5186853975696775e-05,
"loss": 0.0668,
"num_input_tokens_seen": 952640,
"step": 2320
},
{
"epoch": 2.814769975786925,
"grad_norm": 7.228137016296387,
"learning_rate": 4.515564746774265e-05,
"loss": 0.191,
"num_input_tokens_seen": 954656,
"step": 2325
},
{
"epoch": 2.820823244552058,
"grad_norm": 6.490035057067871,
"learning_rate": 4.512435097041412e-05,
"loss": 0.0892,
"num_input_tokens_seen": 956736,
"step": 2330
},
{
"epoch": 2.8268765133171914,
"grad_norm": 3.498931407928467,
"learning_rate": 4.509296462344136e-05,
"loss": 0.0698,
"num_input_tokens_seen": 958816,
"step": 2335
},
{
"epoch": 2.8329297820823243,
"grad_norm": 5.152684211730957,
"learning_rate": 4.50614885669557e-05,
"loss": 0.2331,
"num_input_tokens_seen": 960800,
"step": 2340
},
{
"epoch": 2.8389830508474576,
"grad_norm": 4.12315559387207,
"learning_rate": 4.5029922941489e-05,
"loss": 0.1327,
"num_input_tokens_seen": 962976,
"step": 2345
},
{
"epoch": 2.845036319612591,
"grad_norm": 3.800520896911621,
"learning_rate": 4.499826788797302e-05,
"loss": 0.1845,
"num_input_tokens_seen": 965088,
"step": 2350
},
{
"epoch": 2.851089588377724,
"grad_norm": 3.0858726501464844,
"learning_rate": 4.49665235477388e-05,
"loss": 0.1429,
"num_input_tokens_seen": 967136,
"step": 2355
},
{
"epoch": 2.857142857142857,
"grad_norm": 3.461909055709839,
"learning_rate": 4.493469006251601e-05,
"loss": 0.1257,
"num_input_tokens_seen": 969248,
"step": 2360
},
{
"epoch": 2.86319612590799,
"grad_norm": 0.6151326298713684,
"learning_rate": 4.490276757443233e-05,
"loss": 0.1448,
"num_input_tokens_seen": 971360,
"step": 2365
},
{
"epoch": 2.8692493946731235,
"grad_norm": 0.8009752035140991,
"learning_rate": 4.487075622601281e-05,
"loss": 0.0933,
"num_input_tokens_seen": 973408,
"step": 2370
},
{
"epoch": 2.875302663438257,
"grad_norm": 0.28616681694984436,
"learning_rate": 4.483865616017924e-05,
"loss": 0.1037,
"num_input_tokens_seen": 975392,
"step": 2375
},
{
"epoch": 2.8813559322033897,
"grad_norm": 2.447929859161377,
"learning_rate": 4.480646752024951e-05,
"loss": 0.0563,
"num_input_tokens_seen": 977376,
"step": 2380
},
{
"epoch": 2.887409200968523,
"grad_norm": 0.7191622853279114,
"learning_rate": 4.477419044993697e-05,
"loss": 0.1621,
"num_input_tokens_seen": 979328,
"step": 2385
},
{
"epoch": 2.893462469733656,
"grad_norm": 4.25314474105835,
"learning_rate": 4.474182509334978e-05,
"loss": 0.1238,
"num_input_tokens_seen": 981376,
"step": 2390
},
{
"epoch": 2.8995157384987893,
"grad_norm": 1.7415614128112793,
"learning_rate": 4.470937159499029e-05,
"loss": 0.0546,
"num_input_tokens_seen": 983360,
"step": 2395
},
{
"epoch": 2.9055690072639226,
"grad_norm": 1.0595369338989258,
"learning_rate": 4.467683009975435e-05,
"loss": 0.1113,
"num_input_tokens_seen": 985408,
"step": 2400
},
{
"epoch": 2.9116222760290555,
"grad_norm": 5.065093994140625,
"learning_rate": 4.464420075293072e-05,
"loss": 0.1293,
"num_input_tokens_seen": 987584,
"step": 2405
},
{
"epoch": 2.917675544794189,
"grad_norm": 2.5174288749694824,
"learning_rate": 4.4611483700200374e-05,
"loss": 0.1081,
"num_input_tokens_seen": 989632,
"step": 2410
},
{
"epoch": 2.923728813559322,
"grad_norm": 0.578865110874176,
"learning_rate": 4.457867908763589e-05,
"loss": 0.0653,
"num_input_tokens_seen": 991680,
"step": 2415
},
{
"epoch": 2.929782082324455,
"grad_norm": 0.7825148701667786,
"learning_rate": 4.454578706170075e-05,
"loss": 0.0944,
"num_input_tokens_seen": 993824,
"step": 2420
},
{
"epoch": 2.9358353510895885,
"grad_norm": 0.21404187381267548,
"learning_rate": 4.4512807769248723e-05,
"loss": 0.0885,
"num_input_tokens_seen": 995904,
"step": 2425
},
{
"epoch": 2.9418886198547214,
"grad_norm": 0.4999610483646393,
"learning_rate": 4.447974135752321e-05,
"loss": 0.0637,
"num_input_tokens_seen": 998080,
"step": 2430
},
{
"epoch": 2.9479418886198547,
"grad_norm": 5.155250072479248,
"learning_rate": 4.444658797415656e-05,
"loss": 0.1987,
"num_input_tokens_seen": 1000160,
"step": 2435
},
{
"epoch": 2.9539951573849876,
"grad_norm": 0.5416507124900818,
"learning_rate": 4.441334776716944e-05,
"loss": 0.1704,
"num_input_tokens_seen": 1002368,
"step": 2440
},
{
"epoch": 2.960048426150121,
"grad_norm": 5.383670806884766,
"learning_rate": 4.438002088497015e-05,
"loss": 0.1466,
"num_input_tokens_seen": 1004448,
"step": 2445
},
{
"epoch": 2.9661016949152543,
"grad_norm": 3.468714475631714,
"learning_rate": 4.434660747635396e-05,
"loss": 0.0692,
"num_input_tokens_seen": 1006336,
"step": 2450
},
{
"epoch": 2.9721549636803877,
"grad_norm": 0.6584161520004272,
"learning_rate": 4.4313107690502485e-05,
"loss": 0.1368,
"num_input_tokens_seen": 1008416,
"step": 2455
},
{
"epoch": 2.9782082324455206,
"grad_norm": 0.13571301102638245,
"learning_rate": 4.427952167698298e-05,
"loss": 0.1044,
"num_input_tokens_seen": 1010400,
"step": 2460
},
{
"epoch": 2.9842615012106535,
"grad_norm": 0.7680113315582275,
"learning_rate": 4.4245849585747654e-05,
"loss": 0.1064,
"num_input_tokens_seen": 1012320,
"step": 2465
},
{
"epoch": 2.990314769975787,
"grad_norm": 0.6504201889038086,
"learning_rate": 4.4212091567133083e-05,
"loss": 0.1309,
"num_input_tokens_seen": 1014400,
"step": 2470
},
{
"epoch": 2.99636803874092,
"grad_norm": 0.6503428220748901,
"learning_rate": 4.417824777185943e-05,
"loss": 0.0702,
"num_input_tokens_seen": 1016480,
"step": 2475
},
{
"epoch": 3.0,
"eval_loss": 0.14438338577747345,
"eval_runtime": 4.5701,
"eval_samples_per_second": 80.305,
"eval_steps_per_second": 20.131,
"num_input_tokens_seen": 1017368,
"step": 2478
},
{
"epoch": 3.002421307506053,
"grad_norm": 2.1441495418548584,
"learning_rate": 4.414431835102987e-05,
"loss": 0.0468,
"num_input_tokens_seen": 1018232,
"step": 2480
},
{
"epoch": 3.0084745762711864,
"grad_norm": 0.11217233538627625,
"learning_rate": 4.411030345612984e-05,
"loss": 0.1225,
"num_input_tokens_seen": 1020280,
"step": 2485
},
{
"epoch": 3.0145278450363198,
"grad_norm": 0.7029927372932434,
"learning_rate": 4.407620323902643e-05,
"loss": 0.0755,
"num_input_tokens_seen": 1022328,
"step": 2490
},
{
"epoch": 3.0205811138014527,
"grad_norm": 3.3221864700317383,
"learning_rate": 4.404201785196762e-05,
"loss": 0.1046,
"num_input_tokens_seen": 1024312,
"step": 2495
},
{
"epoch": 3.026634382566586,
"grad_norm": 1.9272534847259521,
"learning_rate": 4.400774744758171e-05,
"loss": 0.094,
"num_input_tokens_seen": 1026392,
"step": 2500
},
{
"epoch": 3.032687651331719,
"grad_norm": 6.2990899085998535,
"learning_rate": 4.397339217887652e-05,
"loss": 0.0966,
"num_input_tokens_seen": 1028344,
"step": 2505
},
{
"epoch": 3.0387409200968523,
"grad_norm": 6.206118583679199,
"learning_rate": 4.393895219923881e-05,
"loss": 0.2425,
"num_input_tokens_seen": 1030520,
"step": 2510
},
{
"epoch": 3.0447941888619856,
"grad_norm": 1.4443256855010986,
"learning_rate": 4.3904427662433534e-05,
"loss": 0.1987,
"num_input_tokens_seen": 1032728,
"step": 2515
},
{
"epoch": 3.0508474576271185,
"grad_norm": 3.545659303665161,
"learning_rate": 4.386981872260317e-05,
"loss": 0.0737,
"num_input_tokens_seen": 1034808,
"step": 2520
},
{
"epoch": 3.056900726392252,
"grad_norm": 3.7366652488708496,
"learning_rate": 4.383512553426703e-05,
"loss": 0.1712,
"num_input_tokens_seen": 1036888,
"step": 2525
},
{
"epoch": 3.062953995157385,
"grad_norm": 3.9347126483917236,
"learning_rate": 4.380034825232059e-05,
"loss": 0.11,
"num_input_tokens_seen": 1039000,
"step": 2530
},
{
"epoch": 3.069007263922518,
"grad_norm": 3.4273524284362793,
"learning_rate": 4.376548703203474e-05,
"loss": 0.1139,
"num_input_tokens_seen": 1040984,
"step": 2535
},
{
"epoch": 3.0750605326876514,
"grad_norm": 1.0495061874389648,
"learning_rate": 4.3730542029055174e-05,
"loss": 0.0579,
"num_input_tokens_seen": 1043000,
"step": 2540
},
{
"epoch": 3.0811138014527844,
"grad_norm": 1.4927022457122803,
"learning_rate": 4.3695513399401646e-05,
"loss": 0.1328,
"num_input_tokens_seen": 1044984,
"step": 2545
},
{
"epoch": 3.0871670702179177,
"grad_norm": 1.1897157430648804,
"learning_rate": 4.366040129946725e-05,
"loss": 0.1041,
"num_input_tokens_seen": 1047096,
"step": 2550
},
{
"epoch": 3.093220338983051,
"grad_norm": 1.3103008270263672,
"learning_rate": 4.362520588601777e-05,
"loss": 0.1656,
"num_input_tokens_seen": 1049208,
"step": 2555
},
{
"epoch": 3.099273607748184,
"grad_norm": 4.3546223640441895,
"learning_rate": 4.3589927316190983e-05,
"loss": 0.2196,
"num_input_tokens_seen": 1051192,
"step": 2560
},
{
"epoch": 3.1053268765133173,
"grad_norm": 2.591219186782837,
"learning_rate": 4.35545657474959e-05,
"loss": 0.1287,
"num_input_tokens_seen": 1053240,
"step": 2565
},
{
"epoch": 3.11138014527845,
"grad_norm": 2.16152024269104,
"learning_rate": 4.351912133781213e-05,
"loss": 0.0659,
"num_input_tokens_seen": 1055384,
"step": 2570
},
{
"epoch": 3.1174334140435835,
"grad_norm": 0.5291240811347961,
"learning_rate": 4.3483594245389106e-05,
"loss": 0.1992,
"num_input_tokens_seen": 1057464,
"step": 2575
},
{
"epoch": 3.123486682808717,
"grad_norm": 3.001479148864746,
"learning_rate": 4.3447984628845464e-05,
"loss": 0.0759,
"num_input_tokens_seen": 1059736,
"step": 2580
},
{
"epoch": 3.12953995157385,
"grad_norm": 2.1946427822113037,
"learning_rate": 4.341229264716825e-05,
"loss": 0.1369,
"num_input_tokens_seen": 1061656,
"step": 2585
},
{
"epoch": 3.135593220338983,
"grad_norm": 3.4312591552734375,
"learning_rate": 4.337651845971227e-05,
"loss": 0.0533,
"num_input_tokens_seen": 1063736,
"step": 2590
},
{
"epoch": 3.141646489104116,
"grad_norm": 0.5884976387023926,
"learning_rate": 4.334066222619933e-05,
"loss": 0.1154,
"num_input_tokens_seen": 1065720,
"step": 2595
},
{
"epoch": 3.1476997578692494,
"grad_norm": 4.081981658935547,
"learning_rate": 4.3304724106717584e-05,
"loss": 0.1374,
"num_input_tokens_seen": 1067736,
"step": 2600
},
{
"epoch": 3.1537530266343827,
"grad_norm": 0.2492610514163971,
"learning_rate": 4.326870426172075e-05,
"loss": 0.0978,
"num_input_tokens_seen": 1069848,
"step": 2605
},
{
"epoch": 3.1598062953995156,
"grad_norm": 1.401228427886963,
"learning_rate": 4.323260285202746e-05,
"loss": 0.0858,
"num_input_tokens_seen": 1071960,
"step": 2610
},
{
"epoch": 3.165859564164649,
"grad_norm": 3.4008891582489014,
"learning_rate": 4.3196420038820475e-05,
"loss": 0.0843,
"num_input_tokens_seen": 1074072,
"step": 2615
},
{
"epoch": 3.171912832929782,
"grad_norm": 3.2857916355133057,
"learning_rate": 4.316015598364603e-05,
"loss": 0.1736,
"num_input_tokens_seen": 1076184,
"step": 2620
},
{
"epoch": 3.1779661016949152,
"grad_norm": 1.1122257709503174,
"learning_rate": 4.312381084841307e-05,
"loss": 0.1213,
"num_input_tokens_seen": 1078360,
"step": 2625
},
{
"epoch": 3.1840193704600486,
"grad_norm": 1.347262978553772,
"learning_rate": 4.308738479539254e-05,
"loss": 0.125,
"num_input_tokens_seen": 1080440,
"step": 2630
},
{
"epoch": 3.1900726392251815,
"grad_norm": 0.6629340052604675,
"learning_rate": 4.305087798721665e-05,
"loss": 0.11,
"num_input_tokens_seen": 1082520,
"step": 2635
},
{
"epoch": 3.196125907990315,
"grad_norm": 4.0958170890808105,
"learning_rate": 4.30142905868782e-05,
"loss": 0.1791,
"num_input_tokens_seen": 1084408,
"step": 2640
},
{
"epoch": 3.2021791767554477,
"grad_norm": 2.998307943344116,
"learning_rate": 4.297762275772975e-05,
"loss": 0.0911,
"num_input_tokens_seen": 1086424,
"step": 2645
},
{
"epoch": 3.208232445520581,
"grad_norm": 3.835343360900879,
"learning_rate": 4.2940874663483005e-05,
"loss": 0.1416,
"num_input_tokens_seen": 1088440,
"step": 2650
},
{
"epoch": 3.2142857142857144,
"grad_norm": 0.17636780440807343,
"learning_rate": 4.2904046468208006e-05,
"loss": 0.1582,
"num_input_tokens_seen": 1090648,
"step": 2655
},
{
"epoch": 3.2203389830508473,
"grad_norm": 5.161961555480957,
"learning_rate": 4.286713833633242e-05,
"loss": 0.1795,
"num_input_tokens_seen": 1092632,
"step": 2660
},
{
"epoch": 3.2263922518159807,
"grad_norm": 0.3390583097934723,
"learning_rate": 4.283015043264084e-05,
"loss": 0.0862,
"num_input_tokens_seen": 1094648,
"step": 2665
},
{
"epoch": 3.232445520581114,
"grad_norm": 2.341804265975952,
"learning_rate": 4.279308292227396e-05,
"loss": 0.1128,
"num_input_tokens_seen": 1096728,
"step": 2670
},
{
"epoch": 3.238498789346247,
"grad_norm": 1.2374261617660522,
"learning_rate": 4.275593597072796e-05,
"loss": 0.0901,
"num_input_tokens_seen": 1098872,
"step": 2675
},
{
"epoch": 3.2445520581113803,
"grad_norm": 3.755679130554199,
"learning_rate": 4.2718709743853654e-05,
"loss": 0.0633,
"num_input_tokens_seen": 1101048,
"step": 2680
},
{
"epoch": 3.250605326876513,
"grad_norm": 4.655717372894287,
"learning_rate": 4.268140440785584e-05,
"loss": 0.1257,
"num_input_tokens_seen": 1103320,
"step": 2685
},
{
"epoch": 3.2566585956416465,
"grad_norm": 2.0559744834899902,
"learning_rate": 4.264402012929247e-05,
"loss": 0.0598,
"num_input_tokens_seen": 1105336,
"step": 2690
},
{
"epoch": 3.26271186440678,
"grad_norm": 2.825394868850708,
"learning_rate": 4.2606557075073996e-05,
"loss": 0.1262,
"num_input_tokens_seen": 1107288,
"step": 2695
},
{
"epoch": 3.2687651331719128,
"grad_norm": 1.9354199171066284,
"learning_rate": 4.256901541246255e-05,
"loss": 0.0741,
"num_input_tokens_seen": 1109368,
"step": 2700
},
{
"epoch": 3.274818401937046,
"grad_norm": 4.32333517074585,
"learning_rate": 4.253139530907124e-05,
"loss": 0.1098,
"num_input_tokens_seen": 1111576,
"step": 2705
},
{
"epoch": 3.280871670702179,
"grad_norm": 0.3646182417869568,
"learning_rate": 4.249369693286341e-05,
"loss": 0.1509,
"num_input_tokens_seen": 1113624,
"step": 2710
},
{
"epoch": 3.2869249394673123,
"grad_norm": 3.2211687564849854,
"learning_rate": 4.245592045215182e-05,
"loss": 0.1527,
"num_input_tokens_seen": 1115512,
"step": 2715
},
{
"epoch": 3.2929782082324457,
"grad_norm": 1.915663242340088,
"learning_rate": 4.2418066035598e-05,
"loss": 0.0569,
"num_input_tokens_seen": 1117464,
"step": 2720
},
{
"epoch": 3.2990314769975786,
"grad_norm": 4.4511590003967285,
"learning_rate": 4.238013385221142e-05,
"loss": 0.1169,
"num_input_tokens_seen": 1119480,
"step": 2725
},
{
"epoch": 3.305084745762712,
"grad_norm": 0.2371293008327484,
"learning_rate": 4.2342124071348744e-05,
"loss": 0.0296,
"num_input_tokens_seen": 1121496,
"step": 2730
},
{
"epoch": 3.3111380145278453,
"grad_norm": 2.90073561668396,
"learning_rate": 4.230403686271309e-05,
"loss": 0.0977,
"num_input_tokens_seen": 1123480,
"step": 2735
},
{
"epoch": 3.317191283292978,
"grad_norm": 6.4930500984191895,
"learning_rate": 4.2265872396353314e-05,
"loss": 0.2057,
"num_input_tokens_seen": 1125528,
"step": 2740
},
{
"epoch": 3.3232445520581115,
"grad_norm": 2.2125489711761475,
"learning_rate": 4.2227630842663136e-05,
"loss": 0.0457,
"num_input_tokens_seen": 1127512,
"step": 2745
},
{
"epoch": 3.3292978208232444,
"grad_norm": 0.3942786455154419,
"learning_rate": 4.21893123723805e-05,
"loss": 0.0164,
"num_input_tokens_seen": 1129592,
"step": 2750
},
{
"epoch": 3.335351089588378,
"grad_norm": 6.583146095275879,
"learning_rate": 4.2150917156586735e-05,
"loss": 0.1122,
"num_input_tokens_seen": 1131576,
"step": 2755
},
{
"epoch": 3.341404358353511,
"grad_norm": 1.0589375495910645,
"learning_rate": 4.211244536670584e-05,
"loss": 0.0896,
"num_input_tokens_seen": 1133784,
"step": 2760
},
{
"epoch": 3.347457627118644,
"grad_norm": 3.002046585083008,
"learning_rate": 4.207389717450368e-05,
"loss": 0.2762,
"num_input_tokens_seen": 1135800,
"step": 2765
},
{
"epoch": 3.3535108958837774,
"grad_norm": 4.337278366088867,
"learning_rate": 4.203527275208723e-05,
"loss": 0.1217,
"num_input_tokens_seen": 1137784,
"step": 2770
},
{
"epoch": 3.3595641646489103,
"grad_norm": 1.5810871124267578,
"learning_rate": 4.199657227190384e-05,
"loss": 0.1831,
"num_input_tokens_seen": 1139896,
"step": 2775
},
{
"epoch": 3.3656174334140436,
"grad_norm": 1.9263957738876343,
"learning_rate": 4.195779590674041e-05,
"loss": 0.0314,
"num_input_tokens_seen": 1142040,
"step": 2780
},
{
"epoch": 3.371670702179177,
"grad_norm": 3.1378190517425537,
"learning_rate": 4.191894382972264e-05,
"loss": 0.083,
"num_input_tokens_seen": 1144088,
"step": 2785
},
{
"epoch": 3.37772397094431,
"grad_norm": 0.04914064332842827,
"learning_rate": 4.188001621431429e-05,
"loss": 0.1129,
"num_input_tokens_seen": 1146232,
"step": 2790
},
{
"epoch": 3.383777239709443,
"grad_norm": 5.580847263336182,
"learning_rate": 4.184101323431636e-05,
"loss": 0.3408,
"num_input_tokens_seen": 1148344,
"step": 2795
},
{
"epoch": 3.389830508474576,
"grad_norm": 4.590537071228027,
"learning_rate": 4.180193506386634e-05,
"loss": 0.1085,
"num_input_tokens_seen": 1150360,
"step": 2800
},
{
"epoch": 3.3958837772397095,
"grad_norm": 2.900740146636963,
"learning_rate": 4.1762781877437406e-05,
"loss": 0.0514,
"num_input_tokens_seen": 1152376,
"step": 2805
},
{
"epoch": 3.401937046004843,
"grad_norm": 4.106755256652832,
"learning_rate": 4.172355384983769e-05,
"loss": 0.2247,
"num_input_tokens_seen": 1154424,
"step": 2810
},
{
"epoch": 3.4079903147699757,
"grad_norm": 4.872042179107666,
"learning_rate": 4.168425115620944e-05,
"loss": 0.1178,
"num_input_tokens_seen": 1156472,
"step": 2815
},
{
"epoch": 3.414043583535109,
"grad_norm": 4.917882442474365,
"learning_rate": 4.164487397202829e-05,
"loss": 0.2427,
"num_input_tokens_seen": 1158552,
"step": 2820
},
{
"epoch": 3.420096852300242,
"grad_norm": 4.278311729431152,
"learning_rate": 4.160542247310244e-05,
"loss": 0.1525,
"num_input_tokens_seen": 1160696,
"step": 2825
},
{
"epoch": 3.4261501210653753,
"grad_norm": 5.004565715789795,
"learning_rate": 4.156589683557189e-05,
"loss": 0.1883,
"num_input_tokens_seen": 1162808,
"step": 2830
},
{
"epoch": 3.4322033898305087,
"grad_norm": 3.080528736114502,
"learning_rate": 4.1526297235907635e-05,
"loss": 0.0894,
"num_input_tokens_seen": 1164728,
"step": 2835
},
{
"epoch": 3.4382566585956416,
"grad_norm": 0.4325386881828308,
"learning_rate": 4.148662385091091e-05,
"loss": 0.1208,
"num_input_tokens_seen": 1166808,
"step": 2840
},
{
"epoch": 3.444309927360775,
"grad_norm": 1.3167212009429932,
"learning_rate": 4.144687685771238e-05,
"loss": 0.0684,
"num_input_tokens_seen": 1168888,
"step": 2845
},
{
"epoch": 3.450363196125908,
"grad_norm": 2.3370537757873535,
"learning_rate": 4.140705643377133e-05,
"loss": 0.1124,
"num_input_tokens_seen": 1170872,
"step": 2850
},
{
"epoch": 3.456416464891041,
"grad_norm": 4.785461902618408,
"learning_rate": 4.1367162756874925e-05,
"loss": 0.1198,
"num_input_tokens_seen": 1172984,
"step": 2855
},
{
"epoch": 3.4624697336561745,
"grad_norm": 5.924464702606201,
"learning_rate": 4.132719600513734e-05,
"loss": 0.092,
"num_input_tokens_seen": 1175032,
"step": 2860
},
{
"epoch": 3.4685230024213074,
"grad_norm": 3.9813077449798584,
"learning_rate": 4.128715635699905e-05,
"loss": 0.1867,
"num_input_tokens_seen": 1177240,
"step": 2865
},
{
"epoch": 3.4745762711864407,
"grad_norm": 0.31910327076911926,
"learning_rate": 4.124704399122597e-05,
"loss": 0.1416,
"num_input_tokens_seen": 1179192,
"step": 2870
},
{
"epoch": 3.4806295399515736,
"grad_norm": 0.5019274950027466,
"learning_rate": 4.120685908690869e-05,
"loss": 0.1544,
"num_input_tokens_seen": 1181112,
"step": 2875
},
{
"epoch": 3.486682808716707,
"grad_norm": 6.107857704162598,
"learning_rate": 4.1166601823461656e-05,
"loss": 0.1366,
"num_input_tokens_seen": 1183128,
"step": 2880
},
{
"epoch": 3.4927360774818403,
"grad_norm": 2.1856138706207275,
"learning_rate": 4.112627238062239e-05,
"loss": 0.0745,
"num_input_tokens_seen": 1185240,
"step": 2885
},
{
"epoch": 3.4987893462469732,
"grad_norm": 0.27053192257881165,
"learning_rate": 4.1085870938450656e-05,
"loss": 0.1314,
"num_input_tokens_seen": 1187320,
"step": 2890
},
{
"epoch": 3.5,
"eval_loss": 0.137603759765625,
"eval_runtime": 4.5756,
"eval_samples_per_second": 80.208,
"eval_steps_per_second": 20.107,
"num_input_tokens_seen": 1187704,
"step": 2891
},
{
"epoch": 3.5048426150121066,
"grad_norm": 1.7873985767364502,
"learning_rate": 4.1045397677327684e-05,
"loss": 0.0986,
"num_input_tokens_seen": 1189400,
"step": 2895
},
{
"epoch": 3.5108958837772395,
"grad_norm": 1.725449562072754,
"learning_rate": 4.1004852777955364e-05,
"loss": 0.1157,
"num_input_tokens_seen": 1191384,
"step": 2900
},
{
"epoch": 3.516949152542373,
"grad_norm": 4.23082160949707,
"learning_rate": 4.096423642135543e-05,
"loss": 0.0537,
"num_input_tokens_seen": 1193368,
"step": 2905
},
{
"epoch": 3.523002421307506,
"grad_norm": 0.3017522692680359,
"learning_rate": 4.0923548788868625e-05,
"loss": 0.0296,
"num_input_tokens_seen": 1195512,
"step": 2910
},
{
"epoch": 3.529055690072639,
"grad_norm": 0.3969971239566803,
"learning_rate": 4.0882790062153957e-05,
"loss": 0.1352,
"num_input_tokens_seen": 1197560,
"step": 2915
},
{
"epoch": 3.5351089588377724,
"grad_norm": 0.6447391510009766,
"learning_rate": 4.084196042318783e-05,
"loss": 0.063,
"num_input_tokens_seen": 1199768,
"step": 2920
},
{
"epoch": 3.5411622276029053,
"grad_norm": 5.6643877029418945,
"learning_rate": 4.080106005426326e-05,
"loss": 0.2772,
"num_input_tokens_seen": 1201848,
"step": 2925
},
{
"epoch": 3.5472154963680387,
"grad_norm": 4.083348751068115,
"learning_rate": 4.076008913798903e-05,
"loss": 0.1208,
"num_input_tokens_seen": 1203896,
"step": 2930
},
{
"epoch": 3.553268765133172,
"grad_norm": 3.979558229446411,
"learning_rate": 4.071904785728894e-05,
"loss": 0.1456,
"num_input_tokens_seen": 1205880,
"step": 2935
},
{
"epoch": 3.559322033898305,
"grad_norm": 8.00076961517334,
"learning_rate": 4.0677936395400906e-05,
"loss": 0.1744,
"num_input_tokens_seen": 1207896,
"step": 2940
},
{
"epoch": 3.5653753026634383,
"grad_norm": 3.3506362438201904,
"learning_rate": 4.063675493587621e-05,
"loss": 0.199,
"num_input_tokens_seen": 1210008,
"step": 2945
},
{
"epoch": 3.571428571428571,
"grad_norm": 6.1021881103515625,
"learning_rate": 4.059550366257864e-05,
"loss": 0.0986,
"num_input_tokens_seen": 1212024,
"step": 2950
},
{
"epoch": 3.5774818401937045,
"grad_norm": 0.7336778044700623,
"learning_rate": 4.055418275968368e-05,
"loss": 0.0768,
"num_input_tokens_seen": 1214040,
"step": 2955
},
{
"epoch": 3.583535108958838,
"grad_norm": 2.5828888416290283,
"learning_rate": 4.0512792411677705e-05,
"loss": 0.0929,
"num_input_tokens_seen": 1216088,
"step": 2960
},
{
"epoch": 3.589588377723971,
"grad_norm": 3.6418821811676025,
"learning_rate": 4.047133280335713e-05,
"loss": 0.1131,
"num_input_tokens_seen": 1218136,
"step": 2965
},
{
"epoch": 3.595641646489104,
"grad_norm": 0.2827266454696655,
"learning_rate": 4.042980411982762e-05,
"loss": 0.092,
"num_input_tokens_seen": 1220248,
"step": 2970
},
{
"epoch": 3.601694915254237,
"grad_norm": 4.693321228027344,
"learning_rate": 4.0388206546503215e-05,
"loss": 0.2923,
"num_input_tokens_seen": 1222360,
"step": 2975
},
{
"epoch": 3.6077481840193704,
"grad_norm": 2.132143020629883,
"learning_rate": 4.0346540269105546e-05,
"loss": 0.1654,
"num_input_tokens_seen": 1224568,
"step": 2980
},
{
"epoch": 3.6138014527845037,
"grad_norm": 2.10683274269104,
"learning_rate": 4.030480547366297e-05,
"loss": 0.1568,
"num_input_tokens_seen": 1226648,
"step": 2985
},
{
"epoch": 3.619854721549637,
"grad_norm": 3.6658689975738525,
"learning_rate": 4.026300234650979e-05,
"loss": 0.1603,
"num_input_tokens_seen": 1228600,
"step": 2990
},
{
"epoch": 3.62590799031477,
"grad_norm": 2.4331119060516357,
"learning_rate": 4.022113107428536e-05,
"loss": 0.0855,
"num_input_tokens_seen": 1230616,
"step": 2995
},
{
"epoch": 3.6319612590799033,
"grad_norm": 3.701345443725586,
"learning_rate": 4.0179191843933286e-05,
"loss": 0.1095,
"num_input_tokens_seen": 1232632,
"step": 3000
},
{
"epoch": 3.638014527845036,
"grad_norm": 1.2965837717056274,
"learning_rate": 4.013718484270061e-05,
"loss": 0.0896,
"num_input_tokens_seen": 1234552,
"step": 3005
},
{
"epoch": 3.6440677966101696,
"grad_norm": 1.7559748888015747,
"learning_rate": 4.009511025813694e-05,
"loss": 0.0422,
"num_input_tokens_seen": 1236728,
"step": 3010
},
{
"epoch": 3.650121065375303,
"grad_norm": 3.9477643966674805,
"learning_rate": 4.005296827809362e-05,
"loss": 0.2595,
"num_input_tokens_seen": 1238776,
"step": 3015
},
{
"epoch": 3.656174334140436,
"grad_norm": 1.9731038808822632,
"learning_rate": 4.001075909072289e-05,
"loss": 0.1161,
"num_input_tokens_seen": 1240856,
"step": 3020
},
{
"epoch": 3.662227602905569,
"grad_norm": 0.5970947742462158,
"learning_rate": 3.9968482884477075e-05,
"loss": 0.0828,
"num_input_tokens_seen": 1242936,
"step": 3025
},
{
"epoch": 3.668280871670702,
"grad_norm": 2.401581048965454,
"learning_rate": 3.992613984810771e-05,
"loss": 0.1905,
"num_input_tokens_seen": 1245080,
"step": 3030
},
{
"epoch": 3.6743341404358354,
"grad_norm": 1.213310956954956,
"learning_rate": 3.988373017066469e-05,
"loss": 0.0652,
"num_input_tokens_seen": 1247192,
"step": 3035
},
{
"epoch": 3.6803874092009687,
"grad_norm": 3.343557119369507,
"learning_rate": 3.984125404149548e-05,
"loss": 0.0579,
"num_input_tokens_seen": 1249240,
"step": 3040
},
{
"epoch": 3.6864406779661016,
"grad_norm": 0.24004943668842316,
"learning_rate": 3.9798711650244194e-05,
"loss": 0.1836,
"num_input_tokens_seen": 1251320,
"step": 3045
},
{
"epoch": 3.692493946731235,
"grad_norm": 6.831732749938965,
"learning_rate": 3.9756103186850825e-05,
"loss": 0.2021,
"num_input_tokens_seen": 1253336,
"step": 3050
},
{
"epoch": 3.698547215496368,
"grad_norm": 3.7986762523651123,
"learning_rate": 3.971342884155033e-05,
"loss": 0.1354,
"num_input_tokens_seen": 1255352,
"step": 3055
},
{
"epoch": 3.7046004842615012,
"grad_norm": 0.4976424276828766,
"learning_rate": 3.9670688804871815e-05,
"loss": 0.0733,
"num_input_tokens_seen": 1257272,
"step": 3060
},
{
"epoch": 3.7106537530266346,
"grad_norm": 1.6615444421768188,
"learning_rate": 3.96278832676377e-05,
"loss": 0.1375,
"num_input_tokens_seen": 1259416,
"step": 3065
},
{
"epoch": 3.7167070217917675,
"grad_norm": 0.4979148209095001,
"learning_rate": 3.958501242096283e-05,
"loss": 0.1439,
"num_input_tokens_seen": 1261496,
"step": 3070
},
{
"epoch": 3.722760290556901,
"grad_norm": 1.3393867015838623,
"learning_rate": 3.954207645625365e-05,
"loss": 0.0512,
"num_input_tokens_seen": 1263480,
"step": 3075
},
{
"epoch": 3.7288135593220337,
"grad_norm": 1.7195483446121216,
"learning_rate": 3.949907556520731e-05,
"loss": 0.1391,
"num_input_tokens_seen": 1265528,
"step": 3080
},
{
"epoch": 3.734866828087167,
"grad_norm": 0.9343760013580322,
"learning_rate": 3.9456009939810886e-05,
"loss": 0.0898,
"num_input_tokens_seen": 1267512,
"step": 3085
},
{
"epoch": 3.7409200968523004,
"grad_norm": 4.527099609375,
"learning_rate": 3.941287977234043e-05,
"loss": 0.1332,
"num_input_tokens_seen": 1269560,
"step": 3090
},
{
"epoch": 3.7469733656174333,
"grad_norm": 2.255335807800293,
"learning_rate": 3.9369685255360175e-05,
"loss": 0.0676,
"num_input_tokens_seen": 1271640,
"step": 3095
},
{
"epoch": 3.7530266343825667,
"grad_norm": 1.6302518844604492,
"learning_rate": 3.9326426581721663e-05,
"loss": 0.0341,
"num_input_tokens_seen": 1273688,
"step": 3100
},
{
"epoch": 3.7590799031476996,
"grad_norm": 0.33657604455947876,
"learning_rate": 3.9283103944562874e-05,
"loss": 0.0306,
"num_input_tokens_seen": 1275768,
"step": 3105
},
{
"epoch": 3.765133171912833,
"grad_norm": 0.6542962193489075,
"learning_rate": 3.923971753730735e-05,
"loss": 0.0519,
"num_input_tokens_seen": 1277752,
"step": 3110
},
{
"epoch": 3.7711864406779663,
"grad_norm": 2.5565290451049805,
"learning_rate": 3.919626755366338e-05,
"loss": 0.1122,
"num_input_tokens_seen": 1279864,
"step": 3115
},
{
"epoch": 3.777239709443099,
"grad_norm": 0.5348601937294006,
"learning_rate": 3.9152754187623086e-05,
"loss": 0.1073,
"num_input_tokens_seen": 1281880,
"step": 3120
},
{
"epoch": 3.7832929782082325,
"grad_norm": 3.66865611076355,
"learning_rate": 3.910917763346156e-05,
"loss": 0.0425,
"num_input_tokens_seen": 1283928,
"step": 3125
},
{
"epoch": 3.7893462469733654,
"grad_norm": 2.6302590370178223,
"learning_rate": 3.906553808573604e-05,
"loss": 0.2901,
"num_input_tokens_seen": 1285944,
"step": 3130
},
{
"epoch": 3.7953995157384988,
"grad_norm": 2.9969611167907715,
"learning_rate": 3.9021835739285e-05,
"loss": 0.1615,
"num_input_tokens_seen": 1287928,
"step": 3135
},
{
"epoch": 3.801452784503632,
"grad_norm": 4.440647602081299,
"learning_rate": 3.897807078922728e-05,
"loss": 0.1891,
"num_input_tokens_seen": 1290008,
"step": 3140
},
{
"epoch": 3.807506053268765,
"grad_norm": 6.937892913818359,
"learning_rate": 3.8934243430961265e-05,
"loss": 0.0881,
"num_input_tokens_seen": 1292120,
"step": 3145
},
{
"epoch": 3.8135593220338984,
"grad_norm": 2.623624086380005,
"learning_rate": 3.889035386016393e-05,
"loss": 0.1615,
"num_input_tokens_seen": 1294040,
"step": 3150
},
{
"epoch": 3.8196125907990313,
"grad_norm": 5.344344615936279,
"learning_rate": 3.8846402272790044e-05,
"loss": 0.1284,
"num_input_tokens_seen": 1296024,
"step": 3155
},
{
"epoch": 3.8256658595641646,
"grad_norm": 3.985133409500122,
"learning_rate": 3.8802388865071246e-05,
"loss": 0.1705,
"num_input_tokens_seen": 1298104,
"step": 3160
},
{
"epoch": 3.831719128329298,
"grad_norm": 1.2340222597122192,
"learning_rate": 3.875831383351519e-05,
"loss": 0.0358,
"num_input_tokens_seen": 1300248,
"step": 3165
},
{
"epoch": 3.837772397094431,
"grad_norm": 2.6467792987823486,
"learning_rate": 3.8714177374904683e-05,
"loss": 0.1947,
"num_input_tokens_seen": 1302104,
"step": 3170
},
{
"epoch": 3.843825665859564,
"grad_norm": 0.4315508306026459,
"learning_rate": 3.866997968629674e-05,
"loss": 0.0413,
"num_input_tokens_seen": 1304056,
"step": 3175
},
{
"epoch": 3.849878934624697,
"grad_norm": 3.8331992626190186,
"learning_rate": 3.86257209650218e-05,
"loss": 0.0932,
"num_input_tokens_seen": 1306104,
"step": 3180
},
{
"epoch": 3.8559322033898304,
"grad_norm": 1.633231282234192,
"learning_rate": 3.858140140868276e-05,
"loss": 0.1438,
"num_input_tokens_seen": 1308056,
"step": 3185
},
{
"epoch": 3.861985472154964,
"grad_norm": 3.581052780151367,
"learning_rate": 3.853702121515416e-05,
"loss": 0.1224,
"num_input_tokens_seen": 1310104,
"step": 3190
},
{
"epoch": 3.8680387409200967,
"grad_norm": 3.2716169357299805,
"learning_rate": 3.849258058258124e-05,
"loss": 0.0858,
"num_input_tokens_seen": 1312152,
"step": 3195
},
{
"epoch": 3.87409200968523,
"grad_norm": 6.2645368576049805,
"learning_rate": 3.84480797093791e-05,
"loss": 0.0981,
"num_input_tokens_seen": 1314328,
"step": 3200
},
{
"epoch": 3.880145278450363,
"grad_norm": 3.6259467601776123,
"learning_rate": 3.8403518794231795e-05,
"loss": 0.0788,
"num_input_tokens_seen": 1316344,
"step": 3205
},
{
"epoch": 3.8861985472154963,
"grad_norm": 2.101606845855713,
"learning_rate": 3.835889803609145e-05,
"loss": 0.0663,
"num_input_tokens_seen": 1318360,
"step": 3210
},
{
"epoch": 3.8922518159806296,
"grad_norm": 0.37834468483924866,
"learning_rate": 3.8314217634177376e-05,
"loss": 0.0548,
"num_input_tokens_seen": 1320376,
"step": 3215
},
{
"epoch": 3.898305084745763,
"grad_norm": 0.7180870771408081,
"learning_rate": 3.826947778797516e-05,
"loss": 0.0425,
"num_input_tokens_seen": 1322616,
"step": 3220
},
{
"epoch": 3.904358353510896,
"grad_norm": 1.2626093626022339,
"learning_rate": 3.822467869723581e-05,
"loss": 0.2914,
"num_input_tokens_seen": 1324664,
"step": 3225
},
{
"epoch": 3.910411622276029,
"grad_norm": 5.424513339996338,
"learning_rate": 3.8179820561974835e-05,
"loss": 0.1202,
"num_input_tokens_seen": 1326616,
"step": 3230
},
{
"epoch": 3.916464891041162,
"grad_norm": 3.26811146736145,
"learning_rate": 3.813490358247137e-05,
"loss": 0.0692,
"num_input_tokens_seen": 1328760,
"step": 3235
},
{
"epoch": 3.9225181598062955,
"grad_norm": 2.555222511291504,
"learning_rate": 3.8089927959267255e-05,
"loss": 0.1256,
"num_input_tokens_seen": 1330968,
"step": 3240
},
{
"epoch": 3.928571428571429,
"grad_norm": 3.6134443283081055,
"learning_rate": 3.8044893893166203e-05,
"loss": 0.1744,
"num_input_tokens_seen": 1332952,
"step": 3245
},
{
"epoch": 3.9346246973365617,
"grad_norm": 3.7331314086914062,
"learning_rate": 3.799980158523279e-05,
"loss": 0.1684,
"num_input_tokens_seen": 1335064,
"step": 3250
},
{
"epoch": 3.940677966101695,
"grad_norm": 5.901867389678955,
"learning_rate": 3.795465123679167e-05,
"loss": 0.1036,
"num_input_tokens_seen": 1337080,
"step": 3255
},
{
"epoch": 3.946731234866828,
"grad_norm": 3.149181365966797,
"learning_rate": 3.790944304942664e-05,
"loss": 0.1296,
"num_input_tokens_seen": 1339096,
"step": 3260
},
{
"epoch": 3.9527845036319613,
"grad_norm": 0.9098186492919922,
"learning_rate": 3.7864177224979696e-05,
"loss": 0.0825,
"num_input_tokens_seen": 1341048,
"step": 3265
},
{
"epoch": 3.9588377723970947,
"grad_norm": 5.158130168914795,
"learning_rate": 3.781885396555019e-05,
"loss": 0.1029,
"num_input_tokens_seen": 1343224,
"step": 3270
},
{
"epoch": 3.9648910411622276,
"grad_norm": 0.13243332505226135,
"learning_rate": 3.777347347349392e-05,
"loss": 0.1416,
"num_input_tokens_seen": 1345272,
"step": 3275
},
{
"epoch": 3.970944309927361,
"grad_norm": 9.198348045349121,
"learning_rate": 3.7728035951422166e-05,
"loss": 0.2062,
"num_input_tokens_seen": 1347416,
"step": 3280
},
{
"epoch": 3.976997578692494,
"grad_norm": 0.32454535365104675,
"learning_rate": 3.7682541602200875e-05,
"loss": 0.0542,
"num_input_tokens_seen": 1349464,
"step": 3285
},
{
"epoch": 3.983050847457627,
"grad_norm": 0.24111558496952057,
"learning_rate": 3.76369906289497e-05,
"loss": 0.1584,
"num_input_tokens_seen": 1351352,
"step": 3290
},
{
"epoch": 3.9891041162227605,
"grad_norm": 1.5734061002731323,
"learning_rate": 3.7591383235041086e-05,
"loss": 0.0471,
"num_input_tokens_seen": 1353368,
"step": 3295
},
{
"epoch": 3.9951573849878934,
"grad_norm": 0.10585124790668488,
"learning_rate": 3.75457196240994e-05,
"loss": 0.1547,
"num_input_tokens_seen": 1355416,
"step": 3300
},
{
"epoch": 4.0,
"eval_loss": 0.13813835382461548,
"eval_runtime": 4.5782,
"eval_samples_per_second": 80.162,
"eval_steps_per_second": 20.095,
"num_input_tokens_seen": 1356744,
"step": 3304
},
{
"epoch": 4.001210653753026,
"grad_norm": 0.9487738609313965,
"learning_rate": 3.7500000000000003e-05,
"loss": 0.1349,
"num_input_tokens_seen": 1357192,
"step": 3305
},
{
"epoch": 4.00726392251816,
"grad_norm": 3.7472715377807617,
"learning_rate": 3.7454224566868327e-05,
"loss": 0.1112,
"num_input_tokens_seen": 1359272,
"step": 3310
},
{
"epoch": 4.013317191283293,
"grad_norm": 4.9403886795043945,
"learning_rate": 3.7408393529078985e-05,
"loss": 0.1103,
"num_input_tokens_seen": 1361480,
"step": 3315
},
{
"epoch": 4.019370460048426,
"grad_norm": 0.8156682848930359,
"learning_rate": 3.7362507091254836e-05,
"loss": 0.1043,
"num_input_tokens_seen": 1363560,
"step": 3320
},
{
"epoch": 4.02542372881356,
"grad_norm": 4.155811786651611,
"learning_rate": 3.7316565458266114e-05,
"loss": 0.241,
"num_input_tokens_seen": 1365640,
"step": 3325
},
{
"epoch": 4.031476997578692,
"grad_norm": 1.49620521068573,
"learning_rate": 3.727056883522945e-05,
"loss": 0.0996,
"num_input_tokens_seen": 1367720,
"step": 3330
},
{
"epoch": 4.0375302663438255,
"grad_norm": 4.452794551849365,
"learning_rate": 3.722451742750701e-05,
"loss": 0.1678,
"num_input_tokens_seen": 1369704,
"step": 3335
},
{
"epoch": 4.043583535108959,
"grad_norm": 4.364172458648682,
"learning_rate": 3.717841144070556e-05,
"loss": 0.1008,
"num_input_tokens_seen": 1371816,
"step": 3340
},
{
"epoch": 4.049636803874092,
"grad_norm": 0.06378893554210663,
"learning_rate": 3.713225108067553e-05,
"loss": 0.0347,
"num_input_tokens_seen": 1373800,
"step": 3345
},
{
"epoch": 4.0556900726392255,
"grad_norm": 0.38104724884033203,
"learning_rate": 3.708603655351012e-05,
"loss": 0.0651,
"num_input_tokens_seen": 1375752,
"step": 3350
},
{
"epoch": 4.061743341404358,
"grad_norm": 4.189276695251465,
"learning_rate": 3.7039768065544395e-05,
"loss": 0.1629,
"num_input_tokens_seen": 1377896,
"step": 3355
},
{
"epoch": 4.067796610169491,
"grad_norm": 1.064452052116394,
"learning_rate": 3.69934458233543e-05,
"loss": 0.0949,
"num_input_tokens_seen": 1379912,
"step": 3360
},
{
"epoch": 4.073849878934625,
"grad_norm": 2.7440671920776367,
"learning_rate": 3.694707003375579e-05,
"loss": 0.1777,
"num_input_tokens_seen": 1381960,
"step": 3365
},
{
"epoch": 4.079903147699758,
"grad_norm": 3.948115825653076,
"learning_rate": 3.690064090380392e-05,
"loss": 0.1963,
"num_input_tokens_seen": 1384072,
"step": 3370
},
{
"epoch": 4.085956416464891,
"grad_norm": 2.6125643253326416,
"learning_rate": 3.685415864079185e-05,
"loss": 0.1138,
"num_input_tokens_seen": 1386152,
"step": 3375
},
{
"epoch": 4.092009685230024,
"grad_norm": 2.7647571563720703,
"learning_rate": 3.680762345225001e-05,
"loss": 0.0476,
"num_input_tokens_seen": 1388168,
"step": 3380
},
{
"epoch": 4.098062953995157,
"grad_norm": 2.3222060203552246,
"learning_rate": 3.676103554594511e-05,
"loss": 0.0609,
"num_input_tokens_seen": 1390152,
"step": 3385
},
{
"epoch": 4.1041162227602905,
"grad_norm": 6.68032693862915,
"learning_rate": 3.671439512987921e-05,
"loss": 0.1748,
"num_input_tokens_seen": 1392168,
"step": 3390
},
{
"epoch": 4.110169491525424,
"grad_norm": 2.3445775508880615,
"learning_rate": 3.666770241228883e-05,
"loss": 0.07,
"num_input_tokens_seen": 1394376,
"step": 3395
},
{
"epoch": 4.116222760290557,
"grad_norm": 0.17334812879562378,
"learning_rate": 3.6620957601644016e-05,
"loss": 0.0244,
"num_input_tokens_seen": 1396520,
"step": 3400
},
{
"epoch": 4.12227602905569,
"grad_norm": 0.5321148037910461,
"learning_rate": 3.657416090664737e-05,
"loss": 0.0809,
"num_input_tokens_seen": 1398600,
"step": 3405
},
{
"epoch": 4.128329297820823,
"grad_norm": 1.0061872005462646,
"learning_rate": 3.652731253623315e-05,
"loss": 0.1107,
"num_input_tokens_seen": 1400584,
"step": 3410
},
{
"epoch": 4.134382566585956,
"grad_norm": 0.31218233704566956,
"learning_rate": 3.648041269956634e-05,
"loss": 0.0652,
"num_input_tokens_seen": 1402760,
"step": 3415
},
{
"epoch": 4.14043583535109,
"grad_norm": 2.9069929122924805,
"learning_rate": 3.6433461606041695e-05,
"loss": 0.1431,
"num_input_tokens_seen": 1404936,
"step": 3420
},
{
"epoch": 4.146489104116223,
"grad_norm": 5.1606926918029785,
"learning_rate": 3.6386459465282824e-05,
"loss": 0.0984,
"num_input_tokens_seen": 1406920,
"step": 3425
},
{
"epoch": 4.1525423728813555,
"grad_norm": 5.275994777679443,
"learning_rate": 3.6339406487141255e-05,
"loss": 0.0691,
"num_input_tokens_seen": 1408840,
"step": 3430
},
{
"epoch": 4.158595641646489,
"grad_norm": 2.359971284866333,
"learning_rate": 3.6292302881695464e-05,
"loss": 0.1333,
"num_input_tokens_seen": 1410696,
"step": 3435
},
{
"epoch": 4.164648910411622,
"grad_norm": 1.0342682600021362,
"learning_rate": 3.6245148859249996e-05,
"loss": 0.1415,
"num_input_tokens_seen": 1412680,
"step": 3440
},
{
"epoch": 4.170702179176756,
"grad_norm": 1.001820683479309,
"learning_rate": 3.619794463033447e-05,
"loss": 0.0941,
"num_input_tokens_seen": 1414728,
"step": 3445
},
{
"epoch": 4.176755447941889,
"grad_norm": 0.8069421052932739,
"learning_rate": 3.6150690405702685e-05,
"loss": 0.0293,
"num_input_tokens_seen": 1416712,
"step": 3450
},
{
"epoch": 4.182808716707021,
"grad_norm": 4.724492073059082,
"learning_rate": 3.6103386396331635e-05,
"loss": 0.0611,
"num_input_tokens_seen": 1418952,
"step": 3455
},
{
"epoch": 4.188861985472155,
"grad_norm": 6.935494422912598,
"learning_rate": 3.605603281342061e-05,
"loss": 0.1149,
"num_input_tokens_seen": 1421032,
"step": 3460
},
{
"epoch": 4.194915254237288,
"grad_norm": 0.4098648130893707,
"learning_rate": 3.6008629868390204e-05,
"loss": 0.1266,
"num_input_tokens_seen": 1423048,
"step": 3465
},
{
"epoch": 4.200968523002421,
"grad_norm": 5.3340020179748535,
"learning_rate": 3.5961177772881434e-05,
"loss": 0.2302,
"num_input_tokens_seen": 1425192,
"step": 3470
},
{
"epoch": 4.207021791767555,
"grad_norm": 0.06073416396975517,
"learning_rate": 3.591367673875472e-05,
"loss": 0.1065,
"num_input_tokens_seen": 1427304,
"step": 3475
},
{
"epoch": 4.213075060532688,
"grad_norm": 3.5133183002471924,
"learning_rate": 3.5866126978089025e-05,
"loss": 0.2221,
"num_input_tokens_seen": 1429288,
"step": 3480
},
{
"epoch": 4.219128329297821,
"grad_norm": 6.023957252502441,
"learning_rate": 3.5818528703180826e-05,
"loss": 0.0864,
"num_input_tokens_seen": 1431400,
"step": 3485
},
{
"epoch": 4.225181598062954,
"grad_norm": 3.456350803375244,
"learning_rate": 3.577088212654322e-05,
"loss": 0.137,
"num_input_tokens_seen": 1433576,
"step": 3490
},
{
"epoch": 4.231234866828087,
"grad_norm": 1.2072936296463013,
"learning_rate": 3.572318746090496e-05,
"loss": 0.0733,
"num_input_tokens_seen": 1435560,
"step": 3495
},
{
"epoch": 4.237288135593221,
"grad_norm": 6.404455184936523,
"learning_rate": 3.5675444919209486e-05,
"loss": 0.1385,
"num_input_tokens_seen": 1437672,
"step": 3500
},
{
"epoch": 4.243341404358354,
"grad_norm": 4.31003999710083,
"learning_rate": 3.5627654714614e-05,
"loss": 0.1378,
"num_input_tokens_seen": 1439656,
"step": 3505
},
{
"epoch": 4.249394673123486,
"grad_norm": 1.0932585000991821,
"learning_rate": 3.557981706048852e-05,
"loss": 0.0267,
"num_input_tokens_seen": 1441608,
"step": 3510
},
{
"epoch": 4.25544794188862,
"grad_norm": 3.007089138031006,
"learning_rate": 3.5531932170414896e-05,
"loss": 0.0983,
"num_input_tokens_seen": 1443624,
"step": 3515
},
{
"epoch": 4.261501210653753,
"grad_norm": 0.3836453855037689,
"learning_rate": 3.5484000258185876e-05,
"loss": 0.0431,
"num_input_tokens_seen": 1445736,
"step": 3520
},
{
"epoch": 4.267554479418886,
"grad_norm": 2.0524728298187256,
"learning_rate": 3.5436021537804144e-05,
"loss": 0.1157,
"num_input_tokens_seen": 1447880,
"step": 3525
},
{
"epoch": 4.27360774818402,
"grad_norm": 4.9584479331970215,
"learning_rate": 3.538799622348139e-05,
"loss": 0.2273,
"num_input_tokens_seen": 1449896,
"step": 3530
},
{
"epoch": 4.279661016949152,
"grad_norm": 0.3631916642189026,
"learning_rate": 3.5339924529637304e-05,
"loss": 0.0961,
"num_input_tokens_seen": 1451880,
"step": 3535
},
{
"epoch": 4.285714285714286,
"grad_norm": 0.1526634246110916,
"learning_rate": 3.529180667089868e-05,
"loss": 0.0849,
"num_input_tokens_seen": 1453992,
"step": 3540
},
{
"epoch": 4.291767554479419,
"grad_norm": 0.08341622352600098,
"learning_rate": 3.52436428620984e-05,
"loss": 0.1471,
"num_input_tokens_seen": 1455912,
"step": 3545
},
{
"epoch": 4.297820823244552,
"grad_norm": 1.2977865934371948,
"learning_rate": 3.5195433318274516e-05,
"loss": 0.1623,
"num_input_tokens_seen": 1458024,
"step": 3550
},
{
"epoch": 4.303874092009686,
"grad_norm": 0.30896905064582825,
"learning_rate": 3.514717825466925e-05,
"loss": 0.0768,
"num_input_tokens_seen": 1459976,
"step": 3555
},
{
"epoch": 4.309927360774818,
"grad_norm": 4.159477710723877,
"learning_rate": 3.509887788672809e-05,
"loss": 0.1655,
"num_input_tokens_seen": 1462120,
"step": 3560
},
{
"epoch": 4.315980629539951,
"grad_norm": 4.378042221069336,
"learning_rate": 3.5050532430098774e-05,
"loss": 0.1571,
"num_input_tokens_seen": 1464104,
"step": 3565
},
{
"epoch": 4.322033898305085,
"grad_norm": 4.206084728240967,
"learning_rate": 3.500214210063035e-05,
"loss": 0.1301,
"num_input_tokens_seen": 1466216,
"step": 3570
},
{
"epoch": 4.328087167070218,
"grad_norm": 3.4586846828460693,
"learning_rate": 3.495370711437221e-05,
"loss": 0.0902,
"num_input_tokens_seen": 1468264,
"step": 3575
},
{
"epoch": 4.3341404358353515,
"grad_norm": 0.12353920936584473,
"learning_rate": 3.490522768757316e-05,
"loss": 0.1492,
"num_input_tokens_seen": 1470408,
"step": 3580
},
{
"epoch": 4.340193704600484,
"grad_norm": 0.6406307220458984,
"learning_rate": 3.485670403668036e-05,
"loss": 0.0404,
"num_input_tokens_seen": 1472392,
"step": 3585
},
{
"epoch": 4.346246973365617,
"grad_norm": 0.753180742263794,
"learning_rate": 3.480813637833846e-05,
"loss": 0.0436,
"num_input_tokens_seen": 1474504,
"step": 3590
},
{
"epoch": 4.352300242130751,
"grad_norm": 0.8588903546333313,
"learning_rate": 3.475952492938859e-05,
"loss": 0.1417,
"num_input_tokens_seen": 1476616,
"step": 3595
},
{
"epoch": 4.358353510895884,
"grad_norm": 2.96315336227417,
"learning_rate": 3.471086990686737e-05,
"loss": 0.0421,
"num_input_tokens_seen": 1478664,
"step": 3600
},
{
"epoch": 4.364406779661017,
"grad_norm": 1.0311739444732666,
"learning_rate": 3.466217152800598e-05,
"loss": 0.2103,
"num_input_tokens_seen": 1480648,
"step": 3605
},
{
"epoch": 4.37046004842615,
"grad_norm": 0.5896651744842529,
"learning_rate": 3.461343001022919e-05,
"loss": 0.0813,
"num_input_tokens_seen": 1482760,
"step": 3610
},
{
"epoch": 4.376513317191283,
"grad_norm": 2.433903932571411,
"learning_rate": 3.456464557115433e-05,
"loss": 0.1196,
"num_input_tokens_seen": 1484744,
"step": 3615
},
{
"epoch": 4.3825665859564165,
"grad_norm": 3.2190709114074707,
"learning_rate": 3.45158184285904e-05,
"loss": 0.0418,
"num_input_tokens_seen": 1486728,
"step": 3620
},
{
"epoch": 4.38861985472155,
"grad_norm": 2.588831663131714,
"learning_rate": 3.446694880053704e-05,
"loss": 0.051,
"num_input_tokens_seen": 1488808,
"step": 3625
},
{
"epoch": 4.394673123486683,
"grad_norm": 4.543102264404297,
"learning_rate": 3.441803690518359e-05,
"loss": 0.1555,
"num_input_tokens_seen": 1490984,
"step": 3630
},
{
"epoch": 4.400726392251816,
"grad_norm": 3.894247055053711,
"learning_rate": 3.4369082960908084e-05,
"loss": 0.1524,
"num_input_tokens_seen": 1493000,
"step": 3635
},
{
"epoch": 4.406779661016949,
"grad_norm": 2.111471176147461,
"learning_rate": 3.432008718627631e-05,
"loss": 0.2373,
"num_input_tokens_seen": 1494920,
"step": 3640
},
{
"epoch": 4.412832929782082,
"grad_norm": 5.577587604522705,
"learning_rate": 3.4271049800040805e-05,
"loss": 0.1239,
"num_input_tokens_seen": 1496904,
"step": 3645
},
{
"epoch": 4.418886198547216,
"grad_norm": 2.5321731567382812,
"learning_rate": 3.42219710211399e-05,
"loss": 0.0717,
"num_input_tokens_seen": 1498792,
"step": 3650
},
{
"epoch": 4.424939467312349,
"grad_norm": 0.10607433319091797,
"learning_rate": 3.417285106869673e-05,
"loss": 0.0888,
"num_input_tokens_seen": 1500840,
"step": 3655
},
{
"epoch": 4.4309927360774815,
"grad_norm": 3.724353551864624,
"learning_rate": 3.4123690162018246e-05,
"loss": 0.0908,
"num_input_tokens_seen": 1502888,
"step": 3660
},
{
"epoch": 4.437046004842615,
"grad_norm": 5.9483208656311035,
"learning_rate": 3.407448852059426e-05,
"loss": 0.0772,
"num_input_tokens_seen": 1504904,
"step": 3665
},
{
"epoch": 4.443099273607748,
"grad_norm": 3.859468936920166,
"learning_rate": 3.4025246364096455e-05,
"loss": 0.1303,
"num_input_tokens_seen": 1506824,
"step": 3670
},
{
"epoch": 4.4491525423728815,
"grad_norm": 1.015526294708252,
"learning_rate": 3.397596391237739e-05,
"loss": 0.141,
"num_input_tokens_seen": 1508872,
"step": 3675
},
{
"epoch": 4.455205811138015,
"grad_norm": 3.720564126968384,
"learning_rate": 3.3926641385469556e-05,
"loss": 0.1819,
"num_input_tokens_seen": 1510824,
"step": 3680
},
{
"epoch": 4.461259079903147,
"grad_norm": 3.61460542678833,
"learning_rate": 3.387727900358435e-05,
"loss": 0.1805,
"num_input_tokens_seen": 1512968,
"step": 3685
},
{
"epoch": 4.467312348668281,
"grad_norm": 5.1266889572143555,
"learning_rate": 3.38278769871111e-05,
"loss": 0.1072,
"num_input_tokens_seen": 1515144,
"step": 3690
},
{
"epoch": 4.473365617433414,
"grad_norm": 2.495168685913086,
"learning_rate": 3.377843555661612e-05,
"loss": 0.064,
"num_input_tokens_seen": 1517192,
"step": 3695
},
{
"epoch": 4.479418886198547,
"grad_norm": 1.2742055654525757,
"learning_rate": 3.372895493284167e-05,
"loss": 0.0776,
"num_input_tokens_seen": 1519400,
"step": 3700
},
{
"epoch": 4.485472154963681,
"grad_norm": 0.7095040678977966,
"learning_rate": 3.367943533670501e-05,
"loss": 0.0758,
"num_input_tokens_seen": 1521416,
"step": 3705
},
{
"epoch": 4.491525423728813,
"grad_norm": 2.5754213333129883,
"learning_rate": 3.3629876989297405e-05,
"loss": 0.1493,
"num_input_tokens_seen": 1523240,
"step": 3710
},
{
"epoch": 4.4975786924939465,
"grad_norm": 1.1021342277526855,
"learning_rate": 3.3580280111883125e-05,
"loss": 0.1357,
"num_input_tokens_seen": 1525288,
"step": 3715
},
{
"epoch": 4.5,
"eval_loss": 0.1436297744512558,
"eval_runtime": 4.6052,
"eval_samples_per_second": 79.693,
"eval_steps_per_second": 19.978,
"num_input_tokens_seen": 1526088,
"step": 3717
},
{
"epoch": 4.50363196125908,
"grad_norm": 1.27449631690979,
"learning_rate": 3.3530644925898465e-05,
"loss": 0.1015,
"num_input_tokens_seen": 1527304,
"step": 3720
},
{
"epoch": 4.509685230024213,
"grad_norm": 4.025327205657959,
"learning_rate": 3.348097165295076e-05,
"loss": 0.1539,
"num_input_tokens_seen": 1529384,
"step": 3725
},
{
"epoch": 4.5157384987893465,
"grad_norm": 0.13018469512462616,
"learning_rate": 3.34312605148174e-05,
"loss": 0.0881,
"num_input_tokens_seen": 1531464,
"step": 3730
},
{
"epoch": 4.521791767554479,
"grad_norm": 5.49751091003418,
"learning_rate": 3.338151173344483e-05,
"loss": 0.1117,
"num_input_tokens_seen": 1533608,
"step": 3735
},
{
"epoch": 4.527845036319612,
"grad_norm": 0.5280150175094604,
"learning_rate": 3.333172553094754e-05,
"loss": 0.0613,
"num_input_tokens_seen": 1535656,
"step": 3740
},
{
"epoch": 4.533898305084746,
"grad_norm": 4.4329986572265625,
"learning_rate": 3.328190212960712e-05,
"loss": 0.1342,
"num_input_tokens_seen": 1537640,
"step": 3745
},
{
"epoch": 4.539951573849879,
"grad_norm": 2.851104974746704,
"learning_rate": 3.323204175187125e-05,
"loss": 0.1218,
"num_input_tokens_seen": 1539592,
"step": 3750
},
{
"epoch": 4.546004842615012,
"grad_norm": 1.7534023523330688,
"learning_rate": 3.318214462035266e-05,
"loss": 0.0491,
"num_input_tokens_seen": 1541576,
"step": 3755
},
{
"epoch": 4.552058111380145,
"grad_norm": 3.9379611015319824,
"learning_rate": 3.3132210957828226e-05,
"loss": 0.1587,
"num_input_tokens_seen": 1543464,
"step": 3760
},
{
"epoch": 4.558111380145278,
"grad_norm": 2.7358548641204834,
"learning_rate": 3.3082240987237875e-05,
"loss": 0.0942,
"num_input_tokens_seen": 1545416,
"step": 3765
},
{
"epoch": 4.5641646489104115,
"grad_norm": 1.8120285272598267,
"learning_rate": 3.3032234931683684e-05,
"loss": 0.045,
"num_input_tokens_seen": 1547432,
"step": 3770
},
{
"epoch": 4.570217917675545,
"grad_norm": 0.10306795686483383,
"learning_rate": 3.2982193014428805e-05,
"loss": 0.0788,
"num_input_tokens_seen": 1549576,
"step": 3775
},
{
"epoch": 4.576271186440678,
"grad_norm": 1.0330625772476196,
"learning_rate": 3.2932115458896515e-05,
"loss": 0.1055,
"num_input_tokens_seen": 1551688,
"step": 3780
},
{
"epoch": 4.582324455205811,
"grad_norm": 0.3044089078903198,
"learning_rate": 3.2882002488669204e-05,
"loss": 0.1335,
"num_input_tokens_seen": 1553672,
"step": 3785
},
{
"epoch": 4.588377723970944,
"grad_norm": 1.4460806846618652,
"learning_rate": 3.28318543274874e-05,
"loss": 0.0837,
"num_input_tokens_seen": 1555720,
"step": 3790
},
{
"epoch": 4.594430992736077,
"grad_norm": 3.097297191619873,
"learning_rate": 3.278167119924872e-05,
"loss": 0.1103,
"num_input_tokens_seen": 1557672,
"step": 3795
},
{
"epoch": 4.600484261501211,
"grad_norm": 0.30317577719688416,
"learning_rate": 3.27314533280069e-05,
"loss": 0.1134,
"num_input_tokens_seen": 1559880,
"step": 3800
},
{
"epoch": 4.606537530266344,
"grad_norm": 3.337770700454712,
"learning_rate": 3.268120093797082e-05,
"loss": 0.1148,
"num_input_tokens_seen": 1561960,
"step": 3805
},
{
"epoch": 4.6125907990314765,
"grad_norm": 4.097595691680908,
"learning_rate": 3.263091425350345e-05,
"loss": 0.0856,
"num_input_tokens_seen": 1563880,
"step": 3810
},
{
"epoch": 4.61864406779661,
"grad_norm": 0.2643558979034424,
"learning_rate": 3.258059349912089e-05,
"loss": 0.098,
"num_input_tokens_seen": 1565896,
"step": 3815
},
{
"epoch": 4.624697336561743,
"grad_norm": 2.8181965351104736,
"learning_rate": 3.253023889949135e-05,
"loss": 0.1251,
"num_input_tokens_seen": 1568040,
"step": 3820
},
{
"epoch": 4.6307506053268765,
"grad_norm": 3.229228973388672,
"learning_rate": 3.247985067943414e-05,
"loss": 0.1232,
"num_input_tokens_seen": 1570056,
"step": 3825
},
{
"epoch": 4.63680387409201,
"grad_norm": 4.2846999168396,
"learning_rate": 3.2429429063918696e-05,
"loss": 0.1687,
"num_input_tokens_seen": 1572168,
"step": 3830
},
{
"epoch": 4.642857142857143,
"grad_norm": 0.9762236475944519,
"learning_rate": 3.2378974278063534e-05,
"loss": 0.0517,
"num_input_tokens_seen": 1574216,
"step": 3835
},
{
"epoch": 4.648910411622276,
"grad_norm": 0.32010525465011597,
"learning_rate": 3.232848654713528e-05,
"loss": 0.0454,
"num_input_tokens_seen": 1576168,
"step": 3840
},
{
"epoch": 4.654963680387409,
"grad_norm": 1.3530986309051514,
"learning_rate": 3.227796609654765e-05,
"loss": 0.157,
"num_input_tokens_seen": 1578152,
"step": 3845
},
{
"epoch": 4.661016949152542,
"grad_norm": 0.9239023327827454,
"learning_rate": 3.222741315186043e-05,
"loss": 0.0574,
"num_input_tokens_seen": 1580104,
"step": 3850
},
{
"epoch": 4.667070217917676,
"grad_norm": 0.5891803503036499,
"learning_rate": 3.217682793877851e-05,
"loss": 0.1585,
"num_input_tokens_seen": 1582056,
"step": 3855
},
{
"epoch": 4.673123486682809,
"grad_norm": 2.4281816482543945,
"learning_rate": 3.212621068315081e-05,
"loss": 0.0981,
"num_input_tokens_seen": 1584136,
"step": 3860
},
{
"epoch": 4.6791767554479415,
"grad_norm": 1.570319652557373,
"learning_rate": 3.207556161096935e-05,
"loss": 0.0666,
"num_input_tokens_seen": 1586184,
"step": 3865
},
{
"epoch": 4.685230024213075,
"grad_norm": 0.6124736070632935,
"learning_rate": 3.202488094836819e-05,
"loss": 0.1518,
"num_input_tokens_seen": 1588296,
"step": 3870
},
{
"epoch": 4.691283292978208,
"grad_norm": 0.177334725856781,
"learning_rate": 3.197416892162242e-05,
"loss": 0.1367,
"num_input_tokens_seen": 1590504,
"step": 3875
},
{
"epoch": 4.697336561743342,
"grad_norm": 3.659196376800537,
"learning_rate": 3.1923425757147175e-05,
"loss": 0.1299,
"num_input_tokens_seen": 1592584,
"step": 3880
},
{
"epoch": 4.703389830508475,
"grad_norm": 2.184091806411743,
"learning_rate": 3.1872651681496604e-05,
"loss": 0.1015,
"num_input_tokens_seen": 1594728,
"step": 3885
},
{
"epoch": 4.709443099273607,
"grad_norm": 3.05389404296875,
"learning_rate": 3.182184692136287e-05,
"loss": 0.0826,
"num_input_tokens_seen": 1596776,
"step": 3890
},
{
"epoch": 4.715496368038741,
"grad_norm": 3.5935373306274414,
"learning_rate": 3.177101170357513e-05,
"loss": 0.1999,
"num_input_tokens_seen": 1598984,
"step": 3895
},
{
"epoch": 4.721549636803874,
"grad_norm": 4.548536777496338,
"learning_rate": 3.1720146255098535e-05,
"loss": 0.0944,
"num_input_tokens_seen": 1601096,
"step": 3900
},
{
"epoch": 4.727602905569007,
"grad_norm": 0.9657927751541138,
"learning_rate": 3.16692508030332e-05,
"loss": 0.1482,
"num_input_tokens_seen": 1603336,
"step": 3905
},
{
"epoch": 4.733656174334141,
"grad_norm": 2.0334603786468506,
"learning_rate": 3.16183255746132e-05,
"loss": 0.0449,
"num_input_tokens_seen": 1605320,
"step": 3910
},
{
"epoch": 4.739709443099273,
"grad_norm": 6.739678859710693,
"learning_rate": 3.156737079720555e-05,
"loss": 0.0969,
"num_input_tokens_seen": 1607304,
"step": 3915
},
{
"epoch": 4.745762711864407,
"grad_norm": 0.7341704964637756,
"learning_rate": 3.151638669830919e-05,
"loss": 0.1329,
"num_input_tokens_seen": 1609384,
"step": 3920
},
{
"epoch": 4.75181598062954,
"grad_norm": 4.675315856933594,
"learning_rate": 3.1465373505554e-05,
"loss": 0.1591,
"num_input_tokens_seen": 1611304,
"step": 3925
},
{
"epoch": 4.757869249394673,
"grad_norm": 1.1919448375701904,
"learning_rate": 3.14143314466997e-05,
"loss": 0.1263,
"num_input_tokens_seen": 1613192,
"step": 3930
},
{
"epoch": 4.763922518159807,
"grad_norm": 2.328058958053589,
"learning_rate": 3.136326074963494e-05,
"loss": 0.0304,
"num_input_tokens_seen": 1615304,
"step": 3935
},
{
"epoch": 4.76997578692494,
"grad_norm": 0.7319985628128052,
"learning_rate": 3.131216164237622e-05,
"loss": 0.1002,
"num_input_tokens_seen": 1617288,
"step": 3940
},
{
"epoch": 4.776029055690072,
"grad_norm": 4.388686180114746,
"learning_rate": 3.1261034353066884e-05,
"loss": 0.1414,
"num_input_tokens_seen": 1619336,
"step": 3945
},
{
"epoch": 4.782082324455206,
"grad_norm": 3.907365322113037,
"learning_rate": 3.1209879109976064e-05,
"loss": 0.0695,
"num_input_tokens_seen": 1621416,
"step": 3950
},
{
"epoch": 4.788135593220339,
"grad_norm": 4.539724349975586,
"learning_rate": 3.115869614149776e-05,
"loss": 0.1006,
"num_input_tokens_seen": 1623432,
"step": 3955
},
{
"epoch": 4.7941888619854724,
"grad_norm": 4.285974979400635,
"learning_rate": 3.1107485676149714e-05,
"loss": 0.2159,
"num_input_tokens_seen": 1625448,
"step": 3960
},
{
"epoch": 4.800242130750606,
"grad_norm": 5.8608551025390625,
"learning_rate": 3.105624794257245e-05,
"loss": 0.1279,
"num_input_tokens_seen": 1627624,
"step": 3965
},
{
"epoch": 4.806295399515738,
"grad_norm": 0.8381302952766418,
"learning_rate": 3.100498316952823e-05,
"loss": 0.1144,
"num_input_tokens_seen": 1629800,
"step": 3970
},
{
"epoch": 4.812348668280872,
"grad_norm": 3.4423251152038574,
"learning_rate": 3.095369158590006e-05,
"loss": 0.0671,
"num_input_tokens_seen": 1631720,
"step": 3975
},
{
"epoch": 4.818401937046005,
"grad_norm": 1.0596593618392944,
"learning_rate": 3.09023734206906e-05,
"loss": 0.1878,
"num_input_tokens_seen": 1633704,
"step": 3980
},
{
"epoch": 4.824455205811138,
"grad_norm": 4.54054069519043,
"learning_rate": 3.085102890302125e-05,
"loss": 0.1463,
"num_input_tokens_seen": 1635656,
"step": 3985
},
{
"epoch": 4.830508474576272,
"grad_norm": 1.3257328271865845,
"learning_rate": 3.079965826213102e-05,
"loss": 0.1182,
"num_input_tokens_seen": 1637736,
"step": 3990
},
{
"epoch": 4.836561743341404,
"grad_norm": 1.2008308172225952,
"learning_rate": 3.074826172737559e-05,
"loss": 0.1531,
"num_input_tokens_seen": 1639816,
"step": 3995
},
{
"epoch": 4.842615012106537,
"grad_norm": 5.070513725280762,
"learning_rate": 3.0696839528226206e-05,
"loss": 0.1027,
"num_input_tokens_seen": 1641736,
"step": 4000
},
{
"epoch": 4.848668280871671,
"grad_norm": 1.366325855255127,
"learning_rate": 3.064539189426874e-05,
"loss": 0.144,
"num_input_tokens_seen": 1643656,
"step": 4005
},
{
"epoch": 4.854721549636804,
"grad_norm": 5.837193489074707,
"learning_rate": 3.059391905520259e-05,
"loss": 0.1821,
"num_input_tokens_seen": 1645736,
"step": 4010
},
{
"epoch": 4.8607748184019375,
"grad_norm": 1.7730399370193481,
"learning_rate": 3.054242124083972e-05,
"loss": 0.0776,
"num_input_tokens_seen": 1647688,
"step": 4015
},
{
"epoch": 4.86682808716707,
"grad_norm": 4.303905487060547,
"learning_rate": 3.0490898681103575e-05,
"loss": 0.1172,
"num_input_tokens_seen": 1649768,
"step": 4020
},
{
"epoch": 4.872881355932203,
"grad_norm": 6.675983905792236,
"learning_rate": 3.0439351606028094e-05,
"loss": 0.1297,
"num_input_tokens_seen": 1651688,
"step": 4025
},
{
"epoch": 4.878934624697337,
"grad_norm": 2.1775729656219482,
"learning_rate": 3.0387780245756655e-05,
"loss": 0.0857,
"num_input_tokens_seen": 1653896,
"step": 4030
},
{
"epoch": 4.88498789346247,
"grad_norm": 0.18741416931152344,
"learning_rate": 3.0336184830541093e-05,
"loss": 0.0326,
"num_input_tokens_seen": 1656008,
"step": 4035
},
{
"epoch": 4.891041162227603,
"grad_norm": 0.9524343013763428,
"learning_rate": 3.028456559074061e-05,
"loss": 0.1001,
"num_input_tokens_seen": 1658088,
"step": 4040
},
{
"epoch": 4.897094430992736,
"grad_norm": 1.6602981090545654,
"learning_rate": 3.0232922756820804e-05,
"loss": 0.0551,
"num_input_tokens_seen": 1660200,
"step": 4045
},
{
"epoch": 4.903147699757869,
"grad_norm": 4.7026262283325195,
"learning_rate": 3.0181256559352587e-05,
"loss": 0.1072,
"num_input_tokens_seen": 1662440,
"step": 4050
},
{
"epoch": 4.9092009685230025,
"grad_norm": 0.43130701780319214,
"learning_rate": 3.0129567229011214e-05,
"loss": 0.0924,
"num_input_tokens_seen": 1664552,
"step": 4055
},
{
"epoch": 4.915254237288136,
"grad_norm": 1.2491116523742676,
"learning_rate": 3.0077854996575184e-05,
"loss": 0.081,
"num_input_tokens_seen": 1666600,
"step": 4060
},
{
"epoch": 4.921307506053269,
"grad_norm": 0.4165017902851105,
"learning_rate": 3.0026120092925293e-05,
"loss": 0.1173,
"num_input_tokens_seen": 1668776,
"step": 4065
},
{
"epoch": 4.927360774818402,
"grad_norm": 2.2347145080566406,
"learning_rate": 2.9974362749043512e-05,
"loss": 0.0687,
"num_input_tokens_seen": 1670952,
"step": 4070
},
{
"epoch": 4.933414043583535,
"grad_norm": 0.06528237462043762,
"learning_rate": 2.9922583196012037e-05,
"loss": 0.0733,
"num_input_tokens_seen": 1673128,
"step": 4075
},
{
"epoch": 4.939467312348668,
"grad_norm": 3.1934924125671387,
"learning_rate": 2.9870781665012204e-05,
"loss": 0.0489,
"num_input_tokens_seen": 1675112,
"step": 4080
},
{
"epoch": 4.945520581113802,
"grad_norm": 3.640748977661133,
"learning_rate": 2.981895838732348e-05,
"loss": 0.1047,
"num_input_tokens_seen": 1677096,
"step": 4085
},
{
"epoch": 4.951573849878935,
"grad_norm": 0.3172272741794586,
"learning_rate": 2.9767113594322426e-05,
"loss": 0.1,
"num_input_tokens_seen": 1679080,
"step": 4090
},
{
"epoch": 4.9576271186440675,
"grad_norm": 3.484344959259033,
"learning_rate": 2.9715247517481655e-05,
"loss": 0.067,
"num_input_tokens_seen": 1681000,
"step": 4095
},
{
"epoch": 4.963680387409201,
"grad_norm": 2.876481294631958,
"learning_rate": 2.96633603883688e-05,
"loss": 0.247,
"num_input_tokens_seen": 1683048,
"step": 4100
},
{
"epoch": 4.969733656174334,
"grad_norm": 2.505605697631836,
"learning_rate": 2.961145243864552e-05,
"loss": 0.0494,
"num_input_tokens_seen": 1685160,
"step": 4105
},
{
"epoch": 4.9757869249394675,
"grad_norm": 9.4348783493042,
"learning_rate": 2.9559523900066395e-05,
"loss": 0.2144,
"num_input_tokens_seen": 1687048,
"step": 4110
},
{
"epoch": 4.981840193704601,
"grad_norm": 2.766404390335083,
"learning_rate": 2.9507575004477955e-05,
"loss": 0.1706,
"num_input_tokens_seen": 1689000,
"step": 4115
},
{
"epoch": 4.987893462469733,
"grad_norm": 3.0030455589294434,
"learning_rate": 2.9455605983817598e-05,
"loss": 0.0957,
"num_input_tokens_seen": 1691112,
"step": 4120
},
{
"epoch": 4.993946731234867,
"grad_norm": 0.21309371292591095,
"learning_rate": 2.9403617070112587e-05,
"loss": 0.122,
"num_input_tokens_seen": 1693160,
"step": 4125
},
{
"epoch": 5.0,
"grad_norm": 11.627579689025879,
"learning_rate": 2.9351608495479004e-05,
"loss": 0.069,
"num_input_tokens_seen": 1694912,
"step": 4130
},
{
"epoch": 5.0,
"eval_loss": 0.14715999364852905,
"eval_runtime": 4.5922,
"eval_samples_per_second": 79.919,
"eval_steps_per_second": 20.034,
"num_input_tokens_seen": 1694912,
"step": 4130
},
{
"epoch": 5.006053268765133,
"grad_norm": 8.168224334716797,
"learning_rate": 2.92995804921207e-05,
"loss": 0.0962,
"num_input_tokens_seen": 1697056,
"step": 4135
},
{
"epoch": 5.012106537530267,
"grad_norm": 4.525173187255859,
"learning_rate": 2.9247533292328273e-05,
"loss": 0.1041,
"num_input_tokens_seen": 1699168,
"step": 4140
},
{
"epoch": 5.018159806295399,
"grad_norm": 0.5518445372581482,
"learning_rate": 2.9195467128478044e-05,
"loss": 0.0151,
"num_input_tokens_seen": 1701152,
"step": 4145
},
{
"epoch": 5.0242130750605325,
"grad_norm": 0.14739976823329926,
"learning_rate": 2.914338223303098e-05,
"loss": 0.1272,
"num_input_tokens_seen": 1703168,
"step": 4150
},
{
"epoch": 5.030266343825666,
"grad_norm": 1.3362171649932861,
"learning_rate": 2.9091278838531695e-05,
"loss": 0.0119,
"num_input_tokens_seen": 1705120,
"step": 4155
},
{
"epoch": 5.036319612590799,
"grad_norm": 2.4070515632629395,
"learning_rate": 2.9039157177607383e-05,
"loss": 0.0939,
"num_input_tokens_seen": 1707200,
"step": 4160
},
{
"epoch": 5.0423728813559325,
"grad_norm": 4.166516304016113,
"learning_rate": 2.8987017482966815e-05,
"loss": 0.1308,
"num_input_tokens_seen": 1709120,
"step": 4165
},
{
"epoch": 5.048426150121065,
"grad_norm": 4.306673526763916,
"learning_rate": 2.893485998739926e-05,
"loss": 0.0451,
"num_input_tokens_seen": 1711200,
"step": 4170
},
{
"epoch": 5.054479418886198,
"grad_norm": 7.965576171875,
"learning_rate": 2.8882684923773458e-05,
"loss": 0.2002,
"num_input_tokens_seen": 1713248,
"step": 4175
},
{
"epoch": 5.060532687651332,
"grad_norm": 2.435213327407837,
"learning_rate": 2.883049252503659e-05,
"loss": 0.0611,
"num_input_tokens_seen": 1715296,
"step": 4180
},
{
"epoch": 5.066585956416465,
"grad_norm": 0.1197163313627243,
"learning_rate": 2.877828302421325e-05,
"loss": 0.1464,
"num_input_tokens_seen": 1717280,
"step": 4185
},
{
"epoch": 5.072639225181598,
"grad_norm": 5.31384801864624,
"learning_rate": 2.872605665440436e-05,
"loss": 0.1242,
"num_input_tokens_seen": 1719136,
"step": 4190
},
{
"epoch": 5.078692493946731,
"grad_norm": 0.12134741991758347,
"learning_rate": 2.8673813648786196e-05,
"loss": 0.1163,
"num_input_tokens_seen": 1721152,
"step": 4195
},
{
"epoch": 5.084745762711864,
"grad_norm": 5.05936861038208,
"learning_rate": 2.862155424060926e-05,
"loss": 0.0711,
"num_input_tokens_seen": 1723328,
"step": 4200
},
{
"epoch": 5.0907990314769975,
"grad_norm": 0.19684794545173645,
"learning_rate": 2.856927866319733e-05,
"loss": 0.0897,
"num_input_tokens_seen": 1725280,
"step": 4205
},
{
"epoch": 5.096852300242131,
"grad_norm": 0.09383846819400787,
"learning_rate": 2.851698714994635e-05,
"loss": 0.0437,
"num_input_tokens_seen": 1727328,
"step": 4210
},
{
"epoch": 5.102905569007264,
"grad_norm": 7.017461776733398,
"learning_rate": 2.8464679934323424e-05,
"loss": 0.0987,
"num_input_tokens_seen": 1729472,
"step": 4215
},
{
"epoch": 5.108958837772397,
"grad_norm": 0.7437957525253296,
"learning_rate": 2.841235724986575e-05,
"loss": 0.0985,
"num_input_tokens_seen": 1731392,
"step": 4220
},
{
"epoch": 5.11501210653753,
"grad_norm": 7.1092705726623535,
"learning_rate": 2.8360019330179604e-05,
"loss": 0.188,
"num_input_tokens_seen": 1733472,
"step": 4225
},
{
"epoch": 5.121065375302663,
"grad_norm": 0.43915751576423645,
"learning_rate": 2.8307666408939278e-05,
"loss": 0.107,
"num_input_tokens_seen": 1735520,
"step": 4230
},
{
"epoch": 5.127118644067797,
"grad_norm": 0.4318813979625702,
"learning_rate": 2.8255298719886043e-05,
"loss": 0.1732,
"num_input_tokens_seen": 1737536,
"step": 4235
},
{
"epoch": 5.13317191283293,
"grad_norm": 4.211883544921875,
"learning_rate": 2.820291649682709e-05,
"loss": 0.1299,
"num_input_tokens_seen": 1739424,
"step": 4240
},
{
"epoch": 5.1392251815980625,
"grad_norm": 0.5891851186752319,
"learning_rate": 2.8150519973634543e-05,
"loss": 0.0775,
"num_input_tokens_seen": 1741536,
"step": 4245
},
{
"epoch": 5.145278450363196,
"grad_norm": 4.266312122344971,
"learning_rate": 2.809810938424432e-05,
"loss": 0.1593,
"num_input_tokens_seen": 1743488,
"step": 4250
},
{
"epoch": 5.151331719128329,
"grad_norm": 2.173311710357666,
"learning_rate": 2.804568496265516e-05,
"loss": 0.1045,
"num_input_tokens_seen": 1745728,
"step": 4255
},
{
"epoch": 5.157384987893463,
"grad_norm": 3.343336820602417,
"learning_rate": 2.799324694292757e-05,
"loss": 0.0858,
"num_input_tokens_seen": 1747808,
"step": 4260
},
{
"epoch": 5.163438256658596,
"grad_norm": 0.05580740049481392,
"learning_rate": 2.7940795559182764e-05,
"loss": 0.0505,
"num_input_tokens_seen": 1749856,
"step": 4265
},
{
"epoch": 5.169491525423728,
"grad_norm": 4.62854528427124,
"learning_rate": 2.788833104560161e-05,
"loss": 0.1006,
"num_input_tokens_seen": 1751904,
"step": 4270
},
{
"epoch": 5.175544794188862,
"grad_norm": 0.49617746472358704,
"learning_rate": 2.7835853636423616e-05,
"loss": 0.0892,
"num_input_tokens_seen": 1753984,
"step": 4275
},
{
"epoch": 5.181598062953995,
"grad_norm": 2.32796311378479,
"learning_rate": 2.7783363565945847e-05,
"loss": 0.0829,
"num_input_tokens_seen": 1756000,
"step": 4280
},
{
"epoch": 5.187651331719128,
"grad_norm": 0.8803606033325195,
"learning_rate": 2.773086106852192e-05,
"loss": 0.0805,
"num_input_tokens_seen": 1758080,
"step": 4285
},
{
"epoch": 5.193704600484262,
"grad_norm": 1.4200032949447632,
"learning_rate": 2.7678346378560903e-05,
"loss": 0.0972,
"num_input_tokens_seen": 1760224,
"step": 4290
},
{
"epoch": 5.199757869249395,
"grad_norm": 5.253544807434082,
"learning_rate": 2.762581973052633e-05,
"loss": 0.0848,
"num_input_tokens_seen": 1762176,
"step": 4295
},
{
"epoch": 5.2058111380145276,
"grad_norm": 0.5915831923484802,
"learning_rate": 2.7573281358935104e-05,
"loss": 0.0243,
"num_input_tokens_seen": 1764352,
"step": 4300
},
{
"epoch": 5.211864406779661,
"grad_norm": 5.024724960327148,
"learning_rate": 2.7520731498356494e-05,
"loss": 0.0917,
"num_input_tokens_seen": 1766432,
"step": 4305
},
{
"epoch": 5.217917675544794,
"grad_norm": 1.844663143157959,
"learning_rate": 2.746817038341103e-05,
"loss": 0.2167,
"num_input_tokens_seen": 1768608,
"step": 4310
},
{
"epoch": 5.223970944309928,
"grad_norm": 0.30476701259613037,
"learning_rate": 2.7415598248769524e-05,
"loss": 0.0132,
"num_input_tokens_seen": 1770752,
"step": 4315
},
{
"epoch": 5.230024213075061,
"grad_norm": 2.587169885635376,
"learning_rate": 2.7363015329151965e-05,
"loss": 0.0206,
"num_input_tokens_seen": 1772832,
"step": 4320
},
{
"epoch": 5.236077481840193,
"grad_norm": 0.4779854118824005,
"learning_rate": 2.73104218593265e-05,
"loss": 0.1323,
"num_input_tokens_seen": 1774912,
"step": 4325
},
{
"epoch": 5.242130750605327,
"grad_norm": 1.2622554302215576,
"learning_rate": 2.7257818074108394e-05,
"loss": 0.2511,
"num_input_tokens_seen": 1777024,
"step": 4330
},
{
"epoch": 5.24818401937046,
"grad_norm": 1.7752158641815186,
"learning_rate": 2.7205204208358947e-05,
"loss": 0.1289,
"num_input_tokens_seen": 1779008,
"step": 4335
},
{
"epoch": 5.254237288135593,
"grad_norm": 6.4176836013793945,
"learning_rate": 2.715258049698446e-05,
"loss": 0.0524,
"num_input_tokens_seen": 1780896,
"step": 4340
},
{
"epoch": 5.260290556900727,
"grad_norm": 3.9696273803710938,
"learning_rate": 2.709994717493523e-05,
"loss": 0.1406,
"num_input_tokens_seen": 1783008,
"step": 4345
},
{
"epoch": 5.266343825665859,
"grad_norm": 0.1831623762845993,
"learning_rate": 2.7047304477204416e-05,
"loss": 0.0092,
"num_input_tokens_seen": 1784992,
"step": 4350
},
{
"epoch": 5.272397094430993,
"grad_norm": 11.523201942443848,
"learning_rate": 2.6994652638827078e-05,
"loss": 0.3058,
"num_input_tokens_seen": 1787008,
"step": 4355
},
{
"epoch": 5.278450363196126,
"grad_norm": 3.319190502166748,
"learning_rate": 2.694199189487906e-05,
"loss": 0.1035,
"num_input_tokens_seen": 1789120,
"step": 4360
},
{
"epoch": 5.284503631961259,
"grad_norm": 4.413753509521484,
"learning_rate": 2.688932248047597e-05,
"loss": 0.1166,
"num_input_tokens_seen": 1791232,
"step": 4365
},
{
"epoch": 5.290556900726393,
"grad_norm": 5.708840370178223,
"learning_rate": 2.683664463077214e-05,
"loss": 0.0982,
"num_input_tokens_seen": 1793440,
"step": 4370
},
{
"epoch": 5.296610169491525,
"grad_norm": 4.179380893707275,
"learning_rate": 2.678395858095955e-05,
"loss": 0.0611,
"num_input_tokens_seen": 1795488,
"step": 4375
},
{
"epoch": 5.302663438256658,
"grad_norm": 4.286114692687988,
"learning_rate": 2.6731264566266795e-05,
"loss": 0.0886,
"num_input_tokens_seen": 1797600,
"step": 4380
},
{
"epoch": 5.308716707021792,
"grad_norm": 4.661886215209961,
"learning_rate": 2.6678562821958043e-05,
"loss": 0.1656,
"num_input_tokens_seen": 1799584,
"step": 4385
},
{
"epoch": 5.314769975786925,
"grad_norm": 0.09037864208221436,
"learning_rate": 2.6625853583331943e-05,
"loss": 0.1248,
"num_input_tokens_seen": 1801440,
"step": 4390
},
{
"epoch": 5.3208232445520585,
"grad_norm": 3.2096681594848633,
"learning_rate": 2.6573137085720638e-05,
"loss": 0.0649,
"num_input_tokens_seen": 1803456,
"step": 4395
},
{
"epoch": 5.326876513317191,
"grad_norm": 1.737039566040039,
"learning_rate": 2.6520413564488672e-05,
"loss": 0.1877,
"num_input_tokens_seen": 1805440,
"step": 4400
},
{
"epoch": 5.332929782082324,
"grad_norm": 3.13798451423645,
"learning_rate": 2.6467683255031918e-05,
"loss": 0.0848,
"num_input_tokens_seen": 1807360,
"step": 4405
},
{
"epoch": 5.338983050847458,
"grad_norm": 3.5494918823242188,
"learning_rate": 2.6414946392776597e-05,
"loss": 0.0934,
"num_input_tokens_seen": 1809344,
"step": 4410
},
{
"epoch": 5.345036319612591,
"grad_norm": 0.5050595998764038,
"learning_rate": 2.636220321317816e-05,
"loss": 0.073,
"num_input_tokens_seen": 1811456,
"step": 4415
},
{
"epoch": 5.351089588377724,
"grad_norm": 3.0784072875976562,
"learning_rate": 2.6309453951720274e-05,
"loss": 0.048,
"num_input_tokens_seen": 1813600,
"step": 4420
},
{
"epoch": 5.357142857142857,
"grad_norm": 5.797350883483887,
"learning_rate": 2.625669884391377e-05,
"loss": 0.2233,
"num_input_tokens_seen": 1815424,
"step": 4425
},
{
"epoch": 5.36319612590799,
"grad_norm": 9.18054485321045,
"learning_rate": 2.6203938125295552e-05,
"loss": 0.1113,
"num_input_tokens_seen": 1817600,
"step": 4430
},
{
"epoch": 5.3692493946731235,
"grad_norm": 5.451303482055664,
"learning_rate": 2.6151172031427597e-05,
"loss": 0.0575,
"num_input_tokens_seen": 1819648,
"step": 4435
},
{
"epoch": 5.375302663438257,
"grad_norm": 0.7396259903907776,
"learning_rate": 2.609840079789588e-05,
"loss": 0.1025,
"num_input_tokens_seen": 1821728,
"step": 4440
},
{
"epoch": 5.38135593220339,
"grad_norm": 5.703775882720947,
"learning_rate": 2.604562466030931e-05,
"loss": 0.0703,
"num_input_tokens_seen": 1823776,
"step": 4445
},
{
"epoch": 5.387409200968523,
"grad_norm": 0.9737942218780518,
"learning_rate": 2.599284385429871e-05,
"loss": 0.0886,
"num_input_tokens_seen": 1825856,
"step": 4450
},
{
"epoch": 5.393462469733656,
"grad_norm": 6.848929405212402,
"learning_rate": 2.594005861551574e-05,
"loss": 0.0506,
"num_input_tokens_seen": 1827936,
"step": 4455
},
{
"epoch": 5.399515738498789,
"grad_norm": 0.08520136773586273,
"learning_rate": 2.588726917963183e-05,
"loss": 0.087,
"num_input_tokens_seen": 1829824,
"step": 4460
},
{
"epoch": 5.405569007263923,
"grad_norm": 0.5107972621917725,
"learning_rate": 2.5834475782337187e-05,
"loss": 0.0144,
"num_input_tokens_seen": 1831936,
"step": 4465
},
{
"epoch": 5.411622276029056,
"grad_norm": 1.4982185363769531,
"learning_rate": 2.578167865933967e-05,
"loss": 0.0483,
"num_input_tokens_seen": 1834048,
"step": 4470
},
{
"epoch": 5.4176755447941884,
"grad_norm": 0.8156539797782898,
"learning_rate": 2.5728878046363785e-05,
"loss": 0.0896,
"num_input_tokens_seen": 1836192,
"step": 4475
},
{
"epoch": 5.423728813559322,
"grad_norm": 4.080778121948242,
"learning_rate": 2.5676074179149635e-05,
"loss": 0.0587,
"num_input_tokens_seen": 1838432,
"step": 4480
},
{
"epoch": 5.429782082324455,
"grad_norm": 6.264119625091553,
"learning_rate": 2.5623267293451826e-05,
"loss": 0.1014,
"num_input_tokens_seen": 1840576,
"step": 4485
},
{
"epoch": 5.4358353510895885,
"grad_norm": 0.09231238812208176,
"learning_rate": 2.5570457625038457e-05,
"loss": 0.0556,
"num_input_tokens_seen": 1842624,
"step": 4490
},
{
"epoch": 5.441888619854722,
"grad_norm": 6.133981704711914,
"learning_rate": 2.551764540969005e-05,
"loss": 0.1426,
"num_input_tokens_seen": 1844576,
"step": 4495
},
{
"epoch": 5.447941888619855,
"grad_norm": 1.5127137899398804,
"learning_rate": 2.5464830883198492e-05,
"loss": 0.119,
"num_input_tokens_seen": 1846560,
"step": 4500
},
{
"epoch": 5.453995157384988,
"grad_norm": 5.255079746246338,
"learning_rate": 2.5412014281365986e-05,
"loss": 0.1274,
"num_input_tokens_seen": 1848576,
"step": 4505
},
{
"epoch": 5.460048426150121,
"grad_norm": 5.694177627563477,
"learning_rate": 2.5359195840004023e-05,
"loss": 0.0501,
"num_input_tokens_seen": 1850688,
"step": 4510
},
{
"epoch": 5.466101694915254,
"grad_norm": 2.639869451522827,
"learning_rate": 2.5306375794932273e-05,
"loss": 0.0547,
"num_input_tokens_seen": 1852800,
"step": 4515
},
{
"epoch": 5.472154963680388,
"grad_norm": 1.6186069250106812,
"learning_rate": 2.52535543819776e-05,
"loss": 0.1454,
"num_input_tokens_seen": 1854720,
"step": 4520
},
{
"epoch": 5.478208232445521,
"grad_norm": 4.176394939422607,
"learning_rate": 2.5200731836972956e-05,
"loss": 0.2223,
"num_input_tokens_seen": 1856768,
"step": 4525
},
{
"epoch": 5.4842615012106535,
"grad_norm": 0.455463171005249,
"learning_rate": 2.5147908395756343e-05,
"loss": 0.033,
"num_input_tokens_seen": 1858848,
"step": 4530
},
{
"epoch": 5.490314769975787,
"grad_norm": 3.566337823867798,
"learning_rate": 2.5095084294169768e-05,
"loss": 0.0516,
"num_input_tokens_seen": 1860896,
"step": 4535
},
{
"epoch": 5.49636803874092,
"grad_norm": 4.8022894859313965,
"learning_rate": 2.5042259768058208e-05,
"loss": 0.2009,
"num_input_tokens_seen": 1862816,
"step": 4540
},
{
"epoch": 5.5,
"eval_loss": 0.15775655210018158,
"eval_runtime": 4.5824,
"eval_samples_per_second": 80.089,
"eval_steps_per_second": 20.077,
"num_input_tokens_seen": 1864000,
"step": 4543
},
{
"epoch": 5.5024213075060535,
"grad_norm": 4.100371360778809,
"learning_rate": 2.4989435053268497e-05,
"loss": 0.1968,
"num_input_tokens_seen": 1864832,
"step": 4545
},
{
"epoch": 5.508474576271187,
"grad_norm": 5.8973541259765625,
"learning_rate": 2.493661038564835e-05,
"loss": 0.1107,
"num_input_tokens_seen": 1866752,
"step": 4550
},
{
"epoch": 5.514527845036319,
"grad_norm": 0.15794266760349274,
"learning_rate": 2.4883786001045238e-05,
"loss": 0.0457,
"num_input_tokens_seen": 1868928,
"step": 4555
},
{
"epoch": 5.520581113801453,
"grad_norm": 3.9451935291290283,
"learning_rate": 2.4830962135305398e-05,
"loss": 0.1607,
"num_input_tokens_seen": 1871072,
"step": 4560
},
{
"epoch": 5.526634382566586,
"grad_norm": 1.0250242948532104,
"learning_rate": 2.4778139024272724e-05,
"loss": 0.0756,
"num_input_tokens_seen": 1873152,
"step": 4565
},
{
"epoch": 5.532687651331719,
"grad_norm": 5.201865196228027,
"learning_rate": 2.4725316903787765e-05,
"loss": 0.1061,
"num_input_tokens_seen": 1875040,
"step": 4570
},
{
"epoch": 5.538740920096853,
"grad_norm": 0.0985790342092514,
"learning_rate": 2.4672496009686622e-05,
"loss": 0.0073,
"num_input_tokens_seen": 1877184,
"step": 4575
},
{
"epoch": 5.544794188861985,
"grad_norm": 2.5414302349090576,
"learning_rate": 2.4619676577799946e-05,
"loss": 0.1435,
"num_input_tokens_seen": 1879232,
"step": 4580
},
{
"epoch": 5.5508474576271185,
"grad_norm": 4.586217403411865,
"learning_rate": 2.4566858843951847e-05,
"loss": 0.1246,
"num_input_tokens_seen": 1881312,
"step": 4585
},
{
"epoch": 5.556900726392252,
"grad_norm": 6.781052589416504,
"learning_rate": 2.451404304395884e-05,
"loss": 0.2671,
"num_input_tokens_seen": 1883328,
"step": 4590
},
{
"epoch": 5.562953995157385,
"grad_norm": 10.360923767089844,
"learning_rate": 2.446122941362883e-05,
"loss": 0.1366,
"num_input_tokens_seen": 1885248,
"step": 4595
},
{
"epoch": 5.5690072639225185,
"grad_norm": 3.757084608078003,
"learning_rate": 2.4408418188760026e-05,
"loss": 0.1358,
"num_input_tokens_seen": 1887328,
"step": 4600
},
{
"epoch": 5.575060532687651,
"grad_norm": 3.5405287742614746,
"learning_rate": 2.435560960513989e-05,
"loss": 0.1358,
"num_input_tokens_seen": 1889408,
"step": 4605
},
{
"epoch": 5.581113801452784,
"grad_norm": 1.0986759662628174,
"learning_rate": 2.4302803898544106e-05,
"loss": 0.1262,
"num_input_tokens_seen": 1891456,
"step": 4610
},
{
"epoch": 5.587167070217918,
"grad_norm": 3.6903698444366455,
"learning_rate": 2.425000130473549e-05,
"loss": 0.0779,
"num_input_tokens_seen": 1893696,
"step": 4615
},
{
"epoch": 5.593220338983051,
"grad_norm": 0.8637160658836365,
"learning_rate": 2.4197202059463e-05,
"loss": 0.1851,
"num_input_tokens_seen": 1895616,
"step": 4620
},
{
"epoch": 5.599273607748184,
"grad_norm": 4.0059895515441895,
"learning_rate": 2.4144406398460594e-05,
"loss": 0.1229,
"num_input_tokens_seen": 1897600,
"step": 4625
},
{
"epoch": 5.605326876513317,
"grad_norm": 5.283796787261963,
"learning_rate": 2.4091614557446267e-05,
"loss": 0.2567,
"num_input_tokens_seen": 1899616,
"step": 4630
},
{
"epoch": 5.61138014527845,
"grad_norm": 0.3470328450202942,
"learning_rate": 2.4038826772120932e-05,
"loss": 0.0737,
"num_input_tokens_seen": 1901568,
"step": 4635
},
{
"epoch": 5.6174334140435835,
"grad_norm": 4.049567222595215,
"learning_rate": 2.398604327816742e-05,
"loss": 0.1776,
"num_input_tokens_seen": 1903616,
"step": 4640
},
{
"epoch": 5.623486682808717,
"grad_norm": 3.8298497200012207,
"learning_rate": 2.3933264311249377e-05,
"loss": 0.1617,
"num_input_tokens_seen": 1905504,
"step": 4645
},
{
"epoch": 5.62953995157385,
"grad_norm": 2.1219868659973145,
"learning_rate": 2.3880490107010255e-05,
"loss": 0.0557,
"num_input_tokens_seen": 1907456,
"step": 4650
},
{
"epoch": 5.635593220338983,
"grad_norm": 2.666740894317627,
"learning_rate": 2.382772090107223e-05,
"loss": 0.0658,
"num_input_tokens_seen": 1909568,
"step": 4655
},
{
"epoch": 5.641646489104116,
"grad_norm": 2.493098258972168,
"learning_rate": 2.3774956929035177e-05,
"loss": 0.0427,
"num_input_tokens_seen": 1911840,
"step": 4660
},
{
"epoch": 5.647699757869249,
"grad_norm": 0.14598090946674347,
"learning_rate": 2.3722198426475593e-05,
"loss": 0.1101,
"num_input_tokens_seen": 1913952,
"step": 4665
},
{
"epoch": 5.653753026634383,
"grad_norm": 4.130605220794678,
"learning_rate": 2.3669445628945542e-05,
"loss": 0.0634,
"num_input_tokens_seen": 1916160,
"step": 4670
},
{
"epoch": 5.659806295399516,
"grad_norm": 1.1681698560714722,
"learning_rate": 2.3616698771971633e-05,
"loss": 0.0881,
"num_input_tokens_seen": 1918144,
"step": 4675
},
{
"epoch": 5.6658595641646485,
"grad_norm": 6.252587795257568,
"learning_rate": 2.356395809105396e-05,
"loss": 0.1455,
"num_input_tokens_seen": 1920160,
"step": 4680
},
{
"epoch": 5.671912832929782,
"grad_norm": 1.4607112407684326,
"learning_rate": 2.3511223821665028e-05,
"loss": 0.1253,
"num_input_tokens_seen": 1922368,
"step": 4685
},
{
"epoch": 5.677966101694915,
"grad_norm": 3.692471742630005,
"learning_rate": 2.3458496199248717e-05,
"loss": 0.1402,
"num_input_tokens_seen": 1924448,
"step": 4690
},
{
"epoch": 5.684019370460049,
"grad_norm": 5.765774726867676,
"learning_rate": 2.340577545921923e-05,
"loss": 0.2013,
"num_input_tokens_seen": 1926592,
"step": 4695
},
{
"epoch": 5.690072639225182,
"grad_norm": 2.151799440383911,
"learning_rate": 2.335306183696006e-05,
"loss": 0.0668,
"num_input_tokens_seen": 1928672,
"step": 4700
},
{
"epoch": 5.696125907990314,
"grad_norm": 4.2460222244262695,
"learning_rate": 2.3300355567822897e-05,
"loss": 0.157,
"num_input_tokens_seen": 1930816,
"step": 4705
},
{
"epoch": 5.702179176755448,
"grad_norm": 1.2913779020309448,
"learning_rate": 2.324765688712661e-05,
"loss": 0.1019,
"num_input_tokens_seen": 1932896,
"step": 4710
},
{
"epoch": 5.708232445520581,
"grad_norm": 0.1220802515745163,
"learning_rate": 2.3194966030156187e-05,
"loss": 0.0852,
"num_input_tokens_seen": 1934944,
"step": 4715
},
{
"epoch": 5.714285714285714,
"grad_norm": 0.9283545017242432,
"learning_rate": 2.31422832321617e-05,
"loss": 0.1064,
"num_input_tokens_seen": 1937152,
"step": 4720
},
{
"epoch": 5.720338983050848,
"grad_norm": 0.4256783723831177,
"learning_rate": 2.308960872835721e-05,
"loss": 0.063,
"num_input_tokens_seen": 1939328,
"step": 4725
},
{
"epoch": 5.72639225181598,
"grad_norm": 1.6901894807815552,
"learning_rate": 2.3036942753919775e-05,
"loss": 0.1151,
"num_input_tokens_seen": 1941312,
"step": 4730
},
{
"epoch": 5.732445520581114,
"grad_norm": 5.385450839996338,
"learning_rate": 2.2984285543988352e-05,
"loss": 0.0663,
"num_input_tokens_seen": 1943264,
"step": 4735
},
{
"epoch": 5.738498789346247,
"grad_norm": 0.5914505124092102,
"learning_rate": 2.2931637333662785e-05,
"loss": 0.0648,
"num_input_tokens_seen": 1945312,
"step": 4740
},
{
"epoch": 5.74455205811138,
"grad_norm": 0.29477110505104065,
"learning_rate": 2.287899835800273e-05,
"loss": 0.0784,
"num_input_tokens_seen": 1947328,
"step": 4745
},
{
"epoch": 5.750605326876514,
"grad_norm": 3.8047056198120117,
"learning_rate": 2.2826368852026597e-05,
"loss": 0.0983,
"num_input_tokens_seen": 1949376,
"step": 4750
},
{
"epoch": 5.756658595641646,
"grad_norm": 3.759469509124756,
"learning_rate": 2.277374905071053e-05,
"loss": 0.2523,
"num_input_tokens_seen": 1951456,
"step": 4755
},
{
"epoch": 5.762711864406779,
"grad_norm": 3.657890796661377,
"learning_rate": 2.2721139188987357e-05,
"loss": 0.0834,
"num_input_tokens_seen": 1953536,
"step": 4760
},
{
"epoch": 5.768765133171913,
"grad_norm": 7.039313316345215,
"learning_rate": 2.26685395017455e-05,
"loss": 0.1593,
"num_input_tokens_seen": 1955744,
"step": 4765
},
{
"epoch": 5.774818401937046,
"grad_norm": 5.334690093994141,
"learning_rate": 2.261595022382799e-05,
"loss": 0.1404,
"num_input_tokens_seen": 1957824,
"step": 4770
},
{
"epoch": 5.780871670702179,
"grad_norm": 2.440227508544922,
"learning_rate": 2.256337159003134e-05,
"loss": 0.0983,
"num_input_tokens_seen": 1960000,
"step": 4775
},
{
"epoch": 5.786924939467312,
"grad_norm": 0.9316068291664124,
"learning_rate": 2.251080383510459e-05,
"loss": 0.0422,
"num_input_tokens_seen": 1962144,
"step": 4780
},
{
"epoch": 5.792978208232445,
"grad_norm": 2.1307058334350586,
"learning_rate": 2.2458247193748155e-05,
"loss": 0.1192,
"num_input_tokens_seen": 1964192,
"step": 4785
},
{
"epoch": 5.799031476997579,
"grad_norm": 4.095328330993652,
"learning_rate": 2.240570190061288e-05,
"loss": 0.1351,
"num_input_tokens_seen": 1966240,
"step": 4790
},
{
"epoch": 5.805084745762712,
"grad_norm": 3.0268330574035645,
"learning_rate": 2.2353168190298915e-05,
"loss": 0.1107,
"num_input_tokens_seen": 1968320,
"step": 4795
},
{
"epoch": 5.811138014527845,
"grad_norm": 0.37829795479774475,
"learning_rate": 2.2300646297354704e-05,
"loss": 0.0993,
"num_input_tokens_seen": 1970304,
"step": 4800
},
{
"epoch": 5.817191283292978,
"grad_norm": 2.7740402221679688,
"learning_rate": 2.224813645627592e-05,
"loss": 0.0937,
"num_input_tokens_seen": 1972320,
"step": 4805
},
{
"epoch": 5.823244552058111,
"grad_norm": 2.5438811779022217,
"learning_rate": 2.2195638901504452e-05,
"loss": 0.0555,
"num_input_tokens_seen": 1974432,
"step": 4810
},
{
"epoch": 5.829297820823244,
"grad_norm": 0.15112103521823883,
"learning_rate": 2.2143153867427305e-05,
"loss": 0.0675,
"num_input_tokens_seen": 1976480,
"step": 4815
},
{
"epoch": 5.835351089588378,
"grad_norm": 4.919175148010254,
"learning_rate": 2.2090681588375594e-05,
"loss": 0.1833,
"num_input_tokens_seen": 1978624,
"step": 4820
},
{
"epoch": 5.841404358353511,
"grad_norm": 0.0790180116891861,
"learning_rate": 2.2038222298623507e-05,
"loss": 0.0329,
"num_input_tokens_seen": 1980736,
"step": 4825
},
{
"epoch": 5.847457627118644,
"grad_norm": 0.06224536895751953,
"learning_rate": 2.1985776232387202e-05,
"loss": 0.0506,
"num_input_tokens_seen": 1982656,
"step": 4830
},
{
"epoch": 5.853510895883777,
"grad_norm": 0.7474290132522583,
"learning_rate": 2.1933343623823814e-05,
"loss": 0.1328,
"num_input_tokens_seen": 1984832,
"step": 4835
},
{
"epoch": 5.85956416464891,
"grad_norm": 4.492368698120117,
"learning_rate": 2.1880924707030407e-05,
"loss": 0.1015,
"num_input_tokens_seen": 1986848,
"step": 4840
},
{
"epoch": 5.865617433414044,
"grad_norm": 1.0196672677993774,
"learning_rate": 2.1828519716042888e-05,
"loss": 0.1972,
"num_input_tokens_seen": 1988992,
"step": 4845
},
{
"epoch": 5.871670702179177,
"grad_norm": 0.20354236662387848,
"learning_rate": 2.177612888483502e-05,
"loss": 0.1323,
"num_input_tokens_seen": 1990976,
"step": 4850
},
{
"epoch": 5.877723970944309,
"grad_norm": 0.2845112681388855,
"learning_rate": 2.1723752447317312e-05,
"loss": 0.0714,
"num_input_tokens_seen": 1992928,
"step": 4855
},
{
"epoch": 5.883777239709443,
"grad_norm": 3.6729049682617188,
"learning_rate": 2.167139063733605e-05,
"loss": 0.2155,
"num_input_tokens_seen": 1995104,
"step": 4860
},
{
"epoch": 5.889830508474576,
"grad_norm": 5.529604434967041,
"learning_rate": 2.161904368867217e-05,
"loss": 0.0664,
"num_input_tokens_seen": 1997312,
"step": 4865
},
{
"epoch": 5.8958837772397095,
"grad_norm": 4.269104957580566,
"learning_rate": 2.1566711835040284e-05,
"loss": 0.0805,
"num_input_tokens_seen": 1999232,
"step": 4870
},
{
"epoch": 5.901937046004843,
"grad_norm": 0.41338273882865906,
"learning_rate": 2.1514395310087596e-05,
"loss": 0.1931,
"num_input_tokens_seen": 2001088,
"step": 4875
},
{
"epoch": 5.907990314769976,
"grad_norm": 0.3836718797683716,
"learning_rate": 2.1462094347392887e-05,
"loss": 0.0544,
"num_input_tokens_seen": 2003232,
"step": 4880
},
{
"epoch": 5.914043583535109,
"grad_norm": 4.488040447235107,
"learning_rate": 2.1409809180465436e-05,
"loss": 0.1114,
"num_input_tokens_seen": 2005312,
"step": 4885
},
{
"epoch": 5.920096852300242,
"grad_norm": 5.718064308166504,
"learning_rate": 2.1357540042744006e-05,
"loss": 0.1378,
"num_input_tokens_seen": 2007392,
"step": 4890
},
{
"epoch": 5.926150121065375,
"grad_norm": 0.7707825899124146,
"learning_rate": 2.1305287167595808e-05,
"loss": 0.1329,
"num_input_tokens_seen": 2009568,
"step": 4895
},
{
"epoch": 5.932203389830509,
"grad_norm": 4.785271644592285,
"learning_rate": 2.1253050788315436e-05,
"loss": 0.0762,
"num_input_tokens_seen": 2011424,
"step": 4900
},
{
"epoch": 5.938256658595642,
"grad_norm": 4.0082783699035645,
"learning_rate": 2.120083113812381e-05,
"loss": 0.0583,
"num_input_tokens_seen": 2013504,
"step": 4905
},
{
"epoch": 5.9443099273607745,
"grad_norm": 7.899497985839844,
"learning_rate": 2.1148628450167203e-05,
"loss": 0.1401,
"num_input_tokens_seen": 2015872,
"step": 4910
},
{
"epoch": 5.950363196125908,
"grad_norm": 3.8147408962249756,
"learning_rate": 2.109644295751612e-05,
"loss": 0.1231,
"num_input_tokens_seen": 2017888,
"step": 4915
},
{
"epoch": 5.956416464891041,
"grad_norm": 4.809561252593994,
"learning_rate": 2.1044274893164316e-05,
"loss": 0.149,
"num_input_tokens_seen": 2019904,
"step": 4920
},
{
"epoch": 5.9624697336561745,
"grad_norm": 1.246935486793518,
"learning_rate": 2.0992124490027727e-05,
"loss": 0.1612,
"num_input_tokens_seen": 2022016,
"step": 4925
},
{
"epoch": 5.968523002421308,
"grad_norm": 3.2453746795654297,
"learning_rate": 2.0939991980943437e-05,
"loss": 0.1673,
"num_input_tokens_seen": 2023936,
"step": 4930
},
{
"epoch": 5.97457627118644,
"grad_norm": 4.5902204513549805,
"learning_rate": 2.088787759866863e-05,
"loss": 0.1408,
"num_input_tokens_seen": 2026080,
"step": 4935
},
{
"epoch": 5.980629539951574,
"grad_norm": 3.014443874359131,
"learning_rate": 2.0835781575879574e-05,
"loss": 0.0181,
"num_input_tokens_seen": 2028064,
"step": 4940
},
{
"epoch": 5.986682808716707,
"grad_norm": 4.182370185852051,
"learning_rate": 2.0783704145170547e-05,
"loss": 0.0733,
"num_input_tokens_seen": 2029920,
"step": 4945
},
{
"epoch": 5.99273607748184,
"grad_norm": 4.959621429443359,
"learning_rate": 2.0731645539052845e-05,
"loss": 0.1935,
"num_input_tokens_seen": 2031776,
"step": 4950
},
{
"epoch": 5.998789346246974,
"grad_norm": 3.4146416187286377,
"learning_rate": 2.067960598995369e-05,
"loss": 0.043,
"num_input_tokens_seen": 2033888,
"step": 4955
},
{
"epoch": 6.0,
"eval_loss": 0.1471845954656601,
"eval_runtime": 4.5703,
"eval_samples_per_second": 80.301,
"eval_steps_per_second": 20.13,
"num_input_tokens_seen": 2033992,
"step": 4956
},
{
"epoch": 6.004842615012106,
"grad_norm": 1.2523798942565918,
"learning_rate": 2.062758573021523e-05,
"loss": 0.1246,
"num_input_tokens_seen": 2035752,
"step": 4960
},
{
"epoch": 6.0108958837772395,
"grad_norm": 1.3420109748840332,
"learning_rate": 2.0575584992093505e-05,
"loss": 0.1767,
"num_input_tokens_seen": 2037896,
"step": 4965
},
{
"epoch": 6.016949152542373,
"grad_norm": 3.164762020111084,
"learning_rate": 2.0523604007757374e-05,
"loss": 0.1195,
"num_input_tokens_seen": 2040008,
"step": 4970
},
{
"epoch": 6.023002421307506,
"grad_norm": 3.989443302154541,
"learning_rate": 2.0471643009287523e-05,
"loss": 0.0898,
"num_input_tokens_seen": 2042056,
"step": 4975
},
{
"epoch": 6.0290556900726395,
"grad_norm": 0.10141967236995697,
"learning_rate": 2.0419702228675395e-05,
"loss": 0.2044,
"num_input_tokens_seen": 2044040,
"step": 4980
},
{
"epoch": 6.035108958837772,
"grad_norm": 0.23426412045955658,
"learning_rate": 2.0367781897822147e-05,
"loss": 0.1074,
"num_input_tokens_seen": 2046056,
"step": 4985
},
{
"epoch": 6.041162227602905,
"grad_norm": 1.5576239824295044,
"learning_rate": 2.031588224853767e-05,
"loss": 0.1107,
"num_input_tokens_seen": 2047944,
"step": 4990
},
{
"epoch": 6.047215496368039,
"grad_norm": 1.8153196573257446,
"learning_rate": 2.0264003512539493e-05,
"loss": 0.0688,
"num_input_tokens_seen": 2050024,
"step": 4995
},
{
"epoch": 6.053268765133172,
"grad_norm": 2.0969276428222656,
"learning_rate": 2.0212145921451787e-05,
"loss": 0.1391,
"num_input_tokens_seen": 2052040,
"step": 5000
},
{
"epoch": 6.059322033898305,
"grad_norm": 0.1364622712135315,
"learning_rate": 2.01603097068043e-05,
"loss": 0.0444,
"num_input_tokens_seen": 2054120,
"step": 5005
},
{
"epoch": 6.065375302663438,
"grad_norm": 3.0799670219421387,
"learning_rate": 2.0108495100031364e-05,
"loss": 0.0513,
"num_input_tokens_seen": 2056296,
"step": 5010
},
{
"epoch": 6.071428571428571,
"grad_norm": 2.7169783115386963,
"learning_rate": 2.0056702332470806e-05,
"loss": 0.0249,
"num_input_tokens_seen": 2058408,
"step": 5015
},
{
"epoch": 6.0774818401937045,
"grad_norm": 5.1847991943359375,
"learning_rate": 2.0004931635362982e-05,
"loss": 0.1118,
"num_input_tokens_seen": 2060456,
"step": 5020
},
{
"epoch": 6.083535108958838,
"grad_norm": 0.45843198895454407,
"learning_rate": 1.995318323984969e-05,
"loss": 0.0542,
"num_input_tokens_seen": 2062408,
"step": 5025
},
{
"epoch": 6.089588377723971,
"grad_norm": 2.184920310974121,
"learning_rate": 1.9901457376973143e-05,
"loss": 0.1114,
"num_input_tokens_seen": 2064392,
"step": 5030
},
{
"epoch": 6.095641646489105,
"grad_norm": 1.2919806241989136,
"learning_rate": 1.9849754277674993e-05,
"loss": 0.0539,
"num_input_tokens_seen": 2066504,
"step": 5035
},
{
"epoch": 6.101694915254237,
"grad_norm": 0.44648459553718567,
"learning_rate": 1.979807417279521e-05,
"loss": 0.0888,
"num_input_tokens_seen": 2068584,
"step": 5040
},
{
"epoch": 6.10774818401937,
"grad_norm": 0.25474312901496887,
"learning_rate": 1.974641729307115e-05,
"loss": 0.0106,
"num_input_tokens_seen": 2070696,
"step": 5045
},
{
"epoch": 6.113801452784504,
"grad_norm": 1.318625807762146,
"learning_rate": 1.9694783869136435e-05,
"loss": 0.1124,
"num_input_tokens_seen": 2072872,
"step": 5050
},
{
"epoch": 6.119854721549637,
"grad_norm": 3.6301968097686768,
"learning_rate": 1.9643174131519986e-05,
"loss": 0.0856,
"num_input_tokens_seen": 2074824,
"step": 5055
},
{
"epoch": 6.12590799031477,
"grad_norm": 2.665332317352295,
"learning_rate": 1.9591588310644967e-05,
"loss": 0.0809,
"num_input_tokens_seen": 2076808,
"step": 5060
},
{
"epoch": 6.131961259079903,
"grad_norm": 1.1560572385787964,
"learning_rate": 1.9540026636827742e-05,
"loss": 0.0592,
"num_input_tokens_seen": 2078856,
"step": 5065
},
{
"epoch": 6.138014527845036,
"grad_norm": 8.742777824401855,
"learning_rate": 1.948848934027689e-05,
"loss": 0.0858,
"num_input_tokens_seen": 2081096,
"step": 5070
},
{
"epoch": 6.1440677966101696,
"grad_norm": 2.756279945373535,
"learning_rate": 1.9436976651092144e-05,
"loss": 0.074,
"num_input_tokens_seen": 2083144,
"step": 5075
},
{
"epoch": 6.150121065375303,
"grad_norm": 0.6714470386505127,
"learning_rate": 1.9385488799263372e-05,
"loss": 0.1086,
"num_input_tokens_seen": 2085096,
"step": 5080
},
{
"epoch": 6.156174334140436,
"grad_norm": 0.19403041899204254,
"learning_rate": 1.9334026014669543e-05,
"loss": 0.0633,
"num_input_tokens_seen": 2087272,
"step": 5085
},
{
"epoch": 6.162227602905569,
"grad_norm": 6.13324499130249,
"learning_rate": 1.9282588527077715e-05,
"loss": 0.2309,
"num_input_tokens_seen": 2089224,
"step": 5090
},
{
"epoch": 6.168280871670702,
"grad_norm": 4.615688323974609,
"learning_rate": 1.9231176566142006e-05,
"loss": 0.1434,
"num_input_tokens_seen": 2091432,
"step": 5095
},
{
"epoch": 6.174334140435835,
"grad_norm": 0.10968305915594101,
"learning_rate": 1.917979036140255e-05,
"loss": 0.0737,
"num_input_tokens_seen": 2093384,
"step": 5100
},
{
"epoch": 6.180387409200969,
"grad_norm": 0.185858353972435,
"learning_rate": 1.9128430142284503e-05,
"loss": 0.1049,
"num_input_tokens_seen": 2095528,
"step": 5105
},
{
"epoch": 6.186440677966102,
"grad_norm": 1.6614528894424438,
"learning_rate": 1.9077096138096992e-05,
"loss": 0.0492,
"num_input_tokens_seen": 2097544,
"step": 5110
},
{
"epoch": 6.1924939467312345,
"grad_norm": 2.876973867416382,
"learning_rate": 1.9025788578032113e-05,
"loss": 0.1631,
"num_input_tokens_seen": 2099592,
"step": 5115
},
{
"epoch": 6.198547215496368,
"grad_norm": 0.698144793510437,
"learning_rate": 1.8974507691163867e-05,
"loss": 0.0246,
"num_input_tokens_seen": 2101448,
"step": 5120
},
{
"epoch": 6.204600484261501,
"grad_norm": 0.8185459971427917,
"learning_rate": 1.892325370644721e-05,
"loss": 0.0946,
"num_input_tokens_seen": 2103528,
"step": 5125
},
{
"epoch": 6.210653753026635,
"grad_norm": 2.75486421585083,
"learning_rate": 1.8872026852716954e-05,
"loss": 0.0369,
"num_input_tokens_seen": 2105640,
"step": 5130
},
{
"epoch": 6.216707021791768,
"grad_norm": 1.2081927061080933,
"learning_rate": 1.8820827358686793e-05,
"loss": 0.0666,
"num_input_tokens_seen": 2107688,
"step": 5135
},
{
"epoch": 6.2227602905569,
"grad_norm": 2.066209316253662,
"learning_rate": 1.8769655452948274e-05,
"loss": 0.1806,
"num_input_tokens_seen": 2109608,
"step": 5140
},
{
"epoch": 6.228813559322034,
"grad_norm": 0.48250189423561096,
"learning_rate": 1.8718511363969733e-05,
"loss": 0.0996,
"num_input_tokens_seen": 2111752,
"step": 5145
},
{
"epoch": 6.234866828087167,
"grad_norm": 2.2784717082977295,
"learning_rate": 1.8667395320095367e-05,
"loss": 0.0652,
"num_input_tokens_seen": 2113928,
"step": 5150
},
{
"epoch": 6.2409200968523,
"grad_norm": 0.3844600319862366,
"learning_rate": 1.8616307549544113e-05,
"loss": 0.1524,
"num_input_tokens_seen": 2115944,
"step": 5155
},
{
"epoch": 6.246973365617434,
"grad_norm": 8.31705093383789,
"learning_rate": 1.85652482804087e-05,
"loss": 0.1151,
"num_input_tokens_seen": 2118120,
"step": 5160
},
{
"epoch": 6.253026634382566,
"grad_norm": 2.3616673946380615,
"learning_rate": 1.85142177406546e-05,
"loss": 0.1494,
"num_input_tokens_seen": 2120136,
"step": 5165
},
{
"epoch": 6.2590799031477,
"grad_norm": 3.0859901905059814,
"learning_rate": 1.8463216158119015e-05,
"loss": 0.1174,
"num_input_tokens_seen": 2122184,
"step": 5170
},
{
"epoch": 6.265133171912833,
"grad_norm": 0.7617553472518921,
"learning_rate": 1.8412243760509867e-05,
"loss": 0.1346,
"num_input_tokens_seen": 2124200,
"step": 5175
},
{
"epoch": 6.271186440677966,
"grad_norm": 6.6624298095703125,
"learning_rate": 1.8361300775404765e-05,
"loss": 0.0971,
"num_input_tokens_seen": 2126280,
"step": 5180
},
{
"epoch": 6.2772397094431,
"grad_norm": 3.2902045249938965,
"learning_rate": 1.8310387430250014e-05,
"loss": 0.0593,
"num_input_tokens_seen": 2128360,
"step": 5185
},
{
"epoch": 6.283292978208232,
"grad_norm": 0.7124584913253784,
"learning_rate": 1.825950395235956e-05,
"loss": 0.1357,
"num_input_tokens_seen": 2130248,
"step": 5190
},
{
"epoch": 6.289346246973365,
"grad_norm": 2.4646220207214355,
"learning_rate": 1.8208650568914033e-05,
"loss": 0.1265,
"num_input_tokens_seen": 2132200,
"step": 5195
},
{
"epoch": 6.295399515738499,
"grad_norm": 4.314565181732178,
"learning_rate": 1.815782750695967e-05,
"loss": 0.1629,
"num_input_tokens_seen": 2134216,
"step": 5200
},
{
"epoch": 6.301452784503632,
"grad_norm": 2.004688262939453,
"learning_rate": 1.810703499340735e-05,
"loss": 0.0434,
"num_input_tokens_seen": 2136392,
"step": 5205
},
{
"epoch": 6.3075060532687655,
"grad_norm": 2.316528558731079,
"learning_rate": 1.8056273255031552e-05,
"loss": 0.1194,
"num_input_tokens_seen": 2138504,
"step": 5210
},
{
"epoch": 6.313559322033898,
"grad_norm": 2.124305248260498,
"learning_rate": 1.8005542518469366e-05,
"loss": 0.0821,
"num_input_tokens_seen": 2140616,
"step": 5215
},
{
"epoch": 6.319612590799031,
"grad_norm": 1.2579299211502075,
"learning_rate": 1.7954843010219446e-05,
"loss": 0.1001,
"num_input_tokens_seen": 2142664,
"step": 5220
},
{
"epoch": 6.325665859564165,
"grad_norm": 4.945680141448975,
"learning_rate": 1.790417495664103e-05,
"loss": 0.0734,
"num_input_tokens_seen": 2144744,
"step": 5225
},
{
"epoch": 6.331719128329298,
"grad_norm": 5.6669206619262695,
"learning_rate": 1.785353858395292e-05,
"loss": 0.3636,
"num_input_tokens_seen": 2146760,
"step": 5230
},
{
"epoch": 6.337772397094431,
"grad_norm": 5.804713726043701,
"learning_rate": 1.7802934118232482e-05,
"loss": 0.0949,
"num_input_tokens_seen": 2148904,
"step": 5235
},
{
"epoch": 6.343825665859564,
"grad_norm": 1.7007763385772705,
"learning_rate": 1.775236178541461e-05,
"loss": 0.0563,
"num_input_tokens_seen": 2151080,
"step": 5240
},
{
"epoch": 6.349878934624697,
"grad_norm": 9.71132755279541,
"learning_rate": 1.7701821811290743e-05,
"loss": 0.1433,
"num_input_tokens_seen": 2153032,
"step": 5245
},
{
"epoch": 6.3559322033898304,
"grad_norm": 4.293256759643555,
"learning_rate": 1.7651314421507843e-05,
"loss": 0.1557,
"num_input_tokens_seen": 2155080,
"step": 5250
},
{
"epoch": 6.361985472154964,
"grad_norm": 5.381481647491455,
"learning_rate": 1.7600839841567395e-05,
"loss": 0.18,
"num_input_tokens_seen": 2157064,
"step": 5255
},
{
"epoch": 6.368038740920097,
"grad_norm": 0.4548739790916443,
"learning_rate": 1.7550398296824395e-05,
"loss": 0.1216,
"num_input_tokens_seen": 2159208,
"step": 5260
},
{
"epoch": 6.37409200968523,
"grad_norm": 6.4153923988342285,
"learning_rate": 1.749999001248635e-05,
"loss": 0.0475,
"num_input_tokens_seen": 2161384,
"step": 5265
},
{
"epoch": 6.380145278450363,
"grad_norm": 2.944775342941284,
"learning_rate": 1.7449615213612264e-05,
"loss": 0.1253,
"num_input_tokens_seen": 2163432,
"step": 5270
},
{
"epoch": 6.386198547215496,
"grad_norm": 1.4213374853134155,
"learning_rate": 1.7399274125111635e-05,
"loss": 0.0373,
"num_input_tokens_seen": 2165640,
"step": 5275
},
{
"epoch": 6.39225181598063,
"grad_norm": 4.871456146240234,
"learning_rate": 1.7348966971743465e-05,
"loss": 0.0991,
"num_input_tokens_seen": 2167656,
"step": 5280
},
{
"epoch": 6.398305084745763,
"grad_norm": 1.1972601413726807,
"learning_rate": 1.729869397811523e-05,
"loss": 0.0719,
"num_input_tokens_seen": 2169672,
"step": 5285
},
{
"epoch": 6.404358353510895,
"grad_norm": 3.45843768119812,
"learning_rate": 1.72484553686819e-05,
"loss": 0.0967,
"num_input_tokens_seen": 2171592,
"step": 5290
},
{
"epoch": 6.410411622276029,
"grad_norm": 2.6961007118225098,
"learning_rate": 1.719825136774494e-05,
"loss": 0.047,
"num_input_tokens_seen": 2173576,
"step": 5295
},
{
"epoch": 6.416464891041162,
"grad_norm": 0.6391634345054626,
"learning_rate": 1.714808219945129e-05,
"loss": 0.0287,
"num_input_tokens_seen": 2175592,
"step": 5300
},
{
"epoch": 6.4225181598062955,
"grad_norm": 0.5544232130050659,
"learning_rate": 1.709794808779234e-05,
"loss": 0.1309,
"num_input_tokens_seen": 2177512,
"step": 5305
},
{
"epoch": 6.428571428571429,
"grad_norm": 5.888014316558838,
"learning_rate": 1.704784925660301e-05,
"loss": 0.0952,
"num_input_tokens_seen": 2179560,
"step": 5310
},
{
"epoch": 6.434624697336561,
"grad_norm": 1.9635401964187622,
"learning_rate": 1.699778592956069e-05,
"loss": 0.1128,
"num_input_tokens_seen": 2181608,
"step": 5315
},
{
"epoch": 6.440677966101695,
"grad_norm": 6.4271240234375,
"learning_rate": 1.6947758330184226e-05,
"loss": 0.0603,
"num_input_tokens_seen": 2183656,
"step": 5320
},
{
"epoch": 6.446731234866828,
"grad_norm": 0.48759642243385315,
"learning_rate": 1.689776668183299e-05,
"loss": 0.0778,
"num_input_tokens_seen": 2185576,
"step": 5325
},
{
"epoch": 6.452784503631961,
"grad_norm": 6.219998836517334,
"learning_rate": 1.6847811207705813e-05,
"loss": 0.0543,
"num_input_tokens_seen": 2187592,
"step": 5330
},
{
"epoch": 6.458837772397095,
"grad_norm": 0.6584652066230774,
"learning_rate": 1.6797892130840036e-05,
"loss": 0.0478,
"num_input_tokens_seen": 2189544,
"step": 5335
},
{
"epoch": 6.464891041162228,
"grad_norm": 0.11968863755464554,
"learning_rate": 1.6748009674110477e-05,
"loss": 0.0868,
"num_input_tokens_seen": 2191496,
"step": 5340
},
{
"epoch": 6.4709443099273605,
"grad_norm": 0.749453604221344,
"learning_rate": 1.669816406022848e-05,
"loss": 0.0866,
"num_input_tokens_seen": 2193416,
"step": 5345
},
{
"epoch": 6.476997578692494,
"grad_norm": 4.262302398681641,
"learning_rate": 1.6648355511740876e-05,
"loss": 0.0848,
"num_input_tokens_seen": 2195432,
"step": 5350
},
{
"epoch": 6.483050847457627,
"grad_norm": 2.831707000732422,
"learning_rate": 1.659858425102902e-05,
"loss": 0.0165,
"num_input_tokens_seen": 2197288,
"step": 5355
},
{
"epoch": 6.4891041162227605,
"grad_norm": 7.086419105529785,
"learning_rate": 1.6548850500307772e-05,
"loss": 0.0853,
"num_input_tokens_seen": 2199368,
"step": 5360
},
{
"epoch": 6.495157384987894,
"grad_norm": 2.674248695373535,
"learning_rate": 1.649915448162455e-05,
"loss": 0.0217,
"num_input_tokens_seen": 2201640,
"step": 5365
},
{
"epoch": 6.5,
"eval_loss": 0.15626245737075806,
"eval_runtime": 4.5725,
"eval_samples_per_second": 80.262,
"eval_steps_per_second": 20.12,
"num_input_tokens_seen": 2203208,
"step": 5369
},
{
"epoch": 6.501210653753026,
"grad_norm": 0.10960067063570023,
"learning_rate": 1.6449496416858284e-05,
"loss": 0.0489,
"num_input_tokens_seen": 2203592,
"step": 5370
},
{
"epoch": 6.50726392251816,
"grad_norm": 0.3896951377391815,
"learning_rate": 1.6399876527718456e-05,
"loss": 0.1401,
"num_input_tokens_seen": 2205640,
"step": 5375
},
{
"epoch": 6.513317191283293,
"grad_norm": 4.460085391998291,
"learning_rate": 1.6350295035744094e-05,
"loss": 0.1424,
"num_input_tokens_seen": 2207752,
"step": 5380
},
{
"epoch": 6.519370460048426,
"grad_norm": 0.08996804803609848,
"learning_rate": 1.6300752162302822e-05,
"loss": 0.0192,
"num_input_tokens_seen": 2209864,
"step": 5385
},
{
"epoch": 6.52542372881356,
"grad_norm": 10.497203826904297,
"learning_rate": 1.625124812858982e-05,
"loss": 0.1851,
"num_input_tokens_seen": 2211944,
"step": 5390
},
{
"epoch": 6.531476997578692,
"grad_norm": 4.926606178283691,
"learning_rate": 1.6201783155626862e-05,
"loss": 0.113,
"num_input_tokens_seen": 2213928,
"step": 5395
},
{
"epoch": 6.5375302663438255,
"grad_norm": 0.7235932946205139,
"learning_rate": 1.615235746426133e-05,
"loss": 0.0963,
"num_input_tokens_seen": 2215944,
"step": 5400
},
{
"epoch": 6.543583535108959,
"grad_norm": 3.8486897945404053,
"learning_rate": 1.6102971275165228e-05,
"loss": 0.0764,
"num_input_tokens_seen": 2217960,
"step": 5405
},
{
"epoch": 6.549636803874092,
"grad_norm": 3.5995168685913086,
"learning_rate": 1.6053624808834188e-05,
"loss": 0.1895,
"num_input_tokens_seen": 2219976,
"step": 5410
},
{
"epoch": 6.5556900726392255,
"grad_norm": 5.618495941162109,
"learning_rate": 1.6004318285586497e-05,
"loss": 0.1285,
"num_input_tokens_seen": 2221992,
"step": 5415
},
{
"epoch": 6.561743341404358,
"grad_norm": 4.392757415771484,
"learning_rate": 1.5955051925562092e-05,
"loss": 0.0823,
"num_input_tokens_seen": 2223976,
"step": 5420
},
{
"epoch": 6.567796610169491,
"grad_norm": 1.3197625875473022,
"learning_rate": 1.590582594872162e-05,
"loss": 0.0533,
"num_input_tokens_seen": 2226088,
"step": 5425
},
{
"epoch": 6.573849878934625,
"grad_norm": 4.43609094619751,
"learning_rate": 1.585664057484539e-05,
"loss": 0.1339,
"num_input_tokens_seen": 2228008,
"step": 5430
},
{
"epoch": 6.579903147699758,
"grad_norm": 0.7666565775871277,
"learning_rate": 1.5807496023532472e-05,
"loss": 0.2157,
"num_input_tokens_seen": 2229928,
"step": 5435
},
{
"epoch": 6.585956416464891,
"grad_norm": 0.46131646633148193,
"learning_rate": 1.5758392514199644e-05,
"loss": 0.0424,
"num_input_tokens_seen": 2231912,
"step": 5440
},
{
"epoch": 6.592009685230024,
"grad_norm": 4.645249843597412,
"learning_rate": 1.5709330266080446e-05,
"loss": 0.0594,
"num_input_tokens_seen": 2233992,
"step": 5445
},
{
"epoch": 6.598062953995157,
"grad_norm": 0.14227555692195892,
"learning_rate": 1.5660309498224225e-05,
"loss": 0.0124,
"num_input_tokens_seen": 2235976,
"step": 5450
},
{
"epoch": 6.6041162227602905,
"grad_norm": 1.6020721197128296,
"learning_rate": 1.5611330429495096e-05,
"loss": 0.0878,
"num_input_tokens_seen": 2238088,
"step": 5455
},
{
"epoch": 6.610169491525424,
"grad_norm": 10.044299125671387,
"learning_rate": 1.556239327857101e-05,
"loss": 0.1516,
"num_input_tokens_seen": 2240136,
"step": 5460
},
{
"epoch": 6.616222760290557,
"grad_norm": 0.4725329875946045,
"learning_rate": 1.551349826394278e-05,
"loss": 0.173,
"num_input_tokens_seen": 2242280,
"step": 5465
},
{
"epoch": 6.622276029055691,
"grad_norm": 5.6281890869140625,
"learning_rate": 1.5464645603913066e-05,
"loss": 0.1576,
"num_input_tokens_seen": 2244360,
"step": 5470
},
{
"epoch": 6.628329297820823,
"grad_norm": 3.5972378253936768,
"learning_rate": 1.5415835516595465e-05,
"loss": 0.0579,
"num_input_tokens_seen": 2246536,
"step": 5475
},
{
"epoch": 6.634382566585956,
"grad_norm": 1.099054217338562,
"learning_rate": 1.5367068219913456e-05,
"loss": 0.1159,
"num_input_tokens_seen": 2248616,
"step": 5480
},
{
"epoch": 6.64043583535109,
"grad_norm": 0.566524863243103,
"learning_rate": 1.5318343931599503e-05,
"loss": 0.0055,
"num_input_tokens_seen": 2250664,
"step": 5485
},
{
"epoch": 6.646489104116223,
"grad_norm": 6.453625202178955,
"learning_rate": 1.5269662869194036e-05,
"loss": 0.0432,
"num_input_tokens_seen": 2252808,
"step": 5490
},
{
"epoch": 6.652542372881356,
"grad_norm": 5.013184070587158,
"learning_rate": 1.5221025250044486e-05,
"loss": 0.0926,
"num_input_tokens_seen": 2254984,
"step": 5495
},
{
"epoch": 6.658595641646489,
"grad_norm": 3.822239875793457,
"learning_rate": 1.517243129130433e-05,
"loss": 0.0233,
"num_input_tokens_seen": 2257064,
"step": 5500
},
{
"epoch": 6.664648910411622,
"grad_norm": 6.671678066253662,
"learning_rate": 1.512388120993212e-05,
"loss": 0.092,
"num_input_tokens_seen": 2259112,
"step": 5505
},
{
"epoch": 6.670702179176756,
"grad_norm": 0.08668724447488785,
"learning_rate": 1.5075375222690496e-05,
"loss": 0.0619,
"num_input_tokens_seen": 2261224,
"step": 5510
},
{
"epoch": 6.676755447941889,
"grad_norm": 3.928685188293457,
"learning_rate": 1.5026913546145232e-05,
"loss": 0.1003,
"num_input_tokens_seen": 2263400,
"step": 5515
},
{
"epoch": 6.682808716707022,
"grad_norm": 0.28091078996658325,
"learning_rate": 1.4978496396664279e-05,
"loss": 0.0754,
"num_input_tokens_seen": 2265416,
"step": 5520
},
{
"epoch": 6.688861985472155,
"grad_norm": 0.40043458342552185,
"learning_rate": 1.4930123990416766e-05,
"loss": 0.122,
"num_input_tokens_seen": 2267560,
"step": 5525
},
{
"epoch": 6.694915254237288,
"grad_norm": 0.0906248465180397,
"learning_rate": 1.4881796543372079e-05,
"loss": 0.0766,
"num_input_tokens_seen": 2269544,
"step": 5530
},
{
"epoch": 6.700968523002421,
"grad_norm": 5.206305027008057,
"learning_rate": 1.4833514271298859e-05,
"loss": 0.0786,
"num_input_tokens_seen": 2271720,
"step": 5535
},
{
"epoch": 6.707021791767555,
"grad_norm": 1.6504778861999512,
"learning_rate": 1.4785277389764046e-05,
"loss": 0.0286,
"num_input_tokens_seen": 2273928,
"step": 5540
},
{
"epoch": 6.713075060532688,
"grad_norm": 7.289839267730713,
"learning_rate": 1.4737086114131943e-05,
"loss": 0.0607,
"num_input_tokens_seen": 2275912,
"step": 5545
},
{
"epoch": 6.719128329297821,
"grad_norm": 3.971864938735962,
"learning_rate": 1.4688940659563225e-05,
"loss": 0.0885,
"num_input_tokens_seen": 2278024,
"step": 5550
},
{
"epoch": 6.725181598062954,
"grad_norm": 0.6193917393684387,
"learning_rate": 1.4640841241013995e-05,
"loss": 0.007,
"num_input_tokens_seen": 2280136,
"step": 5555
},
{
"epoch": 6.731234866828087,
"grad_norm": 1.3996975421905518,
"learning_rate": 1.4592788073234803e-05,
"loss": 0.0323,
"num_input_tokens_seen": 2282248,
"step": 5560
},
{
"epoch": 6.737288135593221,
"grad_norm": 0.15475653111934662,
"learning_rate": 1.4544781370769723e-05,
"loss": 0.0204,
"num_input_tokens_seen": 2284424,
"step": 5565
},
{
"epoch": 6.743341404358354,
"grad_norm": 0.561587929725647,
"learning_rate": 1.4496821347955359e-05,
"loss": 0.144,
"num_input_tokens_seen": 2286344,
"step": 5570
},
{
"epoch": 6.749394673123486,
"grad_norm": 0.6858986020088196,
"learning_rate": 1.444890821891991e-05,
"loss": 0.1286,
"num_input_tokens_seen": 2288456,
"step": 5575
},
{
"epoch": 6.75544794188862,
"grad_norm": 0.9666218161582947,
"learning_rate": 1.4401042197582193e-05,
"loss": 0.0445,
"num_input_tokens_seen": 2290312,
"step": 5580
},
{
"epoch": 6.761501210653753,
"grad_norm": 0.7301199436187744,
"learning_rate": 1.4353223497650731e-05,
"loss": 0.0131,
"num_input_tokens_seen": 2292392,
"step": 5585
},
{
"epoch": 6.767554479418886,
"grad_norm": 3.5780975818634033,
"learning_rate": 1.4305452332622748e-05,
"loss": 0.1459,
"num_input_tokens_seen": 2294472,
"step": 5590
},
{
"epoch": 6.77360774818402,
"grad_norm": 5.3841938972473145,
"learning_rate": 1.4257728915783244e-05,
"loss": 0.0683,
"num_input_tokens_seen": 2296616,
"step": 5595
},
{
"epoch": 6.779661016949152,
"grad_norm": 2.9884018898010254,
"learning_rate": 1.4210053460204023e-05,
"loss": 0.1095,
"num_input_tokens_seen": 2298856,
"step": 5600
},
{
"epoch": 6.785714285714286,
"grad_norm": 5.287160396575928,
"learning_rate": 1.4162426178742788e-05,
"loss": 0.2111,
"num_input_tokens_seen": 2301064,
"step": 5605
},
{
"epoch": 6.791767554479419,
"grad_norm": 5.867886543273926,
"learning_rate": 1.4114847284042132e-05,
"loss": 0.2521,
"num_input_tokens_seen": 2303208,
"step": 5610
},
{
"epoch": 6.797820823244552,
"grad_norm": 6.757856845855713,
"learning_rate": 1.4067316988528617e-05,
"loss": 0.0968,
"num_input_tokens_seen": 2305288,
"step": 5615
},
{
"epoch": 6.803874092009686,
"grad_norm": 3.8400421142578125,
"learning_rate": 1.4019835504411827e-05,
"loss": 0.0542,
"num_input_tokens_seen": 2307304,
"step": 5620
},
{
"epoch": 6.809927360774818,
"grad_norm": 0.3626413643360138,
"learning_rate": 1.3972403043683419e-05,
"loss": 0.0524,
"num_input_tokens_seen": 2309448,
"step": 5625
},
{
"epoch": 6.815980629539951,
"grad_norm": 1.7563663721084595,
"learning_rate": 1.3925019818116164e-05,
"loss": 0.1005,
"num_input_tokens_seen": 2311368,
"step": 5630
},
{
"epoch": 6.822033898305085,
"grad_norm": 7.347732067108154,
"learning_rate": 1.387768603926302e-05,
"loss": 0.2338,
"num_input_tokens_seen": 2313448,
"step": 5635
},
{
"epoch": 6.828087167070218,
"grad_norm": 2.2824957370758057,
"learning_rate": 1.383040191845619e-05,
"loss": 0.176,
"num_input_tokens_seen": 2315464,
"step": 5640
},
{
"epoch": 6.8341404358353515,
"grad_norm": 5.796999454498291,
"learning_rate": 1.378316766680615e-05,
"loss": 0.0912,
"num_input_tokens_seen": 2317480,
"step": 5645
},
{
"epoch": 6.840193704600484,
"grad_norm": 2.5402140617370605,
"learning_rate": 1.373598349520073e-05,
"loss": 0.1809,
"num_input_tokens_seen": 2319432,
"step": 5650
},
{
"epoch": 6.846246973365617,
"grad_norm": 4.791908264160156,
"learning_rate": 1.3688849614304164e-05,
"loss": 0.2282,
"num_input_tokens_seen": 2321512,
"step": 5655
},
{
"epoch": 6.852300242130751,
"grad_norm": 4.897888660430908,
"learning_rate": 1.3641766234556146e-05,
"loss": 0.1905,
"num_input_tokens_seen": 2323560,
"step": 5660
},
{
"epoch": 6.858353510895884,
"grad_norm": 3.5213470458984375,
"learning_rate": 1.3594733566170926e-05,
"loss": 0.0668,
"num_input_tokens_seen": 2325576,
"step": 5665
},
{
"epoch": 6.864406779661017,
"grad_norm": 3.2041327953338623,
"learning_rate": 1.3547751819136309e-05,
"loss": 0.0957,
"num_input_tokens_seen": 2327496,
"step": 5670
},
{
"epoch": 6.87046004842615,
"grad_norm": 2.667874813079834,
"learning_rate": 1.350082120321276e-05,
"loss": 0.0558,
"num_input_tokens_seen": 2329576,
"step": 5675
},
{
"epoch": 6.876513317191283,
"grad_norm": 5.709193706512451,
"learning_rate": 1.3453941927932456e-05,
"loss": 0.2192,
"num_input_tokens_seen": 2331656,
"step": 5680
},
{
"epoch": 6.8825665859564165,
"grad_norm": 5.8245768547058105,
"learning_rate": 1.3407114202598369e-05,
"loss": 0.0549,
"num_input_tokens_seen": 2333800,
"step": 5685
},
{
"epoch": 6.88861985472155,
"grad_norm": 5.625098705291748,
"learning_rate": 1.3360338236283295e-05,
"loss": 0.0654,
"num_input_tokens_seen": 2336008,
"step": 5690
},
{
"epoch": 6.894673123486683,
"grad_norm": 9.947223663330078,
"learning_rate": 1.3313614237828948e-05,
"loss": 0.1191,
"num_input_tokens_seen": 2337928,
"step": 5695
},
{
"epoch": 6.900726392251816,
"grad_norm": 3.8002381324768066,
"learning_rate": 1.3266942415845018e-05,
"loss": 0.189,
"num_input_tokens_seen": 2340008,
"step": 5700
},
{
"epoch": 6.906779661016949,
"grad_norm": 0.2843150198459625,
"learning_rate": 1.3220322978708242e-05,
"loss": 0.2774,
"num_input_tokens_seen": 2342024,
"step": 5705
},
{
"epoch": 6.912832929782082,
"grad_norm": 3.9901978969573975,
"learning_rate": 1.317375613456147e-05,
"loss": 0.0321,
"num_input_tokens_seen": 2343976,
"step": 5710
},
{
"epoch": 6.918886198547216,
"grad_norm": 5.7010416984558105,
"learning_rate": 1.3127242091312752e-05,
"loss": 0.0905,
"num_input_tokens_seen": 2345928,
"step": 5715
},
{
"epoch": 6.924939467312349,
"grad_norm": 9.902931213378906,
"learning_rate": 1.3080781056634373e-05,
"loss": 0.147,
"num_input_tokens_seen": 2347976,
"step": 5720
},
{
"epoch": 6.9309927360774815,
"grad_norm": 5.970206260681152,
"learning_rate": 1.3034373237961983e-05,
"loss": 0.1639,
"num_input_tokens_seen": 2349992,
"step": 5725
},
{
"epoch": 6.937046004842615,
"grad_norm": 9.21063232421875,
"learning_rate": 1.2988018842493604e-05,
"loss": 0.1798,
"num_input_tokens_seen": 2351976,
"step": 5730
},
{
"epoch": 6.943099273607748,
"grad_norm": 5.441684246063232,
"learning_rate": 1.2941718077188758e-05,
"loss": 0.0791,
"num_input_tokens_seen": 2353992,
"step": 5735
},
{
"epoch": 6.9491525423728815,
"grad_norm": 5.263315677642822,
"learning_rate": 1.2895471148767508e-05,
"loss": 0.2302,
"num_input_tokens_seen": 2355976,
"step": 5740
},
{
"epoch": 6.955205811138015,
"grad_norm": 0.33733436465263367,
"learning_rate": 1.2849278263709572e-05,
"loss": 0.1627,
"num_input_tokens_seen": 2358152,
"step": 5745
},
{
"epoch": 6.961259079903147,
"grad_norm": 2.2974305152893066,
"learning_rate": 1.2803139628253364e-05,
"loss": 0.117,
"num_input_tokens_seen": 2360040,
"step": 5750
},
{
"epoch": 6.967312348668281,
"grad_norm": 2.7991981506347656,
"learning_rate": 1.2757055448395092e-05,
"loss": 0.0867,
"num_input_tokens_seen": 2362056,
"step": 5755
},
{
"epoch": 6.973365617433414,
"grad_norm": 5.492253303527832,
"learning_rate": 1.271102592988782e-05,
"loss": 0.1119,
"num_input_tokens_seen": 2364008,
"step": 5760
},
{
"epoch": 6.979418886198547,
"grad_norm": 0.22008445858955383,
"learning_rate": 1.2665051278240602e-05,
"loss": 0.066,
"num_input_tokens_seen": 2366056,
"step": 5765
},
{
"epoch": 6.985472154963681,
"grad_norm": 0.251720666885376,
"learning_rate": 1.2619131698717504e-05,
"loss": 0.1107,
"num_input_tokens_seen": 2368072,
"step": 5770
},
{
"epoch": 6.991525423728813,
"grad_norm": 4.450534820556641,
"learning_rate": 1.2573267396336686e-05,
"loss": 0.1177,
"num_input_tokens_seen": 2370024,
"step": 5775
},
{
"epoch": 6.9975786924939465,
"grad_norm": 0.8103474378585815,
"learning_rate": 1.2527458575869539e-05,
"loss": 0.0205,
"num_input_tokens_seen": 2372008,
"step": 5780
},
{
"epoch": 7.0,
"eval_loss": 0.15040120482444763,
"eval_runtime": 4.5759,
"eval_samples_per_second": 80.204,
"eval_steps_per_second": 20.106,
"num_input_tokens_seen": 2372464,
"step": 5782
},
{
"epoch": 7.00363196125908,
"grad_norm": 4.991612434387207,
"learning_rate": 1.2481705441839756e-05,
"loss": 0.4001,
"num_input_tokens_seen": 2373648,
"step": 5785
},
{
"epoch": 7.009685230024213,
"grad_norm": 0.7985994815826416,
"learning_rate": 1.2436008198522376e-05,
"loss": 0.0437,
"num_input_tokens_seen": 2375696,
"step": 5790
},
{
"epoch": 7.0157384987893465,
"grad_norm": 2.3747811317443848,
"learning_rate": 1.2390367049942916e-05,
"loss": 0.0563,
"num_input_tokens_seen": 2377872,
"step": 5795
},
{
"epoch": 7.021791767554479,
"grad_norm": 6.454826831817627,
"learning_rate": 1.2344782199876431e-05,
"loss": 0.0908,
"num_input_tokens_seen": 2379792,
"step": 5800
},
{
"epoch": 7.027845036319612,
"grad_norm": 0.03636226803064346,
"learning_rate": 1.2299253851846651e-05,
"loss": 0.0884,
"num_input_tokens_seen": 2381744,
"step": 5805
},
{
"epoch": 7.033898305084746,
"grad_norm": 0.8091753721237183,
"learning_rate": 1.2253782209125012e-05,
"loss": 0.1468,
"num_input_tokens_seen": 2383664,
"step": 5810
},
{
"epoch": 7.039951573849879,
"grad_norm": 1.8999519348144531,
"learning_rate": 1.220836747472978e-05,
"loss": 0.0527,
"num_input_tokens_seen": 2385744,
"step": 5815
},
{
"epoch": 7.046004842615012,
"grad_norm": 0.8117446899414062,
"learning_rate": 1.2163009851425137e-05,
"loss": 0.0384,
"num_input_tokens_seen": 2387920,
"step": 5820
},
{
"epoch": 7.052058111380146,
"grad_norm": 0.4487296938896179,
"learning_rate": 1.2117709541720306e-05,
"loss": 0.0764,
"num_input_tokens_seen": 2390000,
"step": 5825
},
{
"epoch": 7.058111380145278,
"grad_norm": 2.3434858322143555,
"learning_rate": 1.2072466747868597e-05,
"loss": 0.1071,
"num_input_tokens_seen": 2392016,
"step": 5830
},
{
"epoch": 7.0641646489104115,
"grad_norm": 7.249058723449707,
"learning_rate": 1.2027281671866531e-05,
"loss": 0.0572,
"num_input_tokens_seen": 2394032,
"step": 5835
},
{
"epoch": 7.070217917675545,
"grad_norm": 0.018748486414551735,
"learning_rate": 1.198215451545293e-05,
"loss": 0.1064,
"num_input_tokens_seen": 2395888,
"step": 5840
},
{
"epoch": 7.076271186440678,
"grad_norm": 4.1508564949035645,
"learning_rate": 1.193708548010804e-05,
"loss": 0.0788,
"num_input_tokens_seen": 2398032,
"step": 5845
},
{
"epoch": 7.0823244552058116,
"grad_norm": 0.01983875036239624,
"learning_rate": 1.1892074767052611e-05,
"loss": 0.0342,
"num_input_tokens_seen": 2400016,
"step": 5850
},
{
"epoch": 7.088377723970944,
"grad_norm": 5.080287456512451,
"learning_rate": 1.1847122577246964e-05,
"loss": 0.0491,
"num_input_tokens_seen": 2402032,
"step": 5855
},
{
"epoch": 7.094430992736077,
"grad_norm": 2.0395522117614746,
"learning_rate": 1.1802229111390157e-05,
"loss": 0.0128,
"num_input_tokens_seen": 2404048,
"step": 5860
},
{
"epoch": 7.100484261501211,
"grad_norm": 5.25778341293335,
"learning_rate": 1.175739456991908e-05,
"loss": 0.1062,
"num_input_tokens_seen": 2406128,
"step": 5865
},
{
"epoch": 7.106537530266344,
"grad_norm": 6.711263656616211,
"learning_rate": 1.1712619153007517e-05,
"loss": 0.1401,
"num_input_tokens_seen": 2408112,
"step": 5870
},
{
"epoch": 7.112590799031477,
"grad_norm": 5.80398416519165,
"learning_rate": 1.166790306056528e-05,
"loss": 0.1134,
"num_input_tokens_seen": 2410160,
"step": 5875
},
{
"epoch": 7.11864406779661,
"grad_norm": 5.206965446472168,
"learning_rate": 1.1623246492237305e-05,
"loss": 0.0945,
"num_input_tokens_seen": 2412112,
"step": 5880
},
{
"epoch": 7.124697336561743,
"grad_norm": 0.28761962056159973,
"learning_rate": 1.1578649647402806e-05,
"loss": 0.0389,
"num_input_tokens_seen": 2414224,
"step": 5885
},
{
"epoch": 7.1307506053268765,
"grad_norm": 2.857328176498413,
"learning_rate": 1.1534112725174306e-05,
"loss": 0.0971,
"num_input_tokens_seen": 2416272,
"step": 5890
},
{
"epoch": 7.13680387409201,
"grad_norm": 2.5085062980651855,
"learning_rate": 1.1489635924396817e-05,
"loss": 0.0905,
"num_input_tokens_seen": 2418224,
"step": 5895
},
{
"epoch": 7.142857142857143,
"grad_norm": 0.11224648356437683,
"learning_rate": 1.1445219443646896e-05,
"loss": 0.0208,
"num_input_tokens_seen": 2420336,
"step": 5900
},
{
"epoch": 7.148910411622276,
"grad_norm": 0.478684663772583,
"learning_rate": 1.1400863481231833e-05,
"loss": 0.0516,
"num_input_tokens_seen": 2422288,
"step": 5905
},
{
"epoch": 7.154963680387409,
"grad_norm": 3.883321523666382,
"learning_rate": 1.1356568235188682e-05,
"loss": 0.0703,
"num_input_tokens_seen": 2424400,
"step": 5910
},
{
"epoch": 7.161016949152542,
"grad_norm": 5.41986083984375,
"learning_rate": 1.1312333903283435e-05,
"loss": 0.1168,
"num_input_tokens_seen": 2426384,
"step": 5915
},
{
"epoch": 7.167070217917676,
"grad_norm": 0.2709377110004425,
"learning_rate": 1.1268160683010096e-05,
"loss": 0.0523,
"num_input_tokens_seen": 2428400,
"step": 5920
},
{
"epoch": 7.173123486682809,
"grad_norm": 5.297135829925537,
"learning_rate": 1.122404877158986e-05,
"loss": 0.1166,
"num_input_tokens_seen": 2430416,
"step": 5925
},
{
"epoch": 7.1791767554479415,
"grad_norm": 0.9408299326896667,
"learning_rate": 1.1179998365970174e-05,
"loss": 0.1321,
"num_input_tokens_seen": 2432464,
"step": 5930
},
{
"epoch": 7.185230024213075,
"grad_norm": 6.506756782531738,
"learning_rate": 1.113600966282386e-05,
"loss": 0.0385,
"num_input_tokens_seen": 2434480,
"step": 5935
},
{
"epoch": 7.191283292978208,
"grad_norm": 0.5841665267944336,
"learning_rate": 1.1092082858548275e-05,
"loss": 0.0832,
"num_input_tokens_seen": 2436496,
"step": 5940
},
{
"epoch": 7.197336561743342,
"grad_norm": 0.013317148201167583,
"learning_rate": 1.1048218149264434e-05,
"loss": 0.0993,
"num_input_tokens_seen": 2438544,
"step": 5945
},
{
"epoch": 7.203389830508475,
"grad_norm": 0.01834205724298954,
"learning_rate": 1.1004415730816083e-05,
"loss": 0.0511,
"num_input_tokens_seen": 2440528,
"step": 5950
},
{
"epoch": 7.209443099273607,
"grad_norm": 7.901388645172119,
"learning_rate": 1.0960675798768871e-05,
"loss": 0.2747,
"num_input_tokens_seen": 2442576,
"step": 5955
},
{
"epoch": 7.215496368038741,
"grad_norm": 6.9650492668151855,
"learning_rate": 1.0916998548409449e-05,
"loss": 0.1191,
"num_input_tokens_seen": 2444560,
"step": 5960
},
{
"epoch": 7.221549636803874,
"grad_norm": 0.1719561368227005,
"learning_rate": 1.0873384174744641e-05,
"loss": 0.0347,
"num_input_tokens_seen": 2446704,
"step": 5965
},
{
"epoch": 7.227602905569007,
"grad_norm": 6.2616095542907715,
"learning_rate": 1.0829832872500523e-05,
"loss": 0.1151,
"num_input_tokens_seen": 2448720,
"step": 5970
},
{
"epoch": 7.233656174334141,
"grad_norm": 5.786017417907715,
"learning_rate": 1.078634483612157e-05,
"loss": 0.1013,
"num_input_tokens_seen": 2450832,
"step": 5975
},
{
"epoch": 7.239709443099273,
"grad_norm": 2.1609995365142822,
"learning_rate": 1.0742920259769792e-05,
"loss": 0.0251,
"num_input_tokens_seen": 2452752,
"step": 5980
},
{
"epoch": 7.245762711864407,
"grad_norm": 3.690556764602661,
"learning_rate": 1.06995593373239e-05,
"loss": 0.1575,
"num_input_tokens_seen": 2454832,
"step": 5985
},
{
"epoch": 7.25181598062954,
"grad_norm": 6.517083644866943,
"learning_rate": 1.0656262262378367e-05,
"loss": 0.0948,
"num_input_tokens_seen": 2456848,
"step": 5990
},
{
"epoch": 7.257869249394673,
"grad_norm": 5.877458095550537,
"learning_rate": 1.0613029228242627e-05,
"loss": 0.1154,
"num_input_tokens_seen": 2459024,
"step": 5995
},
{
"epoch": 7.263922518159807,
"grad_norm": 0.49661943316459656,
"learning_rate": 1.0569860427940179e-05,
"loss": 0.0497,
"num_input_tokens_seen": 2461168,
"step": 6000
},
{
"epoch": 7.269975786924939,
"grad_norm": 4.893890857696533,
"learning_rate": 1.0526756054207737e-05,
"loss": 0.022,
"num_input_tokens_seen": 2463120,
"step": 6005
},
{
"epoch": 7.276029055690072,
"grad_norm": 9.290037155151367,
"learning_rate": 1.0483716299494392e-05,
"loss": 0.1265,
"num_input_tokens_seen": 2465168,
"step": 6010
},
{
"epoch": 7.282082324455206,
"grad_norm": 5.3339314460754395,
"learning_rate": 1.044074135596069e-05,
"loss": 0.165,
"num_input_tokens_seen": 2467248,
"step": 6015
},
{
"epoch": 7.288135593220339,
"grad_norm": 0.101209856569767,
"learning_rate": 1.0397831415477823e-05,
"loss": 0.1604,
"num_input_tokens_seen": 2469200,
"step": 6020
},
{
"epoch": 7.2941888619854724,
"grad_norm": 6.609338283538818,
"learning_rate": 1.0354986669626796e-05,
"loss": 0.0876,
"num_input_tokens_seen": 2471312,
"step": 6025
},
{
"epoch": 7.300242130750606,
"grad_norm": 8.601812362670898,
"learning_rate": 1.0312207309697502e-05,
"loss": 0.0983,
"num_input_tokens_seen": 2473424,
"step": 6030
},
{
"epoch": 7.306295399515738,
"grad_norm": 0.25469648838043213,
"learning_rate": 1.0269493526687915e-05,
"loss": 0.1301,
"num_input_tokens_seen": 2475504,
"step": 6035
},
{
"epoch": 7.312348668280872,
"grad_norm": 1.2198680639266968,
"learning_rate": 1.0226845511303219e-05,
"loss": 0.093,
"num_input_tokens_seen": 2477392,
"step": 6040
},
{
"epoch": 7.318401937046005,
"grad_norm": 6.0976057052612305,
"learning_rate": 1.0184263453954988e-05,
"loss": 0.1598,
"num_input_tokens_seen": 2479504,
"step": 6045
},
{
"epoch": 7.324455205811138,
"grad_norm": 0.11773672699928284,
"learning_rate": 1.0141747544760285e-05,
"loss": 0.1831,
"num_input_tokens_seen": 2481584,
"step": 6050
},
{
"epoch": 7.330508474576272,
"grad_norm": 0.2373896986246109,
"learning_rate": 1.0099297973540852e-05,
"loss": 0.1016,
"num_input_tokens_seen": 2483632,
"step": 6055
},
{
"epoch": 7.336561743341404,
"grad_norm": 0.4794990122318268,
"learning_rate": 1.0056914929822248e-05,
"loss": 0.1583,
"num_input_tokens_seen": 2485680,
"step": 6060
},
{
"epoch": 7.342615012106537,
"grad_norm": 3.694774866104126,
"learning_rate": 1.0014598602832995e-05,
"loss": 0.1414,
"num_input_tokens_seen": 2487824,
"step": 6065
},
{
"epoch": 7.348668280871671,
"grad_norm": 5.654700756072998,
"learning_rate": 9.972349181503773e-06,
"loss": 0.2079,
"num_input_tokens_seen": 2489872,
"step": 6070
},
{
"epoch": 7.354721549636804,
"grad_norm": 2.5101606845855713,
"learning_rate": 9.930166854466516e-06,
"loss": 0.1189,
"num_input_tokens_seen": 2491888,
"step": 6075
},
{
"epoch": 7.3607748184019375,
"grad_norm": 3.4558751583099365,
"learning_rate": 9.888051810053617e-06,
"loss": 0.063,
"num_input_tokens_seen": 2493968,
"step": 6080
},
{
"epoch": 7.36682808716707,
"grad_norm": 1.2189325094223022,
"learning_rate": 9.846004236297052e-06,
"loss": 0.0761,
"num_input_tokens_seen": 2495920,
"step": 6085
},
{
"epoch": 7.372881355932203,
"grad_norm": 2.1722373962402344,
"learning_rate": 9.804024320927604e-06,
"loss": 0.0449,
"num_input_tokens_seen": 2498000,
"step": 6090
},
{
"epoch": 7.378934624697337,
"grad_norm": 6.354626655578613,
"learning_rate": 9.76211225137392e-06,
"loss": 0.0556,
"num_input_tokens_seen": 2499952,
"step": 6095
},
{
"epoch": 7.38498789346247,
"grad_norm": 2.806201934814453,
"learning_rate": 9.720268214761763e-06,
"loss": 0.0527,
"num_input_tokens_seen": 2501968,
"step": 6100
},
{
"epoch": 7.391041162227603,
"grad_norm": 1.7928763628005981,
"learning_rate": 9.678492397913167e-06,
"loss": 0.0513,
"num_input_tokens_seen": 2503984,
"step": 6105
},
{
"epoch": 7.397094430992736,
"grad_norm": 1.5846754312515259,
"learning_rate": 9.636784987345554e-06,
"loss": 0.0094,
"num_input_tokens_seen": 2505968,
"step": 6110
},
{
"epoch": 7.403147699757869,
"grad_norm": 1.3514759540557861,
"learning_rate": 9.595146169270944e-06,
"loss": 0.0665,
"num_input_tokens_seen": 2508080,
"step": 6115
},
{
"epoch": 7.4092009685230025,
"grad_norm": 6.123665809631348,
"learning_rate": 9.553576129595101e-06,
"loss": 0.2015,
"num_input_tokens_seen": 2510160,
"step": 6120
},
{
"epoch": 7.415254237288136,
"grad_norm": 2.2483015060424805,
"learning_rate": 9.512075053916735e-06,
"loss": 0.1365,
"num_input_tokens_seen": 2512304,
"step": 6125
},
{
"epoch": 7.421307506053269,
"grad_norm": 2.598407745361328,
"learning_rate": 9.470643127526627e-06,
"loss": 0.0505,
"num_input_tokens_seen": 2514544,
"step": 6130
},
{
"epoch": 7.427360774818402,
"grad_norm": 0.049765050411224365,
"learning_rate": 9.429280535406834e-06,
"loss": 0.0258,
"num_input_tokens_seen": 2516592,
"step": 6135
},
{
"epoch": 7.433414043583535,
"grad_norm": 4.562868118286133,
"learning_rate": 9.387987462229859e-06,
"loss": 0.1327,
"num_input_tokens_seen": 2518512,
"step": 6140
},
{
"epoch": 7.439467312348668,
"grad_norm": 3.8121907711029053,
"learning_rate": 9.346764092357801e-06,
"loss": 0.0704,
"num_input_tokens_seen": 2520592,
"step": 6145
},
{
"epoch": 7.445520581113802,
"grad_norm": 0.061380598694086075,
"learning_rate": 9.305610609841598e-06,
"loss": 0.1017,
"num_input_tokens_seen": 2522640,
"step": 6150
},
{
"epoch": 7.451573849878935,
"grad_norm": 0.5764665007591248,
"learning_rate": 9.264527198420117e-06,
"loss": 0.0743,
"num_input_tokens_seen": 2524592,
"step": 6155
},
{
"epoch": 7.4576271186440675,
"grad_norm": 7.019524097442627,
"learning_rate": 9.2235140415194e-06,
"loss": 0.1273,
"num_input_tokens_seen": 2526576,
"step": 6160
},
{
"epoch": 7.463680387409201,
"grad_norm": 0.13237552344799042,
"learning_rate": 9.182571322251796e-06,
"loss": 0.0638,
"num_input_tokens_seen": 2528656,
"step": 6165
},
{
"epoch": 7.469733656174334,
"grad_norm": 5.415895462036133,
"learning_rate": 9.141699223415221e-06,
"loss": 0.1563,
"num_input_tokens_seen": 2530864,
"step": 6170
},
{
"epoch": 7.4757869249394675,
"grad_norm": 6.8290557861328125,
"learning_rate": 9.10089792749223e-06,
"loss": 0.0785,
"num_input_tokens_seen": 2532880,
"step": 6175
},
{
"epoch": 7.481840193704601,
"grad_norm": 8.107266426086426,
"learning_rate": 9.06016761664929e-06,
"loss": 0.1175,
"num_input_tokens_seen": 2534864,
"step": 6180
},
{
"epoch": 7.487893462469733,
"grad_norm": 5.507659435272217,
"learning_rate": 9.019508472735958e-06,
"loss": 0.1338,
"num_input_tokens_seen": 2536912,
"step": 6185
},
{
"epoch": 7.493946731234867,
"grad_norm": 4.443814754486084,
"learning_rate": 8.978920677284022e-06,
"loss": 0.0693,
"num_input_tokens_seen": 2538832,
"step": 6190
},
{
"epoch": 7.5,
"grad_norm": 5.766138553619385,
"learning_rate": 8.938404411506732e-06,
"loss": 0.2827,
"num_input_tokens_seen": 2540880,
"step": 6195
},
{
"epoch": 7.5,
"eval_loss": 0.1561979204416275,
"eval_runtime": 4.5842,
"eval_samples_per_second": 80.058,
"eval_steps_per_second": 20.069,
"num_input_tokens_seen": 2540880,
"step": 6195
},
{
"epoch": 7.506053268765133,
"grad_norm": 4.710324287414551,
"learning_rate": 8.897959856297971e-06,
"loss": 0.0657,
"num_input_tokens_seen": 2542832,
"step": 6200
},
{
"epoch": 7.512106537530267,
"grad_norm": 0.08274981379508972,
"learning_rate": 8.857587192231452e-06,
"loss": 0.0517,
"num_input_tokens_seen": 2544784,
"step": 6205
},
{
"epoch": 7.518159806295399,
"grad_norm": 0.5334417223930359,
"learning_rate": 8.817286599559932e-06,
"loss": 0.0684,
"num_input_tokens_seen": 2546832,
"step": 6210
},
{
"epoch": 7.5242130750605325,
"grad_norm": 1.4246139526367188,
"learning_rate": 8.777058258214377e-06,
"loss": 0.0981,
"num_input_tokens_seen": 2549008,
"step": 6215
},
{
"epoch": 7.530266343825666,
"grad_norm": 4.641688346862793,
"learning_rate": 8.736902347803163e-06,
"loss": 0.1203,
"num_input_tokens_seen": 2551120,
"step": 6220
},
{
"epoch": 7.536319612590799,
"grad_norm": 2.8254716396331787,
"learning_rate": 8.696819047611288e-06,
"loss": 0.0808,
"num_input_tokens_seen": 2553168,
"step": 6225
},
{
"epoch": 7.5423728813559325,
"grad_norm": 3.209864377975464,
"learning_rate": 8.65680853659958e-06,
"loss": 0.0453,
"num_input_tokens_seen": 2555184,
"step": 6230
},
{
"epoch": 7.548426150121065,
"grad_norm": 3.2419934272766113,
"learning_rate": 8.616870993403864e-06,
"loss": 0.0221,
"num_input_tokens_seen": 2557264,
"step": 6235
},
{
"epoch": 7.554479418886198,
"grad_norm": 0.510710597038269,
"learning_rate": 8.577006596334191e-06,
"loss": 0.1491,
"num_input_tokens_seen": 2559248,
"step": 6240
},
{
"epoch": 7.560532687651332,
"grad_norm": 2.6407363414764404,
"learning_rate": 8.537215523374038e-06,
"loss": 0.1646,
"num_input_tokens_seen": 2561264,
"step": 6245
},
{
"epoch": 7.566585956416465,
"grad_norm": 5.177677154541016,
"learning_rate": 8.4974979521795e-06,
"loss": 0.1805,
"num_input_tokens_seen": 2563120,
"step": 6250
},
{
"epoch": 7.572639225181598,
"grad_norm": 0.05995741859078407,
"learning_rate": 8.45785406007852e-06,
"loss": 0.2352,
"num_input_tokens_seen": 2565104,
"step": 6255
},
{
"epoch": 7.578692493946731,
"grad_norm": 4.417571544647217,
"learning_rate": 8.418284024070069e-06,
"loss": 0.0918,
"num_input_tokens_seen": 2567216,
"step": 6260
},
{
"epoch": 7.584745762711864,
"grad_norm": 3.200150728225708,
"learning_rate": 8.378788020823394e-06,
"loss": 0.0234,
"num_input_tokens_seen": 2569360,
"step": 6265
},
{
"epoch": 7.5907990314769975,
"grad_norm": 10.74376392364502,
"learning_rate": 8.33936622667719e-06,
"loss": 0.1187,
"num_input_tokens_seen": 2571536,
"step": 6270
},
{
"epoch": 7.596852300242131,
"grad_norm": 0.05067841336131096,
"learning_rate": 8.300018817638825e-06,
"loss": 0.2934,
"num_input_tokens_seen": 2573648,
"step": 6275
},
{
"epoch": 7.602905569007264,
"grad_norm": 0.693376362323761,
"learning_rate": 8.260745969383565e-06,
"loss": 0.0099,
"num_input_tokens_seen": 2575888,
"step": 6280
},
{
"epoch": 7.608958837772397,
"grad_norm": 3.702101230621338,
"learning_rate": 8.221547857253781e-06,
"loss": 0.0783,
"num_input_tokens_seen": 2578032,
"step": 6285
},
{
"epoch": 7.61501210653753,
"grad_norm": 3.041938543319702,
"learning_rate": 8.182424656258178e-06,
"loss": 0.1982,
"num_input_tokens_seen": 2580016,
"step": 6290
},
{
"epoch": 7.621065375302663,
"grad_norm": 0.7098401784896851,
"learning_rate": 8.143376541070993e-06,
"loss": 0.172,
"num_input_tokens_seen": 2582192,
"step": 6295
},
{
"epoch": 7.627118644067797,
"grad_norm": 3.364931106567383,
"learning_rate": 8.104403686031225e-06,
"loss": 0.1141,
"num_input_tokens_seen": 2584208,
"step": 6300
},
{
"epoch": 7.63317191283293,
"grad_norm": 0.41202157735824585,
"learning_rate": 8.06550626514185e-06,
"loss": 0.0881,
"num_input_tokens_seen": 2586160,
"step": 6305
},
{
"epoch": 7.6392251815980625,
"grad_norm": 4.636420726776123,
"learning_rate": 8.026684452069084e-06,
"loss": 0.032,
"num_input_tokens_seen": 2588240,
"step": 6310
},
{
"epoch": 7.645278450363196,
"grad_norm": 3.400038719177246,
"learning_rate": 7.987938420141536e-06,
"loss": 0.0891,
"num_input_tokens_seen": 2590480,
"step": 6315
},
{
"epoch": 7.651331719128329,
"grad_norm": 4.440576553344727,
"learning_rate": 7.949268342349495e-06,
"loss": 0.1652,
"num_input_tokens_seen": 2592400,
"step": 6320
},
{
"epoch": 7.657384987893463,
"grad_norm": 7.222400188446045,
"learning_rate": 7.910674391344129e-06,
"loss": 0.0654,
"num_input_tokens_seen": 2594352,
"step": 6325
},
{
"epoch": 7.663438256658596,
"grad_norm": 7.135024070739746,
"learning_rate": 7.872156739436722e-06,
"loss": 0.1163,
"num_input_tokens_seen": 2596464,
"step": 6330
},
{
"epoch": 7.669491525423728,
"grad_norm": 3.5936965942382812,
"learning_rate": 7.833715558597907e-06,
"loss": 0.1517,
"num_input_tokens_seen": 2598544,
"step": 6335
},
{
"epoch": 7.675544794188862,
"grad_norm": 2.601381540298462,
"learning_rate": 7.795351020456887e-06,
"loss": 0.0728,
"num_input_tokens_seen": 2600656,
"step": 6340
},
{
"epoch": 7.681598062953995,
"grad_norm": 1.4275619983673096,
"learning_rate": 7.757063296300681e-06,
"loss": 0.0455,
"num_input_tokens_seen": 2602832,
"step": 6345
},
{
"epoch": 7.687651331719128,
"grad_norm": 4.785684585571289,
"learning_rate": 7.718852557073366e-06,
"loss": 0.1142,
"num_input_tokens_seen": 2605008,
"step": 6350
},
{
"epoch": 7.693704600484262,
"grad_norm": 1.9919321537017822,
"learning_rate": 7.680718973375287e-06,
"loss": 0.0911,
"num_input_tokens_seen": 2607152,
"step": 6355
},
{
"epoch": 7.699757869249394,
"grad_norm": 4.593861103057861,
"learning_rate": 7.642662715462315e-06,
"loss": 0.0768,
"num_input_tokens_seen": 2609264,
"step": 6360
},
{
"epoch": 7.7058111380145276,
"grad_norm": 1.1676124334335327,
"learning_rate": 7.604683953245076e-06,
"loss": 0.0514,
"num_input_tokens_seen": 2611344,
"step": 6365
},
{
"epoch": 7.711864406779661,
"grad_norm": 9.047789573669434,
"learning_rate": 7.566782856288224e-06,
"loss": 0.1201,
"num_input_tokens_seen": 2613488,
"step": 6370
},
{
"epoch": 7.717917675544794,
"grad_norm": 5.686391830444336,
"learning_rate": 7.5289595938096344e-06,
"loss": 0.1994,
"num_input_tokens_seen": 2615408,
"step": 6375
},
{
"epoch": 7.723970944309928,
"grad_norm": 5.593163967132568,
"learning_rate": 7.4912143346796805e-06,
"loss": 0.1044,
"num_input_tokens_seen": 2617424,
"step": 6380
},
{
"epoch": 7.73002421307506,
"grad_norm": 0.4675053060054779,
"learning_rate": 7.4535472474204645e-06,
"loss": 0.1039,
"num_input_tokens_seen": 2619312,
"step": 6385
},
{
"epoch": 7.736077481840193,
"grad_norm": 4.127689361572266,
"learning_rate": 7.415958500205103e-06,
"loss": 0.1168,
"num_input_tokens_seen": 2621168,
"step": 6390
},
{
"epoch": 7.742130750605327,
"grad_norm": 5.092752933502197,
"learning_rate": 7.37844826085691e-06,
"loss": 0.0661,
"num_input_tokens_seen": 2623152,
"step": 6395
},
{
"epoch": 7.74818401937046,
"grad_norm": 0.2673223614692688,
"learning_rate": 7.341016696848699e-06,
"loss": 0.0332,
"num_input_tokens_seen": 2625328,
"step": 6400
},
{
"epoch": 7.754237288135593,
"grad_norm": 0.20804698765277863,
"learning_rate": 7.303663975302022e-06,
"loss": 0.026,
"num_input_tokens_seen": 2627536,
"step": 6405
},
{
"epoch": 7.760290556900727,
"grad_norm": 0.20616967976093292,
"learning_rate": 7.2663902629864165e-06,
"loss": 0.0101,
"num_input_tokens_seen": 2629616,
"step": 6410
},
{
"epoch": 7.766343825665859,
"grad_norm": 3.6831769943237305,
"learning_rate": 7.229195726318669e-06,
"loss": 0.0538,
"num_input_tokens_seen": 2631696,
"step": 6415
},
{
"epoch": 7.772397094430993,
"grad_norm": 0.7039980292320251,
"learning_rate": 7.192080531362067e-06,
"loss": 0.064,
"num_input_tokens_seen": 2633776,
"step": 6420
},
{
"epoch": 7.778450363196126,
"grad_norm": 3.7811477184295654,
"learning_rate": 7.155044843825651e-06,
"loss": 0.0414,
"num_input_tokens_seen": 2635920,
"step": 6425
},
{
"epoch": 7.784503631961259,
"grad_norm": 6.906268119812012,
"learning_rate": 7.118088829063504e-06,
"loss": 0.1027,
"num_input_tokens_seen": 2637936,
"step": 6430
},
{
"epoch": 7.790556900726393,
"grad_norm": 3.0784387588500977,
"learning_rate": 7.081212652073979e-06,
"loss": 0.0815,
"num_input_tokens_seen": 2639984,
"step": 6435
},
{
"epoch": 7.796610169491525,
"grad_norm": 1.9412931203842163,
"learning_rate": 7.044416477498972e-06,
"loss": 0.1324,
"num_input_tokens_seen": 2642000,
"step": 6440
},
{
"epoch": 7.802663438256658,
"grad_norm": 0.9284259080886841,
"learning_rate": 7.007700469623185e-06,
"loss": 0.0571,
"num_input_tokens_seen": 2644016,
"step": 6445
},
{
"epoch": 7.808716707021792,
"grad_norm": 2.565072774887085,
"learning_rate": 6.971064792373427e-06,
"loss": 0.1923,
"num_input_tokens_seen": 2646000,
"step": 6450
},
{
"epoch": 7.814769975786925,
"grad_norm": 2.1427745819091797,
"learning_rate": 6.934509609317821e-06,
"loss": 0.1555,
"num_input_tokens_seen": 2648048,
"step": 6455
},
{
"epoch": 7.8208232445520585,
"grad_norm": 5.358201026916504,
"learning_rate": 6.898035083665124e-06,
"loss": 0.0466,
"num_input_tokens_seen": 2650064,
"step": 6460
},
{
"epoch": 7.826876513317191,
"grad_norm": 3.4232592582702637,
"learning_rate": 6.861641378263964e-06,
"loss": 0.0734,
"num_input_tokens_seen": 2652016,
"step": 6465
},
{
"epoch": 7.832929782082324,
"grad_norm": 2.332578182220459,
"learning_rate": 6.825328655602153e-06,
"loss": 0.1261,
"num_input_tokens_seen": 2654192,
"step": 6470
},
{
"epoch": 7.838983050847458,
"grad_norm": 0.7169421911239624,
"learning_rate": 6.789097077805917e-06,
"loss": 0.1027,
"num_input_tokens_seen": 2656208,
"step": 6475
},
{
"epoch": 7.845036319612591,
"grad_norm": 5.153744220733643,
"learning_rate": 6.7529468066392015e-06,
"loss": 0.0898,
"num_input_tokens_seen": 2658384,
"step": 6480
},
{
"epoch": 7.851089588377724,
"grad_norm": 0.8772520422935486,
"learning_rate": 6.7168780035029385e-06,
"loss": 0.0203,
"num_input_tokens_seen": 2660528,
"step": 6485
},
{
"epoch": 7.857142857142857,
"grad_norm": 7.855628967285156,
"learning_rate": 6.680890829434325e-06,
"loss": 0.157,
"num_input_tokens_seen": 2662640,
"step": 6490
},
{
"epoch": 7.86319612590799,
"grad_norm": 3.3216187953948975,
"learning_rate": 6.644985445106114e-06,
"loss": 0.0637,
"num_input_tokens_seen": 2664624,
"step": 6495
},
{
"epoch": 7.8692493946731235,
"grad_norm": 0.16674460470676422,
"learning_rate": 6.609162010825881e-06,
"loss": 0.016,
"num_input_tokens_seen": 2666640,
"step": 6500
},
{
"epoch": 7.875302663438257,
"grad_norm": 4.573734283447266,
"learning_rate": 6.573420686535317e-06,
"loss": 0.0683,
"num_input_tokens_seen": 2668592,
"step": 6505
},
{
"epoch": 7.88135593220339,
"grad_norm": 4.325008869171143,
"learning_rate": 6.537761631809533e-06,
"loss": 0.1327,
"num_input_tokens_seen": 2670704,
"step": 6510
},
{
"epoch": 7.8874092009685235,
"grad_norm": 1.69081449508667,
"learning_rate": 6.502185005856312e-06,
"loss": 0.0717,
"num_input_tokens_seen": 2672752,
"step": 6515
},
{
"epoch": 7.893462469733656,
"grad_norm": 1.8181393146514893,
"learning_rate": 6.4666909675154155e-06,
"loss": 0.0724,
"num_input_tokens_seen": 2674864,
"step": 6520
},
{
"epoch": 7.899515738498789,
"grad_norm": 0.46458661556243896,
"learning_rate": 6.431279675257873e-06,
"loss": 0.2251,
"num_input_tokens_seen": 2676944,
"step": 6525
},
{
"epoch": 7.905569007263923,
"grad_norm": 3.5621135234832764,
"learning_rate": 6.395951287185295e-06,
"loss": 0.0576,
"num_input_tokens_seen": 2679024,
"step": 6530
},
{
"epoch": 7.911622276029056,
"grad_norm": 1.915932059288025,
"learning_rate": 6.360705961029126e-06,
"loss": 0.035,
"num_input_tokens_seen": 2680976,
"step": 6535
},
{
"epoch": 7.917675544794189,
"grad_norm": 6.927842140197754,
"learning_rate": 6.325543854149968e-06,
"loss": 0.108,
"num_input_tokens_seen": 2683088,
"step": 6540
},
{
"epoch": 7.923728813559322,
"grad_norm": 0.36779117584228516,
"learning_rate": 6.290465123536876e-06,
"loss": 0.0904,
"num_input_tokens_seen": 2685072,
"step": 6545
},
{
"epoch": 7.929782082324455,
"grad_norm": 0.3823559880256653,
"learning_rate": 6.255469925806643e-06,
"loss": 0.084,
"num_input_tokens_seen": 2687184,
"step": 6550
},
{
"epoch": 7.9358353510895885,
"grad_norm": 1.9843477010726929,
"learning_rate": 6.220558417203132e-06,
"loss": 0.0953,
"num_input_tokens_seen": 2689232,
"step": 6555
},
{
"epoch": 7.941888619854722,
"grad_norm": 0.8949267864227295,
"learning_rate": 6.185730753596539e-06,
"loss": 0.0793,
"num_input_tokens_seen": 2691280,
"step": 6560
},
{
"epoch": 7.947941888619855,
"grad_norm": 0.34748074412345886,
"learning_rate": 6.150987090482715e-06,
"loss": 0.0306,
"num_input_tokens_seen": 2693328,
"step": 6565
},
{
"epoch": 7.953995157384988,
"grad_norm": 2.629479169845581,
"learning_rate": 6.116327582982484e-06,
"loss": 0.1464,
"num_input_tokens_seen": 2695440,
"step": 6570
},
{
"epoch": 7.960048426150121,
"grad_norm": 0.09015600383281708,
"learning_rate": 6.0817523858409245e-06,
"loss": 0.0922,
"num_input_tokens_seen": 2697488,
"step": 6575
},
{
"epoch": 7.966101694915254,
"grad_norm": 4.677306175231934,
"learning_rate": 6.047261653426708e-06,
"loss": 0.1554,
"num_input_tokens_seen": 2699504,
"step": 6580
},
{
"epoch": 7.972154963680388,
"grad_norm": 6.132607460021973,
"learning_rate": 6.012855539731374e-06,
"loss": 0.0634,
"num_input_tokens_seen": 2701456,
"step": 6585
},
{
"epoch": 7.978208232445521,
"grad_norm": 4.510989189147949,
"learning_rate": 5.978534198368691e-06,
"loss": 0.0738,
"num_input_tokens_seen": 2703504,
"step": 6590
},
{
"epoch": 7.9842615012106535,
"grad_norm": 1.4074878692626953,
"learning_rate": 5.944297782573918e-06,
"loss": 0.1079,
"num_input_tokens_seen": 2705616,
"step": 6595
},
{
"epoch": 7.990314769975787,
"grad_norm": 4.238768577575684,
"learning_rate": 5.910146445203154e-06,
"loss": 0.0776,
"num_input_tokens_seen": 2707728,
"step": 6600
},
{
"epoch": 7.99636803874092,
"grad_norm": 0.08241457492113113,
"learning_rate": 5.876080338732643e-06,
"loss": 0.1001,
"num_input_tokens_seen": 2709776,
"step": 6605
},
{
"epoch": 8.0,
"eval_loss": 0.15882021188735962,
"eval_runtime": 4.5817,
"eval_samples_per_second": 80.1,
"eval_steps_per_second": 20.08,
"num_input_tokens_seen": 2710624,
"step": 6608
},
{
"epoch": 8.002421307506053,
"grad_norm": 2.513517141342163,
"learning_rate": 5.842099615258109e-06,
"loss": 0.064,
"num_input_tokens_seen": 2711456,
"step": 6610
},
{
"epoch": 8.008474576271187,
"grad_norm": 0.1411186158657074,
"learning_rate": 5.808204426494054e-06,
"loss": 0.1748,
"num_input_tokens_seen": 2713536,
"step": 6615
},
{
"epoch": 8.01452784503632,
"grad_norm": 0.5461141467094421,
"learning_rate": 5.774394923773088e-06,
"loss": 0.1396,
"num_input_tokens_seen": 2715680,
"step": 6620
},
{
"epoch": 8.020581113801454,
"grad_norm": 2.81977915763855,
"learning_rate": 5.74067125804526e-06,
"loss": 0.0513,
"num_input_tokens_seen": 2717728,
"step": 6625
},
{
"epoch": 8.026634382566586,
"grad_norm": 9.47875690460205,
"learning_rate": 5.70703357987738e-06,
"loss": 0.0797,
"num_input_tokens_seen": 2719712,
"step": 6630
},
{
"epoch": 8.032687651331718,
"grad_norm": 2.103945016860962,
"learning_rate": 5.673482039452363e-06,
"loss": 0.0679,
"num_input_tokens_seen": 2721856,
"step": 6635
},
{
"epoch": 8.038740920096853,
"grad_norm": 10.135988235473633,
"learning_rate": 5.640016786568525e-06,
"loss": 0.1698,
"num_input_tokens_seen": 2724000,
"step": 6640
},
{
"epoch": 8.044794188861985,
"grad_norm": 0.5726349949836731,
"learning_rate": 5.606637970638917e-06,
"loss": 0.1044,
"num_input_tokens_seen": 2725792,
"step": 6645
},
{
"epoch": 8.05084745762712,
"grad_norm": 0.0631587952375412,
"learning_rate": 5.573345740690714e-06,
"loss": 0.1507,
"num_input_tokens_seen": 2727744,
"step": 6650
},
{
"epoch": 8.056900726392252,
"grad_norm": 3.7353861331939697,
"learning_rate": 5.540140245364478e-06,
"loss": 0.0763,
"num_input_tokens_seen": 2729856,
"step": 6655
},
{
"epoch": 8.062953995157384,
"grad_norm": 0.1272682547569275,
"learning_rate": 5.5070216329135365e-06,
"loss": 0.0488,
"num_input_tokens_seen": 2731872,
"step": 6660
},
{
"epoch": 8.069007263922519,
"grad_norm": 0.2341405749320984,
"learning_rate": 5.473990051203298e-06,
"loss": 0.0809,
"num_input_tokens_seen": 2734016,
"step": 6665
},
{
"epoch": 8.075060532687651,
"grad_norm": 1.8176710605621338,
"learning_rate": 5.441045647710627e-06,
"loss": 0.0458,
"num_input_tokens_seen": 2736032,
"step": 6670
},
{
"epoch": 8.081113801452785,
"grad_norm": 7.36388635635376,
"learning_rate": 5.408188569523137e-06,
"loss": 0.1572,
"num_input_tokens_seen": 2737952,
"step": 6675
},
{
"epoch": 8.087167070217918,
"grad_norm": 2.5287160873413086,
"learning_rate": 5.375418963338566e-06,
"loss": 0.0209,
"num_input_tokens_seen": 2739968,
"step": 6680
},
{
"epoch": 8.09322033898305,
"grad_norm": 3.441671133041382,
"learning_rate": 5.342736975464116e-06,
"loss": 0.1182,
"num_input_tokens_seen": 2742016,
"step": 6685
},
{
"epoch": 8.099273607748184,
"grad_norm": 2.918954610824585,
"learning_rate": 5.310142751815792e-06,
"loss": 0.0744,
"num_input_tokens_seen": 2744128,
"step": 6690
},
{
"epoch": 8.105326876513317,
"grad_norm": 0.13784159719944,
"learning_rate": 5.277636437917769e-06,
"loss": 0.0489,
"num_input_tokens_seen": 2746112,
"step": 6695
},
{
"epoch": 8.111380145278451,
"grad_norm": 0.6086945533752441,
"learning_rate": 5.245218178901717e-06,
"loss": 0.0761,
"num_input_tokens_seen": 2748128,
"step": 6700
},
{
"epoch": 8.117433414043584,
"grad_norm": 5.943281650543213,
"learning_rate": 5.212888119506168e-06,
"loss": 0.0651,
"num_input_tokens_seen": 2750176,
"step": 6705
},
{
"epoch": 8.123486682808716,
"grad_norm": 0.10667125880718231,
"learning_rate": 5.180646404075862e-06,
"loss": 0.1376,
"num_input_tokens_seen": 2752256,
"step": 6710
},
{
"epoch": 8.12953995157385,
"grad_norm": 0.2919732630252838,
"learning_rate": 5.1484931765611286e-06,
"loss": 0.037,
"num_input_tokens_seen": 2754208,
"step": 6715
},
{
"epoch": 8.135593220338983,
"grad_norm": 0.1272069811820984,
"learning_rate": 5.116428580517207e-06,
"loss": 0.0543,
"num_input_tokens_seen": 2756352,
"step": 6720
},
{
"epoch": 8.141646489104117,
"grad_norm": 5.189438819885254,
"learning_rate": 5.084452759103603e-06,
"loss": 0.1686,
"num_input_tokens_seen": 2758528,
"step": 6725
},
{
"epoch": 8.14769975786925,
"grad_norm": 1.4413750171661377,
"learning_rate": 5.052565855083511e-06,
"loss": 0.05,
"num_input_tokens_seen": 2760608,
"step": 6730
},
{
"epoch": 8.153753026634382,
"grad_norm": 0.21941789984703064,
"learning_rate": 5.020768010823102e-06,
"loss": 0.0895,
"num_input_tokens_seen": 2762656,
"step": 6735
},
{
"epoch": 8.159806295399516,
"grad_norm": 6.086516857147217,
"learning_rate": 4.98905936829093e-06,
"loss": 0.3365,
"num_input_tokens_seen": 2764672,
"step": 6740
},
{
"epoch": 8.165859564164649,
"grad_norm": 4.337617874145508,
"learning_rate": 4.957440069057281e-06,
"loss": 0.2026,
"num_input_tokens_seen": 2766752,
"step": 6745
},
{
"epoch": 8.171912832929783,
"grad_norm": 0.7985682487487793,
"learning_rate": 4.92591025429357e-06,
"loss": 0.1161,
"num_input_tokens_seen": 2768736,
"step": 6750
},
{
"epoch": 8.177966101694915,
"grad_norm": 0.06866363435983658,
"learning_rate": 4.8944700647716616e-06,
"loss": 0.0773,
"num_input_tokens_seen": 2770752,
"step": 6755
},
{
"epoch": 8.184019370460048,
"grad_norm": 1.1329652070999146,
"learning_rate": 4.863119640863284e-06,
"loss": 0.0797,
"num_input_tokens_seen": 2772928,
"step": 6760
},
{
"epoch": 8.190072639225182,
"grad_norm": 4.368817329406738,
"learning_rate": 4.831859122539381e-06,
"loss": 0.1068,
"num_input_tokens_seen": 2774976,
"step": 6765
},
{
"epoch": 8.196125907990314,
"grad_norm": 4.116227149963379,
"learning_rate": 4.800688649369489e-06,
"loss": 0.0271,
"num_input_tokens_seen": 2777152,
"step": 6770
},
{
"epoch": 8.202179176755449,
"grad_norm": 1.6234118938446045,
"learning_rate": 4.769608360521135e-06,
"loss": 0.0401,
"num_input_tokens_seen": 2779296,
"step": 6775
},
{
"epoch": 8.208232445520581,
"grad_norm": 0.17533978819847107,
"learning_rate": 4.7386183947591815e-06,
"loss": 0.0862,
"num_input_tokens_seen": 2781376,
"step": 6780
},
{
"epoch": 8.214285714285714,
"grad_norm": 6.969298362731934,
"learning_rate": 4.7077188904452255e-06,
"loss": 0.1029,
"num_input_tokens_seen": 2783424,
"step": 6785
},
{
"epoch": 8.220338983050848,
"grad_norm": 4.845013618469238,
"learning_rate": 4.676909985536981e-06,
"loss": 0.0861,
"num_input_tokens_seen": 2785376,
"step": 6790
},
{
"epoch": 8.22639225181598,
"grad_norm": 0.1180926263332367,
"learning_rate": 4.64619181758767e-06,
"loss": 0.0417,
"num_input_tokens_seen": 2787456,
"step": 6795
},
{
"epoch": 8.232445520581114,
"grad_norm": 0.3309163451194763,
"learning_rate": 4.615564523745391e-06,
"loss": 0.0407,
"num_input_tokens_seen": 2789568,
"step": 6800
},
{
"epoch": 8.238498789346247,
"grad_norm": 0.4439374506473541,
"learning_rate": 4.585028240752498e-06,
"loss": 0.0463,
"num_input_tokens_seen": 2791584,
"step": 6805
},
{
"epoch": 8.24455205811138,
"grad_norm": 0.6282607913017273,
"learning_rate": 4.554583104945037e-06,
"loss": 0.0457,
"num_input_tokens_seen": 2793632,
"step": 6810
},
{
"epoch": 8.250605326876514,
"grad_norm": 12.020580291748047,
"learning_rate": 4.524229252252091e-06,
"loss": 0.176,
"num_input_tokens_seen": 2795744,
"step": 6815
},
{
"epoch": 8.256658595641646,
"grad_norm": 0.6101036071777344,
"learning_rate": 4.493966818195191e-06,
"loss": 0.091,
"num_input_tokens_seen": 2797696,
"step": 6820
},
{
"epoch": 8.26271186440678,
"grad_norm": 6.851560592651367,
"learning_rate": 4.463795937887713e-06,
"loss": 0.1245,
"num_input_tokens_seen": 2799744,
"step": 6825
},
{
"epoch": 8.268765133171913,
"grad_norm": 0.21290670335292816,
"learning_rate": 4.433716746034252e-06,
"loss": 0.0675,
"num_input_tokens_seen": 2801824,
"step": 6830
},
{
"epoch": 8.274818401937045,
"grad_norm": 0.04258127510547638,
"learning_rate": 4.40372937693008e-06,
"loss": 0.1002,
"num_input_tokens_seen": 2803872,
"step": 6835
},
{
"epoch": 8.28087167070218,
"grad_norm": 0.23941421508789062,
"learning_rate": 4.3738339644604635e-06,
"loss": 0.1303,
"num_input_tokens_seen": 2805824,
"step": 6840
},
{
"epoch": 8.286924939467312,
"grad_norm": 0.6830506920814514,
"learning_rate": 4.344030642100133e-06,
"loss": 0.0172,
"num_input_tokens_seen": 2807712,
"step": 6845
},
{
"epoch": 8.292978208232446,
"grad_norm": 0.09925970435142517,
"learning_rate": 4.314319542912643e-06,
"loss": 0.0334,
"num_input_tokens_seen": 2809760,
"step": 6850
},
{
"epoch": 8.299031476997579,
"grad_norm": 3.751558542251587,
"learning_rate": 4.284700799549829e-06,
"loss": 0.1723,
"num_input_tokens_seen": 2811872,
"step": 6855
},
{
"epoch": 8.305084745762711,
"grad_norm": 4.737812042236328,
"learning_rate": 4.255174544251147e-06,
"loss": 0.1962,
"num_input_tokens_seen": 2813888,
"step": 6860
},
{
"epoch": 8.311138014527845,
"grad_norm": 3.3046317100524902,
"learning_rate": 4.225740908843146e-06,
"loss": 0.0687,
"num_input_tokens_seen": 2815808,
"step": 6865
},
{
"epoch": 8.317191283292978,
"grad_norm": 0.9242627024650574,
"learning_rate": 4.196400024738831e-06,
"loss": 0.0231,
"num_input_tokens_seen": 2817760,
"step": 6870
},
{
"epoch": 8.323244552058112,
"grad_norm": 4.347622394561768,
"learning_rate": 4.167152022937124e-06,
"loss": 0.0313,
"num_input_tokens_seen": 2819776,
"step": 6875
},
{
"epoch": 8.329297820823244,
"grad_norm": 9.906876564025879,
"learning_rate": 4.137997034022237e-06,
"loss": 0.0923,
"num_input_tokens_seen": 2821824,
"step": 6880
},
{
"epoch": 8.335351089588377,
"grad_norm": 2.6566002368927,
"learning_rate": 4.108935188163096e-06,
"loss": 0.042,
"num_input_tokens_seen": 2823968,
"step": 6885
},
{
"epoch": 8.341404358353511,
"grad_norm": 0.5351431369781494,
"learning_rate": 4.079966615112782e-06,
"loss": 0.0691,
"num_input_tokens_seen": 2825920,
"step": 6890
},
{
"epoch": 8.347457627118644,
"grad_norm": 0.3128598630428314,
"learning_rate": 4.05109144420795e-06,
"loss": 0.0254,
"num_input_tokens_seen": 2828224,
"step": 6895
},
{
"epoch": 8.353510895883778,
"grad_norm": 0.5572378635406494,
"learning_rate": 4.022309804368215e-06,
"loss": 0.0951,
"num_input_tokens_seen": 2830208,
"step": 6900
},
{
"epoch": 8.35956416464891,
"grad_norm": 0.455200731754303,
"learning_rate": 3.993621824095622e-06,
"loss": 0.0888,
"num_input_tokens_seen": 2832160,
"step": 6905
},
{
"epoch": 8.365617433414043,
"grad_norm": 2.736868381500244,
"learning_rate": 3.965027631474036e-06,
"loss": 0.0434,
"num_input_tokens_seen": 2834208,
"step": 6910
},
{
"epoch": 8.371670702179177,
"grad_norm": 5.61325216293335,
"learning_rate": 3.936527354168606e-06,
"loss": 0.0488,
"num_input_tokens_seen": 2836320,
"step": 6915
},
{
"epoch": 8.37772397094431,
"grad_norm": 0.13815198838710785,
"learning_rate": 3.90812111942516e-06,
"loss": 0.0717,
"num_input_tokens_seen": 2838464,
"step": 6920
},
{
"epoch": 8.383777239709444,
"grad_norm": 0.10977087169885635,
"learning_rate": 3.8798090540696495e-06,
"loss": 0.1805,
"num_input_tokens_seen": 2840512,
"step": 6925
},
{
"epoch": 8.389830508474576,
"grad_norm": 1.4197776317596436,
"learning_rate": 3.851591284507591e-06,
"loss": 0.0433,
"num_input_tokens_seen": 2842592,
"step": 6930
},
{
"epoch": 8.39588377723971,
"grad_norm": 4.4109883308410645,
"learning_rate": 3.82346793672351e-06,
"loss": 0.1743,
"num_input_tokens_seen": 2844576,
"step": 6935
},
{
"epoch": 8.401937046004843,
"grad_norm": 0.10218428820371628,
"learning_rate": 3.795439136280346e-06,
"loss": 0.0671,
"num_input_tokens_seen": 2846720,
"step": 6940
},
{
"epoch": 8.407990314769975,
"grad_norm": 2.4533486366271973,
"learning_rate": 3.767505008318914e-06,
"loss": 0.1105,
"num_input_tokens_seen": 2848832,
"step": 6945
},
{
"epoch": 8.41404358353511,
"grad_norm": 0.9111641049385071,
"learning_rate": 3.739665677557341e-06,
"loss": 0.0978,
"num_input_tokens_seen": 2850880,
"step": 6950
},
{
"epoch": 8.420096852300242,
"grad_norm": 0.06459851562976837,
"learning_rate": 3.711921268290533e-06,
"loss": 0.0228,
"num_input_tokens_seen": 2852896,
"step": 6955
},
{
"epoch": 8.426150121065376,
"grad_norm": 0.08616799116134644,
"learning_rate": 3.6842719043895748e-06,
"loss": 0.1082,
"num_input_tokens_seen": 2855008,
"step": 6960
},
{
"epoch": 8.432203389830509,
"grad_norm": 0.536651611328125,
"learning_rate": 3.656717709301194e-06,
"loss": 0.0382,
"num_input_tokens_seen": 2856960,
"step": 6965
},
{
"epoch": 8.438256658595641,
"grad_norm": 5.659543514251709,
"learning_rate": 3.629258806047231e-06,
"loss": 0.1379,
"num_input_tokens_seen": 2859200,
"step": 6970
},
{
"epoch": 8.444309927360775,
"grad_norm": 1.424406886100769,
"learning_rate": 3.60189531722408e-06,
"loss": 0.0356,
"num_input_tokens_seen": 2861280,
"step": 6975
},
{
"epoch": 8.450363196125908,
"grad_norm": 1.0679043531417847,
"learning_rate": 3.5746273650021228e-06,
"loss": 0.013,
"num_input_tokens_seen": 2863392,
"step": 6980
},
{
"epoch": 8.456416464891042,
"grad_norm": 4.228886604309082,
"learning_rate": 3.5474550711252026e-06,
"loss": 0.0767,
"num_input_tokens_seen": 2865344,
"step": 6985
},
{
"epoch": 8.462469733656174,
"grad_norm": 3.4282588958740234,
"learning_rate": 3.5203785569100674e-06,
"loss": 0.1068,
"num_input_tokens_seen": 2867424,
"step": 6990
},
{
"epoch": 8.468523002421307,
"grad_norm": 0.013153821229934692,
"learning_rate": 3.493397943245852e-06,
"loss": 0.0051,
"num_input_tokens_seen": 2869472,
"step": 6995
},
{
"epoch": 8.474576271186441,
"grad_norm": 1.3492847681045532,
"learning_rate": 3.466513350593506e-06,
"loss": 0.083,
"num_input_tokens_seen": 2871680,
"step": 7000
},
{
"epoch": 8.480629539951574,
"grad_norm": 0.17737948894500732,
"learning_rate": 3.439724898985278e-06,
"loss": 0.0866,
"num_input_tokens_seen": 2873760,
"step": 7005
},
{
"epoch": 8.486682808716708,
"grad_norm": 0.177162766456604,
"learning_rate": 3.4130327080241636e-06,
"loss": 0.1555,
"num_input_tokens_seen": 2875712,
"step": 7010
},
{
"epoch": 8.49273607748184,
"grad_norm": 1.8084608316421509,
"learning_rate": 3.3864368968834074e-06,
"loss": 0.0909,
"num_input_tokens_seen": 2877696,
"step": 7015
},
{
"epoch": 8.498789346246973,
"grad_norm": 4.3322858810424805,
"learning_rate": 3.3599375843059193e-06,
"loss": 0.0458,
"num_input_tokens_seen": 2879680,
"step": 7020
},
{
"epoch": 8.5,
"eval_loss": 0.1609281599521637,
"eval_runtime": 4.5734,
"eval_samples_per_second": 80.246,
"eval_steps_per_second": 20.116,
"num_input_tokens_seen": 2880128,
"step": 7021
},
{
"epoch": 8.504842615012107,
"grad_norm": 2.0026533603668213,
"learning_rate": 3.3335348886037815e-06,
"loss": 0.0488,
"num_input_tokens_seen": 2881728,
"step": 7025
},
{
"epoch": 8.51089588377724,
"grad_norm": 0.5009611248970032,
"learning_rate": 3.3072289276576964e-06,
"loss": 0.1506,
"num_input_tokens_seen": 2883744,
"step": 7030
},
{
"epoch": 8.516949152542374,
"grad_norm": 0.33006390929222107,
"learning_rate": 3.281019818916492e-06,
"loss": 0.0674,
"num_input_tokens_seen": 2885792,
"step": 7035
},
{
"epoch": 8.523002421307506,
"grad_norm": 4.603799819946289,
"learning_rate": 3.254907679396574e-06,
"loss": 0.0554,
"num_input_tokens_seen": 2888000,
"step": 7040
},
{
"epoch": 8.529055690072639,
"grad_norm": 7.967130661010742,
"learning_rate": 3.2288926256813846e-06,
"loss": 0.0574,
"num_input_tokens_seen": 2890016,
"step": 7045
},
{
"epoch": 8.535108958837773,
"grad_norm": 6.041900157928467,
"learning_rate": 3.2029747739209247e-06,
"loss": 0.0822,
"num_input_tokens_seen": 2891968,
"step": 7050
},
{
"epoch": 8.541162227602905,
"grad_norm": 2.6833653450012207,
"learning_rate": 3.177154239831223e-06,
"loss": 0.0605,
"num_input_tokens_seen": 2893952,
"step": 7055
},
{
"epoch": 8.54721549636804,
"grad_norm": 6.842079162597656,
"learning_rate": 3.1514311386937917e-06,
"loss": 0.167,
"num_input_tokens_seen": 2895904,
"step": 7060
},
{
"epoch": 8.553268765133172,
"grad_norm": 0.07613459974527359,
"learning_rate": 3.1258055853551487e-06,
"loss": 0.0747,
"num_input_tokens_seen": 2897952,
"step": 7065
},
{
"epoch": 8.559322033898304,
"grad_norm": 0.6995208263397217,
"learning_rate": 3.1002776942262696e-06,
"loss": 0.0433,
"num_input_tokens_seen": 2900064,
"step": 7070
},
{
"epoch": 8.565375302663439,
"grad_norm": 4.104430675506592,
"learning_rate": 3.0748475792821197e-06,
"loss": 0.1335,
"num_input_tokens_seen": 2902432,
"step": 7075
},
{
"epoch": 8.571428571428571,
"grad_norm": 5.85823917388916,
"learning_rate": 3.0495153540611e-06,
"loss": 0.0937,
"num_input_tokens_seen": 2904512,
"step": 7080
},
{
"epoch": 8.577481840193705,
"grad_norm": 1.4460917711257935,
"learning_rate": 3.024281131664569e-06,
"loss": 0.0436,
"num_input_tokens_seen": 2906496,
"step": 7085
},
{
"epoch": 8.583535108958838,
"grad_norm": 0.2464975118637085,
"learning_rate": 2.999145024756325e-06,
"loss": 0.1751,
"num_input_tokens_seen": 2908544,
"step": 7090
},
{
"epoch": 8.58958837772397,
"grad_norm": 6.364808082580566,
"learning_rate": 2.9741071455621245e-06,
"loss": 0.0699,
"num_input_tokens_seen": 2910688,
"step": 7095
},
{
"epoch": 8.595641646489105,
"grad_norm": 0.0688992291688919,
"learning_rate": 2.9491676058691437e-06,
"loss": 0.1593,
"num_input_tokens_seen": 2912832,
"step": 7100
},
{
"epoch": 8.601694915254237,
"grad_norm": 0.6298401355743408,
"learning_rate": 2.924326517025508e-06,
"loss": 0.1751,
"num_input_tokens_seen": 2914816,
"step": 7105
},
{
"epoch": 8.607748184019371,
"grad_norm": 2.046719551086426,
"learning_rate": 2.8995839899397915e-06,
"loss": 0.0117,
"num_input_tokens_seen": 2916928,
"step": 7110
},
{
"epoch": 8.613801452784504,
"grad_norm": 6.791380405426025,
"learning_rate": 2.8749401350805115e-06,
"loss": 0.0582,
"num_input_tokens_seen": 2918912,
"step": 7115
},
{
"epoch": 8.619854721549636,
"grad_norm": 7.068711280822754,
"learning_rate": 2.8503950624756415e-06,
"loss": 0.1209,
"num_input_tokens_seen": 2920896,
"step": 7120
},
{
"epoch": 8.62590799031477,
"grad_norm": 0.11452826112508774,
"learning_rate": 2.825948881712123e-06,
"loss": 0.0824,
"num_input_tokens_seen": 2922976,
"step": 7125
},
{
"epoch": 8.631961259079903,
"grad_norm": 4.309703350067139,
"learning_rate": 2.801601701935369e-06,
"loss": 0.0265,
"num_input_tokens_seen": 2925120,
"step": 7130
},
{
"epoch": 8.638014527845037,
"grad_norm": 5.783072471618652,
"learning_rate": 2.777353631848789e-06,
"loss": 0.2264,
"num_input_tokens_seen": 2927072,
"step": 7135
},
{
"epoch": 8.64406779661017,
"grad_norm": 6.108299255371094,
"learning_rate": 2.7532047797132867e-06,
"loss": 0.1097,
"num_input_tokens_seen": 2929152,
"step": 7140
},
{
"epoch": 8.650121065375302,
"grad_norm": 1.342014193534851,
"learning_rate": 2.7291552533467853e-06,
"loss": 0.0541,
"num_input_tokens_seen": 2931136,
"step": 7145
},
{
"epoch": 8.656174334140436,
"grad_norm": 0.19719941914081573,
"learning_rate": 2.7052051601237473e-06,
"loss": 0.0733,
"num_input_tokens_seen": 2933312,
"step": 7150
},
{
"epoch": 8.662227602905569,
"grad_norm": 5.709043979644775,
"learning_rate": 2.681354606974698e-06,
"loss": 0.1117,
"num_input_tokens_seen": 2935328,
"step": 7155
},
{
"epoch": 8.668280871670703,
"grad_norm": 7.4091901779174805,
"learning_rate": 2.6576037003857414e-06,
"loss": 0.1103,
"num_input_tokens_seen": 2937216,
"step": 7160
},
{
"epoch": 8.674334140435835,
"grad_norm": 4.7495527267456055,
"learning_rate": 2.633952546398083e-06,
"loss": 0.0773,
"num_input_tokens_seen": 2939168,
"step": 7165
},
{
"epoch": 8.680387409200968,
"grad_norm": 6.540948867797852,
"learning_rate": 2.6104012506075692e-06,
"loss": 0.0429,
"num_input_tokens_seen": 2941504,
"step": 7170
},
{
"epoch": 8.686440677966102,
"grad_norm": 5.340511322021484,
"learning_rate": 2.5869499181641916e-06,
"loss": 0.1727,
"num_input_tokens_seen": 2943520,
"step": 7175
},
{
"epoch": 8.692493946731235,
"grad_norm": 9.046024322509766,
"learning_rate": 2.5635986537716538e-06,
"loss": 0.1482,
"num_input_tokens_seen": 2945568,
"step": 7180
},
{
"epoch": 8.698547215496369,
"grad_norm": 6.008487701416016,
"learning_rate": 2.540347561686873e-06,
"loss": 0.0563,
"num_input_tokens_seen": 2947488,
"step": 7185
},
{
"epoch": 8.704600484261501,
"grad_norm": 1.535462498664856,
"learning_rate": 2.5171967457195216e-06,
"loss": 0.2149,
"num_input_tokens_seen": 2949504,
"step": 7190
},
{
"epoch": 8.710653753026634,
"grad_norm": 1.0134094953536987,
"learning_rate": 2.494146309231571e-06,
"loss": 0.1471,
"num_input_tokens_seen": 2951552,
"step": 7195
},
{
"epoch": 8.716707021791768,
"grad_norm": 8.139573097229004,
"learning_rate": 2.471196355136826e-06,
"loss": 0.1541,
"num_input_tokens_seen": 2953632,
"step": 7200
},
{
"epoch": 8.7227602905569,
"grad_norm": 0.31339648365974426,
"learning_rate": 2.4483469859004625e-06,
"loss": 0.1033,
"num_input_tokens_seen": 2955936,
"step": 7205
},
{
"epoch": 8.728813559322035,
"grad_norm": 2.6392133235931396,
"learning_rate": 2.425598303538576e-06,
"loss": 0.0152,
"num_input_tokens_seen": 2958048,
"step": 7210
},
{
"epoch": 8.734866828087167,
"grad_norm": 0.007008890621364117,
"learning_rate": 2.402950409617727e-06,
"loss": 0.0336,
"num_input_tokens_seen": 2960160,
"step": 7215
},
{
"epoch": 8.7409200968523,
"grad_norm": 0.641621470451355,
"learning_rate": 2.380403405254475e-06,
"loss": 0.1413,
"num_input_tokens_seen": 2962208,
"step": 7220
},
{
"epoch": 8.746973365617434,
"grad_norm": 0.4098236858844757,
"learning_rate": 2.35795739111494e-06,
"loss": 0.0205,
"num_input_tokens_seen": 2964320,
"step": 7225
},
{
"epoch": 8.753026634382566,
"grad_norm": 1.6772291660308838,
"learning_rate": 2.335612467414344e-06,
"loss": 0.1145,
"num_input_tokens_seen": 2966272,
"step": 7230
},
{
"epoch": 8.7590799031477,
"grad_norm": 1.6794875860214233,
"learning_rate": 2.313368733916585e-06,
"loss": 0.0659,
"num_input_tokens_seen": 2968288,
"step": 7235
},
{
"epoch": 8.765133171912833,
"grad_norm": 6.174964427947998,
"learning_rate": 2.291226289933751e-06,
"loss": 0.1205,
"num_input_tokens_seen": 2970208,
"step": 7240
},
{
"epoch": 8.771186440677965,
"grad_norm": 0.636359691619873,
"learning_rate": 2.2691852343257157e-06,
"loss": 0.0074,
"num_input_tokens_seen": 2972352,
"step": 7245
},
{
"epoch": 8.7772397094431,
"grad_norm": 1.1045042276382446,
"learning_rate": 2.2472456654996755e-06,
"loss": 0.1758,
"num_input_tokens_seen": 2974368,
"step": 7250
},
{
"epoch": 8.783292978208232,
"grad_norm": 6.679035186767578,
"learning_rate": 2.2254076814097163e-06,
"loss": 0.1391,
"num_input_tokens_seen": 2976288,
"step": 7255
},
{
"epoch": 8.789346246973366,
"grad_norm": 6.327536582946777,
"learning_rate": 2.203671379556388e-06,
"loss": 0.2639,
"num_input_tokens_seen": 2978240,
"step": 7260
},
{
"epoch": 8.795399515738499,
"grad_norm": 6.892092704772949,
"learning_rate": 2.1820368569862444e-06,
"loss": 0.1743,
"num_input_tokens_seen": 2980256,
"step": 7265
},
{
"epoch": 8.801452784503631,
"grad_norm": 4.450016975402832,
"learning_rate": 2.1605042102914227e-06,
"loss": 0.1708,
"num_input_tokens_seen": 2982400,
"step": 7270
},
{
"epoch": 8.807506053268765,
"grad_norm": 7.299108028411865,
"learning_rate": 2.1390735356092206e-06,
"loss": 0.168,
"num_input_tokens_seen": 2984416,
"step": 7275
},
{
"epoch": 8.813559322033898,
"grad_norm": 0.5720050930976868,
"learning_rate": 2.1177449286216565e-06,
"loss": 0.1734,
"num_input_tokens_seen": 2986496,
"step": 7280
},
{
"epoch": 8.819612590799032,
"grad_norm": 2.3085198402404785,
"learning_rate": 2.0965184845550407e-06,
"loss": 0.0285,
"num_input_tokens_seen": 2988512,
"step": 7285
},
{
"epoch": 8.825665859564165,
"grad_norm": 2.5211293697357178,
"learning_rate": 2.075394298179553e-06,
"loss": 0.0939,
"num_input_tokens_seen": 2990560,
"step": 7290
},
{
"epoch": 8.831719128329297,
"grad_norm": 11.825849533081055,
"learning_rate": 2.0543724638088347e-06,
"loss": 0.1561,
"num_input_tokens_seen": 2992768,
"step": 7295
},
{
"epoch": 8.837772397094431,
"grad_norm": 4.789031982421875,
"learning_rate": 2.0334530752995433e-06,
"loss": 0.0939,
"num_input_tokens_seen": 2994784,
"step": 7300
},
{
"epoch": 8.843825665859564,
"grad_norm": 0.1911478191614151,
"learning_rate": 2.01263622605094e-06,
"loss": 0.1379,
"num_input_tokens_seen": 2996896,
"step": 7305
},
{
"epoch": 8.849878934624698,
"grad_norm": 1.790691614151001,
"learning_rate": 1.991922009004485e-06,
"loss": 0.1036,
"num_input_tokens_seen": 2998976,
"step": 7310
},
{
"epoch": 8.85593220338983,
"grad_norm": 0.69582200050354,
"learning_rate": 1.9713105166434042e-06,
"loss": 0.1181,
"num_input_tokens_seen": 3001184,
"step": 7315
},
{
"epoch": 8.861985472154963,
"grad_norm": 5.200904846191406,
"learning_rate": 1.950801840992303e-06,
"loss": 0.1421,
"num_input_tokens_seen": 3003168,
"step": 7320
},
{
"epoch": 8.868038740920097,
"grad_norm": 0.38341599702835083,
"learning_rate": 1.930396073616725e-06,
"loss": 0.1463,
"num_input_tokens_seen": 3005152,
"step": 7325
},
{
"epoch": 8.87409200968523,
"grad_norm": 0.5471208691596985,
"learning_rate": 1.9100933056227593e-06,
"loss": 0.0775,
"num_input_tokens_seen": 3007200,
"step": 7330
},
{
"epoch": 8.880145278450364,
"grad_norm": 0.09188104420900345,
"learning_rate": 1.8898936276566303e-06,
"loss": 0.0454,
"num_input_tokens_seen": 3009280,
"step": 7335
},
{
"epoch": 8.886198547215496,
"grad_norm": 5.813639163970947,
"learning_rate": 1.8697971299043048e-06,
"loss": 0.1039,
"num_input_tokens_seen": 3011360,
"step": 7340
},
{
"epoch": 8.892251815980629,
"grad_norm": 1.5798746347427368,
"learning_rate": 1.8498039020910628e-06,
"loss": 0.0886,
"num_input_tokens_seen": 3013568,
"step": 7345
},
{
"epoch": 8.898305084745763,
"grad_norm": 3.015583038330078,
"learning_rate": 1.8299140334811226e-06,
"loss": 0.0553,
"num_input_tokens_seen": 3015552,
"step": 7350
},
{
"epoch": 8.904358353510895,
"grad_norm": 8.158613204956055,
"learning_rate": 1.8101276128772272e-06,
"loss": 0.1808,
"num_input_tokens_seen": 3017536,
"step": 7355
},
{
"epoch": 8.91041162227603,
"grad_norm": 1.8312183618545532,
"learning_rate": 1.7904447286202607e-06,
"loss": 0.1005,
"num_input_tokens_seen": 3019584,
"step": 7360
},
{
"epoch": 8.916464891041162,
"grad_norm": 5.6660919189453125,
"learning_rate": 1.7708654685888337e-06,
"loss": 0.0898,
"num_input_tokens_seen": 3021728,
"step": 7365
},
{
"epoch": 8.922518159806295,
"grad_norm": 0.28093379735946655,
"learning_rate": 1.7513899201989148e-06,
"loss": 0.0113,
"num_input_tokens_seen": 3023584,
"step": 7370
},
{
"epoch": 8.928571428571429,
"grad_norm": 8.074005126953125,
"learning_rate": 1.7320181704034237e-06,
"loss": 0.1523,
"num_input_tokens_seen": 3025600,
"step": 7375
},
{
"epoch": 8.934624697336561,
"grad_norm": 2.8498823642730713,
"learning_rate": 1.7127503056918542e-06,
"loss": 0.0602,
"num_input_tokens_seen": 3027680,
"step": 7380
},
{
"epoch": 8.940677966101696,
"grad_norm": 0.15879835188388824,
"learning_rate": 1.6935864120898704e-06,
"loss": 0.0356,
"num_input_tokens_seen": 3029856,
"step": 7385
},
{
"epoch": 8.946731234866828,
"grad_norm": 7.197122097015381,
"learning_rate": 1.674526575158944e-06,
"loss": 0.1736,
"num_input_tokens_seen": 3032096,
"step": 7390
},
{
"epoch": 8.95278450363196,
"grad_norm": 0.51031494140625,
"learning_rate": 1.6555708799959547e-06,
"loss": 0.0252,
"num_input_tokens_seen": 3034112,
"step": 7395
},
{
"epoch": 8.958837772397095,
"grad_norm": 0.7516043186187744,
"learning_rate": 1.6367194112328288e-06,
"loss": 0.0302,
"num_input_tokens_seen": 3036096,
"step": 7400
},
{
"epoch": 8.964891041162227,
"grad_norm": 0.05814617499709129,
"learning_rate": 1.617972253036143e-06,
"loss": 0.099,
"num_input_tokens_seen": 3038080,
"step": 7405
},
{
"epoch": 8.970944309927361,
"grad_norm": 0.5936756730079651,
"learning_rate": 1.5993294891067573e-06,
"loss": 0.0268,
"num_input_tokens_seen": 3040064,
"step": 7410
},
{
"epoch": 8.976997578692494,
"grad_norm": 1.6873228549957275,
"learning_rate": 1.580791202679438e-06,
"loss": 0.1383,
"num_input_tokens_seen": 3042048,
"step": 7415
},
{
"epoch": 8.983050847457626,
"grad_norm": 7.362498760223389,
"learning_rate": 1.562357476522497e-06,
"loss": 0.0392,
"num_input_tokens_seen": 3044192,
"step": 7420
},
{
"epoch": 8.98910411622276,
"grad_norm": 1.2069581747055054,
"learning_rate": 1.5440283929374023e-06,
"loss": 0.0879,
"num_input_tokens_seen": 3046112,
"step": 7425
},
{
"epoch": 8.995157384987893,
"grad_norm": 1.0542585849761963,
"learning_rate": 1.5258040337584322e-06,
"loss": 0.1856,
"num_input_tokens_seen": 3048000,
"step": 7430
},
{
"epoch": 9.0,
"eval_loss": 0.16094733774662018,
"eval_runtime": 4.5518,
"eval_samples_per_second": 80.628,
"eval_steps_per_second": 20.212,
"num_input_tokens_seen": 3049392,
"step": 7434
},
{
"epoch": 9.001210653753027,
"grad_norm": 1.983082890510559,
"learning_rate": 1.5076844803522922e-06,
"loss": 0.0777,
"num_input_tokens_seen": 3049808,
"step": 7435
},
{
"epoch": 9.00726392251816,
"grad_norm": 5.638378620147705,
"learning_rate": 1.4896698136177612e-06,
"loss": 0.1153,
"num_input_tokens_seen": 3051792,
"step": 7440
},
{
"epoch": 9.013317191283292,
"grad_norm": 9.8755521774292,
"learning_rate": 1.4717601139853266e-06,
"loss": 0.0989,
"num_input_tokens_seen": 3053776,
"step": 7445
},
{
"epoch": 9.019370460048426,
"grad_norm": 0.596442461013794,
"learning_rate": 1.4539554614168339e-06,
"loss": 0.0655,
"num_input_tokens_seen": 3055856,
"step": 7450
},
{
"epoch": 9.025423728813559,
"grad_norm": 2.234306573867798,
"learning_rate": 1.4362559354051092e-06,
"loss": 0.0259,
"num_input_tokens_seen": 3057840,
"step": 7455
},
{
"epoch": 9.031476997578693,
"grad_norm": 2.3881747722625732,
"learning_rate": 1.4186616149736349e-06,
"loss": 0.2286,
"num_input_tokens_seen": 3059920,
"step": 7460
},
{
"epoch": 9.037530266343826,
"grad_norm": 1.5024694204330444,
"learning_rate": 1.401172578676166e-06,
"loss": 0.1341,
"num_input_tokens_seen": 3061872,
"step": 7465
},
{
"epoch": 9.043583535108958,
"grad_norm": 0.1409914493560791,
"learning_rate": 1.383788904596403e-06,
"loss": 0.0601,
"num_input_tokens_seen": 3063888,
"step": 7470
},
{
"epoch": 9.049636803874092,
"grad_norm": 5.075223922729492,
"learning_rate": 1.3665106703476178e-06,
"loss": 0.1218,
"num_input_tokens_seen": 3065872,
"step": 7475
},
{
"epoch": 9.055690072639225,
"grad_norm": 4.891570568084717,
"learning_rate": 1.349337953072341e-06,
"loss": 0.1693,
"num_input_tokens_seen": 3067888,
"step": 7480
},
{
"epoch": 9.061743341404359,
"grad_norm": 1.4571635723114014,
"learning_rate": 1.3322708294419923e-06,
"loss": 0.0342,
"num_input_tokens_seen": 3069968,
"step": 7485
},
{
"epoch": 9.067796610169491,
"grad_norm": 0.18257658183574677,
"learning_rate": 1.3153093756565426e-06,
"loss": 0.0094,
"num_input_tokens_seen": 3072176,
"step": 7490
},
{
"epoch": 9.073849878934624,
"grad_norm": 0.030617251992225647,
"learning_rate": 1.298453667444169e-06,
"loss": 0.0444,
"num_input_tokens_seen": 3074288,
"step": 7495
},
{
"epoch": 9.079903147699758,
"grad_norm": 4.042976379394531,
"learning_rate": 1.281703780060947e-06,
"loss": 0.0983,
"num_input_tokens_seen": 3076304,
"step": 7500
},
{
"epoch": 9.08595641646489,
"grad_norm": 0.46019235253334045,
"learning_rate": 1.265059788290468e-06,
"loss": 0.0984,
"num_input_tokens_seen": 3078320,
"step": 7505
},
{
"epoch": 9.092009685230025,
"grad_norm": 0.3823973536491394,
"learning_rate": 1.2485217664435418e-06,
"loss": 0.0429,
"num_input_tokens_seen": 3080464,
"step": 7510
},
{
"epoch": 9.098062953995157,
"grad_norm": 1.2773817777633667,
"learning_rate": 1.232089788357843e-06,
"loss": 0.0618,
"num_input_tokens_seen": 3082672,
"step": 7515
},
{
"epoch": 9.104116222760291,
"grad_norm": 6.81667423248291,
"learning_rate": 1.2157639273975979e-06,
"loss": 0.2343,
"num_input_tokens_seen": 3084720,
"step": 7520
},
{
"epoch": 9.110169491525424,
"grad_norm": 3.501394748687744,
"learning_rate": 1.19954425645325e-06,
"loss": 0.1071,
"num_input_tokens_seen": 3086864,
"step": 7525
},
{
"epoch": 9.116222760290556,
"grad_norm": 4.856332778930664,
"learning_rate": 1.183430847941125e-06,
"loss": 0.1255,
"num_input_tokens_seen": 3088880,
"step": 7530
},
{
"epoch": 9.12227602905569,
"grad_norm": 6.885472774505615,
"learning_rate": 1.1674237738031223e-06,
"loss": 0.1039,
"num_input_tokens_seen": 3090960,
"step": 7535
},
{
"epoch": 9.128329297820823,
"grad_norm": 0.33454495668411255,
"learning_rate": 1.1515231055063914e-06,
"loss": 0.1404,
"num_input_tokens_seen": 3093040,
"step": 7540
},
{
"epoch": 9.134382566585957,
"grad_norm": 4.594442367553711,
"learning_rate": 1.135728914043005e-06,
"loss": 0.0211,
"num_input_tokens_seen": 3095024,
"step": 7545
},
{
"epoch": 9.14043583535109,
"grad_norm": 3.8089771270751953,
"learning_rate": 1.120041269929642e-06,
"loss": 0.0929,
"num_input_tokens_seen": 3097008,
"step": 7550
},
{
"epoch": 9.146489104116222,
"grad_norm": 5.433687210083008,
"learning_rate": 1.1044602432072836e-06,
"loss": 0.1376,
"num_input_tokens_seen": 3099184,
"step": 7555
},
{
"epoch": 9.152542372881356,
"grad_norm": 3.2371695041656494,
"learning_rate": 1.0889859034408922e-06,
"loss": 0.0278,
"num_input_tokens_seen": 3101328,
"step": 7560
},
{
"epoch": 9.158595641646489,
"grad_norm": 7.783466815948486,
"learning_rate": 1.0736183197191024e-06,
"loss": 0.0766,
"num_input_tokens_seen": 3103408,
"step": 7565
},
{
"epoch": 9.164648910411623,
"grad_norm": 1.0674914121627808,
"learning_rate": 1.0583575606539108e-06,
"loss": 0.1019,
"num_input_tokens_seen": 3105616,
"step": 7570
},
{
"epoch": 9.170702179176756,
"grad_norm": 0.39573797583580017,
"learning_rate": 1.0432036943803708e-06,
"loss": 0.0384,
"num_input_tokens_seen": 3107536,
"step": 7575
},
{
"epoch": 9.176755447941888,
"grad_norm": 0.9420876502990723,
"learning_rate": 1.0281567885562947e-06,
"loss": 0.1868,
"num_input_tokens_seen": 3109648,
"step": 7580
},
{
"epoch": 9.182808716707022,
"grad_norm": 0.4029005169868469,
"learning_rate": 1.0132169103619444e-06,
"loss": 0.0853,
"num_input_tokens_seen": 3111504,
"step": 7585
},
{
"epoch": 9.188861985472155,
"grad_norm": 5.7842698097229,
"learning_rate": 9.98384126499735e-07,
"loss": 0.0881,
"num_input_tokens_seen": 3113424,
"step": 7590
},
{
"epoch": 9.194915254237289,
"grad_norm": 6.040023326873779,
"learning_rate": 9.836585031939154e-07,
"loss": 0.107,
"num_input_tokens_seen": 3115504,
"step": 7595
},
{
"epoch": 9.200968523002421,
"grad_norm": 0.3490646481513977,
"learning_rate": 9.690401061903249e-07,
"loss": 0.0503,
"num_input_tokens_seen": 3117488,
"step": 7600
},
{
"epoch": 9.207021791767554,
"grad_norm": 4.309531211853027,
"learning_rate": 9.545290007560437e-07,
"loss": 0.0953,
"num_input_tokens_seen": 3119376,
"step": 7605
},
{
"epoch": 9.213075060532688,
"grad_norm": 0.2243291288614273,
"learning_rate": 9.401252516791304e-07,
"loss": 0.1075,
"num_input_tokens_seen": 3121424,
"step": 7610
},
{
"epoch": 9.21912832929782,
"grad_norm": 6.078351020812988,
"learning_rate": 9.258289232683321e-07,
"loss": 0.15,
"num_input_tokens_seen": 3123504,
"step": 7615
},
{
"epoch": 9.225181598062955,
"grad_norm": 0.08517712354660034,
"learning_rate": 9.11640079352788e-07,
"loss": 0.1062,
"num_input_tokens_seen": 3125712,
"step": 7620
},
{
"epoch": 9.231234866828087,
"grad_norm": 3.3524436950683594,
"learning_rate": 8.975587832817545e-07,
"loss": 0.136,
"num_input_tokens_seen": 3127824,
"step": 7625
},
{
"epoch": 9.23728813559322,
"grad_norm": 1.0113067626953125,
"learning_rate": 8.835850979243055e-07,
"loss": 0.081,
"num_input_tokens_seen": 3129936,
"step": 7630
},
{
"epoch": 9.243341404358354,
"grad_norm": 6.101165294647217,
"learning_rate": 8.697190856690685e-07,
"loss": 0.0731,
"num_input_tokens_seen": 3131984,
"step": 7635
},
{
"epoch": 9.249394673123486,
"grad_norm": 7.550969123840332,
"learning_rate": 8.559608084239474e-07,
"loss": 0.0445,
"num_input_tokens_seen": 3134064,
"step": 7640
},
{
"epoch": 9.25544794188862,
"grad_norm": 0.6591706275939941,
"learning_rate": 8.423103276158306e-07,
"loss": 0.054,
"num_input_tokens_seen": 3136176,
"step": 7645
},
{
"epoch": 9.261501210653753,
"grad_norm": 2.5267579555511475,
"learning_rate": 8.287677041903308e-07,
"loss": 0.0777,
"num_input_tokens_seen": 3138288,
"step": 7650
},
{
"epoch": 9.267554479418886,
"grad_norm": 6.8405585289001465,
"learning_rate": 8.15332998611501e-07,
"loss": 0.0748,
"num_input_tokens_seen": 3140272,
"step": 7655
},
{
"epoch": 9.27360774818402,
"grad_norm": 4.520885944366455,
"learning_rate": 8.020062708615745e-07,
"loss": 0.1528,
"num_input_tokens_seen": 3142448,
"step": 7660
},
{
"epoch": 9.279661016949152,
"grad_norm": 1.5813335180282593,
"learning_rate": 7.887875804406946e-07,
"loss": 0.1642,
"num_input_tokens_seen": 3144528,
"step": 7665
},
{
"epoch": 9.285714285714286,
"grad_norm": 7.437231063842773,
"learning_rate": 7.756769863666524e-07,
"loss": 0.0751,
"num_input_tokens_seen": 3146512,
"step": 7670
},
{
"epoch": 9.291767554479419,
"grad_norm": 0.8976998329162598,
"learning_rate": 7.626745471746022e-07,
"loss": 0.0366,
"num_input_tokens_seen": 3148560,
"step": 7675
},
{
"epoch": 9.297820823244551,
"grad_norm": 5.735642910003662,
"learning_rate": 7.497803209168347e-07,
"loss": 0.1955,
"num_input_tokens_seen": 3150640,
"step": 7680
},
{
"epoch": 9.303874092009686,
"grad_norm": 6.1282639503479,
"learning_rate": 7.369943651624938e-07,
"loss": 0.0938,
"num_input_tokens_seen": 3152688,
"step": 7685
},
{
"epoch": 9.309927360774818,
"grad_norm": 8.234983444213867,
"learning_rate": 7.243167369973242e-07,
"loss": 0.1524,
"num_input_tokens_seen": 3154672,
"step": 7690
},
{
"epoch": 9.315980629539952,
"grad_norm": 11.793670654296875,
"learning_rate": 7.117474930234124e-07,
"loss": 0.1478,
"num_input_tokens_seen": 3156656,
"step": 7695
},
{
"epoch": 9.322033898305085,
"grad_norm": 0.2102554440498352,
"learning_rate": 6.992866893589578e-07,
"loss": 0.074,
"num_input_tokens_seen": 3158640,
"step": 7700
},
{
"epoch": 9.328087167070217,
"grad_norm": 1.4114265441894531,
"learning_rate": 6.869343816379825e-07,
"loss": 0.0223,
"num_input_tokens_seen": 3160624,
"step": 7705
},
{
"epoch": 9.334140435835351,
"grad_norm": 1.1425305604934692,
"learning_rate": 6.74690625010116e-07,
"loss": 0.0126,
"num_input_tokens_seen": 3162608,
"step": 7710
},
{
"epoch": 9.340193704600484,
"grad_norm": 0.6410977840423584,
"learning_rate": 6.625554741403333e-07,
"loss": 0.1421,
"num_input_tokens_seen": 3164560,
"step": 7715
},
{
"epoch": 9.346246973365618,
"grad_norm": 6.864235877990723,
"learning_rate": 6.505289832087231e-07,
"loss": 0.1654,
"num_input_tokens_seen": 3166768,
"step": 7720
},
{
"epoch": 9.35230024213075,
"grad_norm": 0.35761308670043945,
"learning_rate": 6.386112059102251e-07,
"loss": 0.1199,
"num_input_tokens_seen": 3168912,
"step": 7725
},
{
"epoch": 9.358353510895883,
"grad_norm": 0.04573501646518707,
"learning_rate": 6.268021954544096e-07,
"loss": 0.0494,
"num_input_tokens_seen": 3170800,
"step": 7730
},
{
"epoch": 9.364406779661017,
"grad_norm": 0.10186677426099777,
"learning_rate": 6.15102004565235e-07,
"loss": 0.041,
"num_input_tokens_seen": 3172784,
"step": 7735
},
{
"epoch": 9.37046004842615,
"grad_norm": 0.06820578128099442,
"learning_rate": 6.035106854808014e-07,
"loss": 0.0475,
"num_input_tokens_seen": 3174928,
"step": 7740
},
{
"epoch": 9.376513317191284,
"grad_norm": 0.32402148842811584,
"learning_rate": 5.920282899531421e-07,
"loss": 0.1466,
"num_input_tokens_seen": 3176976,
"step": 7745
},
{
"epoch": 9.382566585956416,
"grad_norm": 4.519848823547363,
"learning_rate": 5.806548692479624e-07,
"loss": 0.0724,
"num_input_tokens_seen": 3178896,
"step": 7750
},
{
"epoch": 9.388619854721549,
"grad_norm": 1.1942616701126099,
"learning_rate": 5.693904741444267e-07,
"loss": 0.0868,
"num_input_tokens_seen": 3180848,
"step": 7755
},
{
"epoch": 9.394673123486683,
"grad_norm": 0.24089713394641876,
"learning_rate": 5.58235154934944e-07,
"loss": 0.0606,
"num_input_tokens_seen": 3182704,
"step": 7760
},
{
"epoch": 9.400726392251816,
"grad_norm": 0.04828169196844101,
"learning_rate": 5.471889614249104e-07,
"loss": 0.0213,
"num_input_tokens_seen": 3184848,
"step": 7765
},
{
"epoch": 9.40677966101695,
"grad_norm": 4.41227388381958,
"learning_rate": 5.362519429325225e-07,
"loss": 0.0704,
"num_input_tokens_seen": 3186832,
"step": 7770
},
{
"epoch": 9.412832929782082,
"grad_norm": 8.546942710876465,
"learning_rate": 5.254241482885253e-07,
"loss": 0.093,
"num_input_tokens_seen": 3188912,
"step": 7775
},
{
"epoch": 9.418886198547215,
"grad_norm": 5.857113361358643,
"learning_rate": 5.147056258360289e-07,
"loss": 0.2003,
"num_input_tokens_seen": 3191152,
"step": 7780
},
{
"epoch": 9.424939467312349,
"grad_norm": 0.9563597440719604,
"learning_rate": 5.040964234302559e-07,
"loss": 0.0421,
"num_input_tokens_seen": 3193232,
"step": 7785
},
{
"epoch": 9.430992736077481,
"grad_norm": 3.2596330642700195,
"learning_rate": 4.935965884383525e-07,
"loss": 0.0406,
"num_input_tokens_seen": 3195312,
"step": 7790
},
{
"epoch": 9.437046004842616,
"grad_norm": 0.06693828850984573,
"learning_rate": 4.832061677391697e-07,
"loss": 0.0459,
"num_input_tokens_seen": 3197328,
"step": 7795
},
{
"epoch": 9.443099273607748,
"grad_norm": 1.464663028717041,
"learning_rate": 4.729252077230517e-07,
"loss": 0.0572,
"num_input_tokens_seen": 3199280,
"step": 7800
},
{
"epoch": 9.44915254237288,
"grad_norm": 0.9203391671180725,
"learning_rate": 4.6275375429163656e-07,
"loss": 0.0406,
"num_input_tokens_seen": 3201328,
"step": 7805
},
{
"epoch": 9.455205811138015,
"grad_norm": 0.721315860748291,
"learning_rate": 4.526918528576396e-07,
"loss": 0.0139,
"num_input_tokens_seen": 3203344,
"step": 7810
},
{
"epoch": 9.461259079903147,
"grad_norm": 1.4303781986236572,
"learning_rate": 4.427395483446617e-07,
"loss": 0.1111,
"num_input_tokens_seen": 3205488,
"step": 7815
},
{
"epoch": 9.467312348668282,
"grad_norm": 3.8508174419403076,
"learning_rate": 4.328968851869758e-07,
"loss": 0.2106,
"num_input_tokens_seen": 3207504,
"step": 7820
},
{
"epoch": 9.473365617433414,
"grad_norm": 2.1763241291046143,
"learning_rate": 4.231639073293492e-07,
"loss": 0.0216,
"num_input_tokens_seen": 3209712,
"step": 7825
},
{
"epoch": 9.479418886198546,
"grad_norm": 7.907344818115234,
"learning_rate": 4.13540658226827e-07,
"loss": 0.1066,
"num_input_tokens_seen": 3211728,
"step": 7830
},
{
"epoch": 9.48547215496368,
"grad_norm": 7.968501091003418,
"learning_rate": 4.040271808445406e-07,
"loss": 0.2031,
"num_input_tokens_seen": 3213616,
"step": 7835
},
{
"epoch": 9.491525423728813,
"grad_norm": 1.147912859916687,
"learning_rate": 3.94623517657533e-07,
"loss": 0.0102,
"num_input_tokens_seen": 3215536,
"step": 7840
},
{
"epoch": 9.497578692493947,
"grad_norm": 2.5614442825317383,
"learning_rate": 3.8532971065055045e-07,
"loss": 0.0522,
"num_input_tokens_seen": 3217552,
"step": 7845
},
{
"epoch": 9.5,
"eval_loss": 0.16040408611297607,
"eval_runtime": 4.5854,
"eval_samples_per_second": 80.037,
"eval_steps_per_second": 20.064,
"num_input_tokens_seen": 3218352,
"step": 7847
},
{
"epoch": 9.50363196125908,
"grad_norm": 6.244605541229248,
"learning_rate": 3.761458013178648e-07,
"loss": 0.0483,
"num_input_tokens_seen": 3219664,
"step": 7850
},
{
"epoch": 9.509685230024212,
"grad_norm": 0.11520939320325851,
"learning_rate": 3.670718306630766e-07,
"loss": 0.1404,
"num_input_tokens_seen": 3221648,
"step": 7855
},
{
"epoch": 9.515738498789347,
"grad_norm": 0.5533179640769958,
"learning_rate": 3.5810783919895673e-07,
"loss": 0.0363,
"num_input_tokens_seen": 3223632,
"step": 7860
},
{
"epoch": 9.521791767554479,
"grad_norm": 5.129673004150391,
"learning_rate": 3.4925386694723284e-07,
"loss": 0.0456,
"num_input_tokens_seen": 3225616,
"step": 7865
},
{
"epoch": 9.527845036319613,
"grad_norm": 6.019620418548584,
"learning_rate": 3.405099534384393e-07,
"loss": 0.1101,
"num_input_tokens_seen": 3227728,
"step": 7870
},
{
"epoch": 9.533898305084746,
"grad_norm": 0.03893245384097099,
"learning_rate": 3.31876137711723e-07,
"loss": 0.0851,
"num_input_tokens_seen": 3229744,
"step": 7875
},
{
"epoch": 9.539951573849878,
"grad_norm": 9.873018264770508,
"learning_rate": 3.233524583146741e-07,
"loss": 0.1517,
"num_input_tokens_seen": 3231664,
"step": 7880
},
{
"epoch": 9.546004842615012,
"grad_norm": 1.3719098567962646,
"learning_rate": 3.149389533031566e-07,
"loss": 0.1342,
"num_input_tokens_seen": 3233712,
"step": 7885
},
{
"epoch": 9.552058111380145,
"grad_norm": 6.197670936584473,
"learning_rate": 3.066356602411419e-07,
"loss": 0.2945,
"num_input_tokens_seen": 3235728,
"step": 7890
},
{
"epoch": 9.558111380145279,
"grad_norm": 1.7446709871292114,
"learning_rate": 2.984426162005227e-07,
"loss": 0.0866,
"num_input_tokens_seen": 3237712,
"step": 7895
},
{
"epoch": 9.564164648910412,
"grad_norm": 0.39076244831085205,
"learning_rate": 2.903598577609717e-07,
"loss": 0.1491,
"num_input_tokens_seen": 3239664,
"step": 7900
},
{
"epoch": 9.570217917675544,
"grad_norm": 1.8794441223144531,
"learning_rate": 2.823874210097638e-07,
"loss": 0.0952,
"num_input_tokens_seen": 3241776,
"step": 7905
},
{
"epoch": 9.576271186440678,
"grad_norm": 0.4310949444770813,
"learning_rate": 2.745253415416177e-07,
"loss": 0.023,
"num_input_tokens_seen": 3243824,
"step": 7910
},
{
"epoch": 9.58232445520581,
"grad_norm": 0.6770046949386597,
"learning_rate": 2.6677365445852976e-07,
"loss": 0.1038,
"num_input_tokens_seen": 3246000,
"step": 7915
},
{
"epoch": 9.588377723970945,
"grad_norm": 3.035773515701294,
"learning_rate": 2.5913239436964054e-07,
"loss": 0.0265,
"num_input_tokens_seen": 3248112,
"step": 7920
},
{
"epoch": 9.594430992736077,
"grad_norm": 0.0690949484705925,
"learning_rate": 2.5160159539105443e-07,
"loss": 0.0036,
"num_input_tokens_seen": 3250256,
"step": 7925
},
{
"epoch": 9.600484261501212,
"grad_norm": 2.347848892211914,
"learning_rate": 2.441812911456981e-07,
"loss": 0.0254,
"num_input_tokens_seen": 3252368,
"step": 7930
},
{
"epoch": 9.606537530266344,
"grad_norm": 0.4707547426223755,
"learning_rate": 2.3687151476317337e-07,
"loss": 0.0311,
"num_input_tokens_seen": 3254416,
"step": 7935
},
{
"epoch": 9.612590799031477,
"grad_norm": 1.3051100969314575,
"learning_rate": 2.2967229887960186e-07,
"loss": 0.0423,
"num_input_tokens_seen": 3256496,
"step": 7940
},
{
"epoch": 9.61864406779661,
"grad_norm": 4.094726085662842,
"learning_rate": 2.2258367563748884e-07,
"loss": 0.0552,
"num_input_tokens_seen": 3258576,
"step": 7945
},
{
"epoch": 9.624697336561743,
"grad_norm": 0.5801100134849548,
"learning_rate": 2.1560567668556797e-07,
"loss": 0.1695,
"num_input_tokens_seen": 3260496,
"step": 7950
},
{
"epoch": 9.630750605326877,
"grad_norm": 0.6484699249267578,
"learning_rate": 2.0873833317866798e-07,
"loss": 0.077,
"num_input_tokens_seen": 3262608,
"step": 7955
},
{
"epoch": 9.63680387409201,
"grad_norm": 3.7120988368988037,
"learning_rate": 2.019816757775711e-07,
"loss": 0.0599,
"num_input_tokens_seen": 3264592,
"step": 7960
},
{
"epoch": 9.642857142857142,
"grad_norm": 2.2407562732696533,
"learning_rate": 1.9533573464888543e-07,
"loss": 0.0174,
"num_input_tokens_seen": 3266704,
"step": 7965
},
{
"epoch": 9.648910411622277,
"grad_norm": 10.029207229614258,
"learning_rate": 1.8880053946488675e-07,
"loss": 0.1942,
"num_input_tokens_seen": 3268816,
"step": 7970
},
{
"epoch": 9.654963680387409,
"grad_norm": 0.6442005634307861,
"learning_rate": 1.8237611940341291e-07,
"loss": 0.1252,
"num_input_tokens_seen": 3270864,
"step": 7975
},
{
"epoch": 9.661016949152543,
"grad_norm": 10.1329984664917,
"learning_rate": 1.760625031477142e-07,
"loss": 0.2343,
"num_input_tokens_seen": 3272944,
"step": 7980
},
{
"epoch": 9.667070217917676,
"grad_norm": 0.7737054824829102,
"learning_rate": 1.6985971888633935e-07,
"loss": 0.0099,
"num_input_tokens_seen": 3274992,
"step": 7985
},
{
"epoch": 9.673123486682808,
"grad_norm": 0.4194134473800659,
"learning_rate": 1.637677943129967e-07,
"loss": 0.1094,
"num_input_tokens_seen": 3277008,
"step": 7990
},
{
"epoch": 9.679176755447942,
"grad_norm": 1.6003715991973877,
"learning_rate": 1.5778675662643793e-07,
"loss": 0.1485,
"num_input_tokens_seen": 3278928,
"step": 7995
},
{
"epoch": 9.685230024213075,
"grad_norm": 0.3803640604019165,
"learning_rate": 1.5191663253034116e-07,
"loss": 0.0433,
"num_input_tokens_seen": 3280944,
"step": 8000
},
{
"epoch": 9.69128329297821,
"grad_norm": 6.657588005065918,
"learning_rate": 1.461574482331779e-07,
"loss": 0.1022,
"num_input_tokens_seen": 3282960,
"step": 8005
},
{
"epoch": 9.697336561743342,
"grad_norm": 3.707159996032715,
"learning_rate": 1.4050922944811305e-07,
"loss": 0.0805,
"num_input_tokens_seen": 3285008,
"step": 8010
},
{
"epoch": 9.703389830508474,
"grad_norm": 9.373906135559082,
"learning_rate": 1.349720013928718e-07,
"loss": 0.1608,
"num_input_tokens_seen": 3287088,
"step": 8015
},
{
"epoch": 9.709443099273608,
"grad_norm": 6.443699359893799,
"learning_rate": 1.2954578878964507e-07,
"loss": 0.2076,
"num_input_tokens_seen": 3289168,
"step": 8020
},
{
"epoch": 9.71549636803874,
"grad_norm": 6.1982421875,
"learning_rate": 1.2423061586496477e-07,
"loss": 0.1625,
"num_input_tokens_seen": 3291376,
"step": 8025
},
{
"epoch": 9.721549636803875,
"grad_norm": 0.5799146890640259,
"learning_rate": 1.1902650634960378e-07,
"loss": 0.0786,
"num_input_tokens_seen": 3293360,
"step": 8030
},
{
"epoch": 9.727602905569007,
"grad_norm": 0.31346967816352844,
"learning_rate": 1.1393348347846777e-07,
"loss": 0.0491,
"num_input_tokens_seen": 3295344,
"step": 8035
},
{
"epoch": 9.73365617433414,
"grad_norm": 2.450915813446045,
"learning_rate": 1.0895156999048972e-07,
"loss": 0.0661,
"num_input_tokens_seen": 3297392,
"step": 8040
},
{
"epoch": 9.739709443099274,
"grad_norm": 2.6705565452575684,
"learning_rate": 1.0408078812853273e-07,
"loss": 0.1291,
"num_input_tokens_seen": 3299376,
"step": 8045
},
{
"epoch": 9.745762711864407,
"grad_norm": 0.06930442154407501,
"learning_rate": 9.932115963928734e-08,
"loss": 0.0463,
"num_input_tokens_seen": 3301360,
"step": 8050
},
{
"epoch": 9.75181598062954,
"grad_norm": 6.6655402183532715,
"learning_rate": 9.467270577317167e-08,
"loss": 0.0745,
"num_input_tokens_seen": 3303440,
"step": 8055
},
{
"epoch": 9.757869249394673,
"grad_norm": 5.793399810791016,
"learning_rate": 9.013544728424528e-08,
"loss": 0.1174,
"num_input_tokens_seen": 3305552,
"step": 8060
},
{
"epoch": 9.763922518159806,
"grad_norm": 0.10008306801319122,
"learning_rate": 8.570940443010655e-08,
"loss": 0.0508,
"num_input_tokens_seen": 3307728,
"step": 8065
},
{
"epoch": 9.76997578692494,
"grad_norm": 0.7323419451713562,
"learning_rate": 8.139459697181218e-08,
"loss": 0.1837,
"num_input_tokens_seen": 3309776,
"step": 8070
},
{
"epoch": 9.776029055690072,
"grad_norm": 7.605753421783447,
"learning_rate": 7.719104417377443e-08,
"loss": 0.0955,
"num_input_tokens_seen": 3311760,
"step": 8075
},
{
"epoch": 9.782082324455207,
"grad_norm": 5.1498870849609375,
"learning_rate": 7.30987648036946e-08,
"loss": 0.0404,
"num_input_tokens_seen": 3313808,
"step": 8080
},
{
"epoch": 9.788135593220339,
"grad_norm": 4.2535810470581055,
"learning_rate": 6.911777713246581e-08,
"loss": 0.0995,
"num_input_tokens_seen": 3315888,
"step": 8085
},
{
"epoch": 9.794188861985472,
"grad_norm": 3.450693130493164,
"learning_rate": 6.524809893409256e-08,
"loss": 0.0462,
"num_input_tokens_seen": 3318000,
"step": 8090
},
{
"epoch": 9.800242130750606,
"grad_norm": 2.4119420051574707,
"learning_rate": 6.148974748561299e-08,
"loss": 0.0788,
"num_input_tokens_seen": 3320016,
"step": 8095
},
{
"epoch": 9.806295399515738,
"grad_norm": 0.20463275909423828,
"learning_rate": 5.784273956702391e-08,
"loss": 0.136,
"num_input_tokens_seen": 3322096,
"step": 8100
},
{
"epoch": 9.812348668280872,
"grad_norm": 3.652512788772583,
"learning_rate": 5.4307091461205936e-08,
"loss": 0.0674,
"num_input_tokens_seen": 3324176,
"step": 8105
},
{
"epoch": 9.818401937046005,
"grad_norm": 0.506850004196167,
"learning_rate": 5.08828189538485e-08,
"loss": 0.0522,
"num_input_tokens_seen": 3326320,
"step": 8110
},
{
"epoch": 9.824455205811137,
"grad_norm": 7.119671821594238,
"learning_rate": 4.7569937333372115e-08,
"loss": 0.0571,
"num_input_tokens_seen": 3328464,
"step": 8115
},
{
"epoch": 9.830508474576272,
"grad_norm": 0.008174745365977287,
"learning_rate": 4.436846139087847e-08,
"loss": 0.0237,
"num_input_tokens_seen": 3330480,
"step": 8120
},
{
"epoch": 9.836561743341404,
"grad_norm": 1.4426367282867432,
"learning_rate": 4.127840542006711e-08,
"loss": 0.0669,
"num_input_tokens_seen": 3332624,
"step": 8125
},
{
"epoch": 9.842615012106538,
"grad_norm": 2.028547763824463,
"learning_rate": 3.829978321718553e-08,
"loss": 0.0795,
"num_input_tokens_seen": 3334768,
"step": 8130
},
{
"epoch": 9.84866828087167,
"grad_norm": 2.7887542247772217,
"learning_rate": 3.543260808095139e-08,
"loss": 0.0316,
"num_input_tokens_seen": 3336784,
"step": 8135
},
{
"epoch": 9.854721549636803,
"grad_norm": 3.320316791534424,
"learning_rate": 3.267689281250541e-08,
"loss": 0.0624,
"num_input_tokens_seen": 3338832,
"step": 8140
},
{
"epoch": 9.860774818401937,
"grad_norm": 9.407702445983887,
"learning_rate": 3.003264971535857e-08,
"loss": 0.1638,
"num_input_tokens_seen": 3340848,
"step": 8145
},
{
"epoch": 9.86682808716707,
"grad_norm": 0.1599888801574707,
"learning_rate": 2.7499890595314438e-08,
"loss": 0.1581,
"num_input_tokens_seen": 3342960,
"step": 8150
},
{
"epoch": 9.872881355932204,
"grad_norm": 2.643625259399414,
"learning_rate": 2.507862676044137e-08,
"loss": 0.1443,
"num_input_tokens_seen": 3345104,
"step": 8155
},
{
"epoch": 9.878934624697337,
"grad_norm": 2.9549529552459717,
"learning_rate": 2.2768869021014274e-08,
"loss": 0.1202,
"num_input_tokens_seen": 3347024,
"step": 8160
},
{
"epoch": 9.884987893462469,
"grad_norm": 6.08317756652832,
"learning_rate": 2.0570627689459054e-08,
"loss": 0.0827,
"num_input_tokens_seen": 3349200,
"step": 8165
},
{
"epoch": 9.891041162227603,
"grad_norm": 0.14019866287708282,
"learning_rate": 1.848391258031379e-08,
"loss": 0.0283,
"num_input_tokens_seen": 3351248,
"step": 8170
},
{
"epoch": 9.897094430992736,
"grad_norm": 4.22549295425415,
"learning_rate": 1.6508733010184297e-08,
"loss": 0.0468,
"num_input_tokens_seen": 3353488,
"step": 8175
},
{
"epoch": 9.90314769975787,
"grad_norm": 0.9411844611167908,
"learning_rate": 1.4645097797694186e-08,
"loss": 0.015,
"num_input_tokens_seen": 3355440,
"step": 8180
},
{
"epoch": 9.909200968523002,
"grad_norm": 5.702022552490234,
"learning_rate": 1.2893015263459874e-08,
"loss": 0.2151,
"num_input_tokens_seen": 3357296,
"step": 8185
},
{
"epoch": 9.915254237288135,
"grad_norm": 1.883554220199585,
"learning_rate": 1.125249323004618e-08,
"loss": 0.1424,
"num_input_tokens_seen": 3359280,
"step": 8190
},
{
"epoch": 9.92130750605327,
"grad_norm": 5.344539165496826,
"learning_rate": 9.723539021927463e-09,
"loss": 0.243,
"num_input_tokens_seen": 3361328,
"step": 8195
},
{
"epoch": 9.927360774818402,
"grad_norm": 0.5315026640892029,
"learning_rate": 8.306159465459872e-09,
"loss": 0.0859,
"num_input_tokens_seen": 3363344,
"step": 8200
},
{
"epoch": 9.933414043583536,
"grad_norm": 4.213244915008545,
"learning_rate": 7.00036088885081e-09,
"loss": 0.149,
"num_input_tokens_seen": 3365296,
"step": 8205
},
{
"epoch": 9.939467312348668,
"grad_norm": 0.8035793900489807,
"learning_rate": 5.806149122128401e-09,
"loss": 0.0801,
"num_input_tokens_seen": 3367504,
"step": 8210
},
{
"epoch": 9.9455205811138,
"grad_norm": 5.933956146240234,
"learning_rate": 4.723529497113743e-09,
"loss": 0.0346,
"num_input_tokens_seen": 3369616,
"step": 8215
},
{
"epoch": 9.951573849878935,
"grad_norm": 6.575443744659424,
"learning_rate": 3.752506847407023e-09,
"loss": 0.0627,
"num_input_tokens_seen": 3371728,
"step": 8220
},
{
"epoch": 9.957627118644067,
"grad_norm": 3.750946283340454,
"learning_rate": 2.8930855083542096e-09,
"loss": 0.1372,
"num_input_tokens_seen": 3373648,
"step": 8225
},
{
"epoch": 9.963680387409202,
"grad_norm": 0.5259419083595276,
"learning_rate": 2.145269317033183e-09,
"loss": 0.1955,
"num_input_tokens_seen": 3375664,
"step": 8230
},
{
"epoch": 9.969733656174334,
"grad_norm": 6.630932331085205,
"learning_rate": 1.509061612234297e-09,
"loss": 0.1024,
"num_input_tokens_seen": 3377808,
"step": 8235
},
{
"epoch": 9.975786924939467,
"grad_norm": 4.802211761474609,
"learning_rate": 9.844652344492832e-10,
"loss": 0.069,
"num_input_tokens_seen": 3379888,
"step": 8240
},
{
"epoch": 9.9818401937046,
"grad_norm": 5.706575870513916,
"learning_rate": 5.714825258545942e-10,
"loss": 0.155,
"num_input_tokens_seen": 3382064,
"step": 8245
},
{
"epoch": 9.987893462469733,
"grad_norm": 1.3004391193389893,
"learning_rate": 2.7011533030585347e-10,
"loss": 0.0399,
"num_input_tokens_seen": 3384144,
"step": 8250
},
{
"epoch": 9.993946731234868,
"grad_norm": 2.925779342651367,
"learning_rate": 8.036499332397807e-11,
"loss": 0.0171,
"num_input_tokens_seen": 3386160,
"step": 8255
},
{
"epoch": 10.0,
"grad_norm": 1.5842012166976929,
"learning_rate": 2.2323620896269604e-12,
"loss": 0.0796,
"num_input_tokens_seen": 3388032,
"step": 8260
},
{
"epoch": 10.0,
"eval_loss": 0.16043449938297272,
"eval_runtime": 4.5828,
"eval_samples_per_second": 80.082,
"eval_steps_per_second": 20.075,
"num_input_tokens_seen": 3388032,
"step": 8260
},
{
"epoch": 10.0,
"num_input_tokens_seen": 3388032,
"step": 8260,
"total_flos": 1.525669618557911e+17,
"train_loss": 0.11986262800496694,
"train_runtime": 1426.229,
"train_samples_per_second": 23.145,
"train_steps_per_second": 5.791
}
],
"logging_steps": 5,
"max_steps": 8260,
"num_input_tokens_seen": 3388032,
"num_train_epochs": 10,
"save_steps": 413,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.525669618557911e+17,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}