train_copa_789_1760637877 / trainer_state.json
rbelanec's picture
End of training
5f44cdc verified
{
"best_global_step": 1530,
"best_metric": 0.05843701213598251,
"best_model_checkpoint": "saves_multiple/ia3/llama-3-8b-instruct/train_copa_789_1760637877/checkpoint-1530",
"epoch": 20.0,
"eval_steps": 90,
"global_step": 1800,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.05555555555555555,
"grad_norm": 4.65092658996582,
"learning_rate": 1.1111111111111112e-06,
"loss": 0.5534,
"num_input_tokens_seen": 1632,
"step": 5
},
{
"epoch": 0.1111111111111111,
"grad_norm": 4.095375061035156,
"learning_rate": 2.5e-06,
"loss": 0.6235,
"num_input_tokens_seen": 3232,
"step": 10
},
{
"epoch": 0.16666666666666666,
"grad_norm": 4.756162166595459,
"learning_rate": 3.888888888888889e-06,
"loss": 0.6425,
"num_input_tokens_seen": 4832,
"step": 15
},
{
"epoch": 0.2222222222222222,
"grad_norm": 4.172198295593262,
"learning_rate": 5.277777777777778e-06,
"loss": 0.7387,
"num_input_tokens_seen": 6432,
"step": 20
},
{
"epoch": 0.2777777777777778,
"grad_norm": 4.3731584548950195,
"learning_rate": 6.666666666666667e-06,
"loss": 0.681,
"num_input_tokens_seen": 7968,
"step": 25
},
{
"epoch": 0.3333333333333333,
"grad_norm": 4.341258525848389,
"learning_rate": 8.055555555555557e-06,
"loss": 0.518,
"num_input_tokens_seen": 9504,
"step": 30
},
{
"epoch": 0.3888888888888889,
"grad_norm": 3.7260754108428955,
"learning_rate": 9.444444444444445e-06,
"loss": 0.6923,
"num_input_tokens_seen": 11104,
"step": 35
},
{
"epoch": 0.4444444444444444,
"grad_norm": 3.8928024768829346,
"learning_rate": 1.0833333333333334e-05,
"loss": 0.8366,
"num_input_tokens_seen": 12704,
"step": 40
},
{
"epoch": 0.5,
"grad_norm": 3.961286783218384,
"learning_rate": 1.2222222222222222e-05,
"loss": 0.5993,
"num_input_tokens_seen": 14240,
"step": 45
},
{
"epoch": 0.5555555555555556,
"grad_norm": 4.945498943328857,
"learning_rate": 1.3611111111111111e-05,
"loss": 0.6052,
"num_input_tokens_seen": 15808,
"step": 50
},
{
"epoch": 0.6111111111111112,
"grad_norm": 4.686919689178467,
"learning_rate": 1.5e-05,
"loss": 0.72,
"num_input_tokens_seen": 17344,
"step": 55
},
{
"epoch": 0.6666666666666666,
"grad_norm": 4.145110130310059,
"learning_rate": 1.638888888888889e-05,
"loss": 0.5328,
"num_input_tokens_seen": 18912,
"step": 60
},
{
"epoch": 0.7222222222222222,
"grad_norm": 4.903788089752197,
"learning_rate": 1.777777777777778e-05,
"loss": 0.6527,
"num_input_tokens_seen": 20448,
"step": 65
},
{
"epoch": 0.7777777777777778,
"grad_norm": 3.0253090858459473,
"learning_rate": 1.9166666666666667e-05,
"loss": 0.5486,
"num_input_tokens_seen": 21984,
"step": 70
},
{
"epoch": 0.8333333333333334,
"grad_norm": 3.5490806102752686,
"learning_rate": 2.0555555555555555e-05,
"loss": 0.6792,
"num_input_tokens_seen": 23552,
"step": 75
},
{
"epoch": 0.8888888888888888,
"grad_norm": 4.2691826820373535,
"learning_rate": 2.1944444444444445e-05,
"loss": 0.6004,
"num_input_tokens_seen": 25120,
"step": 80
},
{
"epoch": 0.9444444444444444,
"grad_norm": 4.216991901397705,
"learning_rate": 2.3333333333333336e-05,
"loss": 0.5378,
"num_input_tokens_seen": 26656,
"step": 85
},
{
"epoch": 1.0,
"grad_norm": 5.902134895324707,
"learning_rate": 2.4722222222222223e-05,
"loss": 0.6696,
"num_input_tokens_seen": 28192,
"step": 90
},
{
"epoch": 1.0,
"eval_loss": 0.5326165556907654,
"eval_runtime": 0.5121,
"eval_samples_per_second": 78.105,
"eval_steps_per_second": 19.526,
"num_input_tokens_seen": 28192,
"step": 90
},
{
"epoch": 1.0555555555555556,
"grad_norm": 4.175025939941406,
"learning_rate": 2.6111111111111114e-05,
"loss": 0.5202,
"num_input_tokens_seen": 29792,
"step": 95
},
{
"epoch": 1.1111111111111112,
"grad_norm": 3.9834649562835693,
"learning_rate": 2.7500000000000004e-05,
"loss": 0.5988,
"num_input_tokens_seen": 31328,
"step": 100
},
{
"epoch": 1.1666666666666667,
"grad_norm": 5.505035877227783,
"learning_rate": 2.8888888888888888e-05,
"loss": 0.5913,
"num_input_tokens_seen": 32832,
"step": 105
},
{
"epoch": 1.2222222222222223,
"grad_norm": 4.22908878326416,
"learning_rate": 3.0277777777777776e-05,
"loss": 0.548,
"num_input_tokens_seen": 34304,
"step": 110
},
{
"epoch": 1.2777777777777777,
"grad_norm": 4.530610084533691,
"learning_rate": 3.1666666666666666e-05,
"loss": 0.7176,
"num_input_tokens_seen": 35840,
"step": 115
},
{
"epoch": 1.3333333333333333,
"grad_norm": 4.503519058227539,
"learning_rate": 3.3055555555555553e-05,
"loss": 0.8881,
"num_input_tokens_seen": 37376,
"step": 120
},
{
"epoch": 1.3888888888888888,
"grad_norm": 3.23877215385437,
"learning_rate": 3.444444444444445e-05,
"loss": 0.5414,
"num_input_tokens_seen": 38944,
"step": 125
},
{
"epoch": 1.4444444444444444,
"grad_norm": 3.4258248805999756,
"learning_rate": 3.5833333333333335e-05,
"loss": 0.5146,
"num_input_tokens_seen": 40512,
"step": 130
},
{
"epoch": 1.5,
"grad_norm": 1.7767366170883179,
"learning_rate": 3.722222222222222e-05,
"loss": 0.2194,
"num_input_tokens_seen": 42080,
"step": 135
},
{
"epoch": 1.5555555555555556,
"grad_norm": 2.1948487758636475,
"learning_rate": 3.8611111111111116e-05,
"loss": 0.345,
"num_input_tokens_seen": 43680,
"step": 140
},
{
"epoch": 1.6111111111111112,
"grad_norm": 0.2988741099834442,
"learning_rate": 4e-05,
"loss": 0.0985,
"num_input_tokens_seen": 45216,
"step": 145
},
{
"epoch": 1.6666666666666665,
"grad_norm": 1.2166426181793213,
"learning_rate": 4.138888888888889e-05,
"loss": 0.3053,
"num_input_tokens_seen": 46720,
"step": 150
},
{
"epoch": 1.7222222222222223,
"grad_norm": 0.09494301676750183,
"learning_rate": 4.277777777777778e-05,
"loss": 0.02,
"num_input_tokens_seen": 48288,
"step": 155
},
{
"epoch": 1.7777777777777777,
"grad_norm": 1.532537579536438,
"learning_rate": 4.4166666666666665e-05,
"loss": 0.1515,
"num_input_tokens_seen": 49888,
"step": 160
},
{
"epoch": 1.8333333333333335,
"grad_norm": 0.09204932302236557,
"learning_rate": 4.555555555555556e-05,
"loss": 0.0452,
"num_input_tokens_seen": 51424,
"step": 165
},
{
"epoch": 1.8888888888888888,
"grad_norm": 2.5189366340637207,
"learning_rate": 4.6944444444444446e-05,
"loss": 0.2567,
"num_input_tokens_seen": 52992,
"step": 170
},
{
"epoch": 1.9444444444444444,
"grad_norm": 0.16729450225830078,
"learning_rate": 4.8333333333333334e-05,
"loss": 0.0968,
"num_input_tokens_seen": 54592,
"step": 175
},
{
"epoch": 2.0,
"grad_norm": 0.3235429525375366,
"learning_rate": 4.972222222222223e-05,
"loss": 0.1462,
"num_input_tokens_seen": 56192,
"step": 180
},
{
"epoch": 2.0,
"eval_loss": 0.06989195197820663,
"eval_runtime": 0.5163,
"eval_samples_per_second": 77.468,
"eval_steps_per_second": 19.367,
"num_input_tokens_seen": 56192,
"step": 180
},
{
"epoch": 2.0555555555555554,
"grad_norm": 1.0239298343658447,
"learning_rate": 4.9999247861994194e-05,
"loss": 0.0372,
"num_input_tokens_seen": 57760,
"step": 185
},
{
"epoch": 2.111111111111111,
"grad_norm": 0.03282800316810608,
"learning_rate": 4.9996192378909786e-05,
"loss": 0.0526,
"num_input_tokens_seen": 59264,
"step": 190
},
{
"epoch": 2.1666666666666665,
"grad_norm": 0.02693193219602108,
"learning_rate": 4.999078682916774e-05,
"loss": 0.1378,
"num_input_tokens_seen": 60768,
"step": 195
},
{
"epoch": 2.2222222222222223,
"grad_norm": 2.3027944564819336,
"learning_rate": 4.998303172098155e-05,
"loss": 0.1461,
"num_input_tokens_seen": 62272,
"step": 200
},
{
"epoch": 2.2777777777777777,
"grad_norm": 0.21790486574172974,
"learning_rate": 4.997292778346312e-05,
"loss": 0.0839,
"num_input_tokens_seen": 63808,
"step": 205
},
{
"epoch": 2.3333333333333335,
"grad_norm": 1.9243621826171875,
"learning_rate": 4.996047596655418e-05,
"loss": 0.2373,
"num_input_tokens_seen": 65344,
"step": 210
},
{
"epoch": 2.388888888888889,
"grad_norm": 0.5471611022949219,
"learning_rate": 4.994567744093703e-05,
"loss": 0.0171,
"num_input_tokens_seen": 66912,
"step": 215
},
{
"epoch": 2.4444444444444446,
"grad_norm": 1.6998534202575684,
"learning_rate": 4.992853359792444e-05,
"loss": 0.2556,
"num_input_tokens_seen": 68480,
"step": 220
},
{
"epoch": 2.5,
"grad_norm": 0.36938992142677307,
"learning_rate": 4.9909046049328846e-05,
"loss": 0.0088,
"num_input_tokens_seen": 70048,
"step": 225
},
{
"epoch": 2.5555555555555554,
"grad_norm": 0.2923737168312073,
"learning_rate": 4.988721662731083e-05,
"loss": 0.0252,
"num_input_tokens_seen": 71680,
"step": 230
},
{
"epoch": 2.611111111111111,
"grad_norm": 2.4089598655700684,
"learning_rate": 4.9863047384206835e-05,
"loss": 0.2822,
"num_input_tokens_seen": 73280,
"step": 235
},
{
"epoch": 2.6666666666666665,
"grad_norm": 0.03126000985503197,
"learning_rate": 4.983654059233626e-05,
"loss": 0.0374,
"num_input_tokens_seen": 74752,
"step": 240
},
{
"epoch": 2.7222222222222223,
"grad_norm": 0.03382931277155876,
"learning_rate": 4.9807698743787744e-05,
"loss": 0.1257,
"num_input_tokens_seen": 76288,
"step": 245
},
{
"epoch": 2.7777777777777777,
"grad_norm": 0.024398092180490494,
"learning_rate": 4.9776524550184965e-05,
"loss": 0.1005,
"num_input_tokens_seen": 77888,
"step": 250
},
{
"epoch": 2.8333333333333335,
"grad_norm": 0.1415417343378067,
"learning_rate": 4.974302094243164e-05,
"loss": 0.0065,
"num_input_tokens_seen": 79456,
"step": 255
},
{
"epoch": 2.888888888888889,
"grad_norm": 0.4479515552520752,
"learning_rate": 4.970719107043595e-05,
"loss": 0.1297,
"num_input_tokens_seen": 81088,
"step": 260
},
{
"epoch": 2.9444444444444446,
"grad_norm": 0.5779532790184021,
"learning_rate": 4.966903830281449e-05,
"loss": 0.0452,
"num_input_tokens_seen": 82592,
"step": 265
},
{
"epoch": 3.0,
"grad_norm": 1.6592082977294922,
"learning_rate": 4.962856622657541e-05,
"loss": 0.126,
"num_input_tokens_seen": 84192,
"step": 270
},
{
"epoch": 3.0,
"eval_loss": 0.0657103881239891,
"eval_runtime": 0.5198,
"eval_samples_per_second": 76.956,
"eval_steps_per_second": 19.239,
"num_input_tokens_seen": 84192,
"step": 270
},
{
"epoch": 3.0555555555555554,
"grad_norm": 0.03279653564095497,
"learning_rate": 4.9585778646781364e-05,
"loss": 0.0746,
"num_input_tokens_seen": 85696,
"step": 275
},
{
"epoch": 3.111111111111111,
"grad_norm": 0.8626558184623718,
"learning_rate": 4.9540679586191605e-05,
"loss": 0.052,
"num_input_tokens_seen": 87296,
"step": 280
},
{
"epoch": 3.1666666666666665,
"grad_norm": 0.11952415108680725,
"learning_rate": 4.9493273284883854e-05,
"loss": 0.0361,
"num_input_tokens_seen": 88832,
"step": 285
},
{
"epoch": 3.2222222222222223,
"grad_norm": 0.023274501785635948,
"learning_rate": 4.9443564199855666e-05,
"loss": 0.0264,
"num_input_tokens_seen": 90368,
"step": 290
},
{
"epoch": 3.2777777777777777,
"grad_norm": 0.22339174151420593,
"learning_rate": 4.939155700460536e-05,
"loss": 0.0202,
"num_input_tokens_seen": 91968,
"step": 295
},
{
"epoch": 3.3333333333333335,
"grad_norm": 0.9900386333465576,
"learning_rate": 4.933725658869267e-05,
"loss": 0.0386,
"num_input_tokens_seen": 93568,
"step": 300
},
{
"epoch": 3.388888888888889,
"grad_norm": 0.02162184566259384,
"learning_rate": 4.9280668057279014e-05,
"loss": 0.0482,
"num_input_tokens_seen": 95168,
"step": 305
},
{
"epoch": 3.4444444444444446,
"grad_norm": 0.04586295410990715,
"learning_rate": 4.9221796730647516e-05,
"loss": 0.1022,
"num_input_tokens_seen": 96704,
"step": 310
},
{
"epoch": 3.5,
"grad_norm": 0.025104861706495285,
"learning_rate": 4.916064814370287e-05,
"loss": 0.1679,
"num_input_tokens_seen": 98304,
"step": 315
},
{
"epoch": 3.5555555555555554,
"grad_norm": 0.028672898188233376,
"learning_rate": 4.9097228045450864e-05,
"loss": 0.0053,
"num_input_tokens_seen": 99840,
"step": 320
},
{
"epoch": 3.611111111111111,
"grad_norm": 0.24982739984989166,
"learning_rate": 4.9031542398457974e-05,
"loss": 0.0527,
"num_input_tokens_seen": 101376,
"step": 325
},
{
"epoch": 3.6666666666666665,
"grad_norm": 0.42373406887054443,
"learning_rate": 4.896359737829071e-05,
"loss": 0.1674,
"num_input_tokens_seen": 102976,
"step": 330
},
{
"epoch": 3.7222222222222223,
"grad_norm": 1.0080994367599487,
"learning_rate": 4.889339937293508e-05,
"loss": 0.0776,
"num_input_tokens_seen": 104576,
"step": 335
},
{
"epoch": 3.7777777777777777,
"grad_norm": 0.6784046292304993,
"learning_rate": 4.8820954982195905e-05,
"loss": 0.2309,
"num_input_tokens_seen": 106112,
"step": 340
},
{
"epoch": 3.8333333333333335,
"grad_norm": 0.024361956864595413,
"learning_rate": 4.874627101707644e-05,
"loss": 0.1049,
"num_input_tokens_seen": 107712,
"step": 345
},
{
"epoch": 3.888888888888889,
"grad_norm": 0.15874440968036652,
"learning_rate": 4.8669354499137955e-05,
"loss": 0.129,
"num_input_tokens_seen": 109344,
"step": 350
},
{
"epoch": 3.9444444444444446,
"grad_norm": 2.3382439613342285,
"learning_rate": 4.859021265983959e-05,
"loss": 0.1734,
"num_input_tokens_seen": 110944,
"step": 355
},
{
"epoch": 4.0,
"grad_norm": 0.6307403445243835,
"learning_rate": 4.850885293985853e-05,
"loss": 0.0396,
"num_input_tokens_seen": 112544,
"step": 360
},
{
"epoch": 4.0,
"eval_loss": 0.06675007939338684,
"eval_runtime": 0.5188,
"eval_samples_per_second": 77.097,
"eval_steps_per_second": 19.274,
"num_input_tokens_seen": 112544,
"step": 360
},
{
"epoch": 4.055555555555555,
"grad_norm": 0.09657640755176544,
"learning_rate": 4.8425282988390376e-05,
"loss": 0.0751,
"num_input_tokens_seen": 114144,
"step": 365
},
{
"epoch": 4.111111111111111,
"grad_norm": 0.8270570039749146,
"learning_rate": 4.8339510662430046e-05,
"loss": 0.0563,
"num_input_tokens_seen": 115712,
"step": 370
},
{
"epoch": 4.166666666666667,
"grad_norm": 0.6188746094703674,
"learning_rate": 4.825154402603308e-05,
"loss": 0.1947,
"num_input_tokens_seen": 117312,
"step": 375
},
{
"epoch": 4.222222222222222,
"grad_norm": 0.978348970413208,
"learning_rate": 4.816139134955746e-05,
"loss": 0.1331,
"num_input_tokens_seen": 118848,
"step": 380
},
{
"epoch": 4.277777777777778,
"grad_norm": 1.4264014959335327,
"learning_rate": 4.806906110888606e-05,
"loss": 0.0498,
"num_input_tokens_seen": 120416,
"step": 385
},
{
"epoch": 4.333333333333333,
"grad_norm": 0.5848916172981262,
"learning_rate": 4.797456198462979e-05,
"loss": 0.0876,
"num_input_tokens_seen": 122016,
"step": 390
},
{
"epoch": 4.388888888888889,
"grad_norm": 1.7021077871322632,
"learning_rate": 4.7877902861311446e-05,
"loss": 0.1327,
"num_input_tokens_seen": 123584,
"step": 395
},
{
"epoch": 4.444444444444445,
"grad_norm": 0.02586347423493862,
"learning_rate": 4.777909282653042e-05,
"loss": 0.0162,
"num_input_tokens_seen": 125216,
"step": 400
},
{
"epoch": 4.5,
"grad_norm": 0.7212857604026794,
"learning_rate": 4.7678141170108345e-05,
"loss": 0.0556,
"num_input_tokens_seen": 126784,
"step": 405
},
{
"epoch": 4.555555555555555,
"grad_norm": 0.12025772035121918,
"learning_rate": 4.757505738321563e-05,
"loss": 0.0259,
"num_input_tokens_seen": 128352,
"step": 410
},
{
"epoch": 4.611111111111111,
"grad_norm": 0.2885558009147644,
"learning_rate": 4.7469851157479177e-05,
"loss": 0.021,
"num_input_tokens_seen": 129920,
"step": 415
},
{
"epoch": 4.666666666666667,
"grad_norm": 0.09183027595281601,
"learning_rate": 4.736253238407119e-05,
"loss": 0.0269,
"num_input_tokens_seen": 131488,
"step": 420
},
{
"epoch": 4.722222222222222,
"grad_norm": 0.04295544698834419,
"learning_rate": 4.725311115277924e-05,
"loss": 0.1203,
"num_input_tokens_seen": 133056,
"step": 425
},
{
"epoch": 4.777777777777778,
"grad_norm": 0.2513387203216553,
"learning_rate": 4.714159775105765e-05,
"loss": 0.0552,
"num_input_tokens_seen": 134624,
"step": 430
},
{
"epoch": 4.833333333333333,
"grad_norm": 0.2819466292858124,
"learning_rate": 4.70280026630603e-05,
"loss": 0.0788,
"num_input_tokens_seen": 136192,
"step": 435
},
{
"epoch": 4.888888888888889,
"grad_norm": 0.02205601893365383,
"learning_rate": 4.6912336568654925e-05,
"loss": 0.0426,
"num_input_tokens_seen": 137824,
"step": 440
},
{
"epoch": 4.944444444444445,
"grad_norm": 0.248850479722023,
"learning_rate": 4.679461034241906e-05,
"loss": 0.0641,
"num_input_tokens_seen": 139392,
"step": 445
},
{
"epoch": 5.0,
"grad_norm": 0.6842424273490906,
"learning_rate": 4.667483505261762e-05,
"loss": 0.1156,
"num_input_tokens_seen": 140960,
"step": 450
},
{
"epoch": 5.0,
"eval_loss": 0.06348364055156708,
"eval_runtime": 0.5151,
"eval_samples_per_second": 77.651,
"eval_steps_per_second": 19.413,
"num_input_tokens_seen": 140960,
"step": 450
},
{
"epoch": 5.055555555555555,
"grad_norm": 0.03485225886106491,
"learning_rate": 4.655302196016228e-05,
"loss": 0.0857,
"num_input_tokens_seen": 142560,
"step": 455
},
{
"epoch": 5.111111111111111,
"grad_norm": 0.07330955564975739,
"learning_rate": 4.642918251755281e-05,
"loss": 0.0651,
"num_input_tokens_seen": 144096,
"step": 460
},
{
"epoch": 5.166666666666667,
"grad_norm": 0.15627992153167725,
"learning_rate": 4.6303328367800284e-05,
"loss": 0.0184,
"num_input_tokens_seen": 145632,
"step": 465
},
{
"epoch": 5.222222222222222,
"grad_norm": 1.0182912349700928,
"learning_rate": 4.6175471343332485e-05,
"loss": 0.085,
"num_input_tokens_seen": 147232,
"step": 470
},
{
"epoch": 5.277777777777778,
"grad_norm": 0.15309298038482666,
"learning_rate": 4.604562346488144e-05,
"loss": 0.1157,
"num_input_tokens_seen": 148864,
"step": 475
},
{
"epoch": 5.333333333333333,
"grad_norm": 0.08715078979730606,
"learning_rate": 4.591379694035325e-05,
"loss": 0.0295,
"num_input_tokens_seen": 150432,
"step": 480
},
{
"epoch": 5.388888888888889,
"grad_norm": 0.48800748586654663,
"learning_rate": 4.5780004163680365e-05,
"loss": 0.0973,
"num_input_tokens_seen": 152032,
"step": 485
},
{
"epoch": 5.444444444444445,
"grad_norm": 0.9018146395683289,
"learning_rate": 4.5644257713656356e-05,
"loss": 0.0468,
"num_input_tokens_seen": 153600,
"step": 490
},
{
"epoch": 5.5,
"grad_norm": 1.2570453882217407,
"learning_rate": 4.550657035275323e-05,
"loss": 0.0803,
"num_input_tokens_seen": 155200,
"step": 495
},
{
"epoch": 5.555555555555555,
"grad_norm": 0.04507768154144287,
"learning_rate": 4.536695502592162e-05,
"loss": 0.1479,
"num_input_tokens_seen": 156800,
"step": 500
},
{
"epoch": 5.611111111111111,
"grad_norm": 0.02345445193350315,
"learning_rate": 4.522542485937369e-05,
"loss": 0.0097,
"num_input_tokens_seen": 158336,
"step": 505
},
{
"epoch": 5.666666666666667,
"grad_norm": 0.13815099000930786,
"learning_rate": 4.5081993159349056e-05,
"loss": 0.0184,
"num_input_tokens_seen": 159904,
"step": 510
},
{
"epoch": 5.722222222222222,
"grad_norm": 1.6783974170684814,
"learning_rate": 4.493667341086379e-05,
"loss": 0.109,
"num_input_tokens_seen": 161504,
"step": 515
},
{
"epoch": 5.777777777777778,
"grad_norm": 0.03608032688498497,
"learning_rate": 4.478947927644258e-05,
"loss": 0.0146,
"num_input_tokens_seen": 163008,
"step": 520
},
{
"epoch": 5.833333333333333,
"grad_norm": 0.6085768342018127,
"learning_rate": 4.464042459483425e-05,
"loss": 0.0752,
"num_input_tokens_seen": 164512,
"step": 525
},
{
"epoch": 5.888888888888889,
"grad_norm": 0.6818676590919495,
"learning_rate": 4.448952337971064e-05,
"loss": 0.1839,
"num_input_tokens_seen": 166080,
"step": 530
},
{
"epoch": 5.944444444444445,
"grad_norm": 0.025021351873874664,
"learning_rate": 4.43367898183491e-05,
"loss": 0.0553,
"num_input_tokens_seen": 167680,
"step": 535
},
{
"epoch": 6.0,
"grad_norm": 0.016283836215734482,
"learning_rate": 4.418223827029867e-05,
"loss": 0.0155,
"num_input_tokens_seen": 169216,
"step": 540
},
{
"epoch": 6.0,
"eval_loss": 0.06451700627803802,
"eval_runtime": 0.5189,
"eval_samples_per_second": 77.089,
"eval_steps_per_second": 19.272,
"num_input_tokens_seen": 169216,
"step": 540
},
{
"epoch": 6.055555555555555,
"grad_norm": 0.04106647148728371,
"learning_rate": 4.402588326603002e-05,
"loss": 0.0236,
"num_input_tokens_seen": 170784,
"step": 545
},
{
"epoch": 6.111111111111111,
"grad_norm": 0.8891972303390503,
"learning_rate": 4.386773950556931e-05,
"loss": 0.0654,
"num_input_tokens_seen": 172352,
"step": 550
},
{
"epoch": 6.166666666666667,
"grad_norm": 0.3865717649459839,
"learning_rate": 4.3707821857116176e-05,
"loss": 0.0432,
"num_input_tokens_seen": 173920,
"step": 555
},
{
"epoch": 6.222222222222222,
"grad_norm": 0.35376498103141785,
"learning_rate": 4.354614535564588e-05,
"loss": 0.0359,
"num_input_tokens_seen": 175488,
"step": 560
},
{
"epoch": 6.277777777777778,
"grad_norm": 1.738903522491455,
"learning_rate": 4.3382725201495723e-05,
"loss": 0.0785,
"num_input_tokens_seen": 177024,
"step": 565
},
{
"epoch": 6.333333333333333,
"grad_norm": 0.38843443989753723,
"learning_rate": 4.321757675893596e-05,
"loss": 0.0491,
"num_input_tokens_seen": 178624,
"step": 570
},
{
"epoch": 6.388888888888889,
"grad_norm": 0.322394996881485,
"learning_rate": 4.305071555472534e-05,
"loss": 0.1942,
"num_input_tokens_seen": 180096,
"step": 575
},
{
"epoch": 6.444444444444445,
"grad_norm": 0.8024235963821411,
"learning_rate": 4.288215727665129e-05,
"loss": 0.056,
"num_input_tokens_seen": 181600,
"step": 580
},
{
"epoch": 6.5,
"grad_norm": 1.246673583984375,
"learning_rate": 4.2711917772055e-05,
"loss": 0.0492,
"num_input_tokens_seen": 183232,
"step": 585
},
{
"epoch": 6.555555555555555,
"grad_norm": 0.022138679400086403,
"learning_rate": 4.254001304634151e-05,
"loss": 0.0965,
"num_input_tokens_seen": 184768,
"step": 590
},
{
"epoch": 6.611111111111111,
"grad_norm": 0.013323606923222542,
"learning_rate": 4.2366459261474933e-05,
"loss": 0.0785,
"num_input_tokens_seen": 186208,
"step": 595
},
{
"epoch": 6.666666666666667,
"grad_norm": 0.017629526555538177,
"learning_rate": 4.2191272734458955e-05,
"loss": 0.0285,
"num_input_tokens_seen": 187776,
"step": 600
},
{
"epoch": 6.722222222222222,
"grad_norm": 1.184860110282898,
"learning_rate": 4.201446993580276e-05,
"loss": 0.0469,
"num_input_tokens_seen": 189344,
"step": 605
},
{
"epoch": 6.777777777777778,
"grad_norm": 0.07062964886426926,
"learning_rate": 4.183606748797251e-05,
"loss": 0.0205,
"num_input_tokens_seen": 190912,
"step": 610
},
{
"epoch": 6.833333333333333,
"grad_norm": 0.04091072455048561,
"learning_rate": 4.1656082163828566e-05,
"loss": 0.1118,
"num_input_tokens_seen": 192512,
"step": 615
},
{
"epoch": 6.888888888888889,
"grad_norm": 0.05158746615052223,
"learning_rate": 4.147453088504854e-05,
"loss": 0.0199,
"num_input_tokens_seen": 194112,
"step": 620
},
{
"epoch": 6.944444444444445,
"grad_norm": 0.7025681138038635,
"learning_rate": 4.129143072053638e-05,
"loss": 0.1499,
"num_input_tokens_seen": 195712,
"step": 625
},
{
"epoch": 7.0,
"grad_norm": 0.13185453414916992,
"learning_rate": 4.110679888481763e-05,
"loss": 0.0091,
"num_input_tokens_seen": 197248,
"step": 630
},
{
"epoch": 7.0,
"eval_loss": 0.06116596609354019,
"eval_runtime": 0.5186,
"eval_samples_per_second": 77.136,
"eval_steps_per_second": 19.284,
"num_input_tokens_seen": 197248,
"step": 630
},
{
"epoch": 7.055555555555555,
"grad_norm": 0.1526983231306076,
"learning_rate": 4.09206527364209e-05,
"loss": 0.0193,
"num_input_tokens_seen": 198784,
"step": 635
},
{
"epoch": 7.111111111111111,
"grad_norm": 1.433021903038025,
"learning_rate": 4.073300977624594e-05,
"loss": 0.092,
"num_input_tokens_seen": 200352,
"step": 640
},
{
"epoch": 7.166666666666667,
"grad_norm": 0.4112715721130371,
"learning_rate": 4.054388764591822e-05,
"loss": 0.0539,
"num_input_tokens_seen": 201856,
"step": 645
},
{
"epoch": 7.222222222222222,
"grad_norm": 0.022885221987962723,
"learning_rate": 4.035330412613035e-05,
"loss": 0.0623,
"num_input_tokens_seen": 203488,
"step": 650
},
{
"epoch": 7.277777777777778,
"grad_norm": 0.032947391271591187,
"learning_rate": 4.0161277134970345e-05,
"loss": 0.0238,
"num_input_tokens_seen": 205056,
"step": 655
},
{
"epoch": 7.333333333333333,
"grad_norm": 0.25257608294487,
"learning_rate": 3.996782472623705e-05,
"loss": 0.0829,
"num_input_tokens_seen": 206656,
"step": 660
},
{
"epoch": 7.388888888888889,
"grad_norm": 0.2931102216243744,
"learning_rate": 3.977296508774278e-05,
"loss": 0.0418,
"num_input_tokens_seen": 208288,
"step": 665
},
{
"epoch": 7.444444444444445,
"grad_norm": 1.5941400527954102,
"learning_rate": 3.957671653960337e-05,
"loss": 0.1037,
"num_input_tokens_seen": 209856,
"step": 670
},
{
"epoch": 7.5,
"grad_norm": 1.2473227977752686,
"learning_rate": 3.9379097532515725e-05,
"loss": 0.0465,
"num_input_tokens_seen": 211424,
"step": 675
},
{
"epoch": 7.555555555555555,
"grad_norm": 0.053567372262477875,
"learning_rate": 3.918012664602317e-05,
"loss": 0.0115,
"num_input_tokens_seen": 212928,
"step": 680
},
{
"epoch": 7.611111111111111,
"grad_norm": 0.13769978284835815,
"learning_rate": 3.897982258676867e-05,
"loss": 0.0293,
"num_input_tokens_seen": 214496,
"step": 685
},
{
"epoch": 7.666666666666667,
"grad_norm": 0.046278007328510284,
"learning_rate": 3.8778204186736076e-05,
"loss": 0.0137,
"num_input_tokens_seen": 216064,
"step": 690
},
{
"epoch": 7.722222222222222,
"grad_norm": 0.03257855400443077,
"learning_rate": 3.8575290401479586e-05,
"loss": 0.125,
"num_input_tokens_seen": 217600,
"step": 695
},
{
"epoch": 7.777777777777778,
"grad_norm": 1.0644551515579224,
"learning_rate": 3.837110030834161e-05,
"loss": 0.0516,
"num_input_tokens_seen": 219136,
"step": 700
},
{
"epoch": 7.833333333333333,
"grad_norm": 0.1464487761259079,
"learning_rate": 3.8165653104659185e-05,
"loss": 0.0523,
"num_input_tokens_seen": 220704,
"step": 705
},
{
"epoch": 7.888888888888889,
"grad_norm": 0.7324170470237732,
"learning_rate": 3.79589681059591e-05,
"loss": 0.0839,
"num_input_tokens_seen": 222272,
"step": 710
},
{
"epoch": 7.944444444444445,
"grad_norm": 0.17301106452941895,
"learning_rate": 3.775106474414188e-05,
"loss": 0.0171,
"num_input_tokens_seen": 223872,
"step": 715
},
{
"epoch": 8.0,
"grad_norm": 0.8065789937973022,
"learning_rate": 3.75419625656549e-05,
"loss": 0.1638,
"num_input_tokens_seen": 225440,
"step": 720
},
{
"epoch": 8.0,
"eval_loss": 0.06141107156872749,
"eval_runtime": 0.5188,
"eval_samples_per_second": 77.108,
"eval_steps_per_second": 19.277,
"num_input_tokens_seen": 225440,
"step": 720
},
{
"epoch": 8.055555555555555,
"grad_norm": 1.6780734062194824,
"learning_rate": 3.7331681229654635e-05,
"loss": 0.063,
"num_input_tokens_seen": 226912,
"step": 725
},
{
"epoch": 8.11111111111111,
"grad_norm": 0.33701032400131226,
"learning_rate": 3.712024050615843e-05,
"loss": 0.0179,
"num_input_tokens_seen": 228512,
"step": 730
},
{
"epoch": 8.166666666666666,
"grad_norm": 0.5328055024147034,
"learning_rate": 3.690766027418573e-05,
"loss": 0.1383,
"num_input_tokens_seen": 230080,
"step": 735
},
{
"epoch": 8.222222222222221,
"grad_norm": 0.2738179564476013,
"learning_rate": 3.6693960519889106e-05,
"loss": 0.015,
"num_input_tokens_seen": 231616,
"step": 740
},
{
"epoch": 8.277777777777779,
"grad_norm": 0.012103931978344917,
"learning_rate": 3.6479161334675296e-05,
"loss": 0.0592,
"num_input_tokens_seen": 233184,
"step": 745
},
{
"epoch": 8.333333333333334,
"grad_norm": 0.6792635321617126,
"learning_rate": 3.626328291331618e-05,
"loss": 0.0447,
"num_input_tokens_seen": 234688,
"step": 750
},
{
"epoch": 8.38888888888889,
"grad_norm": 0.24234500527381897,
"learning_rate": 3.60463455520502e-05,
"loss": 0.1588,
"num_input_tokens_seen": 236256,
"step": 755
},
{
"epoch": 8.444444444444445,
"grad_norm": 0.2365965098142624,
"learning_rate": 3.582836964667408e-05,
"loss": 0.0256,
"num_input_tokens_seen": 237856,
"step": 760
},
{
"epoch": 8.5,
"grad_norm": 0.1399311125278473,
"learning_rate": 3.560937569062538e-05,
"loss": 0.0035,
"num_input_tokens_seen": 239392,
"step": 765
},
{
"epoch": 8.555555555555555,
"grad_norm": 0.1434163898229599,
"learning_rate": 3.538938427305573e-05,
"loss": 0.1312,
"num_input_tokens_seen": 240960,
"step": 770
},
{
"epoch": 8.61111111111111,
"grad_norm": 0.06928466260433197,
"learning_rate": 3.516841607689501e-05,
"loss": 0.0597,
"num_input_tokens_seen": 242496,
"step": 775
},
{
"epoch": 8.666666666666666,
"grad_norm": 0.7546709775924683,
"learning_rate": 3.494649187690695e-05,
"loss": 0.0451,
"num_input_tokens_seen": 244064,
"step": 780
},
{
"epoch": 8.722222222222221,
"grad_norm": 0.024969231337308884,
"learning_rate": 3.4723632537735846e-05,
"loss": 0.0154,
"num_input_tokens_seen": 245664,
"step": 785
},
{
"epoch": 8.777777777777779,
"grad_norm": 0.20347218215465546,
"learning_rate": 3.449985901194498e-05,
"loss": 0.0257,
"num_input_tokens_seen": 247264,
"step": 790
},
{
"epoch": 8.833333333333334,
"grad_norm": 0.28938475251197815,
"learning_rate": 3.427519233804667e-05,
"loss": 0.0679,
"num_input_tokens_seen": 248832,
"step": 795
},
{
"epoch": 8.88888888888889,
"grad_norm": 0.037665314972400665,
"learning_rate": 3.404965363852437e-05,
"loss": 0.1039,
"num_input_tokens_seen": 250432,
"step": 800
},
{
"epoch": 8.944444444444445,
"grad_norm": 0.040147844702005386,
"learning_rate": 3.382326411784672e-05,
"loss": 0.0276,
"num_input_tokens_seen": 252064,
"step": 805
},
{
"epoch": 9.0,
"grad_norm": 0.014303365722298622,
"learning_rate": 3.359604506047403e-05,
"loss": 0.0195,
"num_input_tokens_seen": 253632,
"step": 810
},
{
"epoch": 9.0,
"eval_loss": 0.06076971814036369,
"eval_runtime": 0.5198,
"eval_samples_per_second": 76.957,
"eval_steps_per_second": 19.239,
"num_input_tokens_seen": 253632,
"step": 810
},
{
"epoch": 9.055555555555555,
"grad_norm": 0.2893994450569153,
"learning_rate": 3.336801782885712e-05,
"loss": 0.1298,
"num_input_tokens_seen": 255232,
"step": 815
},
{
"epoch": 9.11111111111111,
"grad_norm": 0.035851139575242996,
"learning_rate": 3.313920386142892e-05,
"loss": 0.0098,
"num_input_tokens_seen": 256800,
"step": 820
},
{
"epoch": 9.166666666666666,
"grad_norm": 0.010133305564522743,
"learning_rate": 3.290962467058891e-05,
"loss": 0.1042,
"num_input_tokens_seen": 258368,
"step": 825
},
{
"epoch": 9.222222222222221,
"grad_norm": 0.06425736844539642,
"learning_rate": 3.267930184068057e-05,
"loss": 0.0088,
"num_input_tokens_seen": 259872,
"step": 830
},
{
"epoch": 9.277777777777779,
"grad_norm": 0.11569812148809433,
"learning_rate": 3.244825702596205e-05,
"loss": 0.0193,
"num_input_tokens_seen": 261376,
"step": 835
},
{
"epoch": 9.333333333333334,
"grad_norm": 0.671159029006958,
"learning_rate": 3.2216511948570374e-05,
"loss": 0.0403,
"num_input_tokens_seen": 262944,
"step": 840
},
{
"epoch": 9.38888888888889,
"grad_norm": 0.11181628704071045,
"learning_rate": 3.198408839647911e-05,
"loss": 0.0874,
"num_input_tokens_seen": 264544,
"step": 845
},
{
"epoch": 9.444444444444445,
"grad_norm": 0.47002896666526794,
"learning_rate": 3.1751008221450025e-05,
"loss": 0.0231,
"num_input_tokens_seen": 266080,
"step": 850
},
{
"epoch": 9.5,
"grad_norm": 0.45572900772094727,
"learning_rate": 3.151729333697854e-05,
"loss": 0.0398,
"num_input_tokens_seen": 267680,
"step": 855
},
{
"epoch": 9.555555555555555,
"grad_norm": 1.4050266742706299,
"learning_rate": 3.1282965716233594e-05,
"loss": 0.1802,
"num_input_tokens_seen": 269280,
"step": 860
},
{
"epoch": 9.61111111111111,
"grad_norm": 0.6737051010131836,
"learning_rate": 3.104804738999169e-05,
"loss": 0.032,
"num_input_tokens_seen": 270912,
"step": 865
},
{
"epoch": 9.666666666666666,
"grad_norm": 0.06566260010004044,
"learning_rate": 3.0812560444565745e-05,
"loss": 0.0392,
"num_input_tokens_seen": 272512,
"step": 870
},
{
"epoch": 9.722222222222221,
"grad_norm": 0.9122573137283325,
"learning_rate": 3.057652701972848e-05,
"loss": 0.0411,
"num_input_tokens_seen": 274048,
"step": 875
},
{
"epoch": 9.777777777777779,
"grad_norm": 1.3042923212051392,
"learning_rate": 3.0339969306631005e-05,
"loss": 0.0619,
"num_input_tokens_seen": 275648,
"step": 880
},
{
"epoch": 9.833333333333334,
"grad_norm": 0.04545683041214943,
"learning_rate": 3.0102909545716396e-05,
"loss": 0.0641,
"num_input_tokens_seen": 277248,
"step": 885
},
{
"epoch": 9.88888888888889,
"grad_norm": 0.19024275243282318,
"learning_rate": 2.9865370024628775e-05,
"loss": 0.036,
"num_input_tokens_seen": 278816,
"step": 890
},
{
"epoch": 9.944444444444445,
"grad_norm": 0.047417499125003815,
"learning_rate": 2.9627373076117863e-05,
"loss": 0.0168,
"num_input_tokens_seen": 280416,
"step": 895
},
{
"epoch": 10.0,
"grad_norm": 0.0708514004945755,
"learning_rate": 2.9388941075939334e-05,
"loss": 0.0328,
"num_input_tokens_seen": 281984,
"step": 900
},
{
"epoch": 10.0,
"eval_loss": 0.06215012073516846,
"eval_runtime": 0.517,
"eval_samples_per_second": 77.368,
"eval_steps_per_second": 19.342,
"num_input_tokens_seen": 281984,
"step": 900
},
{
"epoch": 10.055555555555555,
"grad_norm": 1.1088037490844727,
"learning_rate": 2.9150096440751107e-05,
"loss": 0.1483,
"num_input_tokens_seen": 283552,
"step": 905
},
{
"epoch": 10.11111111111111,
"grad_norm": 0.13920870423316956,
"learning_rate": 2.8910861626005776e-05,
"loss": 0.0159,
"num_input_tokens_seen": 285120,
"step": 910
},
{
"epoch": 10.166666666666666,
"grad_norm": 0.1359255462884903,
"learning_rate": 2.8671259123839472e-05,
"loss": 0.0437,
"num_input_tokens_seen": 286720,
"step": 915
},
{
"epoch": 10.222222222222221,
"grad_norm": 0.1851315051317215,
"learning_rate": 2.843131146095719e-05,
"loss": 0.0302,
"num_input_tokens_seen": 288288,
"step": 920
},
{
"epoch": 10.277777777777779,
"grad_norm": 0.017424006015062332,
"learning_rate": 2.8191041196514873e-05,
"loss": 0.0206,
"num_input_tokens_seen": 289856,
"step": 925
},
{
"epoch": 10.333333333333334,
"grad_norm": 1.307128667831421,
"learning_rate": 2.795047091999849e-05,
"loss": 0.1355,
"num_input_tokens_seen": 291392,
"step": 930
},
{
"epoch": 10.38888888888889,
"grad_norm": 0.016336191445589066,
"learning_rate": 2.770962324910027e-05,
"loss": 0.0174,
"num_input_tokens_seen": 292864,
"step": 935
},
{
"epoch": 10.444444444444445,
"grad_norm": 0.267801433801651,
"learning_rate": 2.7468520827592197e-05,
"loss": 0.0938,
"num_input_tokens_seen": 294496,
"step": 940
},
{
"epoch": 10.5,
"grad_norm": 0.007512570358812809,
"learning_rate": 2.7227186323197162e-05,
"loss": 0.0801,
"num_input_tokens_seen": 296096,
"step": 945
},
{
"epoch": 10.555555555555555,
"grad_norm": 0.40150490403175354,
"learning_rate": 2.6985642425457757e-05,
"loss": 0.0221,
"num_input_tokens_seen": 297632,
"step": 950
},
{
"epoch": 10.61111111111111,
"grad_norm": 0.025159059092402458,
"learning_rate": 2.674391184360313e-05,
"loss": 0.0133,
"num_input_tokens_seen": 299168,
"step": 955
},
{
"epoch": 10.666666666666666,
"grad_norm": 0.008369713090360165,
"learning_rate": 2.650201730441392e-05,
"loss": 0.0809,
"num_input_tokens_seen": 300736,
"step": 960
},
{
"epoch": 10.722222222222221,
"grad_norm": 1.6953821182250977,
"learning_rate": 2.6259981550085504e-05,
"loss": 0.057,
"num_input_tokens_seen": 302304,
"step": 965
},
{
"epoch": 10.777777777777779,
"grad_norm": 0.007949888706207275,
"learning_rate": 2.60178273360899e-05,
"loss": 0.0275,
"num_input_tokens_seen": 303936,
"step": 970
},
{
"epoch": 10.833333333333334,
"grad_norm": 0.9999786615371704,
"learning_rate": 2.5775577429036345e-05,
"loss": 0.0349,
"num_input_tokens_seen": 305504,
"step": 975
},
{
"epoch": 10.88888888888889,
"grad_norm": 0.0133050000295043,
"learning_rate": 2.553325460453086e-05,
"loss": 0.0128,
"num_input_tokens_seen": 307040,
"step": 980
},
{
"epoch": 10.944444444444445,
"grad_norm": 0.1516413390636444,
"learning_rate": 2.5290881645034932e-05,
"loss": 0.0584,
"num_input_tokens_seen": 308608,
"step": 985
},
{
"epoch": 11.0,
"grad_norm": 0.009461171925067902,
"learning_rate": 2.504848133772358e-05,
"loss": 0.0279,
"num_input_tokens_seen": 310176,
"step": 990
},
{
"epoch": 11.0,
"eval_loss": 0.06158857420086861,
"eval_runtime": 0.5197,
"eval_samples_per_second": 76.973,
"eval_steps_per_second": 19.243,
"num_input_tokens_seen": 310176,
"step": 990
},
{
"epoch": 11.055555555555555,
"grad_norm": 1.5115296840667725,
"learning_rate": 2.4806076472342997e-05,
"loss": 0.0621,
"num_input_tokens_seen": 311712,
"step": 995
},
{
"epoch": 11.11111111111111,
"grad_norm": 0.05599701404571533,
"learning_rate": 2.4563689839067913e-05,
"loss": 0.0141,
"num_input_tokens_seen": 313280,
"step": 1000
},
{
"epoch": 11.166666666666666,
"grad_norm": 0.10229451209306717,
"learning_rate": 2.432134422635893e-05,
"loss": 0.0907,
"num_input_tokens_seen": 314816,
"step": 1005
},
{
"epoch": 11.222222222222221,
"grad_norm": 0.2486252337694168,
"learning_rate": 2.4079062418820002e-05,
"loss": 0.015,
"num_input_tokens_seen": 316384,
"step": 1010
},
{
"epoch": 11.277777777777779,
"grad_norm": 0.03698575124144554,
"learning_rate": 2.3836867195056335e-05,
"loss": 0.0709,
"num_input_tokens_seen": 317984,
"step": 1015
},
{
"epoch": 11.333333333333334,
"grad_norm": 0.07409622520208359,
"learning_rate": 2.3594781325532784e-05,
"loss": 0.0857,
"num_input_tokens_seen": 319552,
"step": 1020
},
{
"epoch": 11.38888888888889,
"grad_norm": 0.5309990644454956,
"learning_rate": 2.3352827570433036e-05,
"loss": 0.0184,
"num_input_tokens_seen": 321184,
"step": 1025
},
{
"epoch": 11.444444444444445,
"grad_norm": 0.579908013343811,
"learning_rate": 2.3111028677519804e-05,
"loss": 0.0399,
"num_input_tokens_seen": 322688,
"step": 1030
},
{
"epoch": 11.5,
"grad_norm": 0.022151276469230652,
"learning_rate": 2.2869407379996088e-05,
"loss": 0.1395,
"num_input_tokens_seen": 324224,
"step": 1035
},
{
"epoch": 11.555555555555555,
"grad_norm": 1.0108978748321533,
"learning_rate": 2.2627986394367938e-05,
"loss": 0.0416,
"num_input_tokens_seen": 325760,
"step": 1040
},
{
"epoch": 11.61111111111111,
"grad_norm": 1.5363922119140625,
"learning_rate": 2.238678841830867e-05,
"loss": 0.0488,
"num_input_tokens_seen": 327264,
"step": 1045
},
{
"epoch": 11.666666666666666,
"grad_norm": 0.5866866111755371,
"learning_rate": 2.2145836128524902e-05,
"loss": 0.0255,
"num_input_tokens_seen": 328832,
"step": 1050
},
{
"epoch": 11.722222222222221,
"grad_norm": 0.2689912021160126,
"learning_rate": 2.1905152178624595e-05,
"loss": 0.0301,
"num_input_tokens_seen": 330464,
"step": 1055
},
{
"epoch": 11.777777777777779,
"grad_norm": 0.5232972502708435,
"learning_rate": 2.1664759196987182e-05,
"loss": 0.0164,
"num_input_tokens_seen": 332064,
"step": 1060
},
{
"epoch": 11.833333333333334,
"grad_norm": 0.07716183364391327,
"learning_rate": 2.1424679784636144e-05,
"loss": 0.0351,
"num_input_tokens_seen": 333664,
"step": 1065
},
{
"epoch": 11.88888888888889,
"grad_norm": 0.031836166977882385,
"learning_rate": 2.118493651311413e-05,
"loss": 0.063,
"num_input_tokens_seen": 335232,
"step": 1070
},
{
"epoch": 11.944444444444445,
"grad_norm": 0.04803520813584328,
"learning_rate": 2.0945551922360818e-05,
"loss": 0.0262,
"num_input_tokens_seen": 336800,
"step": 1075
},
{
"epoch": 12.0,
"grad_norm": 1.2660475969314575,
"learning_rate": 2.070654851859383e-05,
"loss": 0.0731,
"num_input_tokens_seen": 338400,
"step": 1080
},
{
"epoch": 12.0,
"eval_loss": 0.05893953889608383,
"eval_runtime": 0.5187,
"eval_samples_per_second": 77.115,
"eval_steps_per_second": 19.279,
"num_input_tokens_seen": 338400,
"step": 1080
},
{
"epoch": 12.055555555555555,
"grad_norm": 0.30764204263687134,
"learning_rate": 2.0467948772192713e-05,
"loss": 0.1018,
"num_input_tokens_seen": 339968,
"step": 1085
},
{
"epoch": 12.11111111111111,
"grad_norm": 0.33647072315216064,
"learning_rate": 2.022977511558638e-05,
"loss": 0.0977,
"num_input_tokens_seen": 341536,
"step": 1090
},
{
"epoch": 12.166666666666666,
"grad_norm": 0.13600827753543854,
"learning_rate": 1.9992049941144066e-05,
"loss": 0.01,
"num_input_tokens_seen": 343104,
"step": 1095
},
{
"epoch": 12.222222222222221,
"grad_norm": 0.48333939909935,
"learning_rate": 1.9754795599070068e-05,
"loss": 0.0303,
"num_input_tokens_seen": 344640,
"step": 1100
},
{
"epoch": 12.277777777777779,
"grad_norm": 0.03402300924062729,
"learning_rate": 1.9518034395302414e-05,
"loss": 0.1015,
"num_input_tokens_seen": 346272,
"step": 1105
},
{
"epoch": 12.333333333333334,
"grad_norm": 0.15194876492023468,
"learning_rate": 1.9281788589415804e-05,
"loss": 0.0407,
"num_input_tokens_seen": 347808,
"step": 1110
},
{
"epoch": 12.38888888888889,
"grad_norm": 0.19185632467269897,
"learning_rate": 1.9046080392528735e-05,
"loss": 0.0237,
"num_input_tokens_seen": 349344,
"step": 1115
},
{
"epoch": 12.444444444444445,
"grad_norm": 0.250827431678772,
"learning_rate": 1.8810931965215356e-05,
"loss": 0.0122,
"num_input_tokens_seen": 350912,
"step": 1120
},
{
"epoch": 12.5,
"grad_norm": 0.19316622614860535,
"learning_rate": 1.857636541542195e-05,
"loss": 0.0727,
"num_input_tokens_seen": 352480,
"step": 1125
},
{
"epoch": 12.555555555555555,
"grad_norm": 0.9355464577674866,
"learning_rate": 1.8342402796388445e-05,
"loss": 0.0253,
"num_input_tokens_seen": 354048,
"step": 1130
},
{
"epoch": 12.61111111111111,
"grad_norm": 0.007416177541017532,
"learning_rate": 1.8109066104575023e-05,
"loss": 0.0045,
"num_input_tokens_seen": 355648,
"step": 1135
},
{
"epoch": 12.666666666666666,
"grad_norm": 0.9557670950889587,
"learning_rate": 1.7876377277594053e-05,
"loss": 0.033,
"num_input_tokens_seen": 357248,
"step": 1140
},
{
"epoch": 12.722222222222221,
"grad_norm": 0.011720170266926289,
"learning_rate": 1.764435819214762e-05,
"loss": 0.0273,
"num_input_tokens_seen": 358752,
"step": 1145
},
{
"epoch": 12.777777777777779,
"grad_norm": 0.2780408561229706,
"learning_rate": 1.7413030661970742e-05,
"loss": 0.0187,
"num_input_tokens_seen": 360288,
"step": 1150
},
{
"epoch": 12.833333333333334,
"grad_norm": 0.01299345213919878,
"learning_rate": 1.7182416435780454e-05,
"loss": 0.0163,
"num_input_tokens_seen": 361920,
"step": 1155
},
{
"epoch": 12.88888888888889,
"grad_norm": 0.010123813524842262,
"learning_rate": 1.695253719523115e-05,
"loss": 0.146,
"num_input_tokens_seen": 363520,
"step": 1160
},
{
"epoch": 12.944444444444445,
"grad_norm": 1.4064712524414062,
"learning_rate": 1.672341455287605e-05,
"loss": 0.0515,
"num_input_tokens_seen": 365088,
"step": 1165
},
{
"epoch": 13.0,
"grad_norm": 0.013315699063241482,
"learning_rate": 1.649507005013532e-05,
"loss": 0.0441,
"num_input_tokens_seen": 366688,
"step": 1170
},
{
"epoch": 13.0,
"eval_loss": 0.05899040028452873,
"eval_runtime": 0.5203,
"eval_samples_per_second": 76.873,
"eval_steps_per_second": 19.218,
"num_input_tokens_seen": 366688,
"step": 1170
},
{
"epoch": 13.055555555555555,
"grad_norm": 1.644558310508728,
"learning_rate": 1.6267525155270773e-05,
"loss": 0.0723,
"num_input_tokens_seen": 368192,
"step": 1175
},
{
"epoch": 13.11111111111111,
"grad_norm": 0.33990561962127686,
"learning_rate": 1.6040801261367493e-05,
"loss": 0.1022,
"num_input_tokens_seen": 369792,
"step": 1180
},
{
"epoch": 13.166666666666666,
"grad_norm": 0.10223418474197388,
"learning_rate": 1.5814919684322545e-05,
"loss": 0.0832,
"num_input_tokens_seen": 371360,
"step": 1185
},
{
"epoch": 13.222222222222221,
"grad_norm": 0.014776123687624931,
"learning_rate": 1.5589901660840896e-05,
"loss": 0.0552,
"num_input_tokens_seen": 372928,
"step": 1190
},
{
"epoch": 13.277777777777779,
"grad_norm": 0.13640986382961273,
"learning_rate": 1.5365768346438797e-05,
"loss": 0.0299,
"num_input_tokens_seen": 374528,
"step": 1195
},
{
"epoch": 13.333333333333334,
"grad_norm": 0.2882007956504822,
"learning_rate": 1.5142540813454836e-05,
"loss": 0.0198,
"num_input_tokens_seen": 376000,
"step": 1200
},
{
"epoch": 13.38888888888889,
"grad_norm": 0.2431965470314026,
"learning_rate": 1.4920240049068748e-05,
"loss": 0.0179,
"num_input_tokens_seen": 377600,
"step": 1205
},
{
"epoch": 13.444444444444445,
"grad_norm": 0.20751990377902985,
"learning_rate": 1.4698886953328292e-05,
"loss": 0.0664,
"num_input_tokens_seen": 379200,
"step": 1210
},
{
"epoch": 13.5,
"grad_norm": 0.022774390876293182,
"learning_rate": 1.4478502337184274e-05,
"loss": 0.0527,
"num_input_tokens_seen": 380704,
"step": 1215
},
{
"epoch": 13.555555555555555,
"grad_norm": 0.11283931881189346,
"learning_rate": 1.4259106920533955e-05,
"loss": 0.0598,
"num_input_tokens_seen": 382272,
"step": 1220
},
{
"epoch": 13.61111111111111,
"grad_norm": 0.818737268447876,
"learning_rate": 1.4040721330273062e-05,
"loss": 0.0219,
"num_input_tokens_seen": 383776,
"step": 1225
},
{
"epoch": 13.666666666666666,
"grad_norm": 0.33635279536247253,
"learning_rate": 1.3823366098356487e-05,
"loss": 0.0543,
"num_input_tokens_seen": 385344,
"step": 1230
},
{
"epoch": 13.722222222222221,
"grad_norm": 0.5322341322898865,
"learning_rate": 1.3607061659867892e-05,
"loss": 0.015,
"num_input_tokens_seen": 386912,
"step": 1235
},
{
"epoch": 13.777777777777779,
"grad_norm": 0.007620496675372124,
"learning_rate": 1.3391828351098578e-05,
"loss": 0.0776,
"num_input_tokens_seen": 388480,
"step": 1240
},
{
"epoch": 13.833333333333334,
"grad_norm": 0.013653988018631935,
"learning_rate": 1.3177686407635417e-05,
"loss": 0.0359,
"num_input_tokens_seen": 390048,
"step": 1245
},
{
"epoch": 13.88888888888889,
"grad_norm": 0.031345516443252563,
"learning_rate": 1.29646559624584e-05,
"loss": 0.0188,
"num_input_tokens_seen": 391584,
"step": 1250
},
{
"epoch": 13.944444444444445,
"grad_norm": 0.10659394413232803,
"learning_rate": 1.2752757044047827e-05,
"loss": 0.0245,
"num_input_tokens_seen": 393184,
"step": 1255
},
{
"epoch": 14.0,
"grad_norm": 0.27042534947395325,
"learning_rate": 1.2542009574501246e-05,
"loss": 0.0081,
"num_input_tokens_seen": 394752,
"step": 1260
},
{
"epoch": 14.0,
"eval_loss": 0.060102712363004684,
"eval_runtime": 0.5256,
"eval_samples_per_second": 76.098,
"eval_steps_per_second": 19.024,
"num_input_tokens_seen": 394752,
"step": 1260
},
{
"epoch": 14.055555555555555,
"grad_norm": 0.10951791703701019,
"learning_rate": 1.2332433367660442e-05,
"loss": 0.0423,
"num_input_tokens_seen": 396320,
"step": 1265
},
{
"epoch": 14.11111111111111,
"grad_norm": 0.028836172074079514,
"learning_rate": 1.2124048127248644e-05,
"loss": 0.0312,
"num_input_tokens_seen": 397920,
"step": 1270
},
{
"epoch": 14.166666666666666,
"grad_norm": 0.01687157340347767,
"learning_rate": 1.1916873445017982e-05,
"loss": 0.0153,
"num_input_tokens_seen": 399552,
"step": 1275
},
{
"epoch": 14.222222222222221,
"grad_norm": 0.02062399312853813,
"learning_rate": 1.1710928798907556e-05,
"loss": 0.0118,
"num_input_tokens_seen": 401120,
"step": 1280
},
{
"epoch": 14.277777777777779,
"grad_norm": 1.0031055212020874,
"learning_rate": 1.1506233551212186e-05,
"loss": 0.0415,
"num_input_tokens_seen": 402720,
"step": 1285
},
{
"epoch": 14.333333333333334,
"grad_norm": 0.011867760680615902,
"learning_rate": 1.1302806946762004e-05,
"loss": 0.0264,
"num_input_tokens_seen": 404320,
"step": 1290
},
{
"epoch": 14.38888888888889,
"grad_norm": 0.024303486570715904,
"learning_rate": 1.1100668111113166e-05,
"loss": 0.0123,
"num_input_tokens_seen": 405824,
"step": 1295
},
{
"epoch": 14.444444444444445,
"grad_norm": 1.8085826635360718,
"learning_rate": 1.0899836048749645e-05,
"loss": 0.0914,
"num_input_tokens_seen": 407360,
"step": 1300
},
{
"epoch": 14.5,
"grad_norm": 1.3068459033966064,
"learning_rate": 1.0700329641296541e-05,
"loss": 0.1156,
"num_input_tokens_seen": 408928,
"step": 1305
},
{
"epoch": 14.555555555555555,
"grad_norm": 0.3057030141353607,
"learning_rate": 1.0502167645744895e-05,
"loss": 0.0315,
"num_input_tokens_seen": 410496,
"step": 1310
},
{
"epoch": 14.61111111111111,
"grad_norm": 0.23634877800941467,
"learning_rate": 1.0305368692688174e-05,
"loss": 0.0124,
"num_input_tokens_seen": 412032,
"step": 1315
},
{
"epoch": 14.666666666666666,
"grad_norm": 0.15040916204452515,
"learning_rate": 1.01099512845707e-05,
"loss": 0.0142,
"num_input_tokens_seen": 413536,
"step": 1320
},
{
"epoch": 14.722222222222221,
"grad_norm": 0.03269388899207115,
"learning_rate": 9.91593379394811e-06,
"loss": 0.0331,
"num_input_tokens_seen": 415136,
"step": 1325
},
{
"epoch": 14.777777777777779,
"grad_norm": 0.396207332611084,
"learning_rate": 9.723334461760006e-06,
"loss": 0.1666,
"num_input_tokens_seen": 416704,
"step": 1330
},
{
"epoch": 14.833333333333334,
"grad_norm": 0.46826860308647156,
"learning_rate": 9.532171395615036e-06,
"loss": 0.0218,
"num_input_tokens_seen": 418272,
"step": 1335
},
{
"epoch": 14.88888888888889,
"grad_norm": 0.05246613547205925,
"learning_rate": 9.342462568088416e-06,
"loss": 0.0856,
"num_input_tokens_seen": 419808,
"step": 1340
},
{
"epoch": 14.944444444444445,
"grad_norm": 0.008573368191719055,
"learning_rate": 9.154225815032242e-06,
"loss": 0.0261,
"num_input_tokens_seen": 421344,
"step": 1345
},
{
"epoch": 15.0,
"grad_norm": 0.03231954202055931,
"learning_rate": 8.967478833898612e-06,
"loss": 0.0287,
"num_input_tokens_seen": 422912,
"step": 1350
},
{
"epoch": 15.0,
"eval_loss": 0.05984850972890854,
"eval_runtime": 0.517,
"eval_samples_per_second": 77.37,
"eval_steps_per_second": 19.343,
"num_input_tokens_seen": 422912,
"step": 1350
},
{
"epoch": 15.055555555555555,
"grad_norm": 0.8410363793373108,
"learning_rate": 8.78223918207575e-06,
"loss": 0.0211,
"num_input_tokens_seen": 424480,
"step": 1355
},
{
"epoch": 15.11111111111111,
"grad_norm": 0.8376926183700562,
"learning_rate": 8.598524275237322e-06,
"loss": 0.0395,
"num_input_tokens_seen": 426048,
"step": 1360
},
{
"epoch": 15.166666666666666,
"grad_norm": 0.28883492946624756,
"learning_rate": 8.41635138570507e-06,
"loss": 0.0174,
"num_input_tokens_seen": 427648,
"step": 1365
},
{
"epoch": 15.222222222222221,
"grad_norm": 2.5785083770751953,
"learning_rate": 8.235737640824908e-06,
"loss": 0.1227,
"num_input_tokens_seen": 429216,
"step": 1370
},
{
"epoch": 15.277777777777779,
"grad_norm": 0.7716183066368103,
"learning_rate": 8.056700021356694e-06,
"loss": 0.0267,
"num_input_tokens_seen": 430720,
"step": 1375
},
{
"epoch": 15.333333333333334,
"grad_norm": 0.13386765122413635,
"learning_rate": 7.879255359877705e-06,
"loss": 0.0577,
"num_input_tokens_seen": 432320,
"step": 1380
},
{
"epoch": 15.38888888888889,
"grad_norm": 0.18355993926525116,
"learning_rate": 7.703420339200101e-06,
"loss": 0.0522,
"num_input_tokens_seen": 433888,
"step": 1385
},
{
"epoch": 15.444444444444445,
"grad_norm": 0.21487367153167725,
"learning_rate": 7.529211490802498e-06,
"loss": 0.0453,
"num_input_tokens_seen": 435456,
"step": 1390
},
{
"epoch": 15.5,
"grad_norm": 0.4584398567676544,
"learning_rate": 7.3566451932756744e-06,
"loss": 0.0124,
"num_input_tokens_seen": 437024,
"step": 1395
},
{
"epoch": 15.555555555555555,
"grad_norm": 0.16662505269050598,
"learning_rate": 7.185737670782727e-06,
"loss": 0.0052,
"num_input_tokens_seen": 438528,
"step": 1400
},
{
"epoch": 15.61111111111111,
"grad_norm": 0.3519705832004547,
"learning_rate": 7.016504991533726e-06,
"loss": 0.0061,
"num_input_tokens_seen": 440096,
"step": 1405
},
{
"epoch": 15.666666666666666,
"grad_norm": 0.2552229166030884,
"learning_rate": 6.848963066275027e-06,
"loss": 0.021,
"num_input_tokens_seen": 441632,
"step": 1410
},
{
"epoch": 15.722222222222221,
"grad_norm": 0.3842417895793915,
"learning_rate": 6.683127646793411e-06,
"loss": 0.0763,
"num_input_tokens_seen": 443136,
"step": 1415
},
{
"epoch": 15.777777777777779,
"grad_norm": 0.46216481924057007,
"learning_rate": 6.519014324435102e-06,
"loss": 0.0571,
"num_input_tokens_seen": 444736,
"step": 1420
},
{
"epoch": 15.833333333333334,
"grad_norm": 0.012727758847177029,
"learning_rate": 6.356638528639955e-06,
"loss": 0.009,
"num_input_tokens_seen": 446336,
"step": 1425
},
{
"epoch": 15.88888888888889,
"grad_norm": 0.027039675042033195,
"learning_rate": 6.196015525490825e-06,
"loss": 0.1455,
"num_input_tokens_seen": 447936,
"step": 1430
},
{
"epoch": 15.944444444444445,
"grad_norm": 1.3422183990478516,
"learning_rate": 6.037160416278278e-06,
"loss": 0.0474,
"num_input_tokens_seen": 449440,
"step": 1435
},
{
"epoch": 16.0,
"grad_norm": 0.03860368952155113,
"learning_rate": 5.880088136080814e-06,
"loss": 0.0147,
"num_input_tokens_seen": 451008,
"step": 1440
},
{
"epoch": 16.0,
"eval_loss": 0.05932573229074478,
"eval_runtime": 0.52,
"eval_samples_per_second": 76.927,
"eval_steps_per_second": 19.232,
"num_input_tokens_seen": 451008,
"step": 1440
},
{
"epoch": 16.055555555555557,
"grad_norm": 0.550543487071991,
"learning_rate": 5.724813452360736e-06,
"loss": 0.0636,
"num_input_tokens_seen": 452576,
"step": 1445
},
{
"epoch": 16.11111111111111,
"grad_norm": 1.2353315353393555,
"learning_rate": 5.571350963575728e-06,
"loss": 0.0968,
"num_input_tokens_seen": 454112,
"step": 1450
},
{
"epoch": 16.166666666666668,
"grad_norm": 0.019102858379483223,
"learning_rate": 5.4197150978063965e-06,
"loss": 0.014,
"num_input_tokens_seen": 455680,
"step": 1455
},
{
"epoch": 16.22222222222222,
"grad_norm": 0.14581513404846191,
"learning_rate": 5.269920111399732e-06,
"loss": 0.0105,
"num_input_tokens_seen": 457280,
"step": 1460
},
{
"epoch": 16.27777777777778,
"grad_norm": 0.0162353478372097,
"learning_rate": 5.121980087628803e-06,
"loss": 0.0084,
"num_input_tokens_seen": 458848,
"step": 1465
},
{
"epoch": 16.333333333333332,
"grad_norm": 0.17371828854084015,
"learning_rate": 4.975908935368701e-06,
"loss": 0.0291,
"num_input_tokens_seen": 460416,
"step": 1470
},
{
"epoch": 16.38888888888889,
"grad_norm": 0.4158192276954651,
"learning_rate": 4.831720387788827e-06,
"loss": 0.0344,
"num_input_tokens_seen": 462048,
"step": 1475
},
{
"epoch": 16.444444444444443,
"grad_norm": 0.0122003685683012,
"learning_rate": 4.689428001061774e-06,
"loss": 0.0166,
"num_input_tokens_seen": 463584,
"step": 1480
},
{
"epoch": 16.5,
"grad_norm": 0.004614563658833504,
"learning_rate": 4.549045153088813e-06,
"loss": 0.1125,
"num_input_tokens_seen": 465152,
"step": 1485
},
{
"epoch": 16.555555555555557,
"grad_norm": 0.48291656374931335,
"learning_rate": 4.410585042242124e-06,
"loss": 0.0491,
"num_input_tokens_seen": 466720,
"step": 1490
},
{
"epoch": 16.61111111111111,
"grad_norm": 0.3229008913040161,
"learning_rate": 4.274060686123959e-06,
"loss": 0.018,
"num_input_tokens_seen": 468256,
"step": 1495
},
{
"epoch": 16.666666666666668,
"grad_norm": 0.01832473650574684,
"learning_rate": 4.1394849203427284e-06,
"loss": 0.0542,
"num_input_tokens_seen": 469792,
"step": 1500
},
{
"epoch": 16.72222222222222,
"grad_norm": 0.07098887860774994,
"learning_rate": 4.006870397306256e-06,
"loss": 0.0182,
"num_input_tokens_seen": 471360,
"step": 1505
},
{
"epoch": 16.77777777777778,
"grad_norm": 0.02056993544101715,
"learning_rate": 3.876229585032245e-06,
"loss": 0.0149,
"num_input_tokens_seen": 472896,
"step": 1510
},
{
"epoch": 16.833333333333332,
"grad_norm": 0.3502405881881714,
"learning_rate": 3.7475747659760502e-06,
"loss": 0.0944,
"num_input_tokens_seen": 474496,
"step": 1515
},
{
"epoch": 16.88888888888889,
"grad_norm": 0.21067650616168976,
"learning_rate": 3.6209180358759394e-06,
"loss": 0.0155,
"num_input_tokens_seen": 476064,
"step": 1520
},
{
"epoch": 16.944444444444443,
"grad_norm": 0.05252571403980255,
"learning_rate": 3.4962713026158694e-06,
"loss": 0.0264,
"num_input_tokens_seen": 477600,
"step": 1525
},
{
"epoch": 17.0,
"grad_norm": 0.008412295952439308,
"learning_rate": 3.373646285105958e-06,
"loss": 0.0935,
"num_input_tokens_seen": 479104,
"step": 1530
},
{
"epoch": 17.0,
"eval_loss": 0.05843701213598251,
"eval_runtime": 0.5198,
"eval_samples_per_second": 76.957,
"eval_steps_per_second": 19.239,
"num_input_tokens_seen": 479104,
"step": 1530
},
{
"epoch": 17.055555555555557,
"grad_norm": 0.21587851643562317,
"learning_rate": 3.2530545121807145e-06,
"loss": 0.0435,
"num_input_tokens_seen": 480736,
"step": 1535
},
{
"epoch": 17.11111111111111,
"grad_norm": 0.3278423547744751,
"learning_rate": 3.1345073215151066e-06,
"loss": 0.0471,
"num_input_tokens_seen": 482304,
"step": 1540
},
{
"epoch": 17.166666666666668,
"grad_norm": 0.10238057374954224,
"learning_rate": 3.0180158585586397e-06,
"loss": 0.0383,
"num_input_tokens_seen": 483872,
"step": 1545
},
{
"epoch": 17.22222222222222,
"grad_norm": 0.17296543717384338,
"learning_rate": 2.9035910754875136e-06,
"loss": 0.0238,
"num_input_tokens_seen": 485408,
"step": 1550
},
{
"epoch": 17.27777777777778,
"grad_norm": 0.052419859915971756,
"learning_rate": 2.7912437301749026e-06,
"loss": 0.0439,
"num_input_tokens_seen": 486944,
"step": 1555
},
{
"epoch": 17.333333333333332,
"grad_norm": 0.04378189891576767,
"learning_rate": 2.6809843851795357e-06,
"loss": 0.0293,
"num_input_tokens_seen": 488544,
"step": 1560
},
{
"epoch": 17.38888888888889,
"grad_norm": 0.1454276144504547,
"learning_rate": 2.57282340675267e-06,
"loss": 0.0335,
"num_input_tokens_seen": 490144,
"step": 1565
},
{
"epoch": 17.444444444444443,
"grad_norm": 0.029287101700901985,
"learning_rate": 2.4667709638634434e-06,
"loss": 0.0138,
"num_input_tokens_seen": 491680,
"step": 1570
},
{
"epoch": 17.5,
"grad_norm": 0.005346542689949274,
"learning_rate": 2.3628370272428564e-06,
"loss": 0.0756,
"num_input_tokens_seen": 493248,
"step": 1575
},
{
"epoch": 17.555555555555557,
"grad_norm": 0.03685900568962097,
"learning_rate": 2.2610313684463177e-06,
"loss": 0.0251,
"num_input_tokens_seen": 494816,
"step": 1580
},
{
"epoch": 17.61111111111111,
"grad_norm": 0.04156652092933655,
"learning_rate": 2.1613635589349756e-06,
"loss": 0.1015,
"num_input_tokens_seen": 496416,
"step": 1585
},
{
"epoch": 17.666666666666668,
"grad_norm": 0.8290051221847534,
"learning_rate": 2.063842969175847e-06,
"loss": 0.023,
"num_input_tokens_seen": 497952,
"step": 1590
},
{
"epoch": 17.72222222222222,
"grad_norm": 0.049550581723451614,
"learning_rate": 1.968478767760812e-06,
"loss": 0.0533,
"num_input_tokens_seen": 499520,
"step": 1595
},
{
"epoch": 17.77777777777778,
"grad_norm": 0.15296253561973572,
"learning_rate": 1.8752799205445982e-06,
"loss": 0.1171,
"num_input_tokens_seen": 501120,
"step": 1600
},
{
"epoch": 17.833333333333332,
"grad_norm": 0.030097996816039085,
"learning_rate": 1.784255189801895e-06,
"loss": 0.0103,
"num_input_tokens_seen": 502688,
"step": 1605
},
{
"epoch": 17.88888888888889,
"grad_norm": 0.13713747262954712,
"learning_rate": 1.6954131334034922e-06,
"loss": 0.0261,
"num_input_tokens_seen": 504224,
"step": 1610
},
{
"epoch": 17.944444444444443,
"grad_norm": 0.11297009140253067,
"learning_rate": 1.6087621040117157e-06,
"loss": 0.042,
"num_input_tokens_seen": 505824,
"step": 1615
},
{
"epoch": 18.0,
"grad_norm": 0.0667833760380745,
"learning_rate": 1.524310248295152e-06,
"loss": 0.0122,
"num_input_tokens_seen": 507392,
"step": 1620
},
{
"epoch": 18.0,
"eval_loss": 0.06013888865709305,
"eval_runtime": 0.5215,
"eval_samples_per_second": 76.703,
"eval_steps_per_second": 19.176,
"num_input_tokens_seen": 507392,
"step": 1620
},
{
"epoch": 18.055555555555557,
"grad_norm": 0.12299990653991699,
"learning_rate": 1.4420655061626932e-06,
"loss": 0.0209,
"num_input_tokens_seen": 508928,
"step": 1625
},
{
"epoch": 18.11111111111111,
"grad_norm": 0.20416291058063507,
"learning_rate": 1.362035610017079e-06,
"loss": 0.0254,
"num_input_tokens_seen": 510496,
"step": 1630
},
{
"epoch": 18.166666666666668,
"grad_norm": 0.17090032994747162,
"learning_rate": 1.2842280840278997e-06,
"loss": 0.0594,
"num_input_tokens_seen": 512032,
"step": 1635
},
{
"epoch": 18.22222222222222,
"grad_norm": 0.1019781306385994,
"learning_rate": 1.2086502434241865e-06,
"loss": 0.0074,
"num_input_tokens_seen": 513504,
"step": 1640
},
{
"epoch": 18.27777777777778,
"grad_norm": 0.6727584600448608,
"learning_rate": 1.1353091938067023e-06,
"loss": 0.0281,
"num_input_tokens_seen": 515040,
"step": 1645
},
{
"epoch": 18.333333333333332,
"grad_norm": 0.01017342135310173,
"learning_rate": 1.0642118304798442e-06,
"loss": 0.0396,
"num_input_tokens_seen": 516640,
"step": 1650
},
{
"epoch": 18.38888888888889,
"grad_norm": 0.8693645000457764,
"learning_rate": 9.95364837803392e-07,
"loss": 0.0489,
"num_input_tokens_seen": 518208,
"step": 1655
},
{
"epoch": 18.444444444444443,
"grad_norm": 0.8891027569770813,
"learning_rate": 9.287746885640603e-07,
"loss": 0.0264,
"num_input_tokens_seen": 519776,
"step": 1660
},
{
"epoch": 18.5,
"grad_norm": 0.015380547381937504,
"learning_rate": 8.64447643366953e-07,
"loss": 0.0109,
"num_input_tokens_seen": 521344,
"step": 1665
},
{
"epoch": 18.555555555555557,
"grad_norm": 0.007733294274657965,
"learning_rate": 8.023897500469391e-07,
"loss": 0.1493,
"num_input_tokens_seen": 522976,
"step": 1670
},
{
"epoch": 18.61111111111111,
"grad_norm": 0.9056934714317322,
"learning_rate": 7.426068431000882e-07,
"loss": 0.0268,
"num_input_tokens_seen": 524544,
"step": 1675
},
{
"epoch": 18.666666666666668,
"grad_norm": 0.00778258265927434,
"learning_rate": 6.851045431350927e-07,
"loss": 0.0046,
"num_input_tokens_seen": 526112,
"step": 1680
},
{
"epoch": 18.72222222222222,
"grad_norm": 0.9855499267578125,
"learning_rate": 6.298882563448599e-07,
"loss": 0.041,
"num_input_tokens_seen": 527712,
"step": 1685
},
{
"epoch": 18.77777777777778,
"grad_norm": 1.730617642402649,
"learning_rate": 5.769631739982267e-07,
"loss": 0.0657,
"num_input_tokens_seen": 529280,
"step": 1690
},
{
"epoch": 18.833333333333332,
"grad_norm": 0.15165819227695465,
"learning_rate": 5.263342719518921e-07,
"loss": 0.0513,
"num_input_tokens_seen": 530816,
"step": 1695
},
{
"epoch": 18.88888888888889,
"grad_norm": 0.014716766774654388,
"learning_rate": 4.780063101826132e-07,
"loss": 0.0439,
"num_input_tokens_seen": 532448,
"step": 1700
},
{
"epoch": 18.944444444444443,
"grad_norm": 0.15044061839580536,
"learning_rate": 4.319838323396691e-07,
"loss": 0.0397,
"num_input_tokens_seen": 534016,
"step": 1705
},
{
"epoch": 19.0,
"grad_norm": 0.14136840403079987,
"learning_rate": 3.88271165317694e-07,
"loss": 0.0685,
"num_input_tokens_seen": 535584,
"step": 1710
},
{
"epoch": 19.0,
"eval_loss": 0.06143023446202278,
"eval_runtime": 0.5205,
"eval_samples_per_second": 76.847,
"eval_steps_per_second": 19.212,
"num_input_tokens_seen": 535584,
"step": 1710
},
{
"epoch": 19.055555555555557,
"grad_norm": 0.21940721571445465,
"learning_rate": 3.468724188498751e-07,
"loss": 0.1495,
"num_input_tokens_seen": 537152,
"step": 1715
},
{
"epoch": 19.11111111111111,
"grad_norm": 0.17824964225292206,
"learning_rate": 3.077914851215585e-07,
"loss": 0.0176,
"num_input_tokens_seen": 538720,
"step": 1720
},
{
"epoch": 19.166666666666668,
"grad_norm": 0.67586749792099,
"learning_rate": 2.71032038404323e-07,
"loss": 0.0203,
"num_input_tokens_seen": 540256,
"step": 1725
},
{
"epoch": 19.22222222222222,
"grad_norm": 0.008334583602845669,
"learning_rate": 2.365975347105448e-07,
"loss": 0.008,
"num_input_tokens_seen": 541824,
"step": 1730
},
{
"epoch": 19.27777777777778,
"grad_norm": 0.030415670946240425,
"learning_rate": 2.0449121146845774e-07,
"loss": 0.0081,
"num_input_tokens_seen": 543424,
"step": 1735
},
{
"epoch": 19.333333333333332,
"grad_norm": 1.8074421882629395,
"learning_rate": 1.747160872177883e-07,
"loss": 0.0494,
"num_input_tokens_seen": 545024,
"step": 1740
},
{
"epoch": 19.38888888888889,
"grad_norm": 0.3716520071029663,
"learning_rate": 1.472749613259661e-07,
"loss": 0.0258,
"num_input_tokens_seen": 546560,
"step": 1745
},
{
"epoch": 19.444444444444443,
"grad_norm": 0.03234966844320297,
"learning_rate": 1.22170413724923e-07,
"loss": 0.0316,
"num_input_tokens_seen": 548096,
"step": 1750
},
{
"epoch": 19.5,
"grad_norm": 0.0194929838180542,
"learning_rate": 9.940480466855417e-08,
"loss": 0.0542,
"num_input_tokens_seen": 549664,
"step": 1755
},
{
"epoch": 19.555555555555557,
"grad_norm": 0.012016446329653263,
"learning_rate": 7.898027451078982e-08,
"loss": 0.0125,
"num_input_tokens_seen": 551264,
"step": 1760
},
{
"epoch": 19.61111111111111,
"grad_norm": 0.2182668149471283,
"learning_rate": 6.089874350439506e-08,
"loss": 0.0641,
"num_input_tokens_seen": 552800,
"step": 1765
},
{
"epoch": 19.666666666666668,
"grad_norm": 1.409957766532898,
"learning_rate": 4.516191162040051e-08,
"loss": 0.0581,
"num_input_tokens_seen": 554336,
"step": 1770
},
{
"epoch": 19.72222222222222,
"grad_norm": 0.10802645236253738,
"learning_rate": 3.177125838830786e-08,
"loss": 0.0233,
"num_input_tokens_seen": 555872,
"step": 1775
},
{
"epoch": 19.77777777777778,
"grad_norm": 0.21580937504768372,
"learning_rate": 2.0728042756967824e-08,
"loss": 0.0767,
"num_input_tokens_seen": 557504,
"step": 1780
},
{
"epoch": 19.833333333333332,
"grad_norm": 0.03298446908593178,
"learning_rate": 1.2033302976222071e-08,
"loss": 0.0158,
"num_input_tokens_seen": 559040,
"step": 1785
},
{
"epoch": 19.88888888888889,
"grad_norm": 0.00999500136822462,
"learning_rate": 5.687856499297928e-09,
"loss": 0.0513,
"num_input_tokens_seen": 560640,
"step": 1790
},
{
"epoch": 19.944444444444443,
"grad_norm": 1.7499030828475952,
"learning_rate": 1.692299905944883e-09,
"loss": 0.0649,
"num_input_tokens_seen": 562176,
"step": 1795
},
{
"epoch": 20.0,
"grad_norm": 0.4583163857460022,
"learning_rate": 4.700884634611935e-11,
"loss": 0.0292,
"num_input_tokens_seen": 563744,
"step": 1800
},
{
"epoch": 20.0,
"eval_loss": 0.060703717172145844,
"eval_runtime": 0.519,
"eval_samples_per_second": 77.074,
"eval_steps_per_second": 19.269,
"num_input_tokens_seen": 563744,
"step": 1800
},
{
"epoch": 20.0,
"num_input_tokens_seen": 563744,
"step": 1800,
"total_flos": 2.538580254523392e+16,
"train_loss": 0.10106630020671421,
"train_runtime": 280.5961,
"train_samples_per_second": 25.66,
"train_steps_per_second": 6.415
}
],
"logging_steps": 5,
"max_steps": 1800,
"num_input_tokens_seen": 563744,
"num_train_epochs": 20,
"save_steps": 90,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2.538580254523392e+16,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}