openthoughts3_math_100k / trainer_state.json
sedrickkeh's picture
End of training
28cfc23 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 4.980443285528032,
"eval_steps": 500,
"global_step": 955,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.005215123859191656,
"grad_norm": 8.640251838986332,
"learning_rate": 8.333333333333333e-07,
"loss": 1.3109,
"step": 1
},
{
"epoch": 0.010430247718383311,
"grad_norm": 8.50165895125568,
"learning_rate": 1.6666666666666667e-06,
"loss": 1.2827,
"step": 2
},
{
"epoch": 0.01564537157757497,
"grad_norm": 8.535323811460232,
"learning_rate": 2.5e-06,
"loss": 1.2971,
"step": 3
},
{
"epoch": 0.020860495436766623,
"grad_norm": 8.005773755772996,
"learning_rate": 3.3333333333333333e-06,
"loss": 1.2835,
"step": 4
},
{
"epoch": 0.02607561929595828,
"grad_norm": 6.542532290299446,
"learning_rate": 4.166666666666667e-06,
"loss": 1.2424,
"step": 5
},
{
"epoch": 0.03129074315514994,
"grad_norm": 3.513863098870053,
"learning_rate": 5e-06,
"loss": 1.1855,
"step": 6
},
{
"epoch": 0.03650586701434159,
"grad_norm": 2.7215053265085425,
"learning_rate": 5.833333333333334e-06,
"loss": 1.1771,
"step": 7
},
{
"epoch": 0.041720990873533245,
"grad_norm": 6.45835816206034,
"learning_rate": 6.666666666666667e-06,
"loss": 1.1818,
"step": 8
},
{
"epoch": 0.0469361147327249,
"grad_norm": 6.896643054762279,
"learning_rate": 7.500000000000001e-06,
"loss": 1.1927,
"step": 9
},
{
"epoch": 0.05215123859191656,
"grad_norm": 7.106049193713323,
"learning_rate": 8.333333333333334e-06,
"loss": 1.192,
"step": 10
},
{
"epoch": 0.05736636245110821,
"grad_norm": 5.8848429339749755,
"learning_rate": 9.166666666666666e-06,
"loss": 1.1358,
"step": 11
},
{
"epoch": 0.06258148631029987,
"grad_norm": 5.16279034442262,
"learning_rate": 1e-05,
"loss": 1.1187,
"step": 12
},
{
"epoch": 0.06779661016949153,
"grad_norm": 3.4057013372590914,
"learning_rate": 1.0833333333333334e-05,
"loss": 1.1044,
"step": 13
},
{
"epoch": 0.07301173402868318,
"grad_norm": 2.2087972161452516,
"learning_rate": 1.1666666666666668e-05,
"loss": 1.0906,
"step": 14
},
{
"epoch": 0.07822685788787484,
"grad_norm": 2.1528478641437006,
"learning_rate": 1.25e-05,
"loss": 1.0749,
"step": 15
},
{
"epoch": 0.08344198174706649,
"grad_norm": 2.0837468590704984,
"learning_rate": 1.3333333333333333e-05,
"loss": 1.041,
"step": 16
},
{
"epoch": 0.08865710560625815,
"grad_norm": 2.3435856245272064,
"learning_rate": 1.416666666666667e-05,
"loss": 1.0308,
"step": 17
},
{
"epoch": 0.0938722294654498,
"grad_norm": 1.7734881445436932,
"learning_rate": 1.5000000000000002e-05,
"loss": 1.0104,
"step": 18
},
{
"epoch": 0.09908735332464146,
"grad_norm": 1.2852782680220982,
"learning_rate": 1.5833333333333333e-05,
"loss": 1.0159,
"step": 19
},
{
"epoch": 0.10430247718383312,
"grad_norm": 1.4664459559013807,
"learning_rate": 1.6666666666666667e-05,
"loss": 0.9997,
"step": 20
},
{
"epoch": 0.10951760104302477,
"grad_norm": 1.2747674832880032,
"learning_rate": 1.7500000000000002e-05,
"loss": 0.9999,
"step": 21
},
{
"epoch": 0.11473272490221642,
"grad_norm": 1.1325966855930794,
"learning_rate": 1.8333333333333333e-05,
"loss": 0.9845,
"step": 22
},
{
"epoch": 0.11994784876140809,
"grad_norm": 1.178213874446251,
"learning_rate": 1.916666666666667e-05,
"loss": 0.9754,
"step": 23
},
{
"epoch": 0.12516297262059975,
"grad_norm": 1.0070300787437625,
"learning_rate": 2e-05,
"loss": 0.9916,
"step": 24
},
{
"epoch": 0.1303780964797914,
"grad_norm": 1.1301814564159125,
"learning_rate": 2.0833333333333336e-05,
"loss": 0.9535,
"step": 25
},
{
"epoch": 0.13559322033898305,
"grad_norm": 1.1582102527170561,
"learning_rate": 2.1666666666666667e-05,
"loss": 0.9559,
"step": 26
},
{
"epoch": 0.1408083441981747,
"grad_norm": 1.28096007382199,
"learning_rate": 2.25e-05,
"loss": 0.953,
"step": 27
},
{
"epoch": 0.14602346805736635,
"grad_norm": 1.3963901299642703,
"learning_rate": 2.3333333333333336e-05,
"loss": 0.9467,
"step": 28
},
{
"epoch": 0.15123859191655803,
"grad_norm": 1.2008437237669338,
"learning_rate": 2.4166666666666667e-05,
"loss": 0.9319,
"step": 29
},
{
"epoch": 0.15645371577574968,
"grad_norm": 1.4579402450445522,
"learning_rate": 2.5e-05,
"loss": 0.9583,
"step": 30
},
{
"epoch": 0.16166883963494133,
"grad_norm": 1.332654985022459,
"learning_rate": 2.5833333333333336e-05,
"loss": 0.9222,
"step": 31
},
{
"epoch": 0.16688396349413298,
"grad_norm": 1.0118249538528512,
"learning_rate": 2.6666666666666667e-05,
"loss": 0.919,
"step": 32
},
{
"epoch": 0.17209908735332463,
"grad_norm": 1.9244594897055562,
"learning_rate": 2.75e-05,
"loss": 0.9432,
"step": 33
},
{
"epoch": 0.1773142112125163,
"grad_norm": 1.1988103983333642,
"learning_rate": 2.833333333333334e-05,
"loss": 0.9309,
"step": 34
},
{
"epoch": 0.18252933507170796,
"grad_norm": 1.375838275777245,
"learning_rate": 2.9166666666666666e-05,
"loss": 0.9164,
"step": 35
},
{
"epoch": 0.1877444589308996,
"grad_norm": 1.662174214128309,
"learning_rate": 3.0000000000000004e-05,
"loss": 0.9116,
"step": 36
},
{
"epoch": 0.19295958279009126,
"grad_norm": 19.880826748141878,
"learning_rate": 3.0833333333333335e-05,
"loss": 0.9136,
"step": 37
},
{
"epoch": 0.1981747066492829,
"grad_norm": 2.1089494461311897,
"learning_rate": 3.1666666666666666e-05,
"loss": 0.9288,
"step": 38
},
{
"epoch": 0.2033898305084746,
"grad_norm": 1.0754107081344952,
"learning_rate": 3.2500000000000004e-05,
"loss": 0.9173,
"step": 39
},
{
"epoch": 0.20860495436766624,
"grad_norm": 3.9611232218120964,
"learning_rate": 3.3333333333333335e-05,
"loss": 0.9227,
"step": 40
},
{
"epoch": 0.2138200782268579,
"grad_norm": 3.0041123063486306,
"learning_rate": 3.4166666666666666e-05,
"loss": 0.9319,
"step": 41
},
{
"epoch": 0.21903520208604954,
"grad_norm": 2.7847797531368066,
"learning_rate": 3.5000000000000004e-05,
"loss": 0.9233,
"step": 42
},
{
"epoch": 0.2242503259452412,
"grad_norm": 2.2845773931740725,
"learning_rate": 3.5833333333333335e-05,
"loss": 0.9286,
"step": 43
},
{
"epoch": 0.22946544980443284,
"grad_norm": 2.9550586599454363,
"learning_rate": 3.6666666666666666e-05,
"loss": 0.9194,
"step": 44
},
{
"epoch": 0.23468057366362452,
"grad_norm": 1.713180532466099,
"learning_rate": 3.7500000000000003e-05,
"loss": 0.8909,
"step": 45
},
{
"epoch": 0.23989569752281617,
"grad_norm": 3.6245500815283984,
"learning_rate": 3.833333333333334e-05,
"loss": 0.9061,
"step": 46
},
{
"epoch": 0.24511082138200782,
"grad_norm": 2.7780460507158034,
"learning_rate": 3.9166666666666665e-05,
"loss": 0.9165,
"step": 47
},
{
"epoch": 0.2503259452411995,
"grad_norm": 3.3854708521313563,
"learning_rate": 4e-05,
"loss": 0.9076,
"step": 48
},
{
"epoch": 0.25554106910039115,
"grad_norm": 2.788766938646062,
"learning_rate": 4.0833333333333334e-05,
"loss": 0.9086,
"step": 49
},
{
"epoch": 0.2607561929595828,
"grad_norm": 3.769154968802035,
"learning_rate": 4.166666666666667e-05,
"loss": 0.8985,
"step": 50
},
{
"epoch": 0.26597131681877445,
"grad_norm": 3.6676032273552783,
"learning_rate": 4.25e-05,
"loss": 0.9186,
"step": 51
},
{
"epoch": 0.2711864406779661,
"grad_norm": 2.300855067051272,
"learning_rate": 4.3333333333333334e-05,
"loss": 0.8785,
"step": 52
},
{
"epoch": 0.27640156453715775,
"grad_norm": 2.0855212942964156,
"learning_rate": 4.416666666666667e-05,
"loss": 0.9044,
"step": 53
},
{
"epoch": 0.2816166883963494,
"grad_norm": 3.1541260770808144,
"learning_rate": 4.5e-05,
"loss": 0.9134,
"step": 54
},
{
"epoch": 0.28683181225554105,
"grad_norm": 2.3439141139233635,
"learning_rate": 4.5833333333333334e-05,
"loss": 0.8849,
"step": 55
},
{
"epoch": 0.2920469361147327,
"grad_norm": 3.622853136869387,
"learning_rate": 4.666666666666667e-05,
"loss": 0.8968,
"step": 56
},
{
"epoch": 0.29726205997392435,
"grad_norm": 3.3444739650293793,
"learning_rate": 4.75e-05,
"loss": 0.9119,
"step": 57
},
{
"epoch": 0.30247718383311606,
"grad_norm": 2.4056743976018007,
"learning_rate": 4.8333333333333334e-05,
"loss": 0.8978,
"step": 58
},
{
"epoch": 0.3076923076923077,
"grad_norm": 2.0165924530855692,
"learning_rate": 4.916666666666667e-05,
"loss": 0.8842,
"step": 59
},
{
"epoch": 0.31290743155149936,
"grad_norm": 3.199274379857031,
"learning_rate": 5e-05,
"loss": 0.8926,
"step": 60
},
{
"epoch": 0.318122555410691,
"grad_norm": 2.145560466910737,
"learning_rate": 5.0833333333333333e-05,
"loss": 0.8875,
"step": 61
},
{
"epoch": 0.32333767926988266,
"grad_norm": 3.3364603622456923,
"learning_rate": 5.166666666666667e-05,
"loss": 0.8891,
"step": 62
},
{
"epoch": 0.3285528031290743,
"grad_norm": 3.088241155388676,
"learning_rate": 5.25e-05,
"loss": 0.8989,
"step": 63
},
{
"epoch": 0.33376792698826596,
"grad_norm": 2.2703016974586836,
"learning_rate": 5.333333333333333e-05,
"loss": 0.8848,
"step": 64
},
{
"epoch": 0.3389830508474576,
"grad_norm": 2.1056335353722875,
"learning_rate": 5.416666666666667e-05,
"loss": 0.8895,
"step": 65
},
{
"epoch": 0.34419817470664926,
"grad_norm": 2.195474617684906,
"learning_rate": 5.5e-05,
"loss": 0.8855,
"step": 66
},
{
"epoch": 0.3494132985658409,
"grad_norm": 1.8388111784516368,
"learning_rate": 5.583333333333333e-05,
"loss": 0.88,
"step": 67
},
{
"epoch": 0.3546284224250326,
"grad_norm": 2.205537218893776,
"learning_rate": 5.666666666666668e-05,
"loss": 0.8923,
"step": 68
},
{
"epoch": 0.35984354628422427,
"grad_norm": 1.43087259231813,
"learning_rate": 5.75e-05,
"loss": 0.8783,
"step": 69
},
{
"epoch": 0.3650586701434159,
"grad_norm": 3.170207545618331,
"learning_rate": 5.833333333333333e-05,
"loss": 0.8979,
"step": 70
},
{
"epoch": 0.37027379400260757,
"grad_norm": 2.4965183591219686,
"learning_rate": 5.916666666666668e-05,
"loss": 0.887,
"step": 71
},
{
"epoch": 0.3754889178617992,
"grad_norm": 2.18025601493941,
"learning_rate": 6.000000000000001e-05,
"loss": 0.8885,
"step": 72
},
{
"epoch": 0.38070404172099087,
"grad_norm": 2.320595439694094,
"learning_rate": 6.083333333333333e-05,
"loss": 0.8815,
"step": 73
},
{
"epoch": 0.3859191655801825,
"grad_norm": 2.0185818109580467,
"learning_rate": 6.166666666666667e-05,
"loss": 0.8831,
"step": 74
},
{
"epoch": 0.39113428943937417,
"grad_norm": 2.2105832623786745,
"learning_rate": 6.25e-05,
"loss": 0.8666,
"step": 75
},
{
"epoch": 0.3963494132985658,
"grad_norm": 1.9340187658703214,
"learning_rate": 6.333333333333333e-05,
"loss": 0.888,
"step": 76
},
{
"epoch": 0.4015645371577575,
"grad_norm": 1.811786813233452,
"learning_rate": 6.416666666666668e-05,
"loss": 0.889,
"step": 77
},
{
"epoch": 0.4067796610169492,
"grad_norm": 2.0985830559238563,
"learning_rate": 6.500000000000001e-05,
"loss": 0.8945,
"step": 78
},
{
"epoch": 0.41199478487614083,
"grad_norm": 2.7706210707741885,
"learning_rate": 6.583333333333334e-05,
"loss": 0.903,
"step": 79
},
{
"epoch": 0.4172099087353325,
"grad_norm": 2.328367128642788,
"learning_rate": 6.666666666666667e-05,
"loss": 0.886,
"step": 80
},
{
"epoch": 0.42242503259452413,
"grad_norm": 2.7310201710261857,
"learning_rate": 6.75e-05,
"loss": 0.8728,
"step": 81
},
{
"epoch": 0.4276401564537158,
"grad_norm": 2.012060559904938,
"learning_rate": 6.833333333333333e-05,
"loss": 0.8807,
"step": 82
},
{
"epoch": 0.43285528031290743,
"grad_norm": 1.8442023724268992,
"learning_rate": 6.916666666666668e-05,
"loss": 0.8832,
"step": 83
},
{
"epoch": 0.4380704041720991,
"grad_norm": 3.3879560119725225,
"learning_rate": 7.000000000000001e-05,
"loss": 0.8848,
"step": 84
},
{
"epoch": 0.44328552803129073,
"grad_norm": 2.0302673957424457,
"learning_rate": 7.083333333333334e-05,
"loss": 0.8716,
"step": 85
},
{
"epoch": 0.4485006518904824,
"grad_norm": 3.238439147300149,
"learning_rate": 7.166666666666667e-05,
"loss": 0.8896,
"step": 86
},
{
"epoch": 0.45371577574967403,
"grad_norm": 2.718682181237891,
"learning_rate": 7.25e-05,
"loss": 0.902,
"step": 87
},
{
"epoch": 0.4589308996088657,
"grad_norm": 2.4102500849494035,
"learning_rate": 7.333333333333333e-05,
"loss": 0.8865,
"step": 88
},
{
"epoch": 0.4641460234680574,
"grad_norm": 2.057159117257536,
"learning_rate": 7.416666666666668e-05,
"loss": 0.8714,
"step": 89
},
{
"epoch": 0.46936114732724904,
"grad_norm": 2.303058894217337,
"learning_rate": 7.500000000000001e-05,
"loss": 0.875,
"step": 90
},
{
"epoch": 0.4745762711864407,
"grad_norm": 1.853131768077206,
"learning_rate": 7.583333333333334e-05,
"loss": 0.8697,
"step": 91
},
{
"epoch": 0.47979139504563234,
"grad_norm": 1.5089115884426068,
"learning_rate": 7.666666666666668e-05,
"loss": 0.8672,
"step": 92
},
{
"epoch": 0.485006518904824,
"grad_norm": 3.288728672540363,
"learning_rate": 7.75e-05,
"loss": 0.8787,
"step": 93
},
{
"epoch": 0.49022164276401564,
"grad_norm": 2.8218607718403197,
"learning_rate": 7.833333333333333e-05,
"loss": 0.8886,
"step": 94
},
{
"epoch": 0.4954367666232073,
"grad_norm": 1.966995373674031,
"learning_rate": 7.916666666666668e-05,
"loss": 0.8738,
"step": 95
},
{
"epoch": 0.500651890482399,
"grad_norm": 4.193763943813693,
"learning_rate": 8e-05,
"loss": 0.8813,
"step": 96
},
{
"epoch": 0.5058670143415906,
"grad_norm": 2.2253144313432514,
"learning_rate": 7.99997324882088e-05,
"loss": 0.8778,
"step": 97
},
{
"epoch": 0.5110821382007823,
"grad_norm": 4.059040973559077,
"learning_rate": 7.999892995641334e-05,
"loss": 0.8844,
"step": 98
},
{
"epoch": 0.516297262059974,
"grad_norm": 2.814215374527903,
"learning_rate": 7.999759241534794e-05,
"loss": 0.8899,
"step": 99
},
{
"epoch": 0.5215123859191656,
"grad_norm": 3.1446586623159383,
"learning_rate": 7.9995719882903e-05,
"loss": 0.8933,
"step": 100
},
{
"epoch": 0.5267275097783573,
"grad_norm": 2.5831777088882335,
"learning_rate": 7.999331238412474e-05,
"loss": 0.885,
"step": 101
},
{
"epoch": 0.5319426336375489,
"grad_norm": 2.571515738465628,
"learning_rate": 7.99903699512149e-05,
"loss": 0.8773,
"step": 102
},
{
"epoch": 0.5371577574967406,
"grad_norm": 2.028582558881839,
"learning_rate": 7.998689262353024e-05,
"loss": 0.875,
"step": 103
},
{
"epoch": 0.5423728813559322,
"grad_norm": 1.9833279742128762,
"learning_rate": 7.998288044758206e-05,
"loss": 0.8864,
"step": 104
},
{
"epoch": 0.5475880052151239,
"grad_norm": 1.595344120443328,
"learning_rate": 7.99783334770356e-05,
"loss": 0.8772,
"step": 105
},
{
"epoch": 0.5528031290743155,
"grad_norm": 2.3698581665478025,
"learning_rate": 7.997325177270926e-05,
"loss": 0.8587,
"step": 106
},
{
"epoch": 0.5580182529335072,
"grad_norm": 1.8069354168584124,
"learning_rate": 7.996763540257382e-05,
"loss": 0.8786,
"step": 107
},
{
"epoch": 0.5632333767926988,
"grad_norm": 2.8802931859293968,
"learning_rate": 7.996148444175155e-05,
"loss": 0.8651,
"step": 108
},
{
"epoch": 0.5684485006518905,
"grad_norm": 2.163396645739927,
"learning_rate": 7.99547989725152e-05,
"loss": 0.8779,
"step": 109
},
{
"epoch": 0.5736636245110821,
"grad_norm": 1.7103752366008835,
"learning_rate": 7.994757908428683e-05,
"loss": 0.8627,
"step": 110
},
{
"epoch": 0.5788787483702738,
"grad_norm": 2.6236911355349366,
"learning_rate": 7.99398248736367e-05,
"loss": 0.8775,
"step": 111
},
{
"epoch": 0.5840938722294654,
"grad_norm": 1.7695200626233123,
"learning_rate": 7.993153644428198e-05,
"loss": 0.8675,
"step": 112
},
{
"epoch": 0.5893089960886571,
"grad_norm": 3.1189819236023597,
"learning_rate": 7.992271390708529e-05,
"loss": 0.864,
"step": 113
},
{
"epoch": 0.5945241199478487,
"grad_norm": 2.519671466442974,
"learning_rate": 7.991335738005325e-05,
"loss": 0.8753,
"step": 114
},
{
"epoch": 0.5997392438070405,
"grad_norm": 2.445885184903824,
"learning_rate": 7.990346698833493e-05,
"loss": 0.8791,
"step": 115
},
{
"epoch": 0.6049543676662321,
"grad_norm": 2.6107239711114634,
"learning_rate": 7.989304286422016e-05,
"loss": 0.8688,
"step": 116
},
{
"epoch": 0.6101694915254238,
"grad_norm": 2.3280917526561926,
"learning_rate": 7.988208514713773e-05,
"loss": 0.8558,
"step": 117
},
{
"epoch": 0.6153846153846154,
"grad_norm": 2.01951998260186,
"learning_rate": 7.987059398365358e-05,
"loss": 0.8611,
"step": 118
},
{
"epoch": 0.6205997392438071,
"grad_norm": 2.67003693356188,
"learning_rate": 7.98585695274688e-05,
"loss": 0.8599,
"step": 119
},
{
"epoch": 0.6258148631029987,
"grad_norm": 1.3989184556669023,
"learning_rate": 7.984601193941757e-05,
"loss": 0.8619,
"step": 120
},
{
"epoch": 0.6310299869621904,
"grad_norm": 2.9657726548979007,
"learning_rate": 7.983292138746504e-05,
"loss": 0.8533,
"step": 121
},
{
"epoch": 0.636245110821382,
"grad_norm": 2.2111378936066717,
"learning_rate": 7.981929804670505e-05,
"loss": 0.8647,
"step": 122
},
{
"epoch": 0.6414602346805737,
"grad_norm": 1.7933021378995146,
"learning_rate": 7.980514209935783e-05,
"loss": 0.8572,
"step": 123
},
{
"epoch": 0.6466753585397653,
"grad_norm": 2.6146019104563036,
"learning_rate": 7.97904537347675e-05,
"loss": 0.8629,
"step": 124
},
{
"epoch": 0.651890482398957,
"grad_norm": 1.7226667254129786,
"learning_rate": 7.977523314939961e-05,
"loss": 0.8728,
"step": 125
},
{
"epoch": 0.6571056062581486,
"grad_norm": 2.4681433772246804,
"learning_rate": 7.975948054683847e-05,
"loss": 0.8772,
"step": 126
},
{
"epoch": 0.6623207301173403,
"grad_norm": 1.8531960262958524,
"learning_rate": 7.974319613778441e-05,
"loss": 0.8568,
"step": 127
},
{
"epoch": 0.6675358539765319,
"grad_norm": 3.3273564870681067,
"learning_rate": 7.972638014005102e-05,
"loss": 0.8674,
"step": 128
},
{
"epoch": 0.6727509778357236,
"grad_norm": 1.6565232049947463,
"learning_rate": 7.970903277856216e-05,
"loss": 0.8593,
"step": 129
},
{
"epoch": 0.6779661016949152,
"grad_norm": 3.2772231491550157,
"learning_rate": 7.969115428534904e-05,
"loss": 0.8689,
"step": 130
},
{
"epoch": 0.6831812255541069,
"grad_norm": 2.9366197326510175,
"learning_rate": 7.967274489954703e-05,
"loss": 0.8562,
"step": 131
},
{
"epoch": 0.6883963494132985,
"grad_norm": 1.579970961599737,
"learning_rate": 7.965380486739253e-05,
"loss": 0.8422,
"step": 132
},
{
"epoch": 0.6936114732724902,
"grad_norm": 3.1594495489774967,
"learning_rate": 7.963433444221964e-05,
"loss": 0.875,
"step": 133
},
{
"epoch": 0.6988265971316818,
"grad_norm": 2.3403703396896423,
"learning_rate": 7.961433388445676e-05,
"loss": 0.8591,
"step": 134
},
{
"epoch": 0.7040417209908736,
"grad_norm": 1.8764160357986022,
"learning_rate": 7.959380346162314e-05,
"loss": 0.8585,
"step": 135
},
{
"epoch": 0.7092568448500652,
"grad_norm": 2.954668141615744,
"learning_rate": 7.957274344832533e-05,
"loss": 0.8558,
"step": 136
},
{
"epoch": 0.7144719687092569,
"grad_norm": 2.1132431617150647,
"learning_rate": 7.955115412625337e-05,
"loss": 0.8597,
"step": 137
},
{
"epoch": 0.7196870925684485,
"grad_norm": 1.9269531529548758,
"learning_rate": 7.952903578417719e-05,
"loss": 0.8553,
"step": 138
},
{
"epoch": 0.7249022164276402,
"grad_norm": 1.9600506890917169,
"learning_rate": 7.950638871794268e-05,
"loss": 0.8498,
"step": 139
},
{
"epoch": 0.7301173402868318,
"grad_norm": 1.295131341774357,
"learning_rate": 7.948321323046766e-05,
"loss": 0.862,
"step": 140
},
{
"epoch": 0.7353324641460235,
"grad_norm": 2.598262766807541,
"learning_rate": 7.945950963173797e-05,
"loss": 0.8743,
"step": 141
},
{
"epoch": 0.7405475880052151,
"grad_norm": 1.7173227811522496,
"learning_rate": 7.943527823880321e-05,
"loss": 0.857,
"step": 142
},
{
"epoch": 0.7457627118644068,
"grad_norm": 2.778103098366881,
"learning_rate": 7.941051937577255e-05,
"loss": 0.8644,
"step": 143
},
{
"epoch": 0.7509778357235984,
"grad_norm": 1.9721804979121416,
"learning_rate": 7.938523337381036e-05,
"loss": 0.8641,
"step": 144
},
{
"epoch": 0.7561929595827901,
"grad_norm": 2.534467842230536,
"learning_rate": 7.935942057113185e-05,
"loss": 0.8607,
"step": 145
},
{
"epoch": 0.7614080834419817,
"grad_norm": 2.420990126278799,
"learning_rate": 7.933308131299846e-05,
"loss": 0.8532,
"step": 146
},
{
"epoch": 0.7666232073011734,
"grad_norm": 2.0250401094262775,
"learning_rate": 7.93062159517133e-05,
"loss": 0.8463,
"step": 147
},
{
"epoch": 0.771838331160365,
"grad_norm": 1.949074222131184,
"learning_rate": 7.92788248466164e-05,
"loss": 0.8505,
"step": 148
},
{
"epoch": 0.7770534550195567,
"grad_norm": 1.9136794648235018,
"learning_rate": 7.925090836407997e-05,
"loss": 0.8539,
"step": 149
},
{
"epoch": 0.7822685788787483,
"grad_norm": 1.3862902439859852,
"learning_rate": 7.922246687750341e-05,
"loss": 0.8511,
"step": 150
},
{
"epoch": 0.78748370273794,
"grad_norm": 2.238512517144782,
"learning_rate": 7.919350076730836e-05,
"loss": 0.846,
"step": 151
},
{
"epoch": 0.7926988265971316,
"grad_norm": 2.0835619443377134,
"learning_rate": 7.916401042093361e-05,
"loss": 0.8461,
"step": 152
},
{
"epoch": 0.7979139504563233,
"grad_norm": 1.8060011410999197,
"learning_rate": 7.913399623282997e-05,
"loss": 0.833,
"step": 153
},
{
"epoch": 0.803129074315515,
"grad_norm": 2.0703223704187357,
"learning_rate": 7.910345860445487e-05,
"loss": 0.8447,
"step": 154
},
{
"epoch": 0.8083441981747066,
"grad_norm": 2.299135311289539,
"learning_rate": 7.90723979442671e-05,
"loss": 0.8533,
"step": 155
},
{
"epoch": 0.8135593220338984,
"grad_norm": 1.8730518820522357,
"learning_rate": 7.90408146677213e-05,
"loss": 0.8372,
"step": 156
},
{
"epoch": 0.81877444589309,
"grad_norm": 1.5402936247797026,
"learning_rate": 7.900870919726244e-05,
"loss": 0.8408,
"step": 157
},
{
"epoch": 0.8239895697522817,
"grad_norm": 2.378224013864058,
"learning_rate": 7.897608196232007e-05,
"loss": 0.8492,
"step": 158
},
{
"epoch": 0.8292046936114733,
"grad_norm": 1.8241375846314414,
"learning_rate": 7.894293339930272e-05,
"loss": 0.8338,
"step": 159
},
{
"epoch": 0.834419817470665,
"grad_norm": 2.2711975231391826,
"learning_rate": 7.890926395159197e-05,
"loss": 0.8385,
"step": 160
},
{
"epoch": 0.8396349413298566,
"grad_norm": 1.6195285561704085,
"learning_rate": 7.887507406953651e-05,
"loss": 0.8489,
"step": 161
},
{
"epoch": 0.8448500651890483,
"grad_norm": 2.2453846392163768,
"learning_rate": 7.884036421044618e-05,
"loss": 0.8487,
"step": 162
},
{
"epoch": 0.8500651890482399,
"grad_norm": 1.940408392169657,
"learning_rate": 7.880513483858583e-05,
"loss": 0.8398,
"step": 163
},
{
"epoch": 0.8552803129074316,
"grad_norm": 2.2159286823222675,
"learning_rate": 7.876938642516905e-05,
"loss": 0.8492,
"step": 164
},
{
"epoch": 0.8604954367666232,
"grad_norm": 1.4842300453687591,
"learning_rate": 7.873311944835195e-05,
"loss": 0.8376,
"step": 165
},
{
"epoch": 0.8657105606258149,
"grad_norm": 2.491755340230041,
"learning_rate": 7.869633439322674e-05,
"loss": 0.8386,
"step": 166
},
{
"epoch": 0.8709256844850065,
"grad_norm": 1.8127014745278216,
"learning_rate": 7.865903175181521e-05,
"loss": 0.8318,
"step": 167
},
{
"epoch": 0.8761408083441982,
"grad_norm": 1.5105514574695558,
"learning_rate": 7.862121202306217e-05,
"loss": 0.8317,
"step": 168
},
{
"epoch": 0.8813559322033898,
"grad_norm": 2.485520131584514,
"learning_rate": 7.858287571282882e-05,
"loss": 0.8496,
"step": 169
},
{
"epoch": 0.8865710560625815,
"grad_norm": 1.7516963763143856,
"learning_rate": 7.854402333388587e-05,
"loss": 0.8433,
"step": 170
},
{
"epoch": 0.8917861799217731,
"grad_norm": 2.462296826145561,
"learning_rate": 7.850465540590684e-05,
"loss": 0.8537,
"step": 171
},
{
"epoch": 0.8970013037809648,
"grad_norm": 1.1958019189200906,
"learning_rate": 7.846477245546094e-05,
"loss": 0.8323,
"step": 172
},
{
"epoch": 0.9022164276401564,
"grad_norm": 2.6026749193200693,
"learning_rate": 7.842437501600616e-05,
"loss": 0.8516,
"step": 173
},
{
"epoch": 0.9074315514993481,
"grad_norm": 1.7619131106705077,
"learning_rate": 7.838346362788206e-05,
"loss": 0.838,
"step": 174
},
{
"epoch": 0.9126466753585397,
"grad_norm": 2.4386182828768312,
"learning_rate": 7.834203883830259e-05,
"loss": 0.8444,
"step": 175
},
{
"epoch": 0.9178617992177314,
"grad_norm": 2.0969277412774274,
"learning_rate": 7.830010120134873e-05,
"loss": 0.8622,
"step": 176
},
{
"epoch": 0.9230769230769231,
"grad_norm": 1.4692953467652599,
"learning_rate": 7.825765127796108e-05,
"loss": 0.8338,
"step": 177
},
{
"epoch": 0.9282920469361148,
"grad_norm": 1.5923417027165594,
"learning_rate": 7.821468963593242e-05,
"loss": 0.8391,
"step": 178
},
{
"epoch": 0.9335071707953064,
"grad_norm": 2.2009607061094942,
"learning_rate": 7.817121684990004e-05,
"loss": 0.8589,
"step": 179
},
{
"epoch": 0.9387222946544981,
"grad_norm": 1.5116758680331852,
"learning_rate": 7.812723350133805e-05,
"loss": 0.8341,
"step": 180
},
{
"epoch": 0.9439374185136897,
"grad_norm": 2.038874732993399,
"learning_rate": 7.80827401785497e-05,
"loss": 0.8547,
"step": 181
},
{
"epoch": 0.9491525423728814,
"grad_norm": 2.2985530062759514,
"learning_rate": 7.80377374766594e-05,
"loss": 0.8391,
"step": 182
},
{
"epoch": 0.954367666232073,
"grad_norm": 1.257017121878146,
"learning_rate": 7.799222599760481e-05,
"loss": 0.8325,
"step": 183
},
{
"epoch": 0.9595827900912647,
"grad_norm": 2.873446007386103,
"learning_rate": 7.794620635012883e-05,
"loss": 0.853,
"step": 184
},
{
"epoch": 0.9647979139504563,
"grad_norm": 2.0836694911738207,
"learning_rate": 7.789967914977134e-05,
"loss": 0.8554,
"step": 185
},
{
"epoch": 0.970013037809648,
"grad_norm": 2.095712201456811,
"learning_rate": 7.785264501886108e-05,
"loss": 0.8439,
"step": 186
},
{
"epoch": 0.9752281616688396,
"grad_norm": 1.668972201159559,
"learning_rate": 7.78051045865073e-05,
"loss": 0.8442,
"step": 187
},
{
"epoch": 0.9804432855280313,
"grad_norm": 2.6548529934832787,
"learning_rate": 7.77570584885913e-05,
"loss": 0.8509,
"step": 188
},
{
"epoch": 0.9856584093872229,
"grad_norm": 1.6440679506158635,
"learning_rate": 7.770850736775796e-05,
"loss": 0.8367,
"step": 189
},
{
"epoch": 0.9908735332464146,
"grad_norm": 2.895095407046929,
"learning_rate": 7.765945187340715e-05,
"loss": 0.8476,
"step": 190
},
{
"epoch": 0.9960886571056062,
"grad_norm": 2.1880375444766655,
"learning_rate": 7.760989266168503e-05,
"loss": 0.8466,
"step": 191
},
{
"epoch": 1.001303780964798,
"grad_norm": 2.491373495658068,
"learning_rate": 7.755983039547528e-05,
"loss": 1.0483,
"step": 192
},
{
"epoch": 1.0065189048239895,
"grad_norm": 2.1065510338703195,
"learning_rate": 7.750926574439019e-05,
"loss": 0.8317,
"step": 193
},
{
"epoch": 1.0117340286831813,
"grad_norm": 1.6048527323347803,
"learning_rate": 7.745819938476184e-05,
"loss": 0.8309,
"step": 194
},
{
"epoch": 1.0169491525423728,
"grad_norm": 1.3686572978259008,
"learning_rate": 7.740663199963284e-05,
"loss": 0.8207,
"step": 195
},
{
"epoch": 1.0221642764015646,
"grad_norm": 1.6395534429827354,
"learning_rate": 7.73545642787474e-05,
"loss": 0.824,
"step": 196
},
{
"epoch": 1.0273794002607561,
"grad_norm": 1.829519017394267,
"learning_rate": 7.730199691854198e-05,
"loss": 0.8346,
"step": 197
},
{
"epoch": 1.032594524119948,
"grad_norm": 2.086222113236519,
"learning_rate": 7.724893062213602e-05,
"loss": 0.8204,
"step": 198
},
{
"epoch": 1.0378096479791394,
"grad_norm": 1.6543925285010792,
"learning_rate": 7.71953660993225e-05,
"loss": 0.8238,
"step": 199
},
{
"epoch": 1.0430247718383312,
"grad_norm": 1.6861891172176104,
"learning_rate": 7.71413040665585e-05,
"loss": 0.8215,
"step": 200
},
{
"epoch": 1.0482398956975227,
"grad_norm": 2.313274305017442,
"learning_rate": 7.708674524695559e-05,
"loss": 0.8082,
"step": 201
},
{
"epoch": 1.0534550195567145,
"grad_norm": 1.8334252249650502,
"learning_rate": 7.703169037027014e-05,
"loss": 0.824,
"step": 202
},
{
"epoch": 1.058670143415906,
"grad_norm": 1.7126012022750563,
"learning_rate": 7.697614017289357e-05,
"loss": 0.8105,
"step": 203
},
{
"epoch": 1.0638852672750978,
"grad_norm": 1.3173869406989056,
"learning_rate": 7.692009539784255e-05,
"loss": 0.8303,
"step": 204
},
{
"epoch": 1.0691003911342893,
"grad_norm": 2.0262932335464012,
"learning_rate": 7.686355679474898e-05,
"loss": 0.8197,
"step": 205
},
{
"epoch": 1.074315514993481,
"grad_norm": 2.2714312463405397,
"learning_rate": 7.680652511985e-05,
"loss": 0.8196,
"step": 206
},
{
"epoch": 1.0795306388526726,
"grad_norm": 1.5002234528402432,
"learning_rate": 7.674900113597787e-05,
"loss": 0.8141,
"step": 207
},
{
"epoch": 1.0847457627118644,
"grad_norm": 2.059909369219652,
"learning_rate": 7.669098561254983e-05,
"loss": 0.826,
"step": 208
},
{
"epoch": 1.0899608865710562,
"grad_norm": 1.6415840594311673,
"learning_rate": 7.663247932555767e-05,
"loss": 0.8314,
"step": 209
},
{
"epoch": 1.0951760104302477,
"grad_norm": 2.094580841761285,
"learning_rate": 7.65734830575575e-05,
"loss": 0.8405,
"step": 210
},
{
"epoch": 1.1003911342894395,
"grad_norm": 1.4276266650819935,
"learning_rate": 7.651399759765915e-05,
"loss": 0.8172,
"step": 211
},
{
"epoch": 1.105606258148631,
"grad_norm": 2.406763290618582,
"learning_rate": 7.645402374151575e-05,
"loss": 0.8166,
"step": 212
},
{
"epoch": 1.1108213820078228,
"grad_norm": 1.4071963692309266,
"learning_rate": 7.639356229131298e-05,
"loss": 0.8252,
"step": 213
},
{
"epoch": 1.1160365058670143,
"grad_norm": 1.434300709398784,
"learning_rate": 7.633261405575838e-05,
"loss": 0.8342,
"step": 214
},
{
"epoch": 1.121251629726206,
"grad_norm": 2.6026163076276223,
"learning_rate": 7.627117985007052e-05,
"loss": 0.8277,
"step": 215
},
{
"epoch": 1.1264667535853976,
"grad_norm": 1.545445917096449,
"learning_rate": 7.620926049596814e-05,
"loss": 0.8224,
"step": 216
},
{
"epoch": 1.1316818774445894,
"grad_norm": 2.349483330688553,
"learning_rate": 7.61468568216591e-05,
"loss": 0.8317,
"step": 217
},
{
"epoch": 1.136897001303781,
"grad_norm": 1.6493926282653744,
"learning_rate": 7.60839696618293e-05,
"loss": 0.8235,
"step": 218
},
{
"epoch": 1.1421121251629727,
"grad_norm": 2.523447089439544,
"learning_rate": 7.602059985763165e-05,
"loss": 0.8259,
"step": 219
},
{
"epoch": 1.1473272490221642,
"grad_norm": 1.910781874766756,
"learning_rate": 7.595674825667457e-05,
"loss": 0.826,
"step": 220
},
{
"epoch": 1.152542372881356,
"grad_norm": 2.2270989973834032,
"learning_rate": 7.589241571301091e-05,
"loss": 0.8309,
"step": 221
},
{
"epoch": 1.1577574967405475,
"grad_norm": 1.7934683365685906,
"learning_rate": 7.582760308712634e-05,
"loss": 0.8302,
"step": 222
},
{
"epoch": 1.1629726205997393,
"grad_norm": 2.0662257630292062,
"learning_rate": 7.5762311245928e-05,
"loss": 0.8277,
"step": 223
},
{
"epoch": 1.1681877444589308,
"grad_norm": 1.6790104250183842,
"learning_rate": 7.569654106273268e-05,
"loss": 0.8254,
"step": 224
},
{
"epoch": 1.1734028683181226,
"grad_norm": 2.11893755909352,
"learning_rate": 7.563029341725541e-05,
"loss": 0.8183,
"step": 225
},
{
"epoch": 1.1786179921773141,
"grad_norm": 1.5877658958807297,
"learning_rate": 7.55635691955975e-05,
"loss": 0.8232,
"step": 226
},
{
"epoch": 1.1838331160365059,
"grad_norm": 2.147915736111546,
"learning_rate": 7.549636929023471e-05,
"loss": 0.8185,
"step": 227
},
{
"epoch": 1.1890482398956976,
"grad_norm": 1.7400398545071791,
"learning_rate": 7.542869460000544e-05,
"loss": 0.8219,
"step": 228
},
{
"epoch": 1.1942633637548892,
"grad_norm": 1.8017174411766266,
"learning_rate": 7.536054603009856e-05,
"loss": 0.8099,
"step": 229
},
{
"epoch": 1.1994784876140807,
"grad_norm": 1.7295101992953383,
"learning_rate": 7.529192449204137e-05,
"loss": 0.8199,
"step": 230
},
{
"epoch": 1.2046936114732725,
"grad_norm": 1.81480851347367,
"learning_rate": 7.522283090368739e-05,
"loss": 0.8163,
"step": 231
},
{
"epoch": 1.2099087353324642,
"grad_norm": 1.4532183826025098,
"learning_rate": 7.515326618920409e-05,
"loss": 0.8243,
"step": 232
},
{
"epoch": 1.2151238591916558,
"grad_norm": 2.025221417951849,
"learning_rate": 7.508323127906055e-05,
"loss": 0.8104,
"step": 233
},
{
"epoch": 1.2203389830508475,
"grad_norm": 2.839580655118928,
"learning_rate": 7.5012727110015e-05,
"loss": 0.8156,
"step": 234
},
{
"epoch": 1.225554106910039,
"grad_norm": 0.9612813608575648,
"learning_rate": 7.494175462510225e-05,
"loss": 0.8136,
"step": 235
},
{
"epoch": 1.2307692307692308,
"grad_norm": 4.580627208479269,
"learning_rate": 7.487031477362112e-05,
"loss": 0.8363,
"step": 236
},
{
"epoch": 1.2359843546284224,
"grad_norm": 3.3062212088133585,
"learning_rate": 7.479840851112175e-05,
"loss": 0.8457,
"step": 237
},
{
"epoch": 1.2411994784876141,
"grad_norm": 3.9085448168352577,
"learning_rate": 7.47260367993928e-05,
"loss": 0.8369,
"step": 238
},
{
"epoch": 1.2464146023468057,
"grad_norm": 3.2990078411550416,
"learning_rate": 7.465320060644857e-05,
"loss": 0.83,
"step": 239
},
{
"epoch": 1.2516297262059974,
"grad_norm": 3.171914889017843,
"learning_rate": 7.45799009065161e-05,
"loss": 0.8245,
"step": 240
},
{
"epoch": 1.256844850065189,
"grad_norm": 2.6816905524834436,
"learning_rate": 7.450613868002208e-05,
"loss": 0.8363,
"step": 241
},
{
"epoch": 1.2620599739243807,
"grad_norm": 3.0890905030641256,
"learning_rate": 7.443191491357976e-05,
"loss": 0.823,
"step": 242
},
{
"epoch": 1.2672750977835723,
"grad_norm": 2.3490265350478956,
"learning_rate": 7.435723059997581e-05,
"loss": 0.8276,
"step": 243
},
{
"epoch": 1.272490221642764,
"grad_norm": 3.5496378180168957,
"learning_rate": 7.428208673815693e-05,
"loss": 0.8247,
"step": 244
},
{
"epoch": 1.2777053455019556,
"grad_norm": 3.053170271456407,
"learning_rate": 7.420648433321659e-05,
"loss": 0.8306,
"step": 245
},
{
"epoch": 1.2829204693611473,
"grad_norm": 2.387063335590222,
"learning_rate": 7.41304243963815e-05,
"loss": 0.831,
"step": 246
},
{
"epoch": 1.288135593220339,
"grad_norm": 1.8497576227055157,
"learning_rate": 7.405390794499819e-05,
"loss": 0.8066,
"step": 247
},
{
"epoch": 1.2933507170795306,
"grad_norm": 2.9897041114580727,
"learning_rate": 7.397693600251929e-05,
"loss": 0.8134,
"step": 248
},
{
"epoch": 1.2985658409387222,
"grad_norm": 1.980452439525295,
"learning_rate": 7.389950959848992e-05,
"loss": 0.8252,
"step": 249
},
{
"epoch": 1.303780964797914,
"grad_norm": 3.4714237886205748,
"learning_rate": 7.382162976853387e-05,
"loss": 0.8294,
"step": 250
},
{
"epoch": 1.3089960886571057,
"grad_norm": 3.2963021601322535,
"learning_rate": 7.37432975543398e-05,
"loss": 0.8232,
"step": 251
},
{
"epoch": 1.3142112125162972,
"grad_norm": 2.1372765339871806,
"learning_rate": 7.366451400364723e-05,
"loss": 0.8228,
"step": 252
},
{
"epoch": 1.3194263363754888,
"grad_norm": 1.61329278641058,
"learning_rate": 7.358528017023262e-05,
"loss": 0.8149,
"step": 253
},
{
"epoch": 1.3246414602346805,
"grad_norm": 2.149446564090082,
"learning_rate": 7.350559711389518e-05,
"loss": 0.803,
"step": 254
},
{
"epoch": 1.3298565840938723,
"grad_norm": 1.2080706850732208,
"learning_rate": 7.342546590044279e-05,
"loss": 0.8309,
"step": 255
},
{
"epoch": 1.3350717079530638,
"grad_norm": 2.060776426075165,
"learning_rate": 7.334488760167768e-05,
"loss": 0.8218,
"step": 256
},
{
"epoch": 1.3402868318122556,
"grad_norm": 1.4270579486920896,
"learning_rate": 7.326386329538207e-05,
"loss": 0.8482,
"step": 257
},
{
"epoch": 1.3455019556714471,
"grad_norm": 2.320654966360905,
"learning_rate": 7.318239406530386e-05,
"loss": 0.8284,
"step": 258
},
{
"epoch": 1.350717079530639,
"grad_norm": 2.040787679253322,
"learning_rate": 7.3100481001142e-05,
"loss": 0.8436,
"step": 259
},
{
"epoch": 1.3559322033898304,
"grad_norm": 1.450733455499567,
"learning_rate": 7.301812519853203e-05,
"loss": 0.8067,
"step": 260
},
{
"epoch": 1.3611473272490222,
"grad_norm": 1.4473957425879223,
"learning_rate": 7.293532775903137e-05,
"loss": 0.8172,
"step": 261
},
{
"epoch": 1.3663624511082137,
"grad_norm": 1.4567295321246068,
"learning_rate": 7.285208979010458e-05,
"loss": 0.833,
"step": 262
},
{
"epoch": 1.3715775749674055,
"grad_norm": 2.2303402593920234,
"learning_rate": 7.276841240510858e-05,
"loss": 0.8241,
"step": 263
},
{
"epoch": 1.376792698826597,
"grad_norm": 1.3757747352916987,
"learning_rate": 7.26842967232777e-05,
"loss": 0.8106,
"step": 264
},
{
"epoch": 1.3820078226857888,
"grad_norm": 1.960651210566042,
"learning_rate": 7.25997438697088e-05,
"loss": 0.8258,
"step": 265
},
{
"epoch": 1.3872229465449806,
"grad_norm": 1.6973151493940888,
"learning_rate": 7.251475497534615e-05,
"loss": 0.8421,
"step": 266
},
{
"epoch": 1.3924380704041721,
"grad_norm": 1.4278961151807503,
"learning_rate": 7.242933117696628e-05,
"loss": 0.8107,
"step": 267
},
{
"epoch": 1.3976531942633637,
"grad_norm": 1.6765492543318952,
"learning_rate": 7.234347361716291e-05,
"loss": 0.8126,
"step": 268
},
{
"epoch": 1.4028683181225554,
"grad_norm": 1.7901534079828416,
"learning_rate": 7.225718344433149e-05,
"loss": 0.8209,
"step": 269
},
{
"epoch": 1.4080834419817472,
"grad_norm": 1.1879344799911462,
"learning_rate": 7.217046181265394e-05,
"loss": 0.8241,
"step": 270
},
{
"epoch": 1.4132985658409387,
"grad_norm": 2.316592389288362,
"learning_rate": 7.208330988208324e-05,
"loss": 0.8233,
"step": 271
},
{
"epoch": 1.4185136897001303,
"grad_norm": 1.3009721963229537,
"learning_rate": 7.199572881832784e-05,
"loss": 0.8094,
"step": 272
},
{
"epoch": 1.423728813559322,
"grad_norm": 2.068602624965321,
"learning_rate": 7.190771979283608e-05,
"loss": 0.8221,
"step": 273
},
{
"epoch": 1.4289439374185138,
"grad_norm": 1.9940233989402643,
"learning_rate": 7.181928398278058e-05,
"loss": 0.8225,
"step": 274
},
{
"epoch": 1.4341590612777053,
"grad_norm": 1.7328843255622977,
"learning_rate": 7.173042257104243e-05,
"loss": 0.8142,
"step": 275
},
{
"epoch": 1.439374185136897,
"grad_norm": 1.5984687241828228,
"learning_rate": 7.164113674619542e-05,
"loss": 0.8062,
"step": 276
},
{
"epoch": 1.4445893089960886,
"grad_norm": 1.9941450850887774,
"learning_rate": 7.155142770249008e-05,
"loss": 0.8156,
"step": 277
},
{
"epoch": 1.4498044328552804,
"grad_norm": 1.2629214490257985,
"learning_rate": 7.146129663983775e-05,
"loss": 0.8029,
"step": 278
},
{
"epoch": 1.455019556714472,
"grad_norm": 1.5341743253202347,
"learning_rate": 7.137074476379454e-05,
"loss": 0.8184,
"step": 279
},
{
"epoch": 1.4602346805736637,
"grad_norm": 1.205052243506182,
"learning_rate": 7.127977328554518e-05,
"loss": 0.8297,
"step": 280
},
{
"epoch": 1.4654498044328552,
"grad_norm": 1.7584530348059961,
"learning_rate": 7.118838342188683e-05,
"loss": 0.8183,
"step": 281
},
{
"epoch": 1.470664928292047,
"grad_norm": 1.8223656535526023,
"learning_rate": 7.10965763952128e-05,
"loss": 0.8154,
"step": 282
},
{
"epoch": 1.4758800521512385,
"grad_norm": 1.3665387213451412,
"learning_rate": 7.100435343349617e-05,
"loss": 0.8114,
"step": 283
},
{
"epoch": 1.4810951760104303,
"grad_norm": 1.499553247270366,
"learning_rate": 7.091171577027344e-05,
"loss": 0.813,
"step": 284
},
{
"epoch": 1.4863102998696218,
"grad_norm": 2.1007419745906866,
"learning_rate": 7.081866464462798e-05,
"loss": 0.8216,
"step": 285
},
{
"epoch": 1.4915254237288136,
"grad_norm": 1.512999754469945,
"learning_rate": 7.072520130117344e-05,
"loss": 0.8182,
"step": 286
},
{
"epoch": 1.4967405475880051,
"grad_norm": 1.6975647047303697,
"learning_rate": 7.063132699003716e-05,
"loss": 0.8244,
"step": 287
},
{
"epoch": 1.5019556714471969,
"grad_norm": 1.7957380356125672,
"learning_rate": 7.053704296684337e-05,
"loss": 0.8162,
"step": 288
},
{
"epoch": 1.5071707953063886,
"grad_norm": 1.2813149409377043,
"learning_rate": 7.044235049269649e-05,
"loss": 0.8095,
"step": 289
},
{
"epoch": 1.5123859191655802,
"grad_norm": 1.56534754361993,
"learning_rate": 7.034725083416419e-05,
"loss": 0.8258,
"step": 290
},
{
"epoch": 1.5176010430247717,
"grad_norm": 1.7975943131313281,
"learning_rate": 7.025174526326045e-05,
"loss": 0.8025,
"step": 291
},
{
"epoch": 1.5228161668839635,
"grad_norm": 1.1853911402078072,
"learning_rate": 7.015583505742857e-05,
"loss": 0.8249,
"step": 292
},
{
"epoch": 1.5280312907431552,
"grad_norm": 1.8494955443278849,
"learning_rate": 7.005952149952416e-05,
"loss": 0.8378,
"step": 293
},
{
"epoch": 1.5332464146023468,
"grad_norm": 1.3622849536120247,
"learning_rate": 6.996280587779778e-05,
"loss": 0.8354,
"step": 294
},
{
"epoch": 1.5384615384615383,
"grad_norm": 2.0112831346634708,
"learning_rate": 6.986568948587792e-05,
"loss": 0.83,
"step": 295
},
{
"epoch": 1.54367666232073,
"grad_norm": 1.5420042484474124,
"learning_rate": 6.976817362275357e-05,
"loss": 0.8109,
"step": 296
},
{
"epoch": 1.5488917861799218,
"grad_norm": 2.0929864944342804,
"learning_rate": 6.96702595927569e-05,
"loss": 0.8334,
"step": 297
},
{
"epoch": 1.5541069100391134,
"grad_norm": 1.8922406476791735,
"learning_rate": 6.957194870554578e-05,
"loss": 0.806,
"step": 298
},
{
"epoch": 1.559322033898305,
"grad_norm": 1.6880868554685944,
"learning_rate": 6.947324227608628e-05,
"loss": 0.8212,
"step": 299
},
{
"epoch": 1.5645371577574967,
"grad_norm": 1.6346451668072235,
"learning_rate": 6.937414162463509e-05,
"loss": 0.8014,
"step": 300
},
{
"epoch": 1.5697522816166884,
"grad_norm": 1.577939054921778,
"learning_rate": 6.927464807672186e-05,
"loss": 0.8187,
"step": 301
},
{
"epoch": 1.57496740547588,
"grad_norm": 1.4316029714279948,
"learning_rate": 6.917476296313145e-05,
"loss": 0.8046,
"step": 302
},
{
"epoch": 1.5801825293350718,
"grad_norm": 1.5381955264209675,
"learning_rate": 6.907448761988612e-05,
"loss": 0.8077,
"step": 303
},
{
"epoch": 1.5853976531942635,
"grad_norm": 1.0983288794814963,
"learning_rate": 6.897382338822772e-05,
"loss": 0.804,
"step": 304
},
{
"epoch": 1.590612777053455,
"grad_norm": 1.2952124244327796,
"learning_rate": 6.88727716145997e-05,
"loss": 0.8014,
"step": 305
},
{
"epoch": 1.5958279009126466,
"grad_norm": 1.1248306945768556,
"learning_rate": 6.877133365062911e-05,
"loss": 0.8086,
"step": 306
},
{
"epoch": 1.6010430247718384,
"grad_norm": 1.7722044591752057,
"learning_rate": 6.86695108531085e-05,
"loss": 0.8061,
"step": 307
},
{
"epoch": 1.6062581486310301,
"grad_norm": 1.8332828396584284,
"learning_rate": 6.856730458397787e-05,
"loss": 0.8177,
"step": 308
},
{
"epoch": 1.6114732724902217,
"grad_norm": 1.3222660990882034,
"learning_rate": 6.846471621030626e-05,
"loss": 0.8012,
"step": 309
},
{
"epoch": 1.6166883963494132,
"grad_norm": 1.3548943859194524,
"learning_rate": 6.836174710427369e-05,
"loss": 0.8171,
"step": 310
},
{
"epoch": 1.621903520208605,
"grad_norm": 1.2347347259073755,
"learning_rate": 6.825839864315264e-05,
"loss": 0.7839,
"step": 311
},
{
"epoch": 1.6271186440677967,
"grad_norm": 1.4912678731047493,
"learning_rate": 6.815467220928972e-05,
"loss": 0.8004,
"step": 312
},
{
"epoch": 1.6323337679269883,
"grad_norm": 1.0089833155389005,
"learning_rate": 6.805056919008714e-05,
"loss": 0.806,
"step": 313
},
{
"epoch": 1.6375488917861798,
"grad_norm": 1.8923057463186084,
"learning_rate": 6.794609097798414e-05,
"loss": 0.8149,
"step": 314
},
{
"epoch": 1.6427640156453716,
"grad_norm": 1.3911432602736478,
"learning_rate": 6.784123897043841e-05,
"loss": 0.8261,
"step": 315
},
{
"epoch": 1.6479791395045633,
"grad_norm": 1.2299562475949994,
"learning_rate": 6.773601456990739e-05,
"loss": 0.8025,
"step": 316
},
{
"epoch": 1.6531942633637549,
"grad_norm": 1.749738991866774,
"learning_rate": 6.763041918382945e-05,
"loss": 0.8087,
"step": 317
},
{
"epoch": 1.6584093872229464,
"grad_norm": 1.2129819476495074,
"learning_rate": 6.752445422460513e-05,
"loss": 0.8058,
"step": 318
},
{
"epoch": 1.6636245110821382,
"grad_norm": 1.9218816457335506,
"learning_rate": 6.741812110957823e-05,
"loss": 0.8199,
"step": 319
},
{
"epoch": 1.66883963494133,
"grad_norm": 1.5263095695339857,
"learning_rate": 6.731142126101688e-05,
"loss": 0.8098,
"step": 320
},
{
"epoch": 1.6740547588005215,
"grad_norm": 1.7388018103606915,
"learning_rate": 6.720435610609443e-05,
"loss": 0.7964,
"step": 321
},
{
"epoch": 1.6792698826597132,
"grad_norm": 1.1143428753117783,
"learning_rate": 6.709692707687047e-05,
"loss": 0.8026,
"step": 322
},
{
"epoch": 1.684485006518905,
"grad_norm": 1.4518152074151873,
"learning_rate": 6.69891356102716e-05,
"loss": 0.8199,
"step": 323
},
{
"epoch": 1.6897001303780965,
"grad_norm": 1.2302866722544838,
"learning_rate": 6.688098314807221e-05,
"loss": 0.8116,
"step": 324
},
{
"epoch": 1.694915254237288,
"grad_norm": 1.355915097493558,
"learning_rate": 6.677247113687527e-05,
"loss": 0.8184,
"step": 325
},
{
"epoch": 1.7001303780964798,
"grad_norm": 1.0973851180923702,
"learning_rate": 6.666360102809289e-05,
"loss": 0.8066,
"step": 326
},
{
"epoch": 1.7053455019556716,
"grad_norm": 2.034989183783714,
"learning_rate": 6.655437427792698e-05,
"loss": 0.8068,
"step": 327
},
{
"epoch": 1.7105606258148631,
"grad_norm": 1.352963702330486,
"learning_rate": 6.644479234734971e-05,
"loss": 0.8337,
"step": 328
},
{
"epoch": 1.7157757496740547,
"grad_norm": 1.486986240165971,
"learning_rate": 6.6334856702084e-05,
"loss": 0.8142,
"step": 329
},
{
"epoch": 1.7209908735332464,
"grad_norm": 1.2755542839313279,
"learning_rate": 6.622456881258392e-05,
"loss": 0.8224,
"step": 330
},
{
"epoch": 1.7262059973924382,
"grad_norm": 1.4619956893213686,
"learning_rate": 6.6113930154015e-05,
"loss": 0.8077,
"step": 331
},
{
"epoch": 1.7314211212516297,
"grad_norm": 1.7136751212219052,
"learning_rate": 6.600294220623457e-05,
"loss": 0.8089,
"step": 332
},
{
"epoch": 1.7366362451108213,
"grad_norm": 1.1691508723533794,
"learning_rate": 6.589160645377181e-05,
"loss": 0.8192,
"step": 333
},
{
"epoch": 1.741851368970013,
"grad_norm": 2.2934106958814544,
"learning_rate": 6.57799243858081e-05,
"loss": 0.8123,
"step": 334
},
{
"epoch": 1.7470664928292048,
"grad_norm": 1.3479467101781495,
"learning_rate": 6.566789749615691e-05,
"loss": 0.8016,
"step": 335
},
{
"epoch": 1.7522816166883963,
"grad_norm": 1.4715014206564454,
"learning_rate": 6.555552728324394e-05,
"loss": 0.8135,
"step": 336
},
{
"epoch": 1.7574967405475879,
"grad_norm": 2.4679525667464155,
"learning_rate": 6.544281525008703e-05,
"loss": 0.8047,
"step": 337
},
{
"epoch": 1.7627118644067796,
"grad_norm": 1.2701148601075805,
"learning_rate": 6.532976290427611e-05,
"loss": 0.8155,
"step": 338
},
{
"epoch": 1.7679269882659714,
"grad_norm": 2.7926000172133967,
"learning_rate": 6.521637175795292e-05,
"loss": 0.8114,
"step": 339
},
{
"epoch": 1.773142112125163,
"grad_norm": 1.8634719780362607,
"learning_rate": 6.51026433277909e-05,
"loss": 0.8279,
"step": 340
},
{
"epoch": 1.7783572359843545,
"grad_norm": 2.527061321023331,
"learning_rate": 6.498857913497485e-05,
"loss": 0.8308,
"step": 341
},
{
"epoch": 1.7835723598435462,
"grad_norm": 1.7861432214936908,
"learning_rate": 6.487418070518063e-05,
"loss": 0.8237,
"step": 342
},
{
"epoch": 1.788787483702738,
"grad_norm": 2.4741757997509812,
"learning_rate": 6.475944956855463e-05,
"loss": 0.8098,
"step": 343
},
{
"epoch": 1.7940026075619295,
"grad_norm": 2.5470632776776583,
"learning_rate": 6.464438725969348e-05,
"loss": 0.8153,
"step": 344
},
{
"epoch": 1.7992177314211213,
"grad_norm": 1.4497598146752289,
"learning_rate": 6.452899531762338e-05,
"loss": 0.809,
"step": 345
},
{
"epoch": 1.804432855280313,
"grad_norm": 1.6671814059233867,
"learning_rate": 6.44132752857796e-05,
"loss": 0.8206,
"step": 346
},
{
"epoch": 1.8096479791395046,
"grad_norm": 1.4862285893517284,
"learning_rate": 6.429722871198579e-05,
"loss": 0.8125,
"step": 347
},
{
"epoch": 1.8148631029986961,
"grad_norm": 1.4717475694489375,
"learning_rate": 6.418085714843328e-05,
"loss": 0.8028,
"step": 348
},
{
"epoch": 1.820078226857888,
"grad_norm": 1.2961734030931862,
"learning_rate": 6.406416215166035e-05,
"loss": 0.814,
"step": 349
},
{
"epoch": 1.8252933507170797,
"grad_norm": 1.375416590864907,
"learning_rate": 6.394714528253137e-05,
"loss": 0.8084,
"step": 350
},
{
"epoch": 1.8305084745762712,
"grad_norm": 1.1939033449159728,
"learning_rate": 6.382980810621595e-05,
"loss": 0.8054,
"step": 351
},
{
"epoch": 1.8357235984354627,
"grad_norm": 1.905725265353936,
"learning_rate": 6.371215219216801e-05,
"loss": 0.799,
"step": 352
},
{
"epoch": 1.8409387222946545,
"grad_norm": 1.5504247264478312,
"learning_rate": 6.359417911410477e-05,
"loss": 0.8184,
"step": 353
},
{
"epoch": 1.8461538461538463,
"grad_norm": 1.034848478726046,
"learning_rate": 6.347589044998568e-05,
"loss": 0.8058,
"step": 354
},
{
"epoch": 1.8513689700130378,
"grad_norm": 2.3826958255372217,
"learning_rate": 6.335728778199139e-05,
"loss": 0.8066,
"step": 355
},
{
"epoch": 1.8565840938722293,
"grad_norm": 1.3959972285328377,
"learning_rate": 6.323837269650249e-05,
"loss": 0.8217,
"step": 356
},
{
"epoch": 1.861799217731421,
"grad_norm": 2.506085020659937,
"learning_rate": 6.311914678407837e-05,
"loss": 0.8216,
"step": 357
},
{
"epoch": 1.8670143415906129,
"grad_norm": 1.8279565139453056,
"learning_rate": 6.299961163943587e-05,
"loss": 0.8213,
"step": 358
},
{
"epoch": 1.8722294654498044,
"grad_norm": 2.0962030329324204,
"learning_rate": 6.287976886142806e-05,
"loss": 0.8206,
"step": 359
},
{
"epoch": 1.877444589308996,
"grad_norm": 1.70186031477118,
"learning_rate": 6.275962005302273e-05,
"loss": 0.8301,
"step": 360
},
{
"epoch": 1.8826597131681877,
"grad_norm": 2.1788285828805365,
"learning_rate": 6.263916682128104e-05,
"loss": 0.817,
"step": 361
},
{
"epoch": 1.8878748370273795,
"grad_norm": 1.805067718405396,
"learning_rate": 6.251841077733595e-05,
"loss": 0.8158,
"step": 362
},
{
"epoch": 1.893089960886571,
"grad_norm": 2.1305468630587066,
"learning_rate": 6.239735353637076e-05,
"loss": 0.8086,
"step": 363
},
{
"epoch": 1.8983050847457628,
"grad_norm": 1.8535876445861497,
"learning_rate": 6.227599671759745e-05,
"loss": 0.8088,
"step": 364
},
{
"epoch": 1.9035202086049545,
"grad_norm": 2.0851098014276914,
"learning_rate": 6.215434194423499e-05,
"loss": 0.8053,
"step": 365
},
{
"epoch": 1.908735332464146,
"grad_norm": 1.6617436320103747,
"learning_rate": 6.203239084348772e-05,
"loss": 0.8272,
"step": 366
},
{
"epoch": 1.9139504563233376,
"grad_norm": 1.9519893634609065,
"learning_rate": 6.191014504652352e-05,
"loss": 0.8052,
"step": 367
},
{
"epoch": 1.9191655801825294,
"grad_norm": 1.8496089676435563,
"learning_rate": 6.178760618845194e-05,
"loss": 0.8152,
"step": 368
},
{
"epoch": 1.9243807040417211,
"grad_norm": 1.5391906529413488,
"learning_rate": 6.166477590830252e-05,
"loss": 0.8129,
"step": 369
},
{
"epoch": 1.9295958279009127,
"grad_norm": 1.5353016752039255,
"learning_rate": 6.154165584900263e-05,
"loss": 0.7994,
"step": 370
},
{
"epoch": 1.9348109517601042,
"grad_norm": 1.4284130690464092,
"learning_rate": 6.141824765735567e-05,
"loss": 0.8169,
"step": 371
},
{
"epoch": 1.940026075619296,
"grad_norm": 1.3150434684077639,
"learning_rate": 6.129455298401894e-05,
"loss": 0.7936,
"step": 372
},
{
"epoch": 1.9452411994784877,
"grad_norm": 1.3747947345260623,
"learning_rate": 6.117057348348164e-05,
"loss": 0.8007,
"step": 373
},
{
"epoch": 1.9504563233376793,
"grad_norm": 1.0176646734905497,
"learning_rate": 6.104631081404269e-05,
"loss": 0.8085,
"step": 374
},
{
"epoch": 1.9556714471968708,
"grad_norm": 1.773293381624066,
"learning_rate": 6.092176663778851e-05,
"loss": 0.8075,
"step": 375
},
{
"epoch": 1.9608865710560626,
"grad_norm": 1.4739056885963973,
"learning_rate": 6.079694262057094e-05,
"loss": 0.8125,
"step": 376
},
{
"epoch": 1.9661016949152543,
"grad_norm": 1.6264911494594487,
"learning_rate": 6.067184043198476e-05,
"loss": 0.8093,
"step": 377
},
{
"epoch": 1.9713168187744459,
"grad_norm": 1.5269696160976443,
"learning_rate": 6.054646174534552e-05,
"loss": 0.8073,
"step": 378
},
{
"epoch": 1.9765319426336374,
"grad_norm": 1.4672155013368013,
"learning_rate": 6.0420808237667055e-05,
"loss": 0.7873,
"step": 379
},
{
"epoch": 1.9817470664928292,
"grad_norm": 1.6071190706840761,
"learning_rate": 6.029488158963912e-05,
"loss": 0.8185,
"step": 380
},
{
"epoch": 1.986962190352021,
"grad_norm": 1.1200471714942668,
"learning_rate": 6.016868348560488e-05,
"loss": 0.7945,
"step": 381
},
{
"epoch": 1.9921773142112125,
"grad_norm": 1.2021517615584545,
"learning_rate": 6.004221561353838e-05,
"loss": 0.8019,
"step": 382
},
{
"epoch": 1.997392438070404,
"grad_norm": 1.1953523539998745,
"learning_rate": 5.991547966502195e-05,
"loss": 0.8847,
"step": 383
},
{
"epoch": 2.002607561929596,
"grad_norm": 1.2139914038103075,
"learning_rate": 5.978847733522363e-05,
"loss": 0.9192,
"step": 384
},
{
"epoch": 2.0078226857887875,
"grad_norm": 1.2663014964519022,
"learning_rate": 5.9661210322874456e-05,
"loss": 0.7934,
"step": 385
},
{
"epoch": 2.013037809647979,
"grad_norm": 1.7536553812874152,
"learning_rate": 5.953368033024576e-05,
"loss": 0.7716,
"step": 386
},
{
"epoch": 2.0182529335071706,
"grad_norm": 1.3505679779762105,
"learning_rate": 5.940588906312636e-05,
"loss": 0.79,
"step": 387
},
{
"epoch": 2.0234680573663626,
"grad_norm": 1.100812193500458,
"learning_rate": 5.9277838230799816e-05,
"loss": 0.7831,
"step": 388
},
{
"epoch": 2.028683181225554,
"grad_norm": 1.312836314335182,
"learning_rate": 5.9149529546021486e-05,
"loss": 0.7833,
"step": 389
},
{
"epoch": 2.0338983050847457,
"grad_norm": 1.9463142850045887,
"learning_rate": 5.902096472499569e-05,
"loss": 0.7829,
"step": 390
},
{
"epoch": 2.039113428943937,
"grad_norm": 0.7033565907846064,
"learning_rate": 5.889214548735269e-05,
"loss": 0.7842,
"step": 391
},
{
"epoch": 2.044328552803129,
"grad_norm": 1.743510351487419,
"learning_rate": 5.876307355612575e-05,
"loss": 0.7679,
"step": 392
},
{
"epoch": 2.0495436766623207,
"grad_norm": 1.2465470038883548,
"learning_rate": 5.8633750657728033e-05,
"loss": 0.7712,
"step": 393
},
{
"epoch": 2.0547588005215123,
"grad_norm": 2.068655430465104,
"learning_rate": 5.850417852192956e-05,
"loss": 0.7853,
"step": 394
},
{
"epoch": 2.0599739243807043,
"grad_norm": 0.9560005505258112,
"learning_rate": 5.837435888183403e-05,
"loss": 0.785,
"step": 395
},
{
"epoch": 2.065189048239896,
"grad_norm": 1.7062918904685223,
"learning_rate": 5.8244293473855664e-05,
"loss": 0.7849,
"step": 396
},
{
"epoch": 2.0704041720990873,
"grad_norm": 1.623721715677207,
"learning_rate": 5.8113984037695984e-05,
"loss": 0.7857,
"step": 397
},
{
"epoch": 2.075619295958279,
"grad_norm": 1.484064136635307,
"learning_rate": 5.798343231632053e-05,
"loss": 0.7779,
"step": 398
},
{
"epoch": 2.080834419817471,
"grad_norm": 1.34676834959028,
"learning_rate": 5.785264005593553e-05,
"loss": 0.7934,
"step": 399
},
{
"epoch": 2.0860495436766624,
"grad_norm": 1.6199775628522408,
"learning_rate": 5.772160900596456e-05,
"loss": 0.7735,
"step": 400
},
{
"epoch": 2.091264667535854,
"grad_norm": 1.5660844096047937,
"learning_rate": 5.7590340919025204e-05,
"loss": 0.7833,
"step": 401
},
{
"epoch": 2.0964797913950455,
"grad_norm": 1.0421277103166564,
"learning_rate": 5.7458837550905486e-05,
"loss": 0.7837,
"step": 402
},
{
"epoch": 2.1016949152542375,
"grad_norm": 1.7458400592594279,
"learning_rate": 5.73271006605405e-05,
"loss": 0.7855,
"step": 403
},
{
"epoch": 2.106910039113429,
"grad_norm": 1.5186591541100996,
"learning_rate": 5.7195132009988814e-05,
"loss": 0.7853,
"step": 404
},
{
"epoch": 2.1121251629726205,
"grad_norm": 0.8719695266664192,
"learning_rate": 5.706293336440894e-05,
"loss": 0.7896,
"step": 405
},
{
"epoch": 2.117340286831812,
"grad_norm": 1.0314918237610309,
"learning_rate": 5.693050649203568e-05,
"loss": 0.7807,
"step": 406
},
{
"epoch": 2.122555410691004,
"grad_norm": 0.7603723547082539,
"learning_rate": 5.679785316415654e-05,
"loss": 0.7763,
"step": 407
},
{
"epoch": 2.1277705345501956,
"grad_norm": 0.9173514911262751,
"learning_rate": 5.6664975155088004e-05,
"loss": 0.7919,
"step": 408
},
{
"epoch": 2.132985658409387,
"grad_norm": 1.2215202346945515,
"learning_rate": 5.6531874242151746e-05,
"loss": 0.772,
"step": 409
},
{
"epoch": 2.1382007822685787,
"grad_norm": 2.368630198683683,
"learning_rate": 5.639855220565098e-05,
"loss": 0.801,
"step": 410
},
{
"epoch": 2.1434159061277707,
"grad_norm": 1.2419888378957813,
"learning_rate": 5.626501082884654e-05,
"loss": 0.7948,
"step": 411
},
{
"epoch": 2.148631029986962,
"grad_norm": 3.0547619405266446,
"learning_rate": 5.613125189793305e-05,
"loss": 0.7761,
"step": 412
},
{
"epoch": 2.1538461538461537,
"grad_norm": 2.1394143129005987,
"learning_rate": 5.5997277202015085e-05,
"loss": 0.7815,
"step": 413
},
{
"epoch": 2.1590612777053453,
"grad_norm": 3.041921943681318,
"learning_rate": 5.586308853308319e-05,
"loss": 0.7889,
"step": 414
},
{
"epoch": 2.1642764015645373,
"grad_norm": 2.4712686274458853,
"learning_rate": 5.572868768598993e-05,
"loss": 0.7764,
"step": 415
},
{
"epoch": 2.169491525423729,
"grad_norm": 2.5501423533072605,
"learning_rate": 5.5594076458425864e-05,
"loss": 0.7995,
"step": 416
},
{
"epoch": 2.1747066492829203,
"grad_norm": 2.2877612282195416,
"learning_rate": 5.545925665089552e-05,
"loss": 0.7882,
"step": 417
},
{
"epoch": 2.1799217731421123,
"grad_norm": 2.1483606998937144,
"learning_rate": 5.532423006669332e-05,
"loss": 0.7873,
"step": 418
},
{
"epoch": 2.185136897001304,
"grad_norm": 1.7462471775177655,
"learning_rate": 5.518899851187942e-05,
"loss": 0.7964,
"step": 419
},
{
"epoch": 2.1903520208604954,
"grad_norm": 2.756935578457596,
"learning_rate": 5.505356379525559e-05,
"loss": 0.7891,
"step": 420
},
{
"epoch": 2.195567144719687,
"grad_norm": 2.4106523309041314,
"learning_rate": 5.491792772834103e-05,
"loss": 0.7804,
"step": 421
},
{
"epoch": 2.200782268578879,
"grad_norm": 1.8981144967918866,
"learning_rate": 5.478209212534809e-05,
"loss": 0.7796,
"step": 422
},
{
"epoch": 2.2059973924380705,
"grad_norm": 1.487529019552184,
"learning_rate": 5.464605880315803e-05,
"loss": 0.7773,
"step": 423
},
{
"epoch": 2.211212516297262,
"grad_norm": 2.44109555313169,
"learning_rate": 5.4509829581296774e-05,
"loss": 0.777,
"step": 424
},
{
"epoch": 2.2164276401564535,
"grad_norm": 1.9157281312835157,
"learning_rate": 5.4373406281910434e-05,
"loss": 0.7835,
"step": 425
},
{
"epoch": 2.2216427640156455,
"grad_norm": 2.6110440355014477,
"learning_rate": 5.423679072974109e-05,
"loss": 0.7792,
"step": 426
},
{
"epoch": 2.226857887874837,
"grad_norm": 2.3658010170065853,
"learning_rate": 5.4099984752102295e-05,
"loss": 0.7914,
"step": 427
},
{
"epoch": 2.2320730117340286,
"grad_norm": 1.768060696859615,
"learning_rate": 5.396299017885465e-05,
"loss": 0.7789,
"step": 428
},
{
"epoch": 2.23728813559322,
"grad_norm": 1.661248122612779,
"learning_rate": 5.3825808842381345e-05,
"loss": 0.7735,
"step": 429
},
{
"epoch": 2.242503259452412,
"grad_norm": 2.362109109009534,
"learning_rate": 5.3688442577563646e-05,
"loss": 0.7886,
"step": 430
},
{
"epoch": 2.2477183833116037,
"grad_norm": 2.0017993267041905,
"learning_rate": 5.355089322175629e-05,
"loss": 0.7735,
"step": 431
},
{
"epoch": 2.252933507170795,
"grad_norm": 1.985452281861886,
"learning_rate": 5.3413162614763043e-05,
"loss": 0.7904,
"step": 432
},
{
"epoch": 2.2581486310299868,
"grad_norm": 1.8249568105371123,
"learning_rate": 5.327525259881196e-05,
"loss": 0.7792,
"step": 433
},
{
"epoch": 2.2633637548891787,
"grad_norm": 1.987099316108934,
"learning_rate": 5.3137165018530805e-05,
"loss": 0.7718,
"step": 434
},
{
"epoch": 2.2685788787483703,
"grad_norm": 1.7309972017145459,
"learning_rate": 5.299890172092238e-05,
"loss": 0.7737,
"step": 435
},
{
"epoch": 2.273794002607562,
"grad_norm": 2.083040780695706,
"learning_rate": 5.286046455533981e-05,
"loss": 0.7861,
"step": 436
},
{
"epoch": 2.279009126466754,
"grad_norm": 1.7833638732253791,
"learning_rate": 5.27218553734618e-05,
"loss": 0.7688,
"step": 437
},
{
"epoch": 2.2842242503259453,
"grad_norm": 2.2540602964270997,
"learning_rate": 5.2583076029267864e-05,
"loss": 0.7868,
"step": 438
},
{
"epoch": 2.289439374185137,
"grad_norm": 2.1042743671125046,
"learning_rate": 5.2444128379013564e-05,
"loss": 0.7806,
"step": 439
},
{
"epoch": 2.2946544980443284,
"grad_norm": 1.638491731440727,
"learning_rate": 5.2305014281205634e-05,
"loss": 0.7778,
"step": 440
},
{
"epoch": 2.2998696219035204,
"grad_norm": 1.4490203056402047,
"learning_rate": 5.2165735596577146e-05,
"loss": 0.7776,
"step": 441
},
{
"epoch": 2.305084745762712,
"grad_norm": 2.285181885567823,
"learning_rate": 5.202629418806263e-05,
"loss": 0.7767,
"step": 442
},
{
"epoch": 2.3102998696219035,
"grad_norm": 1.9464604198183288,
"learning_rate": 5.1886691920773116e-05,
"loss": 0.7651,
"step": 443
},
{
"epoch": 2.315514993481095,
"grad_norm": 1.7631237894703213,
"learning_rate": 5.174693066197125e-05,
"loss": 0.7726,
"step": 444
},
{
"epoch": 2.320730117340287,
"grad_norm": 1.56838244742276,
"learning_rate": 5.160701228104626e-05,
"loss": 0.7819,
"step": 445
},
{
"epoch": 2.3259452411994785,
"grad_norm": 2.202901692990395,
"learning_rate": 5.146693864948898e-05,
"loss": 0.7856,
"step": 446
},
{
"epoch": 2.33116036505867,
"grad_norm": 1.92108099189494,
"learning_rate": 5.132671164086682e-05,
"loss": 0.775,
"step": 447
},
{
"epoch": 2.3363754889178616,
"grad_norm": 1.760952455824346,
"learning_rate": 5.118633313079869e-05,
"loss": 0.7807,
"step": 448
},
{
"epoch": 2.3415906127770536,
"grad_norm": 1.5869940859426193,
"learning_rate": 5.104580499692992e-05,
"loss": 0.7845,
"step": 449
},
{
"epoch": 2.346805736636245,
"grad_norm": 2.0260440621389364,
"learning_rate": 5.090512911890715e-05,
"loss": 0.7857,
"step": 450
},
{
"epoch": 2.3520208604954367,
"grad_norm": 1.7167308373113537,
"learning_rate": 5.076430737835318e-05,
"loss": 0.7815,
"step": 451
},
{
"epoch": 2.3572359843546282,
"grad_norm": 1.881887746447444,
"learning_rate": 5.062334165884182e-05,
"loss": 0.7811,
"step": 452
},
{
"epoch": 2.36245110821382,
"grad_norm": 1.6824543517537593,
"learning_rate": 5.0482233845872674e-05,
"loss": 0.776,
"step": 453
},
{
"epoch": 2.3676662320730117,
"grad_norm": 1.8755854127797766,
"learning_rate": 5.034098582684595e-05,
"loss": 0.7803,
"step": 454
},
{
"epoch": 2.3728813559322033,
"grad_norm": 1.6777197590484048,
"learning_rate": 5.019959949103715e-05,
"loss": 0.7848,
"step": 455
},
{
"epoch": 2.3780964797913953,
"grad_norm": 1.81809003546367,
"learning_rate": 5.005807672957188e-05,
"loss": 0.7858,
"step": 456
},
{
"epoch": 2.383311603650587,
"grad_norm": 1.6133734346606823,
"learning_rate": 4.9916419435400516e-05,
"loss": 0.7808,
"step": 457
},
{
"epoch": 2.3885267275097783,
"grad_norm": 1.8363482997949565,
"learning_rate": 4.9774629503272874e-05,
"loss": 0.7796,
"step": 458
},
{
"epoch": 2.39374185136897,
"grad_norm": 1.5541908015103185,
"learning_rate": 4.96327088297129e-05,
"loss": 0.7771,
"step": 459
},
{
"epoch": 2.3989569752281614,
"grad_norm": 1.754540259013617,
"learning_rate": 4.949065931299328e-05,
"loss": 0.7858,
"step": 460
},
{
"epoch": 2.4041720990873534,
"grad_norm": 1.539083403172259,
"learning_rate": 4.934848285311002e-05,
"loss": 0.778,
"step": 461
},
{
"epoch": 2.409387222946545,
"grad_norm": 1.8702666142803812,
"learning_rate": 4.920618135175712e-05,
"loss": 0.7815,
"step": 462
},
{
"epoch": 2.4146023468057365,
"grad_norm": 1.644430924102753,
"learning_rate": 4.9063756712301036e-05,
"loss": 0.773,
"step": 463
},
{
"epoch": 2.4198174706649285,
"grad_norm": 1.6756469842658643,
"learning_rate": 4.8921210839755304e-05,
"loss": 0.7737,
"step": 464
},
{
"epoch": 2.42503259452412,
"grad_norm": 1.4353920458126117,
"learning_rate": 4.877854564075499e-05,
"loss": 0.7861,
"step": 465
},
{
"epoch": 2.4302477183833116,
"grad_norm": 1.8834730757505287,
"learning_rate": 4.863576302353125e-05,
"loss": 0.7638,
"step": 466
},
{
"epoch": 2.435462842242503,
"grad_norm": 1.6285082262536497,
"learning_rate": 4.849286489788579e-05,
"loss": 0.7719,
"step": 467
},
{
"epoch": 2.440677966101695,
"grad_norm": 1.6841805396754381,
"learning_rate": 4.834985317516525e-05,
"loss": 0.7833,
"step": 468
},
{
"epoch": 2.4458930899608866,
"grad_norm": 1.5254872325101743,
"learning_rate": 4.8206729768235756e-05,
"loss": 0.7766,
"step": 469
},
{
"epoch": 2.451108213820078,
"grad_norm": 1.5946022371035584,
"learning_rate": 4.8063496591457256e-05,
"loss": 0.7688,
"step": 470
},
{
"epoch": 2.4563233376792697,
"grad_norm": 1.3044832158013293,
"learning_rate": 4.792015556065793e-05,
"loss": 0.7745,
"step": 471
},
{
"epoch": 2.4615384615384617,
"grad_norm": 1.822367193607616,
"learning_rate": 4.777670859310857e-05,
"loss": 0.7839,
"step": 472
},
{
"epoch": 2.466753585397653,
"grad_norm": 1.5933250932152585,
"learning_rate": 4.763315760749695e-05,
"loss": 0.7886,
"step": 473
},
{
"epoch": 2.4719687092568448,
"grad_norm": 1.588436894654484,
"learning_rate": 4.748950452390212e-05,
"loss": 0.7742,
"step": 474
},
{
"epoch": 2.4771838331160367,
"grad_norm": 1.4522634410247943,
"learning_rate": 4.734575126376876e-05,
"loss": 0.7823,
"step": 475
},
{
"epoch": 2.4823989569752283,
"grad_norm": 1.586237253805528,
"learning_rate": 4.7201899749881504e-05,
"loss": 0.7926,
"step": 476
},
{
"epoch": 2.48761408083442,
"grad_norm": 1.3312635236777999,
"learning_rate": 4.705795190633915e-05,
"loss": 0.786,
"step": 477
},
{
"epoch": 2.4928292046936114,
"grad_norm": 1.7378087684763845,
"learning_rate": 4.691390965852893e-05,
"loss": 0.7693,
"step": 478
},
{
"epoch": 2.498044328552803,
"grad_norm": 1.554126764803909,
"learning_rate": 4.676977493310088e-05,
"loss": 0.7911,
"step": 479
},
{
"epoch": 2.503259452411995,
"grad_norm": 1.4897288575191927,
"learning_rate": 4.662554965794192e-05,
"loss": 0.7809,
"step": 480
},
{
"epoch": 2.5084745762711864,
"grad_norm": 1.2742943025314304,
"learning_rate": 4.648123576215011e-05,
"loss": 0.7811,
"step": 481
},
{
"epoch": 2.513689700130378,
"grad_norm": 1.714780336353614,
"learning_rate": 4.633683517600891e-05,
"loss": 0.7707,
"step": 482
},
{
"epoch": 2.51890482398957,
"grad_norm": 1.514593536382248,
"learning_rate": 4.61923498309613e-05,
"loss": 0.7817,
"step": 483
},
{
"epoch": 2.5241199478487615,
"grad_norm": 1.4058404125363677,
"learning_rate": 4.604778165958392e-05,
"loss": 0.7691,
"step": 484
},
{
"epoch": 2.529335071707953,
"grad_norm": 1.1465871672288677,
"learning_rate": 4.590313259556132e-05,
"loss": 0.7635,
"step": 485
},
{
"epoch": 2.5345501955671446,
"grad_norm": 1.6402163570623103,
"learning_rate": 4.575840457366001e-05,
"loss": 0.7876,
"step": 486
},
{
"epoch": 2.539765319426336,
"grad_norm": 1.3178513882789409,
"learning_rate": 4.561359952970259e-05,
"loss": 0.7664,
"step": 487
},
{
"epoch": 2.544980443285528,
"grad_norm": 1.6022661486448093,
"learning_rate": 4.546871940054191e-05,
"loss": 0.7905,
"step": 488
},
{
"epoch": 2.5501955671447196,
"grad_norm": 1.4090335919146642,
"learning_rate": 4.5323766124035115e-05,
"loss": 0.7819,
"step": 489
},
{
"epoch": 2.555410691003911,
"grad_norm": 1.324862817997592,
"learning_rate": 4.5178741639017736e-05,
"loss": 0.7787,
"step": 490
},
{
"epoch": 2.560625814863103,
"grad_norm": 1.1347550661741326,
"learning_rate": 4.503364788527774e-05,
"loss": 0.7831,
"step": 491
},
{
"epoch": 2.5658409387222947,
"grad_norm": 1.5760247018602986,
"learning_rate": 4.488848680352965e-05,
"loss": 0.7796,
"step": 492
},
{
"epoch": 2.5710560625814862,
"grad_norm": 1.3090981394784642,
"learning_rate": 4.4743260335388516e-05,
"loss": 0.7724,
"step": 493
},
{
"epoch": 2.576271186440678,
"grad_norm": 1.3548998817619167,
"learning_rate": 4.459797042334395e-05,
"loss": 0.781,
"step": 494
},
{
"epoch": 2.5814863102998697,
"grad_norm": 1.2234336308095453,
"learning_rate": 4.445261901073419e-05,
"loss": 0.7743,
"step": 495
},
{
"epoch": 2.5867014341590613,
"grad_norm": 1.2594794059995964,
"learning_rate": 4.430720804172008e-05,
"loss": 0.7878,
"step": 496
},
{
"epoch": 2.591916558018253,
"grad_norm": 1.0475475763925866,
"learning_rate": 4.416173946125906e-05,
"loss": 0.7851,
"step": 497
},
{
"epoch": 2.5971316818774444,
"grad_norm": 1.3034377416715606,
"learning_rate": 4.401621521507914e-05,
"loss": 0.7661,
"step": 498
},
{
"epoch": 2.6023468057366363,
"grad_norm": 1.0178177147229814,
"learning_rate": 4.387063724965292e-05,
"loss": 0.768,
"step": 499
},
{
"epoch": 2.607561929595828,
"grad_norm": 1.4358534072422087,
"learning_rate": 4.372500751217153e-05,
"loss": 0.7886,
"step": 500
},
{
"epoch": 2.6127770534550194,
"grad_norm": 1.150853401699166,
"learning_rate": 4.357932795051852e-05,
"loss": 0.7934,
"step": 501
},
{
"epoch": 2.6179921773142114,
"grad_norm": 1.1810592463532203,
"learning_rate": 4.3433600513243965e-05,
"loss": 0.7563,
"step": 502
},
{
"epoch": 2.623207301173403,
"grad_norm": 1.0034045366911863,
"learning_rate": 4.328782714953823e-05,
"loss": 0.79,
"step": 503
},
{
"epoch": 2.6284224250325945,
"grad_norm": 1.0980643139376323,
"learning_rate": 4.3142009809205986e-05,
"loss": 0.7861,
"step": 504
},
{
"epoch": 2.633637548891786,
"grad_norm": 0.8881990742171157,
"learning_rate": 4.2996150442640163e-05,
"loss": 0.7758,
"step": 505
},
{
"epoch": 2.6388526727509776,
"grad_norm": 0.8675797674117819,
"learning_rate": 4.2850251000795735e-05,
"loss": 0.7801,
"step": 506
},
{
"epoch": 2.6440677966101696,
"grad_norm": 0.7883090264802906,
"learning_rate": 4.270431343516379e-05,
"loss": 0.7745,
"step": 507
},
{
"epoch": 2.649282920469361,
"grad_norm": 0.7583525803187019,
"learning_rate": 4.2558339697745297e-05,
"loss": 0.7827,
"step": 508
},
{
"epoch": 2.6544980443285526,
"grad_norm": 0.7762877342404452,
"learning_rate": 4.2412331741025045e-05,
"loss": 0.7967,
"step": 509
},
{
"epoch": 2.6597131681877446,
"grad_norm": 0.7014761796032362,
"learning_rate": 4.226629151794555e-05,
"loss": 0.7858,
"step": 510
},
{
"epoch": 2.664928292046936,
"grad_norm": 0.7396414732409832,
"learning_rate": 4.2120220981880875e-05,
"loss": 0.7794,
"step": 511
},
{
"epoch": 2.6701434159061277,
"grad_norm": 0.6998054830397707,
"learning_rate": 4.197412208661058e-05,
"loss": 0.7777,
"step": 512
},
{
"epoch": 2.6753585397653197,
"grad_norm": 0.6386491344974894,
"learning_rate": 4.182799678629351e-05,
"loss": 0.7823,
"step": 513
},
{
"epoch": 2.680573663624511,
"grad_norm": 0.5877328765456284,
"learning_rate": 4.168184703544171e-05,
"loss": 0.7744,
"step": 514
},
{
"epoch": 2.6857887874837028,
"grad_norm": 0.5920983354896419,
"learning_rate": 4.153567478889426e-05,
"loss": 0.7769,
"step": 515
},
{
"epoch": 2.6910039113428943,
"grad_norm": 0.4867319202645289,
"learning_rate": 4.138948200179115e-05,
"loss": 0.7807,
"step": 516
},
{
"epoch": 2.696219035202086,
"grad_norm": 0.5746181168372155,
"learning_rate": 4.124327062954707e-05,
"loss": 0.7708,
"step": 517
},
{
"epoch": 2.701434159061278,
"grad_norm": 0.35394623638423245,
"learning_rate": 4.1097042627825325e-05,
"loss": 0.7823,
"step": 518
},
{
"epoch": 2.7066492829204694,
"grad_norm": 0.6245548216599335,
"learning_rate": 4.095079995251168e-05,
"loss": 0.784,
"step": 519
},
{
"epoch": 2.711864406779661,
"grad_norm": 0.3951096697035697,
"learning_rate": 4.080454455968812e-05,
"loss": 0.7711,
"step": 520
},
{
"epoch": 2.717079530638853,
"grad_norm": 0.5759028178578062,
"learning_rate": 4.065827840560674e-05,
"loss": 0.7773,
"step": 521
},
{
"epoch": 2.7222946544980444,
"grad_norm": 0.39680693057913397,
"learning_rate": 4.0512003446663576e-05,
"loss": 0.7736,
"step": 522
},
{
"epoch": 2.727509778357236,
"grad_norm": 0.49858076462377626,
"learning_rate": 4.0365721639372476e-05,
"loss": 0.7884,
"step": 523
},
{
"epoch": 2.7327249022164275,
"grad_norm": 0.4117350165519155,
"learning_rate": 4.021943494033882e-05,
"loss": 0.7707,
"step": 524
},
{
"epoch": 2.737940026075619,
"grad_norm": 0.4515042047115184,
"learning_rate": 4.007314530623348e-05,
"loss": 0.7778,
"step": 525
},
{
"epoch": 2.743155149934811,
"grad_norm": 0.42547404001436967,
"learning_rate": 3.9926854693766536e-05,
"loss": 0.7688,
"step": 526
},
{
"epoch": 2.7483702737940026,
"grad_norm": 0.35069455268010413,
"learning_rate": 3.978056505966119e-05,
"loss": 0.7744,
"step": 527
},
{
"epoch": 2.753585397653194,
"grad_norm": 0.4444789741021684,
"learning_rate": 3.963427836062753e-05,
"loss": 0.775,
"step": 528
},
{
"epoch": 2.758800521512386,
"grad_norm": 0.43329608911430956,
"learning_rate": 3.948799655333644e-05,
"loss": 0.7748,
"step": 529
},
{
"epoch": 2.7640156453715776,
"grad_norm": 0.4209440665741905,
"learning_rate": 3.9341721594393276e-05,
"loss": 0.7762,
"step": 530
},
{
"epoch": 2.769230769230769,
"grad_norm": 0.4322531684473173,
"learning_rate": 3.919545544031189e-05,
"loss": 0.7878,
"step": 531
},
{
"epoch": 2.774445893089961,
"grad_norm": 0.4720335982176997,
"learning_rate": 3.904920004748833e-05,
"loss": 0.777,
"step": 532
},
{
"epoch": 2.7796610169491527,
"grad_norm": 0.37675833736688985,
"learning_rate": 3.8902957372174675e-05,
"loss": 0.7728,
"step": 533
},
{
"epoch": 2.7848761408083442,
"grad_norm": 0.3961055004093581,
"learning_rate": 3.8756729370452936e-05,
"loss": 0.7745,
"step": 534
},
{
"epoch": 2.7900912646675358,
"grad_norm": 0.39787870628829797,
"learning_rate": 3.8610517998208866e-05,
"loss": 0.779,
"step": 535
},
{
"epoch": 2.7953063885267273,
"grad_norm": 0.366576904221954,
"learning_rate": 3.846432521110574e-05,
"loss": 0.7823,
"step": 536
},
{
"epoch": 2.8005215123859193,
"grad_norm": 0.39669396154921643,
"learning_rate": 3.831815296455829e-05,
"loss": 0.7776,
"step": 537
},
{
"epoch": 2.805736636245111,
"grad_norm": 0.5485537263803792,
"learning_rate": 3.817200321370649e-05,
"loss": 0.7821,
"step": 538
},
{
"epoch": 2.8109517601043024,
"grad_norm": 0.5046596554915849,
"learning_rate": 3.802587791338943e-05,
"loss": 0.757,
"step": 539
},
{
"epoch": 2.8161668839634943,
"grad_norm": 0.4297088800078496,
"learning_rate": 3.787977901811914e-05,
"loss": 0.7819,
"step": 540
},
{
"epoch": 2.821382007822686,
"grad_norm": 0.2792196466948258,
"learning_rate": 3.7733708482054476e-05,
"loss": 0.7908,
"step": 541
},
{
"epoch": 2.8265971316818774,
"grad_norm": 0.3389535808547031,
"learning_rate": 3.758766825897497e-05,
"loss": 0.7789,
"step": 542
},
{
"epoch": 2.831812255541069,
"grad_norm": 0.33504887483040935,
"learning_rate": 3.7441660302254724e-05,
"loss": 0.7641,
"step": 543
},
{
"epoch": 2.8370273794002605,
"grad_norm": 0.35811016749224023,
"learning_rate": 3.729568656483623e-05,
"loss": 0.7711,
"step": 544
},
{
"epoch": 2.8422425032594525,
"grad_norm": 0.3304033008238814,
"learning_rate": 3.714974899920428e-05,
"loss": 0.781,
"step": 545
},
{
"epoch": 2.847457627118644,
"grad_norm": 0.42274779776394655,
"learning_rate": 3.7003849557359863e-05,
"loss": 0.7738,
"step": 546
},
{
"epoch": 2.8526727509778356,
"grad_norm": 0.33608657379975826,
"learning_rate": 3.685799019079402e-05,
"loss": 0.7858,
"step": 547
},
{
"epoch": 2.8578878748370276,
"grad_norm": 0.4402897321734762,
"learning_rate": 3.6712172850461785e-05,
"loss": 0.7679,
"step": 548
},
{
"epoch": 2.863102998696219,
"grad_norm": 0.29934112633294296,
"learning_rate": 3.6566399486756055e-05,
"loss": 0.7889,
"step": 549
},
{
"epoch": 2.8683181225554106,
"grad_norm": 0.38015963993433893,
"learning_rate": 3.642067204948149e-05,
"loss": 0.7695,
"step": 550
},
{
"epoch": 2.8735332464146026,
"grad_norm": 0.38644710923105946,
"learning_rate": 3.627499248782849e-05,
"loss": 0.7745,
"step": 551
},
{
"epoch": 2.878748370273794,
"grad_norm": 0.44283035266374077,
"learning_rate": 3.612936275034709e-05,
"loss": 0.7829,
"step": 552
},
{
"epoch": 2.8839634941329857,
"grad_norm": 0.426691177545099,
"learning_rate": 3.598378478492087e-05,
"loss": 0.7775,
"step": 553
},
{
"epoch": 2.8891786179921772,
"grad_norm": 0.41202751066509746,
"learning_rate": 3.583826053874096e-05,
"loss": 0.7792,
"step": 554
},
{
"epoch": 2.8943937418513688,
"grad_norm": 0.40337608056255875,
"learning_rate": 3.5692791958279924e-05,
"loss": 0.7737,
"step": 555
},
{
"epoch": 2.8996088657105608,
"grad_norm": 0.42109772590116307,
"learning_rate": 3.554738098926582e-05,
"loss": 0.7757,
"step": 556
},
{
"epoch": 2.9048239895697523,
"grad_norm": 0.6458006850893471,
"learning_rate": 3.540202957665606e-05,
"loss": 0.7775,
"step": 557
},
{
"epoch": 2.910039113428944,
"grad_norm": 0.5129008593035667,
"learning_rate": 3.525673966461149e-05,
"loss": 0.7793,
"step": 558
},
{
"epoch": 2.915254237288136,
"grad_norm": 0.37986325064356524,
"learning_rate": 3.5111513196470356e-05,
"loss": 0.7706,
"step": 559
},
{
"epoch": 2.9204693611473274,
"grad_norm": 0.3127213742675476,
"learning_rate": 3.4966352114722264e-05,
"loss": 0.779,
"step": 560
},
{
"epoch": 2.925684485006519,
"grad_norm": 0.44150964872249004,
"learning_rate": 3.482125836098227e-05,
"loss": 0.7726,
"step": 561
},
{
"epoch": 2.9308996088657104,
"grad_norm": 0.37918325106659184,
"learning_rate": 3.46762338759649e-05,
"loss": 0.7763,
"step": 562
},
{
"epoch": 2.936114732724902,
"grad_norm": 0.34801331615309594,
"learning_rate": 3.4531280599458096e-05,
"loss": 0.7792,
"step": 563
},
{
"epoch": 2.941329856584094,
"grad_norm": 0.3417071688239163,
"learning_rate": 3.438640047029741e-05,
"loss": 0.783,
"step": 564
},
{
"epoch": 2.9465449804432855,
"grad_norm": 0.4027724323607999,
"learning_rate": 3.424159542634001e-05,
"loss": 0.7799,
"step": 565
},
{
"epoch": 2.951760104302477,
"grad_norm": 0.359846745274617,
"learning_rate": 3.409686740443869e-05,
"loss": 0.7874,
"step": 566
},
{
"epoch": 2.956975228161669,
"grad_norm": 0.39877444581949706,
"learning_rate": 3.395221834041608e-05,
"loss": 0.7796,
"step": 567
},
{
"epoch": 2.9621903520208606,
"grad_norm": 0.31627637641393286,
"learning_rate": 3.3807650169038714e-05,
"loss": 0.7828,
"step": 568
},
{
"epoch": 2.967405475880052,
"grad_norm": 0.2915329688447508,
"learning_rate": 3.366316482399111e-05,
"loss": 0.7611,
"step": 569
},
{
"epoch": 2.9726205997392436,
"grad_norm": 0.3003806981610255,
"learning_rate": 3.351876423784991e-05,
"loss": 0.7753,
"step": 570
},
{
"epoch": 2.9778357235984356,
"grad_norm": 0.32825206856441885,
"learning_rate": 3.33744503420581e-05,
"loss": 0.7805,
"step": 571
},
{
"epoch": 2.983050847457627,
"grad_norm": 0.28811623986201,
"learning_rate": 3.323022506689913e-05,
"loss": 0.7787,
"step": 572
},
{
"epoch": 2.9882659713168187,
"grad_norm": 0.3007386340259681,
"learning_rate": 3.308609034147109e-05,
"loss": 0.7693,
"step": 573
},
{
"epoch": 2.9934810951760102,
"grad_norm": 0.28601621207061495,
"learning_rate": 3.294204809366088e-05,
"loss": 0.7811,
"step": 574
},
{
"epoch": 2.9986962190352022,
"grad_norm": 0.38527843914278115,
"learning_rate": 3.27981002501185e-05,
"loss": 0.8962,
"step": 575
},
{
"epoch": 3.0039113428943938,
"grad_norm": 0.3891353205481252,
"learning_rate": 3.265424873623125e-05,
"loss": 0.8279,
"step": 576
},
{
"epoch": 3.0091264667535853,
"grad_norm": 0.34677276173333527,
"learning_rate": 3.251049547609789e-05,
"loss": 0.7617,
"step": 577
},
{
"epoch": 3.014341590612777,
"grad_norm": 0.3887463573130963,
"learning_rate": 3.2366842392503065e-05,
"loss": 0.7404,
"step": 578
},
{
"epoch": 3.019556714471969,
"grad_norm": 0.45850333580212727,
"learning_rate": 3.222329140689144e-05,
"loss": 0.7509,
"step": 579
},
{
"epoch": 3.0247718383311604,
"grad_norm": 0.42139454484056155,
"learning_rate": 3.207984443934208e-05,
"loss": 0.7474,
"step": 580
},
{
"epoch": 3.029986962190352,
"grad_norm": 0.32213156347235194,
"learning_rate": 3.193650340854275e-05,
"loss": 0.7449,
"step": 581
},
{
"epoch": 3.035202086049544,
"grad_norm": 0.2882489783740716,
"learning_rate": 3.179327023176426e-05,
"loss": 0.7565,
"step": 582
},
{
"epoch": 3.0404172099087354,
"grad_norm": 0.31398533972182546,
"learning_rate": 3.1650146824834765e-05,
"loss": 0.7663,
"step": 583
},
{
"epoch": 3.045632333767927,
"grad_norm": 0.42483377951795925,
"learning_rate": 3.1507135102114224e-05,
"loss": 0.7524,
"step": 584
},
{
"epoch": 3.0508474576271185,
"grad_norm": 0.4203137878607459,
"learning_rate": 3.1364236976468755e-05,
"loss": 0.7476,
"step": 585
},
{
"epoch": 3.0560625814863105,
"grad_norm": 0.3536224112892052,
"learning_rate": 3.122145435924502e-05,
"loss": 0.7479,
"step": 586
},
{
"epoch": 3.061277705345502,
"grad_norm": 0.3355710189906583,
"learning_rate": 3.107878916024471e-05,
"loss": 0.7475,
"step": 587
},
{
"epoch": 3.0664928292046936,
"grad_norm": 0.3578880245402906,
"learning_rate": 3.093624328769898e-05,
"loss": 0.7542,
"step": 588
},
{
"epoch": 3.071707953063885,
"grad_norm": 0.3807896298844658,
"learning_rate": 3.079381864824289e-05,
"loss": 0.7617,
"step": 589
},
{
"epoch": 3.076923076923077,
"grad_norm": 0.37153630367778145,
"learning_rate": 3.065151714688998e-05,
"loss": 0.7581,
"step": 590
},
{
"epoch": 3.0821382007822686,
"grad_norm": 0.41123425605078595,
"learning_rate": 3.0509340687006735e-05,
"loss": 0.7538,
"step": 591
},
{
"epoch": 3.08735332464146,
"grad_norm": 0.33981481049719536,
"learning_rate": 3.0367291170287102e-05,
"loss": 0.7524,
"step": 592
},
{
"epoch": 3.0925684485006517,
"grad_norm": 0.3651990332335353,
"learning_rate": 3.0225370496727125e-05,
"loss": 0.7584,
"step": 593
},
{
"epoch": 3.0977835723598437,
"grad_norm": 0.2918824973106232,
"learning_rate": 3.0083580564599484e-05,
"loss": 0.7563,
"step": 594
},
{
"epoch": 3.1029986962190352,
"grad_norm": 0.3423081259716029,
"learning_rate": 2.9941923270428125e-05,
"loss": 0.7471,
"step": 595
},
{
"epoch": 3.1082138200782268,
"grad_norm": 0.3604784090698713,
"learning_rate": 2.9800400508962854e-05,
"loss": 0.7458,
"step": 596
},
{
"epoch": 3.1134289439374183,
"grad_norm": 0.44107810877163156,
"learning_rate": 2.9659014173154053e-05,
"loss": 0.761,
"step": 597
},
{
"epoch": 3.1186440677966103,
"grad_norm": 0.36068201923449394,
"learning_rate": 2.9517766154127332e-05,
"loss": 0.7495,
"step": 598
},
{
"epoch": 3.123859191655802,
"grad_norm": 0.32775307660721575,
"learning_rate": 2.93766583411582e-05,
"loss": 0.7463,
"step": 599
},
{
"epoch": 3.1290743155149934,
"grad_norm": 0.4150847775843986,
"learning_rate": 2.923569262164684e-05,
"loss": 0.7459,
"step": 600
},
{
"epoch": 3.134289439374185,
"grad_norm": 0.3453247625813097,
"learning_rate": 2.909487088109287e-05,
"loss": 0.756,
"step": 601
},
{
"epoch": 3.139504563233377,
"grad_norm": 0.3884807540864722,
"learning_rate": 2.8954195003070106e-05,
"loss": 0.7469,
"step": 602
},
{
"epoch": 3.1447196870925684,
"grad_norm": 0.32406875922905615,
"learning_rate": 2.8813666869201323e-05,
"loss": 0.7492,
"step": 603
},
{
"epoch": 3.14993481095176,
"grad_norm": 0.3124660936977556,
"learning_rate": 2.867328835913319e-05,
"loss": 0.7412,
"step": 604
},
{
"epoch": 3.155149934810952,
"grad_norm": 0.402564278458608,
"learning_rate": 2.853306135051103e-05,
"loss": 0.7628,
"step": 605
},
{
"epoch": 3.1603650586701435,
"grad_norm": 0.3914753370029631,
"learning_rate": 2.8392987718953748e-05,
"loss": 0.7467,
"step": 606
},
{
"epoch": 3.165580182529335,
"grad_norm": 0.36094618715771837,
"learning_rate": 2.8253069338028756e-05,
"loss": 0.7528,
"step": 607
},
{
"epoch": 3.1707953063885266,
"grad_norm": 0.4160569360480524,
"learning_rate": 2.8113308079226897e-05,
"loss": 0.753,
"step": 608
},
{
"epoch": 3.1760104302477186,
"grad_norm": 0.36846901981179736,
"learning_rate": 2.7973705811937387e-05,
"loss": 0.762,
"step": 609
},
{
"epoch": 3.18122555410691,
"grad_norm": 0.3181951199229672,
"learning_rate": 2.7834264403422864e-05,
"loss": 0.7581,
"step": 610
},
{
"epoch": 3.1864406779661016,
"grad_norm": 0.28604043043751265,
"learning_rate": 2.769498571879438e-05,
"loss": 0.7591,
"step": 611
},
{
"epoch": 3.191655801825293,
"grad_norm": 0.2634604899316907,
"learning_rate": 2.7555871620986453e-05,
"loss": 0.7554,
"step": 612
},
{
"epoch": 3.196870925684485,
"grad_norm": 0.3039718480846403,
"learning_rate": 2.7416923970732146e-05,
"loss": 0.7588,
"step": 613
},
{
"epoch": 3.2020860495436767,
"grad_norm": 0.288103576580876,
"learning_rate": 2.7278144626538212e-05,
"loss": 0.7487,
"step": 614
},
{
"epoch": 3.2073011734028682,
"grad_norm": 0.32239905156935256,
"learning_rate": 2.71395354446602e-05,
"loss": 0.7459,
"step": 615
},
{
"epoch": 3.21251629726206,
"grad_norm": 0.29806440969996567,
"learning_rate": 2.7001098279077623e-05,
"loss": 0.7548,
"step": 616
},
{
"epoch": 3.2177314211212518,
"grad_norm": 0.25998375637060805,
"learning_rate": 2.6862834981469198e-05,
"loss": 0.7486,
"step": 617
},
{
"epoch": 3.2229465449804433,
"grad_norm": 0.39237992094856566,
"learning_rate": 2.6724747401188055e-05,
"loss": 0.7532,
"step": 618
},
{
"epoch": 3.228161668839635,
"grad_norm": 0.2548892235121485,
"learning_rate": 2.6586837385236966e-05,
"loss": 0.7484,
"step": 619
},
{
"epoch": 3.2333767926988264,
"grad_norm": 0.3225804950825967,
"learning_rate": 2.6449106778243708e-05,
"loss": 0.7529,
"step": 620
},
{
"epoch": 3.2385919165580184,
"grad_norm": 0.45928173129349054,
"learning_rate": 2.631155742243637e-05,
"loss": 0.7549,
"step": 621
},
{
"epoch": 3.24380704041721,
"grad_norm": 0.3308854647358001,
"learning_rate": 2.6174191157618652e-05,
"loss": 0.7554,
"step": 622
},
{
"epoch": 3.2490221642764014,
"grad_norm": 0.2795489427507208,
"learning_rate": 2.6037009821145346e-05,
"loss": 0.7593,
"step": 623
},
{
"epoch": 3.2542372881355934,
"grad_norm": 0.4430769575631773,
"learning_rate": 2.5900015247897714e-05,
"loss": 0.7545,
"step": 624
},
{
"epoch": 3.259452411994785,
"grad_norm": 0.33341725027604363,
"learning_rate": 2.576320927025892e-05,
"loss": 0.753,
"step": 625
},
{
"epoch": 3.2646675358539765,
"grad_norm": 0.33122719602795536,
"learning_rate": 2.562659371808959e-05,
"loss": 0.7436,
"step": 626
},
{
"epoch": 3.269882659713168,
"grad_norm": 0.3963689954079041,
"learning_rate": 2.549017041870325e-05,
"loss": 0.7676,
"step": 627
},
{
"epoch": 3.27509778357236,
"grad_norm": 0.3538575496536462,
"learning_rate": 2.535394119684197e-05,
"loss": 0.7548,
"step": 628
},
{
"epoch": 3.2803129074315516,
"grad_norm": 0.24662943382751912,
"learning_rate": 2.521790787465193e-05,
"loss": 0.7517,
"step": 629
},
{
"epoch": 3.285528031290743,
"grad_norm": 0.39858795017503024,
"learning_rate": 2.5082072271658982e-05,
"loss": 0.7482,
"step": 630
},
{
"epoch": 3.2907431551499347,
"grad_norm": 0.24293852218374687,
"learning_rate": 2.494643620474442e-05,
"loss": 0.7445,
"step": 631
},
{
"epoch": 3.2959582790091266,
"grad_norm": 0.3499814647894016,
"learning_rate": 2.4811001488120598e-05,
"loss": 0.7555,
"step": 632
},
{
"epoch": 3.301173402868318,
"grad_norm": 0.27161500960796325,
"learning_rate": 2.4675769933306696e-05,
"loss": 0.7568,
"step": 633
},
{
"epoch": 3.3063885267275097,
"grad_norm": 0.2881787773315299,
"learning_rate": 2.4540743349104485e-05,
"loss": 0.7442,
"step": 634
},
{
"epoch": 3.3116036505867013,
"grad_norm": 0.2740026111414901,
"learning_rate": 2.440592354157415e-05,
"loss": 0.7579,
"step": 635
},
{
"epoch": 3.3168187744458932,
"grad_norm": 0.2605377452426752,
"learning_rate": 2.427131231401008e-05,
"loss": 0.7518,
"step": 636
},
{
"epoch": 3.3220338983050848,
"grad_norm": 0.2482525542273476,
"learning_rate": 2.413691146691681e-05,
"loss": 0.7615,
"step": 637
},
{
"epoch": 3.3272490221642763,
"grad_norm": 0.33206747294953354,
"learning_rate": 2.400272279798492e-05,
"loss": 0.7498,
"step": 638
},
{
"epoch": 3.332464146023468,
"grad_norm": 0.29080523675344855,
"learning_rate": 2.3868748102066964e-05,
"loss": 0.7457,
"step": 639
},
{
"epoch": 3.33767926988266,
"grad_norm": 0.31636040314912883,
"learning_rate": 2.3734989171153475e-05,
"loss": 0.7605,
"step": 640
},
{
"epoch": 3.3428943937418514,
"grad_norm": 0.3319918116811582,
"learning_rate": 2.3601447794349035e-05,
"loss": 0.7601,
"step": 641
},
{
"epoch": 3.348109517601043,
"grad_norm": 0.24931361371615676,
"learning_rate": 2.346812575784826e-05,
"loss": 0.752,
"step": 642
},
{
"epoch": 3.353324641460235,
"grad_norm": 0.30718347134905344,
"learning_rate": 2.333502484491202e-05,
"loss": 0.7616,
"step": 643
},
{
"epoch": 3.3585397653194264,
"grad_norm": 0.282987877244827,
"learning_rate": 2.3202146835843458e-05,
"loss": 0.7462,
"step": 644
},
{
"epoch": 3.363754889178618,
"grad_norm": 0.24823892221787283,
"learning_rate": 2.3069493507964328e-05,
"loss": 0.7508,
"step": 645
},
{
"epoch": 3.3689700130378095,
"grad_norm": 0.2747764846839421,
"learning_rate": 2.2937066635591082e-05,
"loss": 0.757,
"step": 646
},
{
"epoch": 3.374185136897001,
"grad_norm": 0.2496756674179984,
"learning_rate": 2.2804867990011186e-05,
"loss": 0.7578,
"step": 647
},
{
"epoch": 3.379400260756193,
"grad_norm": 0.2395621072609135,
"learning_rate": 2.26728993394595e-05,
"loss": 0.765,
"step": 648
},
{
"epoch": 3.3846153846153846,
"grad_norm": 0.25006905141016617,
"learning_rate": 2.2541162449094517e-05,
"loss": 0.7558,
"step": 649
},
{
"epoch": 3.389830508474576,
"grad_norm": 0.22593196210127933,
"learning_rate": 2.2409659080974792e-05,
"loss": 0.7441,
"step": 650
},
{
"epoch": 3.395045632333768,
"grad_norm": 0.25905615560125267,
"learning_rate": 2.2278390994035437e-05,
"loss": 0.759,
"step": 651
},
{
"epoch": 3.4002607561929596,
"grad_norm": 0.21460137257291464,
"learning_rate": 2.214735994406449e-05,
"loss": 0.7356,
"step": 652
},
{
"epoch": 3.405475880052151,
"grad_norm": 0.2844041034563339,
"learning_rate": 2.2016567683679474e-05,
"loss": 0.7611,
"step": 653
},
{
"epoch": 3.4106910039113427,
"grad_norm": 0.2193094602636308,
"learning_rate": 2.188601596230402e-05,
"loss": 0.7628,
"step": 654
},
{
"epoch": 3.4159061277705347,
"grad_norm": 0.2688886990907944,
"learning_rate": 2.1755706526144346e-05,
"loss": 0.7555,
"step": 655
},
{
"epoch": 3.4211212516297262,
"grad_norm": 0.22979387277456514,
"learning_rate": 2.1625641118165993e-05,
"loss": 0.7506,
"step": 656
},
{
"epoch": 3.426336375488918,
"grad_norm": 0.2575836540395883,
"learning_rate": 2.1495821478070463e-05,
"loss": 0.7454,
"step": 657
},
{
"epoch": 3.4315514993481093,
"grad_norm": 0.22113251166352366,
"learning_rate": 2.1366249342271983e-05,
"loss": 0.7607,
"step": 658
},
{
"epoch": 3.4367666232073013,
"grad_norm": 0.2541994114755118,
"learning_rate": 2.123692644387427e-05,
"loss": 0.7529,
"step": 659
},
{
"epoch": 3.441981747066493,
"grad_norm": 0.17984867947865277,
"learning_rate": 2.110785451264733e-05,
"loss": 0.7584,
"step": 660
},
{
"epoch": 3.4471968709256844,
"grad_norm": 0.2623173957981238,
"learning_rate": 2.0979035275004326e-05,
"loss": 0.7589,
"step": 661
},
{
"epoch": 3.4524119947848764,
"grad_norm": 0.17573507602846267,
"learning_rate": 2.0850470453978527e-05,
"loss": 0.7575,
"step": 662
},
{
"epoch": 3.457627118644068,
"grad_norm": 0.23281049598480352,
"learning_rate": 2.07221617692002e-05,
"loss": 0.7495,
"step": 663
},
{
"epoch": 3.4628422425032594,
"grad_norm": 0.19792669337262203,
"learning_rate": 2.0594110936873646e-05,
"loss": 0.7554,
"step": 664
},
{
"epoch": 3.468057366362451,
"grad_norm": 0.2348813479171187,
"learning_rate": 2.0466319669754257e-05,
"loss": 0.7459,
"step": 665
},
{
"epoch": 3.4732724902216425,
"grad_norm": 0.17805434883842863,
"learning_rate": 2.033878967712556e-05,
"loss": 0.7482,
"step": 666
},
{
"epoch": 3.4784876140808345,
"grad_norm": 0.2324152396008409,
"learning_rate": 2.0211522664776378e-05,
"loss": 0.7627,
"step": 667
},
{
"epoch": 3.483702737940026,
"grad_norm": 0.21569790923016002,
"learning_rate": 2.0084520334978066e-05,
"loss": 0.7475,
"step": 668
},
{
"epoch": 3.4889178617992176,
"grad_norm": 0.22787366870500253,
"learning_rate": 1.9957784386461642e-05,
"loss": 0.7539,
"step": 669
},
{
"epoch": 3.4941329856584096,
"grad_norm": 0.19346838428327817,
"learning_rate": 1.9831316514395125e-05,
"loss": 0.7482,
"step": 670
},
{
"epoch": 3.499348109517601,
"grad_norm": 0.22386565166114242,
"learning_rate": 1.9705118410360888e-05,
"loss": 0.7516,
"step": 671
},
{
"epoch": 3.5045632333767927,
"grad_norm": 0.19045550096072955,
"learning_rate": 1.9579191762332962e-05,
"loss": 0.7524,
"step": 672
},
{
"epoch": 3.509778357235984,
"grad_norm": 0.2568792722114023,
"learning_rate": 1.9453538254654492e-05,
"loss": 0.7608,
"step": 673
},
{
"epoch": 3.514993481095176,
"grad_norm": 0.2309953251505408,
"learning_rate": 1.9328159568015253e-05,
"loss": 0.761,
"step": 674
},
{
"epoch": 3.5202086049543677,
"grad_norm": 0.20090449798296708,
"learning_rate": 1.920305737942908e-05,
"loss": 0.7498,
"step": 675
},
{
"epoch": 3.5254237288135593,
"grad_norm": 0.1995433496289568,
"learning_rate": 1.9078233362211488e-05,
"loss": 0.7635,
"step": 676
},
{
"epoch": 3.530638852672751,
"grad_norm": 0.20023193260506214,
"learning_rate": 1.895368918595733e-05,
"loss": 0.7497,
"step": 677
},
{
"epoch": 3.5358539765319428,
"grad_norm": 0.19621170906491942,
"learning_rate": 1.882942651651835e-05,
"loss": 0.7434,
"step": 678
},
{
"epoch": 3.5410691003911343,
"grad_norm": 0.20574042364789344,
"learning_rate": 1.8705447015981056e-05,
"loss": 0.7436,
"step": 679
},
{
"epoch": 3.546284224250326,
"grad_norm": 0.18782121755639544,
"learning_rate": 1.858175234264434e-05,
"loss": 0.7556,
"step": 680
},
{
"epoch": 3.551499348109518,
"grad_norm": 0.1936150072481402,
"learning_rate": 1.845834415099736e-05,
"loss": 0.757,
"step": 681
},
{
"epoch": 3.5567144719687094,
"grad_norm": 0.1926949621185055,
"learning_rate": 1.833522409169748e-05,
"loss": 0.756,
"step": 682
},
{
"epoch": 3.561929595827901,
"grad_norm": 0.20476384841537432,
"learning_rate": 1.8212393811548074e-05,
"loss": 0.7341,
"step": 683
},
{
"epoch": 3.5671447196870925,
"grad_norm": 0.1696138534253153,
"learning_rate": 1.8089854953476503e-05,
"loss": 0.7521,
"step": 684
},
{
"epoch": 3.572359843546284,
"grad_norm": 0.22686575356319796,
"learning_rate": 1.796760915651229e-05,
"loss": 0.7621,
"step": 685
},
{
"epoch": 3.577574967405476,
"grad_norm": 0.18860960425351292,
"learning_rate": 1.784565805576503e-05,
"loss": 0.7537,
"step": 686
},
{
"epoch": 3.5827900912646675,
"grad_norm": 0.19640065748239693,
"learning_rate": 1.7724003282402567e-05,
"loss": 0.7592,
"step": 687
},
{
"epoch": 3.588005215123859,
"grad_norm": 0.18973579313916158,
"learning_rate": 1.760264646362926e-05,
"loss": 0.7477,
"step": 688
},
{
"epoch": 3.593220338983051,
"grad_norm": 0.20298759538425024,
"learning_rate": 1.7481589222664076e-05,
"loss": 0.7486,
"step": 689
},
{
"epoch": 3.5984354628422426,
"grad_norm": 0.17790150008707342,
"learning_rate": 1.736083317871898e-05,
"loss": 0.7458,
"step": 690
},
{
"epoch": 3.603650586701434,
"grad_norm": 0.20575599062083585,
"learning_rate": 1.7240379946977283e-05,
"loss": 0.7544,
"step": 691
},
{
"epoch": 3.6088657105606257,
"grad_norm": 0.17855398739046716,
"learning_rate": 1.712023113857195e-05,
"loss": 0.756,
"step": 692
},
{
"epoch": 3.614080834419817,
"grad_norm": 0.1947379717329557,
"learning_rate": 1.700038836056413e-05,
"loss": 0.7547,
"step": 693
},
{
"epoch": 3.619295958279009,
"grad_norm": 0.17740265903389008,
"learning_rate": 1.6880853215921642e-05,
"loss": 0.7532,
"step": 694
},
{
"epoch": 3.6245110821382007,
"grad_norm": 0.18082866727173452,
"learning_rate": 1.676162730349751e-05,
"loss": 0.7466,
"step": 695
},
{
"epoch": 3.6297262059973923,
"grad_norm": 0.18963190902387492,
"learning_rate": 1.6642712218008613e-05,
"loss": 0.742,
"step": 696
},
{
"epoch": 3.6349413298565842,
"grad_norm": 0.18716923070682592,
"learning_rate": 1.6524109550014323e-05,
"loss": 0.7515,
"step": 697
},
{
"epoch": 3.640156453715776,
"grad_norm": 0.20784228257641144,
"learning_rate": 1.6405820885895242e-05,
"loss": 0.7578,
"step": 698
},
{
"epoch": 3.6453715775749673,
"grad_norm": 0.1920361900700138,
"learning_rate": 1.6287847807832e-05,
"loss": 0.7538,
"step": 699
},
{
"epoch": 3.6505867014341593,
"grad_norm": 0.1652320731622616,
"learning_rate": 1.617019189378407e-05,
"loss": 0.7511,
"step": 700
},
{
"epoch": 3.655801825293351,
"grad_norm": 0.19431451990268614,
"learning_rate": 1.6052854717468644e-05,
"loss": 0.7515,
"step": 701
},
{
"epoch": 3.6610169491525424,
"grad_norm": 0.19087250562535932,
"learning_rate": 1.5935837848339667e-05,
"loss": 0.7568,
"step": 702
},
{
"epoch": 3.666232073011734,
"grad_norm": 0.20029152301458664,
"learning_rate": 1.581914285156673e-05,
"loss": 0.7465,
"step": 703
},
{
"epoch": 3.6714471968709255,
"grad_norm": 0.18770482135472025,
"learning_rate": 1.5702771288014206e-05,
"loss": 0.7516,
"step": 704
},
{
"epoch": 3.6766623207301175,
"grad_norm": 0.17463919163255975,
"learning_rate": 1.5586724714220397e-05,
"loss": 0.7498,
"step": 705
},
{
"epoch": 3.681877444589309,
"grad_norm": 0.1942555556221587,
"learning_rate": 1.5471004682376626e-05,
"loss": 0.743,
"step": 706
},
{
"epoch": 3.6870925684485005,
"grad_norm": 0.1946562516538543,
"learning_rate": 1.535561274030652e-05,
"loss": 0.7569,
"step": 707
},
{
"epoch": 3.6923076923076925,
"grad_norm": 0.17582331062620302,
"learning_rate": 1.5240550431445376e-05,
"loss": 0.7519,
"step": 708
},
{
"epoch": 3.697522816166884,
"grad_norm": 0.1861426528732765,
"learning_rate": 1.5125819294819391e-05,
"loss": 0.7482,
"step": 709
},
{
"epoch": 3.7027379400260756,
"grad_norm": 0.18269562874012577,
"learning_rate": 1.5011420865025143e-05,
"loss": 0.7523,
"step": 710
},
{
"epoch": 3.707953063885267,
"grad_norm": 0.18506228544951506,
"learning_rate": 1.4897356672209112e-05,
"loss": 0.7547,
"step": 711
},
{
"epoch": 3.7131681877444587,
"grad_norm": 0.20416348855073826,
"learning_rate": 1.4783628242047097e-05,
"loss": 0.7502,
"step": 712
},
{
"epoch": 3.7183833116036507,
"grad_norm": 0.16997764410348623,
"learning_rate": 1.467023709572391e-05,
"loss": 0.7512,
"step": 713
},
{
"epoch": 3.723598435462842,
"grad_norm": 0.18897244039596858,
"learning_rate": 1.4557184749912981e-05,
"loss": 0.7473,
"step": 714
},
{
"epoch": 3.7288135593220337,
"grad_norm": 0.19172065409291622,
"learning_rate": 1.4444472716756074e-05,
"loss": 0.7497,
"step": 715
},
{
"epoch": 3.7340286831812257,
"grad_norm": 0.171487618996453,
"learning_rate": 1.433210250384311e-05,
"loss": 0.7526,
"step": 716
},
{
"epoch": 3.7392438070404173,
"grad_norm": 0.17442752265085087,
"learning_rate": 1.4220075614191924e-05,
"loss": 0.7515,
"step": 717
},
{
"epoch": 3.744458930899609,
"grad_norm": 0.17508720890471086,
"learning_rate": 1.4108393546228194e-05,
"loss": 0.7578,
"step": 718
},
{
"epoch": 3.749674054758801,
"grad_norm": 0.19876602903764187,
"learning_rate": 1.3997057793765452e-05,
"loss": 0.7483,
"step": 719
},
{
"epoch": 3.7548891786179923,
"grad_norm": 0.17214365448338131,
"learning_rate": 1.3886069845985013e-05,
"loss": 0.7568,
"step": 720
},
{
"epoch": 3.760104302477184,
"grad_norm": 0.1920697005442559,
"learning_rate": 1.3775431187416097e-05,
"loss": 0.7504,
"step": 721
},
{
"epoch": 3.7653194263363754,
"grad_norm": 0.18503507230287503,
"learning_rate": 1.3665143297916018e-05,
"loss": 0.7636,
"step": 722
},
{
"epoch": 3.770534550195567,
"grad_norm": 0.19252969937704967,
"learning_rate": 1.3555207652650308e-05,
"loss": 0.7526,
"step": 723
},
{
"epoch": 3.775749674054759,
"grad_norm": 0.1554286223570067,
"learning_rate": 1.3445625722073023e-05,
"loss": 0.7512,
"step": 724
},
{
"epoch": 3.7809647979139505,
"grad_norm": 0.16978176727895614,
"learning_rate": 1.3336398971907114e-05,
"loss": 0.7556,
"step": 725
},
{
"epoch": 3.786179921773142,
"grad_norm": 0.15542348945118697,
"learning_rate": 1.3227528863124745e-05,
"loss": 0.7512,
"step": 726
},
{
"epoch": 3.791395045632334,
"grad_norm": 0.16399605746487972,
"learning_rate": 1.3119016851927798e-05,
"loss": 0.756,
"step": 727
},
{
"epoch": 3.7966101694915255,
"grad_norm": 0.16707283374852486,
"learning_rate": 1.301086438972842e-05,
"loss": 0.7475,
"step": 728
},
{
"epoch": 3.801825293350717,
"grad_norm": 0.15983694259778197,
"learning_rate": 1.2903072923129547e-05,
"loss": 0.7474,
"step": 729
},
{
"epoch": 3.8070404172099086,
"grad_norm": 0.14876964520461528,
"learning_rate": 1.2795643893905575e-05,
"loss": 0.7471,
"step": 730
},
{
"epoch": 3.8122555410691,
"grad_norm": 0.19924804420654435,
"learning_rate": 1.2688578738983131e-05,
"loss": 0.7464,
"step": 731
},
{
"epoch": 3.817470664928292,
"grad_norm": 0.19191423801767207,
"learning_rate": 1.2581878890421777e-05,
"loss": 0.7562,
"step": 732
},
{
"epoch": 3.8226857887874837,
"grad_norm": 0.20765129848679256,
"learning_rate": 1.2475545775394879e-05,
"loss": 0.7565,
"step": 733
},
{
"epoch": 3.827900912646675,
"grad_norm": 0.1697957267489159,
"learning_rate": 1.2369580816170563e-05,
"loss": 0.7586,
"step": 734
},
{
"epoch": 3.833116036505867,
"grad_norm": 0.1716215315213691,
"learning_rate": 1.2263985430092618e-05,
"loss": 0.7524,
"step": 735
},
{
"epoch": 3.8383311603650587,
"grad_norm": 0.18944814607450408,
"learning_rate": 1.2158761029561582e-05,
"loss": 0.7547,
"step": 736
},
{
"epoch": 3.8435462842242503,
"grad_norm": 0.14989091598529683,
"learning_rate": 1.2053909022015863e-05,
"loss": 0.7446,
"step": 737
},
{
"epoch": 3.8487614080834422,
"grad_norm": 0.1858141017872855,
"learning_rate": 1.194943080991287e-05,
"loss": 0.7503,
"step": 738
},
{
"epoch": 3.853976531942634,
"grad_norm": 0.14994883200639647,
"learning_rate": 1.1845327790710276e-05,
"loss": 0.747,
"step": 739
},
{
"epoch": 3.8591916558018253,
"grad_norm": 0.14941004189650142,
"learning_rate": 1.1741601356847365e-05,
"loss": 0.7541,
"step": 740
},
{
"epoch": 3.864406779661017,
"grad_norm": 0.16071992101640972,
"learning_rate": 1.1638252895726327e-05,
"loss": 0.7496,
"step": 741
},
{
"epoch": 3.8696219035202084,
"grad_norm": 0.1374205892222837,
"learning_rate": 1.1535283789693756e-05,
"loss": 0.753,
"step": 742
},
{
"epoch": 3.8748370273794004,
"grad_norm": 0.1394433360371365,
"learning_rate": 1.1432695416022158e-05,
"loss": 0.7535,
"step": 743
},
{
"epoch": 3.880052151238592,
"grad_norm": 0.16866985633081483,
"learning_rate": 1.1330489146891494e-05,
"loss": 0.7517,
"step": 744
},
{
"epoch": 3.8852672750977835,
"grad_norm": 0.15730043046261316,
"learning_rate": 1.1228666349370897e-05,
"loss": 0.7474,
"step": 745
},
{
"epoch": 3.8904823989569755,
"grad_norm": 0.1848086766193644,
"learning_rate": 1.112722838540031e-05,
"loss": 0.7615,
"step": 746
},
{
"epoch": 3.895697522816167,
"grad_norm": 0.16625357168861077,
"learning_rate": 1.1026176611772286e-05,
"loss": 0.7488,
"step": 747
},
{
"epoch": 3.9009126466753585,
"grad_norm": 0.1629255000824365,
"learning_rate": 1.0925512380113892e-05,
"loss": 0.7577,
"step": 748
},
{
"epoch": 3.90612777053455,
"grad_norm": 0.18675770122521296,
"learning_rate": 1.0825237036868575e-05,
"loss": 0.7531,
"step": 749
},
{
"epoch": 3.9113428943937416,
"grad_norm": 0.15755029689979122,
"learning_rate": 1.0725351923278144e-05,
"loss": 0.7548,
"step": 750
},
{
"epoch": 3.9165580182529336,
"grad_norm": 0.18913533326276255,
"learning_rate": 1.0625858375364917e-05,
"loss": 0.754,
"step": 751
},
{
"epoch": 3.921773142112125,
"grad_norm": 0.18170716935082054,
"learning_rate": 1.0526757723913735e-05,
"loss": 0.7441,
"step": 752
},
{
"epoch": 3.9269882659713167,
"grad_norm": 0.13608488064261914,
"learning_rate": 1.042805129445423e-05,
"loss": 0.7508,
"step": 753
},
{
"epoch": 3.9322033898305087,
"grad_norm": 0.17632467806971025,
"learning_rate": 1.0329740407243105e-05,
"loss": 0.7747,
"step": 754
},
{
"epoch": 3.9374185136897,
"grad_norm": 0.17072961754308208,
"learning_rate": 1.0231826377246437e-05,
"loss": 0.7462,
"step": 755
},
{
"epoch": 3.9426336375488917,
"grad_norm": 0.16478645139827036,
"learning_rate": 1.0134310514122082e-05,
"loss": 0.7563,
"step": 756
},
{
"epoch": 3.9478487614080837,
"grad_norm": 0.17542376854887573,
"learning_rate": 1.0037194122202227e-05,
"loss": 0.7567,
"step": 757
},
{
"epoch": 3.9530638852672753,
"grad_norm": 0.17159268136488617,
"learning_rate": 9.940478500475858e-06,
"loss": 0.7503,
"step": 758
},
{
"epoch": 3.958279009126467,
"grad_norm": 0.15300026372913983,
"learning_rate": 9.844164942571424e-06,
"loss": 0.7523,
"step": 759
},
{
"epoch": 3.9634941329856583,
"grad_norm": 0.159680621058249,
"learning_rate": 9.748254736739571e-06,
"loss": 0.757,
"step": 760
},
{
"epoch": 3.96870925684485,
"grad_norm": 0.15768593487660648,
"learning_rate": 9.652749165835828e-06,
"loss": 0.7458,
"step": 761
},
{
"epoch": 3.973924380704042,
"grad_norm": 0.16943127341342135,
"learning_rate": 9.557649507303508e-06,
"loss": 0.7456,
"step": 762
},
{
"epoch": 3.9791395045632334,
"grad_norm": 0.14842747623637131,
"learning_rate": 9.462957033156632e-06,
"loss": 0.7497,
"step": 763
},
{
"epoch": 3.984354628422425,
"grad_norm": 0.1501263870664962,
"learning_rate": 9.368673009962852e-06,
"loss": 0.7514,
"step": 764
},
{
"epoch": 3.989569752281617,
"grad_norm": 0.17717065342340108,
"learning_rate": 9.274798698826557e-06,
"loss": 0.7531,
"step": 765
},
{
"epoch": 3.9947848761408085,
"grad_norm": 0.15866607398462376,
"learning_rate": 9.181335355372028e-06,
"loss": 0.7512,
"step": 766
},
{
"epoch": 4.0,
"grad_norm": 0.17952277081095838,
"learning_rate": 9.088284229726572e-06,
"loss": 0.9412,
"step": 767
},
{
"epoch": 4.005215123859192,
"grad_norm": 0.23274451897753248,
"learning_rate": 8.995646566503838e-06,
"loss": 0.7375,
"step": 768
},
{
"epoch": 4.010430247718383,
"grad_norm": 0.17288112228894242,
"learning_rate": 8.90342360478723e-06,
"loss": 0.7373,
"step": 769
},
{
"epoch": 4.015645371577575,
"grad_norm": 0.20174270563988336,
"learning_rate": 8.81161657811318e-06,
"loss": 0.7453,
"step": 770
},
{
"epoch": 4.020860495436767,
"grad_norm": 0.1726715461295719,
"learning_rate": 8.720226714454831e-06,
"loss": 0.7421,
"step": 771
},
{
"epoch": 4.026075619295958,
"grad_norm": 0.20664612564398913,
"learning_rate": 8.629255236205475e-06,
"loss": 0.7393,
"step": 772
},
{
"epoch": 4.03129074315515,
"grad_norm": 0.183032821290067,
"learning_rate": 8.538703360162267e-06,
"loss": 0.7262,
"step": 773
},
{
"epoch": 4.036505867014341,
"grad_norm": 0.19162204178734604,
"learning_rate": 8.44857229750994e-06,
"loss": 0.7229,
"step": 774
},
{
"epoch": 4.041720990873533,
"grad_norm": 0.19714648419474126,
"learning_rate": 8.3588632538046e-06,
"loss": 0.7406,
"step": 775
},
{
"epoch": 4.046936114732725,
"grad_norm": 0.18562646027448648,
"learning_rate": 8.269577428957571e-06,
"loss": 0.7375,
"step": 776
},
{
"epoch": 4.052151238591916,
"grad_norm": 0.19319081667416735,
"learning_rate": 8.180716017219433e-06,
"loss": 0.7315,
"step": 777
},
{
"epoch": 4.057366362451108,
"grad_norm": 0.17933639021135425,
"learning_rate": 8.092280207163941e-06,
"loss": 0.7439,
"step": 778
},
{
"epoch": 4.0625814863103,
"grad_norm": 0.16796265908006924,
"learning_rate": 8.004271181672178e-06,
"loss": 0.7219,
"step": 779
},
{
"epoch": 4.067796610169491,
"grad_norm": 0.16506690334266957,
"learning_rate": 7.916690117916772e-06,
"loss": 0.7409,
"step": 780
},
{
"epoch": 4.073011734028683,
"grad_norm": 0.16443977378096947,
"learning_rate": 7.829538187346077e-06,
"loss": 0.7377,
"step": 781
},
{
"epoch": 4.078226857887874,
"grad_norm": 0.1997012117535286,
"learning_rate": 7.742816555668532e-06,
"loss": 0.7399,
"step": 782
},
{
"epoch": 4.083441981747066,
"grad_norm": 0.14695927923208477,
"learning_rate": 7.656526382837106e-06,
"loss": 0.7327,
"step": 783
},
{
"epoch": 4.088657105606258,
"grad_norm": 0.1799749838620325,
"learning_rate": 7.570668823033727e-06,
"loss": 0.7309,
"step": 784
},
{
"epoch": 4.0938722294654495,
"grad_norm": 0.17989297489241943,
"learning_rate": 7.4852450246538685e-06,
"loss": 0.7343,
"step": 785
},
{
"epoch": 4.0990873533246415,
"grad_norm": 0.15506025585209593,
"learning_rate": 7.40025613029121e-06,
"loss": 0.736,
"step": 786
},
{
"epoch": 4.1043024771838335,
"grad_norm": 0.18105126244534,
"learning_rate": 7.315703276722317e-06,
"loss": 0.7342,
"step": 787
},
{
"epoch": 4.1095176010430245,
"grad_norm": 0.19050717050961422,
"learning_rate": 7.231587594891438e-06,
"loss": 0.7458,
"step": 788
},
{
"epoch": 4.1147327249022165,
"grad_norm": 0.14216335950493533,
"learning_rate": 7.147910209895435e-06,
"loss": 0.7318,
"step": 789
},
{
"epoch": 4.1199478487614085,
"grad_norm": 0.1655844728014387,
"learning_rate": 7.064672240968638e-06,
"loss": 0.7442,
"step": 790
},
{
"epoch": 4.1251629726206,
"grad_norm": 0.14883381171151777,
"learning_rate": 6.9818748014679785e-06,
"loss": 0.7403,
"step": 791
},
{
"epoch": 4.130378096479792,
"grad_norm": 0.1602685057342838,
"learning_rate": 6.899518998858017e-06,
"loss": 0.7405,
"step": 792
},
{
"epoch": 4.135593220338983,
"grad_norm": 0.1649659601993552,
"learning_rate": 6.81760593469615e-06,
"loss": 0.7346,
"step": 793
},
{
"epoch": 4.140808344198175,
"grad_norm": 0.12921893028933829,
"learning_rate": 6.7361367046179325e-06,
"loss": 0.7303,
"step": 794
},
{
"epoch": 4.146023468057367,
"grad_norm": 0.12489877860585255,
"learning_rate": 6.655112398322332e-06,
"loss": 0.731,
"step": 795
},
{
"epoch": 4.151238591916558,
"grad_norm": 0.17428621687246298,
"learning_rate": 6.574534099557204e-06,
"loss": 0.7307,
"step": 796
},
{
"epoch": 4.15645371577575,
"grad_norm": 0.1251987161714777,
"learning_rate": 6.49440288610482e-06,
"loss": 0.7275,
"step": 797
},
{
"epoch": 4.161668839634942,
"grad_norm": 0.15533418379259104,
"learning_rate": 6.414719829767406e-06,
"loss": 0.7339,
"step": 798
},
{
"epoch": 4.166883963494133,
"grad_norm": 0.16236363314856642,
"learning_rate": 6.335485996352782e-06,
"loss": 0.7418,
"step": 799
},
{
"epoch": 4.172099087353325,
"grad_norm": 0.13801981121095086,
"learning_rate": 6.256702445660221e-06,
"loss": 0.7304,
"step": 800
},
{
"epoch": 4.177314211212516,
"grad_norm": 0.14142425026968236,
"learning_rate": 6.178370231466142e-06,
"loss": 0.743,
"step": 801
},
{
"epoch": 4.182529335071708,
"grad_norm": 0.16216801700272557,
"learning_rate": 6.100490401510089e-06,
"loss": 0.7437,
"step": 802
},
{
"epoch": 4.1877444589309,
"grad_norm": 0.14030637701048476,
"learning_rate": 6.0230639974807206e-06,
"loss": 0.7383,
"step": 803
},
{
"epoch": 4.192959582790091,
"grad_norm": 0.13754640300037912,
"learning_rate": 5.946092055001824e-06,
"loss": 0.7329,
"step": 804
},
{
"epoch": 4.198174706649283,
"grad_norm": 0.16139415971113086,
"learning_rate": 5.869575603618507e-06,
"loss": 0.7417,
"step": 805
},
{
"epoch": 4.203389830508475,
"grad_norm": 0.15626356140455866,
"learning_rate": 5.793515666783429e-06,
"loss": 0.7337,
"step": 806
},
{
"epoch": 4.208604954367666,
"grad_norm": 0.12177414306535114,
"learning_rate": 5.717913261843078e-06,
"loss": 0.733,
"step": 807
},
{
"epoch": 4.213820078226858,
"grad_norm": 0.12926025156980447,
"learning_rate": 5.642769400024199e-06,
"loss": 0.7422,
"step": 808
},
{
"epoch": 4.219035202086049,
"grad_norm": 0.13704261747895924,
"learning_rate": 5.568085086420247e-06,
"loss": 0.7404,
"step": 809
},
{
"epoch": 4.224250325945241,
"grad_norm": 0.13842188890961235,
"learning_rate": 5.4938613199779334e-06,
"loss": 0.7334,
"step": 810
},
{
"epoch": 4.229465449804433,
"grad_norm": 0.13979962035636995,
"learning_rate": 5.420099093483911e-06,
"loss": 0.7222,
"step": 811
},
{
"epoch": 4.234680573663624,
"grad_norm": 0.12565163857223183,
"learning_rate": 5.346799393551436e-06,
"loss": 0.7307,
"step": 812
},
{
"epoch": 4.239895697522816,
"grad_norm": 0.12679243834545414,
"learning_rate": 5.273963200607206e-06,
"loss": 0.7387,
"step": 813
},
{
"epoch": 4.245110821382008,
"grad_norm": 0.13813903814909753,
"learning_rate": 5.201591488878257e-06,
"loss": 0.7251,
"step": 814
},
{
"epoch": 4.250325945241199,
"grad_norm": 0.1285263356419264,
"learning_rate": 5.129685226378893e-06,
"loss": 0.7234,
"step": 815
},
{
"epoch": 4.255541069100391,
"grad_norm": 0.12820199717697753,
"learning_rate": 5.058245374897763e-06,
"loss": 0.7271,
"step": 816
},
{
"epoch": 4.260756192959583,
"grad_norm": 0.12495602397125814,
"learning_rate": 4.987272889985009e-06,
"loss": 0.7393,
"step": 817
},
{
"epoch": 4.265971316818774,
"grad_norm": 0.15151715317554978,
"learning_rate": 4.916768720939451e-06,
"loss": 0.7332,
"step": 818
},
{
"epoch": 4.271186440677966,
"grad_norm": 0.14747133461014475,
"learning_rate": 4.846733810795914e-06,
"loss": 0.7297,
"step": 819
},
{
"epoch": 4.276401564537157,
"grad_norm": 0.14100514701144612,
"learning_rate": 4.777169096312624e-06,
"loss": 0.7276,
"step": 820
},
{
"epoch": 4.281616688396349,
"grad_norm": 0.1646113420856527,
"learning_rate": 4.708075507958642e-06,
"loss": 0.7281,
"step": 821
},
{
"epoch": 4.286831812255541,
"grad_norm": 0.13983624281016735,
"learning_rate": 4.639453969901442e-06,
"loss": 0.7276,
"step": 822
},
{
"epoch": 4.292046936114732,
"grad_norm": 0.13237794200876413,
"learning_rate": 4.571305399994561e-06,
"loss": 0.7409,
"step": 823
},
{
"epoch": 4.297262059973924,
"grad_norm": 0.180777223722747,
"learning_rate": 4.503630709765294e-06,
"loss": 0.741,
"step": 824
},
{
"epoch": 4.302477183833116,
"grad_norm": 0.1367694718240071,
"learning_rate": 4.436430804402521e-06,
"loss": 0.7331,
"step": 825
},
{
"epoch": 4.3076923076923075,
"grad_norm": 0.15547588436169807,
"learning_rate": 4.3697065827446e-06,
"loss": 0.7376,
"step": 826
},
{
"epoch": 4.3129074315514995,
"grad_norm": 0.17342743882348932,
"learning_rate": 4.303458937267326e-06,
"loss": 0.7267,
"step": 827
},
{
"epoch": 4.318122555410691,
"grad_norm": 0.16682655867364918,
"learning_rate": 4.237688754072022e-06,
"loss": 0.7309,
"step": 828
},
{
"epoch": 4.3233376792698825,
"grad_norm": 0.16450623559503522,
"learning_rate": 4.172396912873659e-06,
"loss": 0.7342,
"step": 829
},
{
"epoch": 4.3285528031290745,
"grad_norm": 0.19380979384775532,
"learning_rate": 4.107584286989097e-06,
"loss": 0.7392,
"step": 830
},
{
"epoch": 4.333767926988266,
"grad_norm": 0.18721173933947008,
"learning_rate": 4.043251743325436e-06,
"loss": 0.7303,
"step": 831
},
{
"epoch": 4.338983050847458,
"grad_norm": 0.15835333882238062,
"learning_rate": 3.979400142368368e-06,
"loss": 0.7301,
"step": 832
},
{
"epoch": 4.34419817470665,
"grad_norm": 0.18385561342339599,
"learning_rate": 3.916030338170696e-06,
"loss": 0.7374,
"step": 833
},
{
"epoch": 4.349413298565841,
"grad_norm": 0.2022205872843893,
"learning_rate": 3.8531431783409165e-06,
"loss": 0.7387,
"step": 834
},
{
"epoch": 4.354628422425033,
"grad_norm": 0.15316197738525242,
"learning_rate": 3.7907395040318685e-06,
"loss": 0.7385,
"step": 835
},
{
"epoch": 4.359843546284225,
"grad_norm": 0.1500484412205804,
"learning_rate": 3.7288201499294797e-06,
"loss": 0.7268,
"step": 836
},
{
"epoch": 4.365058670143416,
"grad_norm": 0.14189535258541974,
"learning_rate": 3.667385944241626e-06,
"loss": 0.7453,
"step": 837
},
{
"epoch": 4.370273794002608,
"grad_norm": 0.15984630251802126,
"learning_rate": 3.606437708687023e-06,
"loss": 0.7395,
"step": 838
},
{
"epoch": 4.375488917861799,
"grad_norm": 0.1642822923840156,
"learning_rate": 3.545976258484247e-06,
"loss": 0.7354,
"step": 839
},
{
"epoch": 4.380704041720991,
"grad_norm": 0.13221533894356902,
"learning_rate": 3.4860024023408577e-06,
"loss": 0.7459,
"step": 840
},
{
"epoch": 4.385919165580183,
"grad_norm": 0.18318603173559903,
"learning_rate": 3.426516942442524e-06,
"loss": 0.7509,
"step": 841
},
{
"epoch": 4.391134289439374,
"grad_norm": 0.17040203525878148,
"learning_rate": 3.3675206744423394e-06,
"loss": 0.7375,
"step": 842
},
{
"epoch": 4.396349413298566,
"grad_norm": 0.157205271311182,
"learning_rate": 3.309014387450189e-06,
"loss": 0.7302,
"step": 843
},
{
"epoch": 4.401564537157758,
"grad_norm": 0.15260420463119376,
"learning_rate": 3.2509988640221414e-06,
"loss": 0.7407,
"step": 844
},
{
"epoch": 4.406779661016949,
"grad_norm": 0.16154440296062286,
"learning_rate": 3.1934748801500184e-06,
"loss": 0.7255,
"step": 845
},
{
"epoch": 4.411994784876141,
"grad_norm": 0.14388929124722757,
"learning_rate": 3.136443205251034e-06,
"loss": 0.7369,
"step": 846
},
{
"epoch": 4.417209908735332,
"grad_norm": 0.14793139417234433,
"learning_rate": 3.0799046021574574e-06,
"loss": 0.7337,
"step": 847
},
{
"epoch": 4.422425032594524,
"grad_norm": 0.14906886650426293,
"learning_rate": 3.0238598271064323e-06,
"loss": 0.7319,
"step": 848
},
{
"epoch": 4.427640156453716,
"grad_norm": 0.18885017044608088,
"learning_rate": 2.9683096297298753e-06,
"loss": 0.7226,
"step": 849
},
{
"epoch": 4.432855280312907,
"grad_norm": 0.16698890562230234,
"learning_rate": 2.9132547530444254e-06,
"loss": 0.74,
"step": 850
},
{
"epoch": 4.438070404172099,
"grad_norm": 0.12976275259498832,
"learning_rate": 2.8586959334415064e-06,
"loss": 0.7325,
"step": 851
},
{
"epoch": 4.443285528031291,
"grad_norm": 0.14807620276113978,
"learning_rate": 2.8046339006775115e-06,
"loss": 0.7197,
"step": 852
},
{
"epoch": 4.448500651890482,
"grad_norm": 0.15235231363332793,
"learning_rate": 2.751069377863993e-06,
"loss": 0.7302,
"step": 853
},
{
"epoch": 4.453715775749674,
"grad_norm": 0.15202060711926385,
"learning_rate": 2.698003081458018e-06,
"loss": 0.7289,
"step": 854
},
{
"epoch": 4.458930899608866,
"grad_norm": 0.12630616368200856,
"learning_rate": 2.6454357212526026e-06,
"loss": 0.7279,
"step": 855
},
{
"epoch": 4.464146023468057,
"grad_norm": 0.11739628579126643,
"learning_rate": 2.5933680003671625e-06,
"loss": 0.7335,
"step": 856
},
{
"epoch": 4.469361147327249,
"grad_norm": 0.1434952804888338,
"learning_rate": 2.541800615238175e-06,
"loss": 0.7266,
"step": 857
},
{
"epoch": 4.47457627118644,
"grad_norm": 0.176277045965935,
"learning_rate": 2.4907342556098125e-06,
"loss": 0.7269,
"step": 858
},
{
"epoch": 4.479791395045632,
"grad_norm": 0.18491464302255675,
"learning_rate": 2.440169604524738e-06,
"loss": 0.7453,
"step": 859
},
{
"epoch": 4.485006518904824,
"grad_norm": 0.12415001369259007,
"learning_rate": 2.390107338314982e-06,
"loss": 0.7319,
"step": 860
},
{
"epoch": 4.490221642764015,
"grad_norm": 0.10478588253479926,
"learning_rate": 2.340548126592861e-06,
"loss": 0.737,
"step": 861
},
{
"epoch": 4.495436766623207,
"grad_norm": 0.14165809120057415,
"learning_rate": 2.2914926322420515e-06,
"loss": 0.7343,
"step": 862
},
{
"epoch": 4.500651890482399,
"grad_norm": 0.11590106390711405,
"learning_rate": 2.242941511408714e-06,
"loss": 0.737,
"step": 863
},
{
"epoch": 4.50586701434159,
"grad_norm": 0.12634034846215603,
"learning_rate": 2.1948954134927103e-06,
"loss": 0.7377,
"step": 864
},
{
"epoch": 4.511082138200782,
"grad_norm": 0.11268712032715957,
"learning_rate": 2.147354981138925e-06,
"loss": 0.7257,
"step": 865
},
{
"epoch": 4.5162972620599735,
"grad_norm": 0.11916219994594991,
"learning_rate": 2.1003208502286744e-06,
"loss": 0.729,
"step": 866
},
{
"epoch": 4.5215123859191655,
"grad_norm": 0.13119722935564196,
"learning_rate": 2.0537936498711898e-06,
"loss": 0.7316,
"step": 867
},
{
"epoch": 4.5267275097783575,
"grad_norm": 0.12883622307786546,
"learning_rate": 2.007774002395193e-06,
"loss": 0.7337,
"step": 868
},
{
"epoch": 4.531942633637549,
"grad_norm": 0.10574304590110693,
"learning_rate": 1.9622625233406143e-06,
"loss": 0.7268,
"step": 869
},
{
"epoch": 4.5371577574967406,
"grad_norm": 0.11189460238334994,
"learning_rate": 1.9172598214503146e-06,
"loss": 0.736,
"step": 870
},
{
"epoch": 4.5423728813559325,
"grad_norm": 0.12169459450452168,
"learning_rate": 1.8727664986619576e-06,
"loss": 0.7359,
"step": 871
},
{
"epoch": 4.547588005215124,
"grad_norm": 0.11602787776789947,
"learning_rate": 1.8287831500999775e-06,
"loss": 0.733,
"step": 872
},
{
"epoch": 4.552803129074316,
"grad_norm": 0.11542784974890338,
"learning_rate": 1.7853103640675852e-06,
"loss": 0.7274,
"step": 873
},
{
"epoch": 4.558018252933508,
"grad_norm": 0.10042034550407807,
"learning_rate": 1.74234872203892e-06,
"loss": 0.7346,
"step": 874
},
{
"epoch": 4.563233376792699,
"grad_norm": 0.11267524138272358,
"learning_rate": 1.699898798651285e-06,
"loss": 0.7355,
"step": 875
},
{
"epoch": 4.568448500651891,
"grad_norm": 0.14020374405591798,
"learning_rate": 1.6579611616974212e-06,
"loss": 0.7323,
"step": 876
},
{
"epoch": 4.573663624511082,
"grad_norm": 0.1329010216092144,
"learning_rate": 1.6165363721179471e-06,
"loss": 0.7333,
"step": 877
},
{
"epoch": 4.578878748370274,
"grad_norm": 0.1111745646565014,
"learning_rate": 1.575624983993853e-06,
"loss": 0.7198,
"step": 878
},
{
"epoch": 4.584093872229466,
"grad_norm": 0.12252637821649354,
"learning_rate": 1.5352275445390752e-06,
"loss": 0.7265,
"step": 879
},
{
"epoch": 4.589308996088657,
"grad_norm": 0.10588650578235918,
"learning_rate": 1.4953445940931688e-06,
"loss": 0.7328,
"step": 880
},
{
"epoch": 4.594524119947849,
"grad_norm": 0.10242904609486453,
"learning_rate": 1.4559766661141317e-06,
"loss": 0.7305,
"step": 881
},
{
"epoch": 4.599739243807041,
"grad_norm": 0.11484087780107081,
"learning_rate": 1.4171242871711922e-06,
"loss": 0.7432,
"step": 882
},
{
"epoch": 4.604954367666232,
"grad_norm": 0.10614866491620119,
"learning_rate": 1.3787879769378277e-06,
"loss": 0.7317,
"step": 883
},
{
"epoch": 4.610169491525424,
"grad_norm": 0.11216621248162341,
"learning_rate": 1.3409682481848063e-06,
"loss": 0.7379,
"step": 884
},
{
"epoch": 4.615384615384615,
"grad_norm": 0.09993770692863328,
"learning_rate": 1.3036656067732679e-06,
"loss": 0.739,
"step": 885
},
{
"epoch": 4.620599739243807,
"grad_norm": 0.09930632699157292,
"learning_rate": 1.2668805516480577e-06,
"loss": 0.743,
"step": 886
},
{
"epoch": 4.625814863102999,
"grad_norm": 0.1008961163196706,
"learning_rate": 1.2306135748309633e-06,
"loss": 0.746,
"step": 887
},
{
"epoch": 4.63102998696219,
"grad_norm": 0.1255524490616452,
"learning_rate": 1.1948651614141783e-06,
"loss": 0.7317,
"step": 888
},
{
"epoch": 4.636245110821382,
"grad_norm": 0.10155914604030952,
"learning_rate": 1.1596357895538213e-06,
"loss": 0.7213,
"step": 889
},
{
"epoch": 4.641460234680574,
"grad_norm": 0.10501816031109035,
"learning_rate": 1.1249259304634986e-06,
"loss": 0.7398,
"step": 890
},
{
"epoch": 4.646675358539765,
"grad_norm": 0.13819236460373796,
"learning_rate": 1.0907360484080409e-06,
"loss": 0.7368,
"step": 891
},
{
"epoch": 4.651890482398957,
"grad_norm": 0.10830515113482954,
"learning_rate": 1.0570666006972875e-06,
"loss": 0.7543,
"step": 892
},
{
"epoch": 4.657105606258149,
"grad_norm": 0.10654993027327486,
"learning_rate": 1.023918037679943e-06,
"loss": 0.7243,
"step": 893
},
{
"epoch": 4.66232073011734,
"grad_norm": 0.10012139135715376,
"learning_rate": 9.912908027375745e-07,
"loss": 0.7373,
"step": 894
},
{
"epoch": 4.667535853976532,
"grad_norm": 0.0950421289561811,
"learning_rate": 9.591853322787003e-07,
"loss": 0.7412,
"step": 895
},
{
"epoch": 4.672750977835723,
"grad_norm": 0.09904664343265647,
"learning_rate": 9.276020557329101e-07,
"loss": 0.729,
"step": 896
},
{
"epoch": 4.677966101694915,
"grad_norm": 0.1051253462879674,
"learning_rate": 8.965413955451363e-07,
"loss": 0.7329,
"step": 897
},
{
"epoch": 4.683181225554107,
"grad_norm": 0.10272032718312339,
"learning_rate": 8.660037671700405e-07,
"loss": 0.7387,
"step": 898
},
{
"epoch": 4.688396349413298,
"grad_norm": 0.09813662944001211,
"learning_rate": 8.359895790663963e-07,
"loss": 0.7377,
"step": 899
},
{
"epoch": 4.69361147327249,
"grad_norm": 0.10536763446890161,
"learning_rate": 8.064992326916577e-07,
"loss": 0.7353,
"step": 900
},
{
"epoch": 4.698826597131681,
"grad_norm": 0.10236023154880665,
"learning_rate": 7.775331224966076e-07,
"loss": 0.7272,
"step": 901
},
{
"epoch": 4.704041720990873,
"grad_norm": 0.09414675460502064,
"learning_rate": 7.49091635920034e-07,
"loss": 0.7292,
"step": 902
},
{
"epoch": 4.709256844850065,
"grad_norm": 0.0943501127919075,
"learning_rate": 7.211751533835998e-07,
"loss": 0.7416,
"step": 903
},
{
"epoch": 4.7144719687092564,
"grad_norm": 0.1021023935601688,
"learning_rate": 6.937840482867142e-07,
"loss": 0.7337,
"step": 904
},
{
"epoch": 4.719687092568448,
"grad_norm": 0.09881423844437276,
"learning_rate": 6.669186870015454e-07,
"loss": 0.7324,
"step": 905
},
{
"epoch": 4.72490221642764,
"grad_norm": 0.09987028627197758,
"learning_rate": 6.405794288681577e-07,
"loss": 0.7374,
"step": 906
},
{
"epoch": 4.7301173402868315,
"grad_norm": 0.09574731953052935,
"learning_rate": 6.147666261896445e-07,
"loss": 0.7326,
"step": 907
},
{
"epoch": 4.7353324641460235,
"grad_norm": 0.09726177695264816,
"learning_rate": 5.894806242274565e-07,
"loss": 0.7375,
"step": 908
},
{
"epoch": 4.7405475880052155,
"grad_norm": 0.09663506260906218,
"learning_rate": 5.647217611967914e-07,
"loss": 0.7366,
"step": 909
},
{
"epoch": 4.745762711864407,
"grad_norm": 0.09898203274690626,
"learning_rate": 5.404903682620299e-07,
"loss": 0.7282,
"step": 910
},
{
"epoch": 4.7509778357235986,
"grad_norm": 0.09888749932985194,
"learning_rate": 5.167867695323426e-07,
"loss": 0.7234,
"step": 911
},
{
"epoch": 4.7561929595827905,
"grad_norm": 0.09677300838881348,
"learning_rate": 4.936112820573335e-07,
"loss": 0.734,
"step": 912
},
{
"epoch": 4.761408083441982,
"grad_norm": 0.09338358111347181,
"learning_rate": 4.709642158228134e-07,
"loss": 0.7319,
"step": 913
},
{
"epoch": 4.766623207301174,
"grad_norm": 0.09475677222529419,
"learning_rate": 4.4884587374663727e-07,
"loss": 0.7345,
"step": 914
},
{
"epoch": 4.771838331160365,
"grad_norm": 0.09660842094296682,
"learning_rate": 4.2725655167468626e-07,
"loss": 0.7328,
"step": 915
},
{
"epoch": 4.777053455019557,
"grad_norm": 0.09522844691697419,
"learning_rate": 4.061965383768529e-07,
"loss": 0.7395,
"step": 916
},
{
"epoch": 4.782268578878749,
"grad_norm": 0.09658310921334537,
"learning_rate": 3.856661155432484e-07,
"loss": 0.7282,
"step": 917
},
{
"epoch": 4.78748370273794,
"grad_norm": 0.09209868347782292,
"learning_rate": 3.656655577803703e-07,
"loss": 0.7321,
"step": 918
},
{
"epoch": 4.792698826597132,
"grad_norm": 0.09809162403374437,
"learning_rate": 3.4619513260746973e-07,
"loss": 0.721,
"step": 919
},
{
"epoch": 4.797913950456323,
"grad_norm": 0.09728296962426303,
"learning_rate": 3.2725510045297225e-07,
"loss": 0.7383,
"step": 920
},
{
"epoch": 4.803129074315515,
"grad_norm": 0.0982190410036559,
"learning_rate": 3.0884571465096936e-07,
"loss": 0.7356,
"step": 921
},
{
"epoch": 4.808344198174707,
"grad_norm": 0.09138351849890962,
"learning_rate": 2.909672214378434e-07,
"loss": 0.7375,
"step": 922
},
{
"epoch": 4.813559322033898,
"grad_norm": 0.08931700971308501,
"learning_rate": 2.736198599489903e-07,
"loss": 0.7399,
"step": 923
},
{
"epoch": 4.81877444589309,
"grad_norm": 0.09559406138911163,
"learning_rate": 2.568038622155955e-07,
"loss": 0.7509,
"step": 924
},
{
"epoch": 4.823989569752282,
"grad_norm": 0.09150057561208538,
"learning_rate": 2.4051945316153846e-07,
"loss": 0.7284,
"step": 925
},
{
"epoch": 4.829204693611473,
"grad_norm": 0.09349065890924081,
"learning_rate": 2.2476685060039527e-07,
"loss": 0.7299,
"step": 926
},
{
"epoch": 4.834419817470665,
"grad_norm": 0.09556536713909304,
"learning_rate": 2.0954626523251197e-07,
"loss": 0.7276,
"step": 927
},
{
"epoch": 4.839634941329857,
"grad_norm": 0.09471713249250732,
"learning_rate": 1.948579006421847e-07,
"loss": 0.7428,
"step": 928
},
{
"epoch": 4.844850065189048,
"grad_norm": 0.09109503005212934,
"learning_rate": 1.8070195329495944e-07,
"loss": 0.7401,
"step": 929
},
{
"epoch": 4.85006518904824,
"grad_norm": 0.09359860047319951,
"learning_rate": 1.6707861253497214e-07,
"loss": 0.7409,
"step": 930
},
{
"epoch": 4.855280312907432,
"grad_norm": 0.09252775154456866,
"learning_rate": 1.539880605824351e-07,
"loss": 0.7362,
"step": 931
},
{
"epoch": 4.860495436766623,
"grad_norm": 0.09910390005741515,
"learning_rate": 1.4143047253120322e-07,
"loss": 0.7332,
"step": 932
},
{
"epoch": 4.865710560625815,
"grad_norm": 0.0900655265782037,
"learning_rate": 1.2940601634642059e-07,
"loss": 0.7233,
"step": 933
},
{
"epoch": 4.870925684485006,
"grad_norm": 0.09409378931281938,
"learning_rate": 1.1791485286227311e-07,
"loss": 0.7395,
"step": 934
},
{
"epoch": 4.876140808344198,
"grad_norm": 0.10015850940263663,
"learning_rate": 1.0695713577984824e-07,
"loss": 0.7231,
"step": 935
},
{
"epoch": 4.88135593220339,
"grad_norm": 0.0903358412925811,
"learning_rate": 9.653301166507422e-08,
"loss": 0.7351,
"step": 936
},
{
"epoch": 4.886571056062581,
"grad_norm": 0.0902382557022594,
"learning_rate": 8.664261994675738e-08,
"loss": 0.7333,
"step": 937
},
{
"epoch": 4.891786179921773,
"grad_norm": 0.09428682569360029,
"learning_rate": 7.728609291471678e-08,
"loss": 0.7353,
"step": 938
},
{
"epoch": 4.897001303780964,
"grad_norm": 0.09406554806150552,
"learning_rate": 6.846355571801688e-08,
"loss": 0.7387,
"step": 939
},
{
"epoch": 4.902216427640156,
"grad_norm": 0.10257884244588195,
"learning_rate": 6.017512636329325e-08,
"loss": 0.7327,
"step": 940
},
{
"epoch": 4.907431551499348,
"grad_norm": 0.08993907897520984,
"learning_rate": 5.242091571318053e-08,
"loss": 0.7323,
"step": 941
},
{
"epoch": 4.912646675358539,
"grad_norm": 0.09752580757548222,
"learning_rate": 4.5201027484811365e-08,
"loss": 0.7369,
"step": 942
},
{
"epoch": 4.917861799217731,
"grad_norm": 0.09295306846482325,
"learning_rate": 3.851555824844866e-08,
"loss": 0.7269,
"step": 943
},
{
"epoch": 4.923076923076923,
"grad_norm": 0.09090791000944588,
"learning_rate": 3.23645974261888e-08,
"loss": 0.72,
"step": 944
},
{
"epoch": 4.9282920469361144,
"grad_norm": 0.0941176803792802,
"learning_rate": 2.674822729075377e-08,
"loss": 0.7302,
"step": 945
},
{
"epoch": 4.933507170795306,
"grad_norm": 0.09476591232745427,
"learning_rate": 2.166652296440752e-08,
"loss": 0.7389,
"step": 946
},
{
"epoch": 4.938722294654498,
"grad_norm": 0.09487039421752298,
"learning_rate": 1.7119552417943496e-08,
"loss": 0.7384,
"step": 947
},
{
"epoch": 4.9439374185136895,
"grad_norm": 0.09350946727423737,
"learning_rate": 1.3107376469769783e-08,
"loss": 0.7352,
"step": 948
},
{
"epoch": 4.9491525423728815,
"grad_norm": 0.09008974287657726,
"learning_rate": 9.630048785105318e-09,
"loss": 0.7202,
"step": 949
},
{
"epoch": 4.9543676662320735,
"grad_norm": 0.09424907250830968,
"learning_rate": 6.687615875264897e-09,
"loss": 0.7352,
"step": 950
},
{
"epoch": 4.959582790091265,
"grad_norm": 0.08996524126321254,
"learning_rate": 4.280117097015257e-09,
"loss": 0.7342,
"step": 951
},
{
"epoch": 4.9647979139504566,
"grad_norm": 0.09830845644976817,
"learning_rate": 2.407584652073247e-09,
"loss": 0.7357,
"step": 952
},
{
"epoch": 4.970013037809648,
"grad_norm": 0.0969409044778751,
"learning_rate": 1.0700435866706216e-09,
"loss": 0.7212,
"step": 953
},
{
"epoch": 4.97522816166884,
"grad_norm": 0.09285007352436725,
"learning_rate": 2.6751179119877124e-10,
"loss": 0.7353,
"step": 954
},
{
"epoch": 4.980443285528032,
"grad_norm": 0.09324700428805709,
"learning_rate": 0.0,
"loss": 0.7386,
"step": 955
},
{
"epoch": 4.980443285528032,
"step": 955,
"total_flos": 1.9663195796108476e+19,
"train_loss": 0.7990808349005215,
"train_runtime": 49935.3477,
"train_samples_per_second": 9.826,
"train_steps_per_second": 0.019
}
],
"logging_steps": 1.0,
"max_steps": 955,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.9663195796108476e+19,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}