SVRC / trainer_state.json
lol
Upload folder using huggingface_hub
56413dc verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.990202482037884,
"eval_steps": 500,
"global_step": 381,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.007838014369693011,
"grad_norm": 3.012711763381958,
"learning_rate": 1.282051282051282e-06,
"loss": 0.5662,
"step": 1
},
{
"epoch": 0.015676028739386023,
"grad_norm": 3.1244778633117676,
"learning_rate": 2.564102564102564e-06,
"loss": 0.5951,
"step": 2
},
{
"epoch": 0.023514043109079032,
"grad_norm": 3.0221784114837646,
"learning_rate": 3.846153846153847e-06,
"loss": 0.5891,
"step": 3
},
{
"epoch": 0.031352057478772045,
"grad_norm": 2.6240241527557373,
"learning_rate": 5.128205128205128e-06,
"loss": 0.5812,
"step": 4
},
{
"epoch": 0.039190071848465055,
"grad_norm": 1.920555591583252,
"learning_rate": 6.41025641025641e-06,
"loss": 0.54,
"step": 5
},
{
"epoch": 0.047028086218158065,
"grad_norm": 1.4635577201843262,
"learning_rate": 7.692307692307694e-06,
"loss": 0.5043,
"step": 6
},
{
"epoch": 0.054866100587851074,
"grad_norm": 1.418139100074768,
"learning_rate": 8.974358974358976e-06,
"loss": 0.491,
"step": 7
},
{
"epoch": 0.06270411495754409,
"grad_norm": 1.3632681369781494,
"learning_rate": 1.0256410256410256e-05,
"loss": 0.4744,
"step": 8
},
{
"epoch": 0.0705421293272371,
"grad_norm": 1.7710790634155273,
"learning_rate": 1.153846153846154e-05,
"loss": 0.4128,
"step": 9
},
{
"epoch": 0.07838014369693011,
"grad_norm": 1.3074016571044922,
"learning_rate": 1.282051282051282e-05,
"loss": 0.3723,
"step": 10
},
{
"epoch": 0.08621815806662313,
"grad_norm": 0.8031821250915527,
"learning_rate": 1.4102564102564104e-05,
"loss": 0.3585,
"step": 11
},
{
"epoch": 0.09405617243631613,
"grad_norm": 1.116591453552246,
"learning_rate": 1.5384615384615387e-05,
"loss": 0.3427,
"step": 12
},
{
"epoch": 0.10189418680600915,
"grad_norm": 0.970167875289917,
"learning_rate": 1.6666666666666667e-05,
"loss": 0.3333,
"step": 13
},
{
"epoch": 0.10973220117570215,
"grad_norm": 0.8765995502471924,
"learning_rate": 1.794871794871795e-05,
"loss": 0.3051,
"step": 14
},
{
"epoch": 0.11757021554539517,
"grad_norm": 0.7228055000305176,
"learning_rate": 1.923076923076923e-05,
"loss": 0.2992,
"step": 15
},
{
"epoch": 0.12540822991508818,
"grad_norm": 0.6161549687385559,
"learning_rate": 2.0512820512820512e-05,
"loss": 0.2884,
"step": 16
},
{
"epoch": 0.13324624428478118,
"grad_norm": 0.5537469983100891,
"learning_rate": 2.1794871794871795e-05,
"loss": 0.2829,
"step": 17
},
{
"epoch": 0.1410842586544742,
"grad_norm": 0.5031456351280212,
"learning_rate": 2.307692307692308e-05,
"loss": 0.2651,
"step": 18
},
{
"epoch": 0.14892227302416722,
"grad_norm": 0.49946239590644836,
"learning_rate": 2.435897435897436e-05,
"loss": 0.2594,
"step": 19
},
{
"epoch": 0.15676028739386022,
"grad_norm": 0.434522420167923,
"learning_rate": 2.564102564102564e-05,
"loss": 0.2491,
"step": 20
},
{
"epoch": 0.16459830176355322,
"grad_norm": 0.5493834614753723,
"learning_rate": 2.6923076923076923e-05,
"loss": 0.2474,
"step": 21
},
{
"epoch": 0.17243631613324625,
"grad_norm": 0.40903612971305847,
"learning_rate": 2.8205128205128207e-05,
"loss": 0.2534,
"step": 22
},
{
"epoch": 0.18027433050293926,
"grad_norm": 0.46355342864990234,
"learning_rate": 2.948717948717949e-05,
"loss": 0.2547,
"step": 23
},
{
"epoch": 0.18811234487263226,
"grad_norm": 0.38935673236846924,
"learning_rate": 3.0769230769230774e-05,
"loss": 0.238,
"step": 24
},
{
"epoch": 0.1959503592423253,
"grad_norm": 0.43248051404953003,
"learning_rate": 3.205128205128206e-05,
"loss": 0.243,
"step": 25
},
{
"epoch": 0.2037883736120183,
"grad_norm": 0.3940970301628113,
"learning_rate": 3.3333333333333335e-05,
"loss": 0.2365,
"step": 26
},
{
"epoch": 0.2116263879817113,
"grad_norm": 0.39611291885375977,
"learning_rate": 3.461538461538462e-05,
"loss": 0.2299,
"step": 27
},
{
"epoch": 0.2194644023514043,
"grad_norm": 0.36743637919425964,
"learning_rate": 3.58974358974359e-05,
"loss": 0.235,
"step": 28
},
{
"epoch": 0.22730241672109733,
"grad_norm": 0.35339489579200745,
"learning_rate": 3.717948717948718e-05,
"loss": 0.224,
"step": 29
},
{
"epoch": 0.23514043109079033,
"grad_norm": 0.33060234785079956,
"learning_rate": 3.846153846153846e-05,
"loss": 0.2276,
"step": 30
},
{
"epoch": 0.24297844546048333,
"grad_norm": 0.3493947386741638,
"learning_rate": 3.974358974358974e-05,
"loss": 0.228,
"step": 31
},
{
"epoch": 0.25081645983017636,
"grad_norm": 0.34183913469314575,
"learning_rate": 4.1025641025641023e-05,
"loss": 0.2245,
"step": 32
},
{
"epoch": 0.25865447419986937,
"grad_norm": 0.3201765716075897,
"learning_rate": 4.230769230769231e-05,
"loss": 0.2303,
"step": 33
},
{
"epoch": 0.26649248856956237,
"grad_norm": 0.3946562707424164,
"learning_rate": 4.358974358974359e-05,
"loss": 0.2266,
"step": 34
},
{
"epoch": 0.27433050293925537,
"grad_norm": 0.36500218510627747,
"learning_rate": 4.4871794871794874e-05,
"loss": 0.2103,
"step": 35
},
{
"epoch": 0.2821685173089484,
"grad_norm": 0.3446497321128845,
"learning_rate": 4.615384615384616e-05,
"loss": 0.2309,
"step": 36
},
{
"epoch": 0.29000653167864143,
"grad_norm": 0.3353784382343292,
"learning_rate": 4.7435897435897435e-05,
"loss": 0.2294,
"step": 37
},
{
"epoch": 0.29784454604833444,
"grad_norm": 0.3502955138683319,
"learning_rate": 4.871794871794872e-05,
"loss": 0.2279,
"step": 38
},
{
"epoch": 0.30568256041802744,
"grad_norm": 0.3561626076698303,
"learning_rate": 5e-05,
"loss": 0.2289,
"step": 39
},
{
"epoch": 0.31352057478772044,
"grad_norm": 0.3842853009700775,
"learning_rate": 4.999894523795781e-05,
"loss": 0.2257,
"step": 40
},
{
"epoch": 0.32135858915741344,
"grad_norm": 0.4691725969314575,
"learning_rate": 4.999578104083307e-05,
"loss": 0.2227,
"step": 41
},
{
"epoch": 0.32919660352710645,
"grad_norm": 0.4886467754840851,
"learning_rate": 4.999050767562379e-05,
"loss": 0.2187,
"step": 42
},
{
"epoch": 0.3370346178967995,
"grad_norm": 0.38050761818885803,
"learning_rate": 4.998312558730159e-05,
"loss": 0.2301,
"step": 43
},
{
"epoch": 0.3448726322664925,
"grad_norm": 0.33100268244743347,
"learning_rate": 4.997363539877422e-05,
"loss": 0.2289,
"step": 44
},
{
"epoch": 0.3527106466361855,
"grad_norm": 0.36914992332458496,
"learning_rate": 4.996203791083291e-05,
"loss": 0.2301,
"step": 45
},
{
"epoch": 0.3605486610058785,
"grad_norm": 0.32373785972595215,
"learning_rate": 4.994833410208487e-05,
"loss": 0.2212,
"step": 46
},
{
"epoch": 0.3683866753755715,
"grad_norm": 0.3918166160583496,
"learning_rate": 4.993252512887069e-05,
"loss": 0.229,
"step": 47
},
{
"epoch": 0.3762246897452645,
"grad_norm": 0.3503532111644745,
"learning_rate": 4.991461232516675e-05,
"loss": 0.2323,
"step": 48
},
{
"epoch": 0.3840627041149575,
"grad_norm": 0.3073519766330719,
"learning_rate": 4.9894597202472696e-05,
"loss": 0.2248,
"step": 49
},
{
"epoch": 0.3919007184846506,
"grad_norm": 0.30389830470085144,
"learning_rate": 4.9872481449683844e-05,
"loss": 0.234,
"step": 50
},
{
"epoch": 0.3997387328543436,
"grad_norm": 0.3088521957397461,
"learning_rate": 4.984826693294874e-05,
"loss": 0.232,
"step": 51
},
{
"epoch": 0.4075767472240366,
"grad_norm": 0.2747111916542053,
"learning_rate": 4.982195569551162e-05,
"loss": 0.238,
"step": 52
},
{
"epoch": 0.4154147615937296,
"grad_norm": 0.33293288946151733,
"learning_rate": 4.979354995754006e-05,
"loss": 0.2407,
"step": 53
},
{
"epoch": 0.4232527759634226,
"grad_norm": 0.3552456796169281,
"learning_rate": 4.976305211593758e-05,
"loss": 0.2422,
"step": 54
},
{
"epoch": 0.4310907903331156,
"grad_norm": 0.29555025696754456,
"learning_rate": 4.9730464744141445e-05,
"loss": 0.2311,
"step": 55
},
{
"epoch": 0.4389288047028086,
"grad_norm": 0.314091295003891,
"learning_rate": 4.969579059190549e-05,
"loss": 0.2329,
"step": 56
},
{
"epoch": 0.44676681907250165,
"grad_norm": 0.30469128489494324,
"learning_rate": 4.965903258506806e-05,
"loss": 0.2278,
"step": 57
},
{
"epoch": 0.45460483344219466,
"grad_norm": 0.32766035199165344,
"learning_rate": 4.962019382530521e-05,
"loss": 0.2341,
"step": 58
},
{
"epoch": 0.46244284781188766,
"grad_norm": 0.29133275151252747,
"learning_rate": 4.957927758986888e-05,
"loss": 0.2274,
"step": 59
},
{
"epoch": 0.47028086218158066,
"grad_norm": 0.36449989676475525,
"learning_rate": 4.953628733131045e-05,
"loss": 0.2332,
"step": 60
},
{
"epoch": 0.47811887655127366,
"grad_norm": 0.2951260805130005,
"learning_rate": 4.949122667718935e-05,
"loss": 0.2408,
"step": 61
},
{
"epoch": 0.48595689092096667,
"grad_norm": 0.35234174132347107,
"learning_rate": 4.944409942976699e-05,
"loss": 0.2372,
"step": 62
},
{
"epoch": 0.4937949052906597,
"grad_norm": 0.33611589670181274,
"learning_rate": 4.9394909565685894e-05,
"loss": 0.2377,
"step": 63
},
{
"epoch": 0.5016329196603527,
"grad_norm": 0.32448795437812805,
"learning_rate": 4.93436612356342e-05,
"loss": 0.2374,
"step": 64
},
{
"epoch": 0.5094709340300457,
"grad_norm": 0.3616539537906647,
"learning_rate": 4.929035876399535e-05,
"loss": 0.2347,
"step": 65
},
{
"epoch": 0.5173089483997387,
"grad_norm": 0.3518874943256378,
"learning_rate": 4.923500664848326e-05,
"loss": 0.2368,
"step": 66
},
{
"epoch": 0.5251469627694317,
"grad_norm": 0.28374969959259033,
"learning_rate": 4.917760955976277e-05,
"loss": 0.2278,
"step": 67
},
{
"epoch": 0.5329849771391247,
"grad_norm": 0.3445357382297516,
"learning_rate": 4.9118172341055516e-05,
"loss": 0.2251,
"step": 68
},
{
"epoch": 0.5408229915088177,
"grad_norm": 0.31107786297798157,
"learning_rate": 4.905670000773126e-05,
"loss": 0.2288,
"step": 69
},
{
"epoch": 0.5486610058785107,
"grad_norm": 0.29442402720451355,
"learning_rate": 4.899319774688473e-05,
"loss": 0.2373,
"step": 70
},
{
"epoch": 0.5564990202482037,
"grad_norm": 0.33253952860832214,
"learning_rate": 4.892767091689786e-05,
"loss": 0.2363,
"step": 71
},
{
"epoch": 0.5643370346178967,
"grad_norm": 0.28732171654701233,
"learning_rate": 4.886012504698769e-05,
"loss": 0.2336,
"step": 72
},
{
"epoch": 0.5721750489875899,
"grad_norm": 0.3209660053253174,
"learning_rate": 4.87905658367398e-05,
"loss": 0.2444,
"step": 73
},
{
"epoch": 0.5800130633572829,
"grad_norm": 0.3354484438896179,
"learning_rate": 4.871899915562736e-05,
"loss": 0.2286,
"step": 74
},
{
"epoch": 0.5878510777269759,
"grad_norm": 0.34029415249824524,
"learning_rate": 4.864543104251587e-05,
"loss": 0.2375,
"step": 75
},
{
"epoch": 0.5956890920966689,
"grad_norm": 0.305865615606308,
"learning_rate": 4.856986770515358e-05,
"loss": 0.2397,
"step": 76
},
{
"epoch": 0.6035271064663619,
"grad_norm": 0.31168901920318604,
"learning_rate": 4.849231551964771e-05,
"loss": 0.2407,
"step": 77
},
{
"epoch": 0.6113651208360549,
"grad_norm": 0.3203716576099396,
"learning_rate": 4.841278102992637e-05,
"loss": 0.2305,
"step": 78
},
{
"epoch": 0.6192031352057479,
"grad_norm": 0.301705539226532,
"learning_rate": 4.833127094718643e-05,
"loss": 0.2228,
"step": 79
},
{
"epoch": 0.6270411495754409,
"grad_norm": 0.31093651056289673,
"learning_rate": 4.82477921493272e-05,
"loss": 0.2282,
"step": 80
},
{
"epoch": 0.6348791639451339,
"grad_norm": 0.30084165930747986,
"learning_rate": 4.8162351680370044e-05,
"loss": 0.2252,
"step": 81
},
{
"epoch": 0.6427171783148269,
"grad_norm": 0.2620241641998291,
"learning_rate": 4.8074956749864075e-05,
"loss": 0.2401,
"step": 82
},
{
"epoch": 0.6505551926845199,
"grad_norm": 0.2853231728076935,
"learning_rate": 4.79856147322777e-05,
"loss": 0.2346,
"step": 83
},
{
"epoch": 0.6583932070542129,
"grad_norm": 0.28441861271858215,
"learning_rate": 4.789433316637644e-05,
"loss": 0.2393,
"step": 84
},
{
"epoch": 0.6662312214239059,
"grad_norm": 0.2583891749382019,
"learning_rate": 4.7801119754586766e-05,
"loss": 0.2328,
"step": 85
},
{
"epoch": 0.674069235793599,
"grad_norm": 0.3212043344974518,
"learning_rate": 4.7705982362346164e-05,
"loss": 0.2336,
"step": 86
},
{
"epoch": 0.681907250163292,
"grad_norm": 0.29511553049087524,
"learning_rate": 4.760892901743944e-05,
"loss": 0.223,
"step": 87
},
{
"epoch": 0.689745264532985,
"grad_norm": 0.2569979429244995,
"learning_rate": 4.750996790932134e-05,
"loss": 0.227,
"step": 88
},
{
"epoch": 0.697583278902678,
"grad_norm": 0.3018832504749298,
"learning_rate": 4.7409107388425504e-05,
"loss": 0.2276,
"step": 89
},
{
"epoch": 0.705421293272371,
"grad_norm": 0.3047080636024475,
"learning_rate": 4.730635596545985e-05,
"loss": 0.2246,
"step": 90
},
{
"epoch": 0.713259307642064,
"grad_norm": 0.2749479413032532,
"learning_rate": 4.7201722310688445e-05,
"loss": 0.2259,
"step": 91
},
{
"epoch": 0.721097322011757,
"grad_norm": 0.27786287665367126,
"learning_rate": 4.709521525319986e-05,
"loss": 0.2294,
"step": 92
},
{
"epoch": 0.72893533638145,
"grad_norm": 0.26435744762420654,
"learning_rate": 4.698684378016222e-05,
"loss": 0.2241,
"step": 93
},
{
"epoch": 0.736773350751143,
"grad_norm": 0.26848137378692627,
"learning_rate": 4.6876617036064844e-05,
"loss": 0.2278,
"step": 94
},
{
"epoch": 0.744611365120836,
"grad_norm": 0.2701055407524109,
"learning_rate": 4.676454432194656e-05,
"loss": 0.2369,
"step": 95
},
{
"epoch": 0.752449379490529,
"grad_norm": 0.2705919146537781,
"learning_rate": 4.665063509461097e-05,
"loss": 0.2262,
"step": 96
},
{
"epoch": 0.760287393860222,
"grad_norm": 0.2773014307022095,
"learning_rate": 4.6534898965828405e-05,
"loss": 0.2355,
"step": 97
},
{
"epoch": 0.768125408229915,
"grad_norm": 0.23928780853748322,
"learning_rate": 4.6417345701524915e-05,
"loss": 0.2273,
"step": 98
},
{
"epoch": 0.7759634225996082,
"grad_norm": 0.26342830061912537,
"learning_rate": 4.629798522095818e-05,
"loss": 0.2293,
"step": 99
},
{
"epoch": 0.7838014369693012,
"grad_norm": 0.2792154550552368,
"learning_rate": 4.617682759588055e-05,
"loss": 0.2331,
"step": 100
},
{
"epoch": 0.7916394513389942,
"grad_norm": 0.2701910436153412,
"learning_rate": 4.6053883049689145e-05,
"loss": 0.2278,
"step": 101
},
{
"epoch": 0.7994774657086872,
"grad_norm": 0.2860325276851654,
"learning_rate": 4.592916195656322e-05,
"loss": 0.2342,
"step": 102
},
{
"epoch": 0.8073154800783802,
"grad_norm": 0.3034144639968872,
"learning_rate": 4.580267484058876e-05,
"loss": 0.2351,
"step": 103
},
{
"epoch": 0.8151534944480732,
"grad_norm": 0.28156647086143494,
"learning_rate": 4.5674432374870455e-05,
"loss": 0.2322,
"step": 104
},
{
"epoch": 0.8229915088177662,
"grad_norm": 0.2784689962863922,
"learning_rate": 4.554444538063113e-05,
"loss": 0.2361,
"step": 105
},
{
"epoch": 0.8308295231874592,
"grad_norm": 0.2945846617221832,
"learning_rate": 4.5412724826298576e-05,
"loss": 0.2302,
"step": 106
},
{
"epoch": 0.8386675375571522,
"grad_norm": 0.25669312477111816,
"learning_rate": 4.5279281826580056e-05,
"loss": 0.2327,
"step": 107
},
{
"epoch": 0.8465055519268452,
"grad_norm": 0.2950827181339264,
"learning_rate": 4.514412764152446e-05,
"loss": 0.2332,
"step": 108
},
{
"epoch": 0.8543435662965382,
"grad_norm": 0.2892044186592102,
"learning_rate": 4.5007273675572104e-05,
"loss": 0.2286,
"step": 109
},
{
"epoch": 0.8621815806662312,
"grad_norm": 0.27470967173576355,
"learning_rate": 4.48687314765925e-05,
"loss": 0.2364,
"step": 110
},
{
"epoch": 0.8700195950359242,
"grad_norm": 0.27829453349113464,
"learning_rate": 4.4728512734909844e-05,
"loss": 0.2304,
"step": 111
},
{
"epoch": 0.8778576094056172,
"grad_norm": 0.29203999042510986,
"learning_rate": 4.4586629282316654e-05,
"loss": 0.2363,
"step": 112
},
{
"epoch": 0.8856956237753103,
"grad_norm": 0.24676302075386047,
"learning_rate": 4.444309309107535e-05,
"loss": 0.2215,
"step": 113
},
{
"epoch": 0.8935336381450033,
"grad_norm": 0.27994513511657715,
"learning_rate": 4.4297916272908024e-05,
"loss": 0.232,
"step": 114
},
{
"epoch": 0.9013716525146963,
"grad_norm": 0.2659918963909149,
"learning_rate": 4.415111107797445e-05,
"loss": 0.2323,
"step": 115
},
{
"epoch": 0.9092096668843893,
"grad_norm": 0.2813761234283447,
"learning_rate": 4.400268989383841e-05,
"loss": 0.2294,
"step": 116
},
{
"epoch": 0.9170476812540823,
"grad_norm": 0.27070480585098267,
"learning_rate": 4.385266524442241e-05,
"loss": 0.2254,
"step": 117
},
{
"epoch": 0.9248856956237753,
"grad_norm": 0.2846162021160126,
"learning_rate": 4.370104978895089e-05,
"loss": 0.2324,
"step": 118
},
{
"epoch": 0.9327237099934683,
"grad_norm": 0.257036954164505,
"learning_rate": 4.3547856320882044e-05,
"loss": 0.23,
"step": 119
},
{
"epoch": 0.9405617243631613,
"grad_norm": 0.28918588161468506,
"learning_rate": 4.3393097766828293e-05,
"loss": 0.2234,
"step": 120
},
{
"epoch": 0.9483997387328543,
"grad_norm": 0.23960325121879578,
"learning_rate": 4.3236787185465525e-05,
"loss": 0.2222,
"step": 121
},
{
"epoch": 0.9562377531025473,
"grad_norm": 0.26537278294563293,
"learning_rate": 4.307893776643117e-05,
"loss": 0.2285,
"step": 122
},
{
"epoch": 0.9640757674722403,
"grad_norm": 0.23799960315227509,
"learning_rate": 4.2919562829211283e-05,
"loss": 0.2214,
"step": 123
},
{
"epoch": 0.9719137818419333,
"grad_norm": 0.2599596083164215,
"learning_rate": 4.27586758220166e-05,
"loss": 0.2246,
"step": 124
},
{
"epoch": 0.9797517962116263,
"grad_norm": 0.25664016604423523,
"learning_rate": 4.259629032064779e-05,
"loss": 0.2223,
"step": 125
},
{
"epoch": 0.9875898105813194,
"grad_norm": 0.24727745354175568,
"learning_rate": 4.2432420027349886e-05,
"loss": 0.2225,
"step": 126
},
{
"epoch": 0.9954278249510125,
"grad_norm": 0.2626395523548126,
"learning_rate": 4.226707876965611e-05,
"loss": 0.2333,
"step": 127
},
{
"epoch": 1.0052253429131286,
"grad_norm": 0.4331790506839752,
"learning_rate": 4.210028049922108e-05,
"loss": 0.3128,
"step": 128
},
{
"epoch": 1.0130633572828216,
"grad_norm": 0.31224802136421204,
"learning_rate": 4.193203929064353e-05,
"loss": 0.1265,
"step": 129
},
{
"epoch": 1.0209013716525146,
"grad_norm": 0.34126603603363037,
"learning_rate": 4.176236934027873e-05,
"loss": 0.1188,
"step": 130
},
{
"epoch": 1.0287393860222076,
"grad_norm": 0.3540596067905426,
"learning_rate": 4.159128496504053e-05,
"loss": 0.1149,
"step": 131
},
{
"epoch": 1.0365774003919008,
"grad_norm": 0.2591104209423065,
"learning_rate": 4.141880060119336e-05,
"loss": 0.1109,
"step": 132
},
{
"epoch": 1.0444154147615938,
"grad_norm": 0.33954110741615295,
"learning_rate": 4.1244930803134e-05,
"loss": 0.1307,
"step": 133
},
{
"epoch": 1.0522534291312868,
"grad_norm": 0.2748555839061737,
"learning_rate": 4.1069690242163484e-05,
"loss": 0.1246,
"step": 134
},
{
"epoch": 1.0600914435009798,
"grad_norm": 0.29179254174232483,
"learning_rate": 4.089309370524921e-05,
"loss": 0.1144,
"step": 135
},
{
"epoch": 1.0679294578706728,
"grad_norm": 0.3250279724597931,
"learning_rate": 4.071515609377705e-05,
"loss": 0.1363,
"step": 136
},
{
"epoch": 1.0757674722403658,
"grad_norm": 0.24014687538146973,
"learning_rate": 4.053589242229412e-05,
"loss": 0.1044,
"step": 137
},
{
"epoch": 1.0836054866100588,
"grad_norm": 0.2841418981552124,
"learning_rate": 4.03553178172417e-05,
"loss": 0.1151,
"step": 138
},
{
"epoch": 1.0914435009797518,
"grad_norm": 0.25916630029678345,
"learning_rate": 4.0173447515678916e-05,
"loss": 0.1178,
"step": 139
},
{
"epoch": 1.0992815153494448,
"grad_norm": 0.24618621170520782,
"learning_rate": 3.999029686399704e-05,
"loss": 0.1078,
"step": 140
},
{
"epoch": 1.1071195297191379,
"grad_norm": 0.3014475703239441,
"learning_rate": 3.9805881316624506e-05,
"loss": 0.1277,
"step": 141
},
{
"epoch": 1.1149575440888309,
"grad_norm": 0.2629724442958832,
"learning_rate": 3.962021643472284e-05,
"loss": 0.1168,
"step": 142
},
{
"epoch": 1.1227955584585239,
"grad_norm": 0.26206281781196594,
"learning_rate": 3.9433317884873664e-05,
"loss": 0.117,
"step": 143
},
{
"epoch": 1.1306335728282169,
"grad_norm": 0.2575000524520874,
"learning_rate": 3.9245201437756654e-05,
"loss": 0.1107,
"step": 144
},
{
"epoch": 1.1384715871979099,
"grad_norm": 0.2927393913269043,
"learning_rate": 3.905588296681886e-05,
"loss": 0.1216,
"step": 145
},
{
"epoch": 1.1463096015676029,
"grad_norm": 0.23872609436511993,
"learning_rate": 3.8865378446935216e-05,
"loss": 0.1067,
"step": 146
},
{
"epoch": 1.1541476159372959,
"grad_norm": 0.2771260738372803,
"learning_rate": 3.867370395306068e-05,
"loss": 0.1118,
"step": 147
},
{
"epoch": 1.1619856303069889,
"grad_norm": 0.2423640787601471,
"learning_rate": 3.848087565887365e-05,
"loss": 0.1179,
"step": 148
},
{
"epoch": 1.1698236446766819,
"grad_norm": 0.2449909895658493,
"learning_rate": 3.82869098354114e-05,
"loss": 0.1247,
"step": 149
},
{
"epoch": 1.1776616590463749,
"grad_norm": 0.2508186101913452,
"learning_rate": 3.8091822849696954e-05,
"loss": 0.1186,
"step": 150
},
{
"epoch": 1.1854996734160679,
"grad_norm": 0.22804874181747437,
"learning_rate": 3.7895631163358105e-05,
"loss": 0.1214,
"step": 151
},
{
"epoch": 1.1933376877857609,
"grad_norm": 0.23902134597301483,
"learning_rate": 3.769835133123836e-05,
"loss": 0.1118,
"step": 152
},
{
"epoch": 1.2011757021554539,
"grad_norm": 0.22123728692531586,
"learning_rate": 3.7500000000000003e-05,
"loss": 0.1081,
"step": 153
},
{
"epoch": 1.2090137165251469,
"grad_norm": 0.22215275466442108,
"learning_rate": 3.7300593906719464e-05,
"loss": 0.1115,
"step": 154
},
{
"epoch": 1.21685173089484,
"grad_norm": 0.21011732518672943,
"learning_rate": 3.7100149877474974e-05,
"loss": 0.1105,
"step": 155
},
{
"epoch": 1.224689745264533,
"grad_norm": 0.25005996227264404,
"learning_rate": 3.689868482592684e-05,
"loss": 0.1143,
"step": 156
},
{
"epoch": 1.232527759634226,
"grad_norm": 0.22056043148040771,
"learning_rate": 3.66962157518902e-05,
"loss": 0.1113,
"step": 157
},
{
"epoch": 1.240365774003919,
"grad_norm": 0.226328045129776,
"learning_rate": 3.6492759739900564e-05,
"loss": 0.1153,
"step": 158
},
{
"epoch": 1.248203788373612,
"grad_norm": 0.21589796245098114,
"learning_rate": 3.628833395777224e-05,
"loss": 0.1034,
"step": 159
},
{
"epoch": 1.256041802743305,
"grad_norm": 0.2250795215368271,
"learning_rate": 3.608295565514965e-05,
"loss": 0.116,
"step": 160
},
{
"epoch": 1.263879817112998,
"grad_norm": 0.21993276476860046,
"learning_rate": 3.587664216205183e-05,
"loss": 0.1217,
"step": 161
},
{
"epoch": 1.2717178314826911,
"grad_norm": 0.21394269168376923,
"learning_rate": 3.5669410887410095e-05,
"loss": 0.1119,
"step": 162
},
{
"epoch": 1.2795558458523841,
"grad_norm": 0.20264342427253723,
"learning_rate": 3.546127931759903e-05,
"loss": 0.1076,
"step": 163
},
{
"epoch": 1.2873938602220771,
"grad_norm": 0.23118674755096436,
"learning_rate": 3.5252265014961006e-05,
"loss": 0.1189,
"step": 164
},
{
"epoch": 1.2952318745917701,
"grad_norm": 0.215042844414711,
"learning_rate": 3.504238561632424e-05,
"loss": 0.1096,
"step": 165
},
{
"epoch": 1.3030698889614631,
"grad_norm": 0.21117202937602997,
"learning_rate": 3.483165883151458e-05,
"loss": 0.1201,
"step": 166
},
{
"epoch": 1.3109079033311561,
"grad_norm": 0.21316753327846527,
"learning_rate": 3.4620102441861143e-05,
"loss": 0.1072,
"step": 167
},
{
"epoch": 1.3187459177008491,
"grad_norm": 0.21038144826889038,
"learning_rate": 3.4407734298695895e-05,
"loss": 0.1149,
"step": 168
},
{
"epoch": 1.3265839320705421,
"grad_norm": 0.2167028933763504,
"learning_rate": 3.4194572321847336e-05,
"loss": 0.1075,
"step": 169
},
{
"epoch": 1.3344219464402352,
"grad_norm": 0.21729543805122375,
"learning_rate": 3.398063449812844e-05,
"loss": 0.1056,
"step": 170
},
{
"epoch": 1.3422599608099282,
"grad_norm": 0.20175161957740784,
"learning_rate": 3.376593887981887e-05,
"loss": 0.1053,
"step": 171
},
{
"epoch": 1.3500979751796212,
"grad_norm": 0.21248085796833038,
"learning_rate": 3.355050358314172e-05,
"loss": 0.1059,
"step": 172
},
{
"epoch": 1.3579359895493142,
"grad_norm": 0.21043576300144196,
"learning_rate": 3.333434678673489e-05,
"loss": 0.1051,
"step": 173
},
{
"epoch": 1.3657740039190072,
"grad_norm": 0.21573218703269958,
"learning_rate": 3.311748673011709e-05,
"loss": 0.1168,
"step": 174
},
{
"epoch": 1.3736120182887002,
"grad_norm": 0.20607827603816986,
"learning_rate": 3.289994171214882e-05,
"loss": 0.1071,
"step": 175
},
{
"epoch": 1.3814500326583932,
"grad_norm": 0.20888157188892365,
"learning_rate": 3.268173008948826e-05,
"loss": 0.1066,
"step": 176
},
{
"epoch": 1.3892880470280862,
"grad_norm": 0.20718783140182495,
"learning_rate": 3.246287027504237e-05,
"loss": 0.1034,
"step": 177
},
{
"epoch": 1.3971260613977792,
"grad_norm": 0.21539223194122314,
"learning_rate": 3.224338073641312e-05,
"loss": 0.1284,
"step": 178
},
{
"epoch": 1.4049640757674722,
"grad_norm": 0.2096419632434845,
"learning_rate": 3.202327999433924e-05,
"loss": 0.108,
"step": 179
},
{
"epoch": 1.4128020901371652,
"grad_norm": 0.19958484172821045,
"learning_rate": 3.180258662113338e-05,
"loss": 0.1027,
"step": 180
},
{
"epoch": 1.4206401045068582,
"grad_norm": 0.20901361107826233,
"learning_rate": 3.158131923911498e-05,
"loss": 0.1025,
"step": 181
},
{
"epoch": 1.4284781188765514,
"grad_norm": 0.22095659375190735,
"learning_rate": 3.135949651903891e-05,
"loss": 0.1082,
"step": 182
},
{
"epoch": 1.4363161332462444,
"grad_norm": 0.20527766644954681,
"learning_rate": 3.1137137178519985e-05,
"loss": 0.1114,
"step": 183
},
{
"epoch": 1.4441541476159374,
"grad_norm": 0.21945808827877045,
"learning_rate": 3.091425998045356e-05,
"loss": 0.1178,
"step": 184
},
{
"epoch": 1.4519921619856304,
"grad_norm": 0.20016823709011078,
"learning_rate": 3.069088373143234e-05,
"loss": 0.1078,
"step": 185
},
{
"epoch": 1.4598301763553234,
"grad_norm": 0.20746341347694397,
"learning_rate": 3.04670272801594e-05,
"loss": 0.1134,
"step": 186
},
{
"epoch": 1.4676681907250164,
"grad_norm": 0.23339882493019104,
"learning_rate": 3.0242709515857758e-05,
"loss": 0.1176,
"step": 187
},
{
"epoch": 1.4755062050947094,
"grad_norm": 0.21998368203639984,
"learning_rate": 3.001794936667648e-05,
"loss": 0.118,
"step": 188
},
{
"epoch": 1.4833442194644024,
"grad_norm": 0.2124100923538208,
"learning_rate": 2.9792765798093465e-05,
"loss": 0.1028,
"step": 189
},
{
"epoch": 1.4911822338340954,
"grad_norm": 0.22203968465328217,
"learning_rate": 2.9567177811315178e-05,
"loss": 0.1127,
"step": 190
},
{
"epoch": 1.4990202482037884,
"grad_norm": 0.2147323042154312,
"learning_rate": 2.9341204441673266e-05,
"loss": 0.108,
"step": 191
},
{
"epoch": 1.5068582625734814,
"grad_norm": 0.22974567115306854,
"learning_rate": 2.9114864757018352e-05,
"loss": 0.1063,
"step": 192
},
{
"epoch": 1.5146962769431744,
"grad_norm": 0.21040025353431702,
"learning_rate": 2.8888177856111083e-05,
"loss": 0.1073,
"step": 193
},
{
"epoch": 1.5225342913128674,
"grad_norm": 0.20959196984767914,
"learning_rate": 2.8661162867010543e-05,
"loss": 0.0989,
"step": 194
},
{
"epoch": 1.5303723056825604,
"grad_norm": 0.2233189344406128,
"learning_rate": 2.8433838945460205e-05,
"loss": 0.1153,
"step": 195
},
{
"epoch": 1.5382103200522534,
"grad_norm": 0.21911266446113586,
"learning_rate": 2.820622527327158e-05,
"loss": 0.1052,
"step": 196
},
{
"epoch": 1.5460483344219464,
"grad_norm": 0.22261589765548706,
"learning_rate": 2.797834105670559e-05,
"loss": 0.1109,
"step": 197
},
{
"epoch": 1.5538863487916394,
"grad_norm": 0.19887444376945496,
"learning_rate": 2.7750205524851986e-05,
"loss": 0.1092,
"step": 198
},
{
"epoch": 1.5617243631613325,
"grad_norm": 0.20779380202293396,
"learning_rate": 2.752183792800671e-05,
"loss": 0.1076,
"step": 199
},
{
"epoch": 1.5695623775310255,
"grad_norm": 0.2083713710308075,
"learning_rate": 2.729325753604759e-05,
"loss": 0.1135,
"step": 200
},
{
"epoch": 1.5774003919007185,
"grad_norm": 0.2003626525402069,
"learning_rate": 2.7064483636808313e-05,
"loss": 0.0968,
"step": 201
},
{
"epoch": 1.5852384062704115,
"grad_norm": 0.20525719225406647,
"learning_rate": 2.6835535534450884e-05,
"loss": 0.1042,
"step": 202
},
{
"epoch": 1.5930764206401045,
"grad_norm": 0.21359968185424805,
"learning_rate": 2.6606432547836757e-05,
"loss": 0.1126,
"step": 203
},
{
"epoch": 1.6009144350097975,
"grad_norm": 0.2056104987859726,
"learning_rate": 2.6377194008896637e-05,
"loss": 0.1073,
"step": 204
},
{
"epoch": 1.6087524493794905,
"grad_norm": 0.19534572958946228,
"learning_rate": 2.6147839260999295e-05,
"loss": 0.113,
"step": 205
},
{
"epoch": 1.6165904637491835,
"grad_norm": 0.21906475722789764,
"learning_rate": 2.5918387657319316e-05,
"loss": 0.1098,
"step": 206
},
{
"epoch": 1.6244284781188765,
"grad_norm": 0.19928069412708282,
"learning_rate": 2.5688858559204053e-05,
"loss": 0.1059,
"step": 207
},
{
"epoch": 1.6322664924885695,
"grad_norm": 0.19612881541252136,
"learning_rate": 2.5459271334539934e-05,
"loss": 0.0951,
"step": 208
},
{
"epoch": 1.6401045068582625,
"grad_norm": 0.2001960426568985,
"learning_rate": 2.5229645356118163e-05,
"loss": 0.1039,
"step": 209
},
{
"epoch": 1.6479425212279555,
"grad_norm": 0.20145660638809204,
"learning_rate": 2.5e-05,
"loss": 0.1049,
"step": 210
},
{
"epoch": 1.6557805355976485,
"grad_norm": 0.18295256793498993,
"learning_rate": 2.4770354643881843e-05,
"loss": 0.0985,
"step": 211
},
{
"epoch": 1.6636185499673415,
"grad_norm": 0.2013918161392212,
"learning_rate": 2.4540728665460065e-05,
"loss": 0.1065,
"step": 212
},
{
"epoch": 1.6714565643370345,
"grad_norm": 0.20308274030685425,
"learning_rate": 2.4311141440795953e-05,
"loss": 0.1205,
"step": 213
},
{
"epoch": 1.6792945787067275,
"grad_norm": 0.19719456136226654,
"learning_rate": 2.4081612342680694e-05,
"loss": 0.1027,
"step": 214
},
{
"epoch": 1.6871325930764205,
"grad_norm": 0.19175922870635986,
"learning_rate": 2.3852160739000707e-05,
"loss": 0.1,
"step": 215
},
{
"epoch": 1.6949706074461135,
"grad_norm": 0.20407816767692566,
"learning_rate": 2.3622805991103362e-05,
"loss": 0.102,
"step": 216
},
{
"epoch": 1.7028086218158065,
"grad_norm": 0.20707584917545319,
"learning_rate": 2.3393567452163252e-05,
"loss": 0.1139,
"step": 217
},
{
"epoch": 1.7106466361854997,
"grad_norm": 0.18942616879940033,
"learning_rate": 2.3164464465549118e-05,
"loss": 0.091,
"step": 218
},
{
"epoch": 1.7184846505551927,
"grad_norm": 0.2033768594264984,
"learning_rate": 2.2935516363191693e-05,
"loss": 0.1093,
"step": 219
},
{
"epoch": 1.7263226649248857,
"grad_norm": 0.2145160436630249,
"learning_rate": 2.270674246395241e-05,
"loss": 0.1083,
"step": 220
},
{
"epoch": 1.7341606792945787,
"grad_norm": 0.19836893677711487,
"learning_rate": 2.2478162071993298e-05,
"loss": 0.1011,
"step": 221
},
{
"epoch": 1.7419986936642717,
"grad_norm": 0.20721815526485443,
"learning_rate": 2.224979447514802e-05,
"loss": 0.0975,
"step": 222
},
{
"epoch": 1.7498367080339647,
"grad_norm": 0.20113082230091095,
"learning_rate": 2.202165894329441e-05,
"loss": 0.1086,
"step": 223
},
{
"epoch": 1.7576747224036577,
"grad_norm": 0.19542041420936584,
"learning_rate": 2.179377472672842e-05,
"loss": 0.1029,
"step": 224
},
{
"epoch": 1.7655127367733507,
"grad_norm": 0.19994117319583893,
"learning_rate": 2.1566161054539798e-05,
"loss": 0.1071,
"step": 225
},
{
"epoch": 1.7733507511430437,
"grad_norm": 0.20020049810409546,
"learning_rate": 2.1338837132989466e-05,
"loss": 0.0995,
"step": 226
},
{
"epoch": 1.7811887655127367,
"grad_norm": 0.2069658637046814,
"learning_rate": 2.111182214388893e-05,
"loss": 0.1077,
"step": 227
},
{
"epoch": 1.7890267798824298,
"grad_norm": 0.21685247123241425,
"learning_rate": 2.088513524298165e-05,
"loss": 0.1084,
"step": 228
},
{
"epoch": 1.7968647942521228,
"grad_norm": 0.19922901690006256,
"learning_rate": 2.0658795558326743e-05,
"loss": 0.0999,
"step": 229
},
{
"epoch": 1.8047028086218158,
"grad_norm": 0.20437341928482056,
"learning_rate": 2.043282218868483e-05,
"loss": 0.1076,
"step": 230
},
{
"epoch": 1.812540822991509,
"grad_norm": 0.19708654284477234,
"learning_rate": 2.0207234201906547e-05,
"loss": 0.0971,
"step": 231
},
{
"epoch": 1.820378837361202,
"grad_norm": 0.21140620112419128,
"learning_rate": 1.9982050633323522e-05,
"loss": 0.1046,
"step": 232
},
{
"epoch": 1.828216851730895,
"grad_norm": 0.19234000146389008,
"learning_rate": 1.9757290484142244e-05,
"loss": 0.0957,
"step": 233
},
{
"epoch": 1.836054866100588,
"grad_norm": 0.2157324254512787,
"learning_rate": 1.9532972719840607e-05,
"loss": 0.1082,
"step": 234
},
{
"epoch": 1.843892880470281,
"grad_norm": 0.19539855420589447,
"learning_rate": 1.9309116268567674e-05,
"loss": 0.1014,
"step": 235
},
{
"epoch": 1.851730894839974,
"grad_norm": 0.1897420883178711,
"learning_rate": 1.908574001954644e-05,
"loss": 0.094,
"step": 236
},
{
"epoch": 1.859568909209667,
"grad_norm": 0.20616742968559265,
"learning_rate": 1.8862862821480025e-05,
"loss": 0.1118,
"step": 237
},
{
"epoch": 1.86740692357936,
"grad_norm": 0.1957971155643463,
"learning_rate": 1.864050348096109e-05,
"loss": 0.0948,
"step": 238
},
{
"epoch": 1.875244937949053,
"grad_norm": 0.18928657472133636,
"learning_rate": 1.8418680760885027e-05,
"loss": 0.0963,
"step": 239
},
{
"epoch": 1.883082952318746,
"grad_norm": 0.20447275042533875,
"learning_rate": 1.819741337886662e-05,
"loss": 0.096,
"step": 240
},
{
"epoch": 1.890920966688439,
"grad_norm": 0.19518423080444336,
"learning_rate": 1.797672000566077e-05,
"loss": 0.1042,
"step": 241
},
{
"epoch": 1.898758981058132,
"grad_norm": 0.18361036479473114,
"learning_rate": 1.775661926358689e-05,
"loss": 0.0975,
"step": 242
},
{
"epoch": 1.906596995427825,
"grad_norm": 0.19658760726451874,
"learning_rate": 1.7537129724957642e-05,
"loss": 0.1063,
"step": 243
},
{
"epoch": 1.914435009797518,
"grad_norm": 0.194893017411232,
"learning_rate": 1.7318269910511736e-05,
"loss": 0.1007,
"step": 244
},
{
"epoch": 1.922273024167211,
"grad_norm": 0.19551898539066315,
"learning_rate": 1.710005828785119e-05,
"loss": 0.0989,
"step": 245
},
{
"epoch": 1.930111038536904,
"grad_norm": 0.19419115781784058,
"learning_rate": 1.6882513269882917e-05,
"loss": 0.0894,
"step": 246
},
{
"epoch": 1.937949052906597,
"grad_norm": 0.19363057613372803,
"learning_rate": 1.666565321326512e-05,
"loss": 0.0954,
"step": 247
},
{
"epoch": 1.94578706727629,
"grad_norm": 0.18952466547489166,
"learning_rate": 1.6449496416858284e-05,
"loss": 0.0906,
"step": 248
},
{
"epoch": 1.953625081645983,
"grad_norm": 0.19003801047801971,
"learning_rate": 1.6234061120181142e-05,
"loss": 0.0982,
"step": 249
},
{
"epoch": 1.961463096015676,
"grad_norm": 0.18380698561668396,
"learning_rate": 1.601936550187157e-05,
"loss": 0.099,
"step": 250
},
{
"epoch": 1.969301110385369,
"grad_norm": 0.1997416615486145,
"learning_rate": 1.5805427678152677e-05,
"loss": 0.1027,
"step": 251
},
{
"epoch": 1.977139124755062,
"grad_norm": 0.18836817145347595,
"learning_rate": 1.5592265701304114e-05,
"loss": 0.0983,
"step": 252
},
{
"epoch": 1.984977139124755,
"grad_norm": 0.2013559639453888,
"learning_rate": 1.5379897558138862e-05,
"loss": 0.102,
"step": 253
},
{
"epoch": 1.992815153494448,
"grad_norm": 0.19445432722568512,
"learning_rate": 1.5168341168485423e-05,
"loss": 0.0953,
"step": 254
},
{
"epoch": 2.002612671456564,
"grad_norm": 0.3252091705799103,
"learning_rate": 1.495761438367577e-05,
"loss": 0.1459,
"step": 255
},
{
"epoch": 2.010450685826257,
"grad_norm": 0.20789092779159546,
"learning_rate": 1.4747734985039e-05,
"loss": 0.0344,
"step": 256
},
{
"epoch": 2.01828870019595,
"grad_norm": 0.17120763659477234,
"learning_rate": 1.4538720682400969e-05,
"loss": 0.0347,
"step": 257
},
{
"epoch": 2.026126714565643,
"grad_norm": 0.14296689629554749,
"learning_rate": 1.433058911258991e-05,
"loss": 0.0296,
"step": 258
},
{
"epoch": 2.033964728935336,
"grad_norm": 0.18135866522789001,
"learning_rate": 1.4123357837948175e-05,
"loss": 0.0332,
"step": 259
},
{
"epoch": 2.041802743305029,
"grad_norm": 0.20270270109176636,
"learning_rate": 1.3917044344850355e-05,
"loss": 0.0322,
"step": 260
},
{
"epoch": 2.049640757674722,
"grad_norm": 0.1781613826751709,
"learning_rate": 1.3711666042227772e-05,
"loss": 0.0309,
"step": 261
},
{
"epoch": 2.057478772044415,
"grad_norm": 0.17559246718883514,
"learning_rate": 1.350724026009944e-05,
"loss": 0.0333,
"step": 262
},
{
"epoch": 2.065316786414108,
"grad_norm": 0.14222192764282227,
"learning_rate": 1.330378424810981e-05,
"loss": 0.0309,
"step": 263
},
{
"epoch": 2.0731548007838017,
"grad_norm": 0.15568028390407562,
"learning_rate": 1.3101315174073162e-05,
"loss": 0.0324,
"step": 264
},
{
"epoch": 2.0809928151534947,
"grad_norm": 0.15495392680168152,
"learning_rate": 1.2899850122525037e-05,
"loss": 0.0329,
"step": 265
},
{
"epoch": 2.0888308295231877,
"grad_norm": 0.16096574068069458,
"learning_rate": 1.2699406093280547e-05,
"loss": 0.0317,
"step": 266
},
{
"epoch": 2.0966688438928807,
"grad_norm": 0.13694337010383606,
"learning_rate": 1.2500000000000006e-05,
"loss": 0.0282,
"step": 267
},
{
"epoch": 2.1045068582625737,
"grad_norm": 0.13896653056144714,
"learning_rate": 1.2301648668761647e-05,
"loss": 0.0289,
"step": 268
},
{
"epoch": 2.1123448726322667,
"grad_norm": 0.15309906005859375,
"learning_rate": 1.2104368836641908e-05,
"loss": 0.0321,
"step": 269
},
{
"epoch": 2.1201828870019597,
"grad_norm": 0.16871266067028046,
"learning_rate": 1.1908177150303055e-05,
"loss": 0.0335,
"step": 270
},
{
"epoch": 2.1280209013716527,
"grad_norm": 0.15097489953041077,
"learning_rate": 1.1713090164588607e-05,
"loss": 0.0302,
"step": 271
},
{
"epoch": 2.1358589157413457,
"grad_norm": 0.1583205759525299,
"learning_rate": 1.1519124341126346e-05,
"loss": 0.0337,
"step": 272
},
{
"epoch": 2.1436969301110387,
"grad_norm": 0.1436772346496582,
"learning_rate": 1.1326296046939333e-05,
"loss": 0.0296,
"step": 273
},
{
"epoch": 2.1515349444807317,
"grad_norm": 0.1444602757692337,
"learning_rate": 1.113462155306478e-05,
"loss": 0.0296,
"step": 274
},
{
"epoch": 2.1593729588504247,
"grad_norm": 0.13352134823799133,
"learning_rate": 1.0944117033181151e-05,
"loss": 0.0287,
"step": 275
},
{
"epoch": 2.1672109732201177,
"grad_norm": 0.1418055146932602,
"learning_rate": 1.0754798562243345e-05,
"loss": 0.0294,
"step": 276
},
{
"epoch": 2.1750489875898107,
"grad_norm": 0.12928378582000732,
"learning_rate": 1.0566682115126344e-05,
"loss": 0.0274,
"step": 277
},
{
"epoch": 2.1828870019595037,
"grad_norm": 0.13986484706401825,
"learning_rate": 1.037978356527716e-05,
"loss": 0.0328,
"step": 278
},
{
"epoch": 2.1907250163291967,
"grad_norm": 0.13403278589248657,
"learning_rate": 1.0194118683375503e-05,
"loss": 0.0296,
"step": 279
},
{
"epoch": 2.1985630306988897,
"grad_norm": 0.14188171923160553,
"learning_rate": 1.0009703136002957e-05,
"loss": 0.0299,
"step": 280
},
{
"epoch": 2.2064010450685827,
"grad_norm": 0.13912995159626007,
"learning_rate": 9.826552484321087e-06,
"loss": 0.03,
"step": 281
},
{
"epoch": 2.2142390594382757,
"grad_norm": 0.14268967509269714,
"learning_rate": 9.644682182758306e-06,
"loss": 0.0293,
"step": 282
},
{
"epoch": 2.2220770738079687,
"grad_norm": 0.14812886714935303,
"learning_rate": 9.464107577705886e-06,
"loss": 0.0269,
"step": 283
},
{
"epoch": 2.2299150881776617,
"grad_norm": 0.1425481140613556,
"learning_rate": 9.284843906222948e-06,
"loss": 0.0274,
"step": 284
},
{
"epoch": 2.2377531025473547,
"grad_norm": 0.13893434405326843,
"learning_rate": 9.106906294750805e-06,
"loss": 0.0292,
"step": 285
},
{
"epoch": 2.2455911169170477,
"grad_norm": 0.13851140439510345,
"learning_rate": 8.930309757836517e-06,
"loss": 0.0321,
"step": 286
},
{
"epoch": 2.2534291312867407,
"grad_norm": 0.13696281611919403,
"learning_rate": 8.755069196866014e-06,
"loss": 0.0286,
"step": 287
},
{
"epoch": 2.2612671456564337,
"grad_norm": 0.13719762861728668,
"learning_rate": 8.581199398806641e-06,
"loss": 0.0314,
"step": 288
},
{
"epoch": 2.2691051600261267,
"grad_norm": 0.1380627155303955,
"learning_rate": 8.40871503495947e-06,
"loss": 0.0309,
"step": 289
},
{
"epoch": 2.2769431743958197,
"grad_norm": 0.13105949759483337,
"learning_rate": 8.237630659721277e-06,
"loss": 0.0259,
"step": 290
},
{
"epoch": 2.2847811887655127,
"grad_norm": 0.1306370496749878,
"learning_rate": 8.067960709356478e-06,
"loss": 0.0264,
"step": 291
},
{
"epoch": 2.2926192031352057,
"grad_norm": 0.13892598450183868,
"learning_rate": 7.899719500778923e-06,
"loss": 0.0308,
"step": 292
},
{
"epoch": 2.3004572175048987,
"grad_norm": 0.13053514063358307,
"learning_rate": 7.732921230343892e-06,
"loss": 0.0276,
"step": 293
},
{
"epoch": 2.3082952318745917,
"grad_norm": 0.1372438371181488,
"learning_rate": 7.5675799726501155e-06,
"loss": 0.0271,
"step": 294
},
{
"epoch": 2.3161332462442847,
"grad_norm": 0.1358174979686737,
"learning_rate": 7.403709679352217e-06,
"loss": 0.027,
"step": 295
},
{
"epoch": 2.3239712606139777,
"grad_norm": 0.13224904239177704,
"learning_rate": 7.2413241779834e-06,
"loss": 0.027,
"step": 296
},
{
"epoch": 2.3318092749836707,
"grad_norm": 0.13056500256061554,
"learning_rate": 7.080437170788723e-06,
"loss": 0.0282,
"step": 297
},
{
"epoch": 2.3396472893533637,
"grad_norm": 0.13113614916801453,
"learning_rate": 6.921062233568831e-06,
"loss": 0.028,
"step": 298
},
{
"epoch": 2.3474853037230567,
"grad_norm": 0.13239531219005585,
"learning_rate": 6.763212814534484e-06,
"loss": 0.0272,
"step": 299
},
{
"epoch": 2.3553233180927498,
"grad_norm": 0.12943539023399353,
"learning_rate": 6.606902233171711e-06,
"loss": 0.0247,
"step": 300
},
{
"epoch": 2.3631613324624428,
"grad_norm": 0.12735775113105774,
"learning_rate": 6.452143679117964e-06,
"loss": 0.0263,
"step": 301
},
{
"epoch": 2.3709993468321358,
"grad_norm": 0.13229192793369293,
"learning_rate": 6.298950211049115e-06,
"loss": 0.0283,
"step": 302
},
{
"epoch": 2.3788373612018288,
"grad_norm": 0.12901271879673004,
"learning_rate": 6.147334755577596e-06,
"loss": 0.0258,
"step": 303
},
{
"epoch": 2.3866753755715218,
"grad_norm": 0.13257119059562683,
"learning_rate": 5.997310106161589e-06,
"loss": 0.0279,
"step": 304
},
{
"epoch": 2.3945133899412148,
"grad_norm": 0.12502893805503845,
"learning_rate": 5.848888922025553e-06,
"loss": 0.0259,
"step": 305
},
{
"epoch": 2.4023514043109078,
"grad_norm": 0.13395585119724274,
"learning_rate": 5.702083727091978e-06,
"loss": 0.0286,
"step": 306
},
{
"epoch": 2.4101894186806008,
"grad_norm": 0.12500248849391937,
"learning_rate": 5.556906908924655e-06,
"loss": 0.0266,
"step": 307
},
{
"epoch": 2.4180274330502938,
"grad_norm": 0.13586992025375366,
"learning_rate": 5.413370717683347e-06,
"loss": 0.0302,
"step": 308
},
{
"epoch": 2.4258654474199868,
"grad_norm": 0.13480009138584137,
"learning_rate": 5.271487265090163e-06,
"loss": 0.0281,
"step": 309
},
{
"epoch": 2.43370346178968,
"grad_norm": 0.14151740074157715,
"learning_rate": 5.131268523407509e-06,
"loss": 0.0291,
"step": 310
},
{
"epoch": 2.441541476159373,
"grad_norm": 0.13139696419239044,
"learning_rate": 4.992726324427901e-06,
"loss": 0.0283,
"step": 311
},
{
"epoch": 2.449379490529066,
"grad_norm": 0.12287548929452896,
"learning_rate": 4.855872358475547e-06,
"loss": 0.0243,
"step": 312
},
{
"epoch": 2.4572175048987592,
"grad_norm": 0.12606504559516907,
"learning_rate": 4.720718173419947e-06,
"loss": 0.0271,
"step": 313
},
{
"epoch": 2.465055519268452,
"grad_norm": 0.1324881911277771,
"learning_rate": 4.587275173701428e-06,
"loss": 0.0281,
"step": 314
},
{
"epoch": 2.4728935336381452,
"grad_norm": 0.13066086173057556,
"learning_rate": 4.4555546193688735e-06,
"loss": 0.029,
"step": 315
},
{
"epoch": 2.480731548007838,
"grad_norm": 0.12259192019701004,
"learning_rate": 4.3255676251295456e-06,
"loss": 0.0248,
"step": 316
},
{
"epoch": 2.4885695623775312,
"grad_norm": 0.13243158161640167,
"learning_rate": 4.19732515941125e-06,
"loss": 0.0282,
"step": 317
},
{
"epoch": 2.496407576747224,
"grad_norm": 0.13366110622882843,
"learning_rate": 4.070838043436786e-06,
"loss": 0.0295,
"step": 318
},
{
"epoch": 2.5042455911169172,
"grad_norm": 0.1202232763171196,
"learning_rate": 3.94611695031086e-06,
"loss": 0.0241,
"step": 319
},
{
"epoch": 2.51208360548661,
"grad_norm": 0.13794013857841492,
"learning_rate": 3.8231724041194515e-06,
"loss": 0.0284,
"step": 320
},
{
"epoch": 2.5199216198563033,
"grad_norm": 0.13122816383838654,
"learning_rate": 3.7020147790418263e-06,
"loss": 0.0261,
"step": 321
},
{
"epoch": 2.527759634225996,
"grad_norm": 0.12636157870292664,
"learning_rate": 3.582654298475091e-06,
"loss": 0.024,
"step": 322
},
{
"epoch": 2.5355976485956893,
"grad_norm": 0.12592478096485138,
"learning_rate": 3.4651010341716028e-06,
"loss": 0.0261,
"step": 323
},
{
"epoch": 2.5434356629653823,
"grad_norm": 0.1321692168712616,
"learning_rate": 3.3493649053890326e-06,
"loss": 0.0255,
"step": 324
},
{
"epoch": 2.5512736773350753,
"grad_norm": 0.12444434314966202,
"learning_rate": 3.2354556780534424e-06,
"loss": 0.0243,
"step": 325
},
{
"epoch": 2.5591116917047683,
"grad_norm": 0.1310444325208664,
"learning_rate": 3.1233829639351563e-06,
"loss": 0.0264,
"step": 326
},
{
"epoch": 2.5669497060744613,
"grad_norm": 0.13162964582443237,
"learning_rate": 3.013156219837776e-06,
"loss": 0.0266,
"step": 327
},
{
"epoch": 2.5747877204441543,
"grad_norm": 0.12017495930194855,
"learning_rate": 2.9047847468001404e-06,
"loss": 0.0248,
"step": 328
},
{
"epoch": 2.5826257348138473,
"grad_norm": 0.13093486428260803,
"learning_rate": 2.7982776893115627e-06,
"loss": 0.0279,
"step": 329
},
{
"epoch": 2.5904637491835403,
"grad_norm": 0.12869121134281158,
"learning_rate": 2.6936440345401493e-06,
"loss": 0.0252,
"step": 330
},
{
"epoch": 2.5983017635532333,
"grad_norm": 0.1348036676645279,
"learning_rate": 2.5908926115744997e-06,
"loss": 0.0281,
"step": 331
},
{
"epoch": 2.6061397779229263,
"grad_norm": 0.12896090745925903,
"learning_rate": 2.4900320906786593e-06,
"loss": 0.0267,
"step": 332
},
{
"epoch": 2.6139777922926193,
"grad_norm": 0.1246756836771965,
"learning_rate": 2.391070982560564e-06,
"loss": 0.0256,
"step": 333
},
{
"epoch": 2.6218158066623123,
"grad_norm": 0.12264394015073776,
"learning_rate": 2.2940176376538445e-06,
"loss": 0.0235,
"step": 334
},
{
"epoch": 2.6296538210320053,
"grad_norm": 0.12561777234077454,
"learning_rate": 2.1988802454132403e-06,
"loss": 0.025,
"step": 335
},
{
"epoch": 2.6374918354016983,
"grad_norm": 0.11878997832536697,
"learning_rate": 2.1056668336235622e-06,
"loss": 0.0247,
"step": 336
},
{
"epoch": 2.6453298497713913,
"grad_norm": 0.12836302816867828,
"learning_rate": 2.0143852677223075e-06,
"loss": 0.0285,
"step": 337
},
{
"epoch": 2.6531678641410843,
"grad_norm": 0.12017780542373657,
"learning_rate": 1.9250432501359354e-06,
"loss": 0.025,
"step": 338
},
{
"epoch": 2.6610058785107773,
"grad_norm": 0.11886154115200043,
"learning_rate": 1.837648319629956e-06,
"loss": 0.0244,
"step": 339
},
{
"epoch": 2.6688438928804703,
"grad_norm": 0.13172321021556854,
"learning_rate": 1.7522078506728074e-06,
"loss": 0.0269,
"step": 340
},
{
"epoch": 2.6766819072501633,
"grad_norm": 0.12933161854743958,
"learning_rate": 1.6687290528135723e-06,
"loss": 0.027,
"step": 341
},
{
"epoch": 2.6845199216198563,
"grad_norm": 0.13184285163879395,
"learning_rate": 1.5872189700736339e-06,
"loss": 0.0266,
"step": 342
},
{
"epoch": 2.6923579359895493,
"grad_norm": 0.12135521322488785,
"learning_rate": 1.5076844803522922e-06,
"loss": 0.0236,
"step": 343
},
{
"epoch": 2.7001959503592423,
"grad_norm": 0.12548145651817322,
"learning_rate": 1.4301322948464147e-06,
"loss": 0.0259,
"step": 344
},
{
"epoch": 2.7080339647289353,
"grad_norm": 0.12587113678455353,
"learning_rate": 1.3545689574841342e-06,
"loss": 0.0282,
"step": 345
},
{
"epoch": 2.7158719790986283,
"grad_norm": 0.12320797145366669,
"learning_rate": 1.2810008443726456e-06,
"loss": 0.0251,
"step": 346
},
{
"epoch": 2.7237099934683213,
"grad_norm": 0.12896938621997833,
"learning_rate": 1.2094341632602064e-06,
"loss": 0.0258,
"step": 347
},
{
"epoch": 2.7315480078380143,
"grad_norm": 0.12183412909507751,
"learning_rate": 1.1398749530123127e-06,
"loss": 0.0243,
"step": 348
},
{
"epoch": 2.7393860222077073,
"grad_norm": 0.12513582408428192,
"learning_rate": 1.0723290831021471e-06,
"loss": 0.0254,
"step": 349
},
{
"epoch": 2.7472240365774003,
"grad_norm": 0.13293574750423431,
"learning_rate": 1.006802253115277e-06,
"loss": 0.0287,
"step": 350
},
{
"epoch": 2.7550620509470933,
"grad_norm": 0.11619393527507782,
"learning_rate": 9.432999922687396e-07,
"loss": 0.0229,
"step": 351
},
{
"epoch": 2.7629000653167863,
"grad_norm": 0.12378672510385513,
"learning_rate": 8.818276589444896e-07,
"loss": 0.0247,
"step": 352
},
{
"epoch": 2.7707380796864793,
"grad_norm": 0.1200721487402916,
"learning_rate": 8.223904402372334e-07,
"loss": 0.0244,
"step": 353
},
{
"epoch": 2.7785760940561723,
"grad_norm": 0.11517145484685898,
"learning_rate": 7.649933515167407e-07,
"loss": 0.0228,
"step": 354
},
{
"epoch": 2.7864141084258653,
"grad_norm": 0.12590059638023376,
"learning_rate": 7.096412360046545e-07,
"loss": 0.0257,
"step": 355
},
{
"epoch": 2.7942521227955583,
"grad_norm": 0.12404926866292953,
"learning_rate": 6.563387643658076e-07,
"loss": 0.0252,
"step": 356
},
{
"epoch": 2.8020901371652513,
"grad_norm": 0.12552441656589508,
"learning_rate": 6.050904343141095e-07,
"loss": 0.026,
"step": 357
},
{
"epoch": 2.8099281515349444,
"grad_norm": 0.12259657680988312,
"learning_rate": 5.5590057023302e-07,
"loss": 0.0254,
"step": 358
},
{
"epoch": 2.8177661659046374,
"grad_norm": 0.11798401176929474,
"learning_rate": 5.087733228106517e-07,
"loss": 0.024,
"step": 359
},
{
"epoch": 2.8256041802743304,
"grad_norm": 0.12366917729377747,
"learning_rate": 4.637126686895532e-07,
"loss": 0.0257,
"step": 360
},
{
"epoch": 2.8334421946440234,
"grad_norm": 0.1320623904466629,
"learning_rate": 4.207224101311247e-07,
"loss": 0.0277,
"step": 361
},
{
"epoch": 2.8412802090137164,
"grad_norm": 0.11438736319541931,
"learning_rate": 3.7980617469479953e-07,
"loss": 0.0214,
"step": 362
},
{
"epoch": 2.8491182233834094,
"grad_norm": 0.11790072917938232,
"learning_rate": 3.4096741493194197e-07,
"loss": 0.0238,
"step": 363
},
{
"epoch": 2.856956237753103,
"grad_norm": 0.12108506262302399,
"learning_rate": 3.0420940809451624e-07,
"loss": 0.0241,
"step": 364
},
{
"epoch": 2.8647942521227954,
"grad_norm": 0.12375527620315552,
"learning_rate": 2.6953525585855234e-07,
"loss": 0.0251,
"step": 365
},
{
"epoch": 2.872632266492489,
"grad_norm": 0.12097339332103729,
"learning_rate": 2.3694788406241896e-07,
"loss": 0.025,
"step": 366
},
{
"epoch": 2.8804702808621814,
"grad_norm": 0.12990431487560272,
"learning_rate": 2.064500424599436e-07,
"loss": 0.0257,
"step": 367
},
{
"epoch": 2.888308295231875,
"grad_norm": 0.11965355277061462,
"learning_rate": 1.7804430448837839e-07,
"loss": 0.0267,
"step": 368
},
{
"epoch": 2.8961463096015674,
"grad_norm": 0.12657780945301056,
"learning_rate": 1.517330670512629e-07,
"loss": 0.0257,
"step": 369
},
{
"epoch": 2.903984323971261,
"grad_norm": 0.12143310904502869,
"learning_rate": 1.27518550316158e-07,
"loss": 0.0242,
"step": 370
},
{
"epoch": 2.9118223383409534,
"grad_norm": 0.16820454597473145,
"learning_rate": 1.0540279752731253e-07,
"loss": 0.0241,
"step": 371
},
{
"epoch": 2.919660352710647,
"grad_norm": 0.12241283059120178,
"learning_rate": 8.538767483325383e-08,
"loss": 0.0235,
"step": 372
},
{
"epoch": 2.9274983670803394,
"grad_norm": 0.13118112087249756,
"learning_rate": 6.747487112931661e-08,
"loss": 0.0281,
"step": 373
},
{
"epoch": 2.935336381450033,
"grad_norm": 0.11441458761692047,
"learning_rate": 5.166589791513465e-08,
"loss": 0.023,
"step": 374
},
{
"epoch": 2.9431743958197254,
"grad_norm": 0.12388182431459427,
"learning_rate": 3.796208916709565e-08,
"loss": 0.0266,
"step": 375
},
{
"epoch": 2.951012410189419,
"grad_norm": 0.12555594742298126,
"learning_rate": 2.636460122578399e-08,
"loss": 0.0236,
"step": 376
},
{
"epoch": 2.958850424559112,
"grad_norm": 0.1175137609243393,
"learning_rate": 1.6874412698408836e-08,
"loss": 0.0221,
"step": 377
},
{
"epoch": 2.966688438928805,
"grad_norm": 0.12019700556993484,
"learning_rate": 9.492324376214612e-09,
"loss": 0.0237,
"step": 378
},
{
"epoch": 2.974526453298498,
"grad_norm": 0.12358499318361282,
"learning_rate": 4.218959166932268e-09,
"loss": 0.0248,
"step": 379
},
{
"epoch": 2.982364467668191,
"grad_norm": 0.13017946481704712,
"learning_rate": 1.0547620421907934e-09,
"loss": 0.0284,
"step": 380
},
{
"epoch": 2.990202482037884,
"grad_norm": 0.11907056719064713,
"learning_rate": 0.0,
"loss": 0.0251,
"step": 381
}
],
"logging_steps": 1,
"max_steps": 381,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1405397234876416.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}